AI Agents Revolution

High level overview of AI Agents and evolutions.
LLM, Agents
Published

October 23, 2024

Modified

January 29, 2025

👶 AI Agents Revolution

AI Agents

In the context of Large Language Models (LLMs), AI Agents are autonomous software entities designed to extend the capabilities of LLMs.

Recently , we have seen several agent frameworks such as AutoGen, CrewAI and LangGraph and examples such as AutoGPT and BabyAGI.

These solutions leverage LLM capabilities and typically consist of the following steps:

  • 📄 Task decomposition: break complex tasks into smaller, manageable steps.
  • ⚒️ Tool integration: LLM interacting with various tools and APIs to gather data from the environment, such as web scraping tools, calculator and python interpreter.
  • 📚 Memory: used to retain information from past interactions, allowing to provide more context information.
  • 💱 Autonomous execution: once plan is established, the agent can execute the steps autonomously.

These steps make AI agents powerful as they extend the functionality of LLMs beyond simple generation tasks towards autonomous task execution. For example, ReAct (Reasoning and Acting) is a common methodology applied to AI agents to improve performance by leveraging reasoning and acting capabilities.

ReAct

Picture by Shunyu Yao - LLM Agents - Brief History and Overview

What has changed?

So far, AI agent frameworks have needed to communicate with the outside world through tool integration in the form of APIs. However, today Anthropic announced a new approach—LLMs performing tasks directly on your computer.

This is possible using computer use capability available in Claude 3.5 sonnet and Claude 3.5 Haiku.

Available today on the API, developers can direct Claude to use computers the way people do—by looking at a screen, moving a cursor, clicking buttons, and typing text. Claude 3.5 Sonnet is the first frontier AI model to offer computer use in public beta. At this stage, it is still experimental — at times cumbersome and error-prone. We’re releasing computer use early for feedback from developers, and expect the capability to improve rapidly over time.

According to Anthropic, instead of creating specific tools to help Claude complete individual tasks, we’re teaching it general computer skills—allowing it to use a wide range of standard tools and software programs designed for people.

Examples of Anthropic’s computer use capability include:

  • Creating an entire website on the user’s computer and even fixes bugs in the code [here].
  • Using data from user’s computer and online data to fill out forms [here].
  • Orchestrating a multi-step task by searching the web, using native applications, and creating a plan with the resulting information [here].

Figure below shows the prompt for the example “Using data from user’s computer and online data to fill out forms” from youtube video: Computer use for automating operations.

Anthropic’s Computer Use

Picture by Anthropic’s new computer use tool from Youtube Computer use for automating operations video.

🔭 Things to Consider while building AI Agents that works

This section is based on my notes from the presentation given by Manjeet Singh, Salesforce on the SMALLCON conference by Predibase in Dec. 2024.

Important things to consider while building Production-ready AI agents based on steps provided above:

  • 📄 Task decomposition: Topics
  • ⚒️ Tool integration: Sources, knowledge quality, guardrails
  • 📚 Memory: User context, page context
  • 💱 Autonomous execution: actions, intent, latency

In addition, as highlight by Manjeet Singh, Salesforce customer trust on the agent is important, and common points below need to be considered:

Data privacy and security concerns: e.g. legal

​lack of confidence in accuracy: “how this will behave in production? I am getting different answer each time

​negative past experience: “I tried before and it did not work

​ROI and cost concerns: “how to calculate business value?”

image-20250110082320573

Picture by Manjeet Singh, Salesforce

As in in software development best practice, the agent development follows similar steps, which are:

➡️ Ideate ➡️ Setup ➡️ Configure ➡️ Test ➡️ Deploy ➡️ Monitor 🔁

image-20250110083443875

Picture by Manjeet Singh, Salesforce

⚒️ Test: toolset you need

Testing and evaluation during agent application development is very important.

As in all generative AI usecase, the journey normally looks like this:

➡️ Select a pre-trained LLM (e.g. gpt-4o or claude 3.5)

➡️ Prompt Engineering (improve prompt with few shots)

➡️ Prompt Engineering + RAG (longer prompts, slower inference)

➡️ Advanced techniques (advanced RAG, finetuning - improve speed, quality and cost)

Here is an example from Manjeet Singh, Salesforce to atomate agent evaluation.

image-20250110084935905

Picture by Manjeet Singh, Salesforce

Below is an example of evaluation while considering a RAG (Retrieval Augmented Generation) pipeline.

image-20250110085416224

Picture by Manjeet Singh, Salesforce

👩‍🏫 How good are those agents today?

Agent Evaluations

Picture by Philipp Schimid

References: