Robots AtlasRobots Atlas

AI Agent

Embodies the Agentic AI paradigm as a concrete executable artifact - a single, autonomous instance with a goal, access to tools, memory, and a control loop that can be deployed, monitored, and versioned as a production software component.

Category
Abstraction level
Operation level
Coding automation Multi-file editing Autonomous task resolutionComputer user agents – agents controlling the mouse, keyboard, and browser.Customer support agents with access to a knowledge base and actions.Researching and synthesizing information from multiple sourcesData analysis agents – iterative code execution and result visualizationVirtual assistants – scheduling, email management, calendar organization, shopping.High-level planning for robotic agents controlling robot actions

The agent receives a goal from the user and definitions of available tools (JSON/OpenAPI/MCP schemas) and system instructions (role, security policies). In each iteration of the loop: (1) the model analyzes the current context and decides on the next action - invoke a tool, ask a question, or terminate; (2) the host executes the chosen tool and returns the result; (3) the result is appended to the context as an observation; (4) the model decides whether to continue. The loop ends when the model determines the goal has been achieved, the max_steps limit is reached, or an error state requiring escalation to a human is detected. The agent can maintain short-term memory within the context window and long-term memory in an external storage (vector database, key-value store).

A single LLM invocation is unable to handle open-ended tasks, where the number of steps is not known in advance, requiring interaction with the environment, access to current data, code execution, or iterative result verification. The AI Agent solves this problem by embedding the model in a control loop with access to tools and memory, enabling autonomous end-to-end task execution.

01

Base language model

Reasoning and control loop engine

The agent's reasoning and decision engine. Generates plans, selects tools, interprets results, and decides termination. Typically an LLM post-trained with RLHF and tool-use, optionally a reasoning model (CoT with a dedicated reasoning token budget).

Chat LLM (Claude, GPT, Gemini)Reasoning model (o-series, R1)
02

Tool Interface

Bridge between the agent and the external world.

Modular

Definitions of callable functions with their schemas (JSON Schema, OpenAPI) and documentation. Anthropic calls this the Agent-Computer Interface (ACI) β€” care in designing it is critical for agent reliability. Often exposed through Model Context Protocol.

Function calling APIModel Context Protocol (MCP)
03

Memory System

Persistence of task state and context

Modular

Short-term memory (conversation history, tool results in context) and optional long-term memory (vector store, key-value store, episodic structures) across sessions. Determines coherence and personalization in long-running tasks.

04

Control Loop

Reasoning-action cycle executor

Modular

Mechanism executing iterations: fetch context β†’ call model β†’ parse decision β†’ execute tool β†’ update context β†’ check termination. Manages limits (max_steps, time/token budgets) and detects infinite loops.

05

System Prompt

Defines the agent's identity and operational boundaries.

Constant instruction defining the agent's identity, goal, scope of responsibility, safety rules, response format, and termination criteria. First line of defense against misbehavior and prompt injection.

06

Guardrails / Safeguards

Safety and compliance oversight

Modular

Filters and validators operating before inference (input sanitization), during (tool call schema validation), and after (output control, PII redaction, blocking irreversible actions). Critical for production safety.

07

Evaluation and Observability

Production observability and quality

Modular

Step logging, traces (LangSmith, Arize, Helicone), metrics (success rate, tool error rate, average steps, cost per task), and automated evaluations against test sets. Essential for production maintenance of an agent.

Parallelism

Conditionally parallel

Parallelism is most often achieved inter-sessionally (multiple agents for different tasks) or in orchestrator-workers patterns (one orchestrator delegates to multiple worker agents simultaneously).

Paradigm

Conditional

Input dependent

An agent is distinct from a workflow, where the path is predefined in the code and the LLM only executes specific steps, whereas in an agent, the LLM controls the entire process.

Poziom autonomii

Critical
  • proposal_onlyAgent prepares, human executes.
  • confirm_irreversibleAutonomy with confirmation required for irreversible actions.
  • full_autonomy

Scope of decisions the agent makes without human approval β€” from proposal-only mode to full autonomy with rollback.

Toolkit

Critical
  • web_search + code_exec
  • file_read + file_write + bash + git
  • browser_use (computer use)

List of callable functions available to the agent. Defines the action space and is the strongest predictor of agent behavior.

Maximum Number of Steps

Standard
  • 10Short tasks, low budget.
  • 50-200Coding agents, long-term research.

Hard limit on loop iterations before forced termination. Safeguards against cost overrun and infinite loops.

Memory Strategy

Standard
  • in_context_only
  • in_context + summarization
  • in_context + vector_memory

How context is managed across steps and sessions: context window only, summarization, vector store, episodic structures.

Cost / Token Budget

Standard
  • 10k_tokens
  • 1M_tokens / 5_USD

Maximum computational cost or token count per agent run. Critical for production deployments with outcome-based billing.

Human fallback strategy

Standard
  • before_irreversible_actions
  • after_3_consecutive_failures

When the agent escalates to a human: never, on demand, after N failed steps, before irreversible action, based on uncertainty signals.

Common pitfalls

Poorly Designed Agent-Computer Interface (ACI)
CRITICAL

Unclear tool names, missing examples, ambiguous parameters β€” the same problems that affect junior developers affect the model. Anthropic reports spending more time optimizing tools than the agent prompt itself.

Each tool definition should include usage examples, edge cases, input format, and boundaries with other tools. Use absolute paths, clear types, and poka-yoke (error prevention through argument structure). Test iteratively on real-world examples.

Hallucinations in Action (Execution Claims)
CRITICAL

The agent may claim to have performed an action it didn't actually execute, or invoke tools with fabricated parameters β€” particularly dangerous in multi-step pipelines where errors propagate.

Validate all tool calls and schema; enforce ground truth through execution results; do not rely on the model's internal declarations about the system state β€” verify the actual state using tools.

Infinite loops and compounding errors
HIGH

Without a hard max_steps and repetition detection, the agent can loop indefinitely, generating wrong steps based on previously wrong observations. Costs grow linearly with the number of steps.

Set explicit max_steps, detect repeating action signatures, use an evaluator-optimizer to enforce termination conditions, and log costs in real-time.

Tool result prompt injection
CRITICAL

Malicious instructions embedded in web pages, documents, or emails the agent reads can hijack its behavior by impersonating system instructions.

Structurally isolate untrusted content by using role labeling and sandbox tags, require explicit user confirmation for actions resulting from observed content, and apply content filtering.

Irreversible actions taken without human oversight
CRITICAL

An agent with access to write-capable tools (delete, send_email, db_write, payment) can cause real damage based on flawed reasoning. Consequences may be irreversible.

Grant the agent minimal privileges, prefer reversible operations such as soft delete and drafts instead of send, and require human-in-the-loop before taking irreversible actions.

Context Window Overflow
HIGH

Accumulated action history and tool results can exceed the model's context window, causing silent truncation of earlier steps and loss of relevant information.

Implement compacting or summarizing of history, offload tool results to external storage with a reference, and monitor token budget per step.

Building an agent where a workflow would suffice
MEDIUM

Anthropic strongly recommends: don't build an agent when the task has a known, predefined structure. A workflow is cheaper, faster, more predictable, and easier to debug than an agent.

First, try a solution with a single LLM call plus retrieval, then workflow with chain, routing, or parallelization; introduce an agent only when the task is open-ended and the number of steps is unpredictable.

1995

BDI Architectures – foundation of agent theory

Russell and Norvig formalize rational agents; Belief-Desire-Intention architectures emerge (Rao and Georgeff). The canon is defined: an agent perceives the environment and takes goal-oriented actions.

2022

ReAct – LLM as Loop Reasoning-Action Engine

breakthrough

Yao et al. (2022) demonstrate that LLMs can interleave Chain-of-Thought with tool calls in a single loop. The practical definition of an LLM-based AI Agent.

2023

AI Agents (Autonomous Agents) - First Wave of Autonomous Agents

Virally popular implementations show autonomous GPT-4 agents performing multi-step tasks. Despite limited reliability, they show the potential and popularize the term.

2023

Function calling in main model APIs

breakthrough

OpenAI (June 2023) introduces function calling in GPT-4; Anthropic and Google follow. First-class agent support at the level of commercial model APIs.

2024

Building Effective Agents – Formal Definition of Agent vs Workflow

breakthrough

Anthropic publishes (December 2024) guidelines distinguishing agents from workflows and five composition patterns. The canonical definition of an agent: a system in which an LLM dynamically directs its own process.

2024

Computer use – screen-controlling agents

Anthropic introduces Computer Use in Claude (October 2024) β€” the agent clicks, types, and moves the mouse like a human. OpenAI's Operator (2025) follows. Opens a class of GUI-driven agents independent of APIs.

2025

Model Context Protocol – standard for agent tools interoperability

Anthropic releases MCP as an open standard for connecting LLMs to external tool servers. Enables an ecosystem of tools portable across model providers.

2026

Commercialization of agents in the Agents-as-a-Service model

Sierra (March 2026) announces the Agents-as-a-Service paradigm β€” customers buy outcomes delivered by an agent rather than SaaS access. Agents become the unit of product delivery, not just a technical library.

GPU Tensor CoresPRIMARY

Base LLM inference dominates the agent's cost and latency; GPUs with tensor cores are the standard for all modern production-grade models.

TPUGOOD

Google deploys Gemini-based agents on TPUs; comparable throughput and cost to GPUs for most workloads.

Hardware agnosticGOOD

The control loop, tool parsing, and orchestration layer itself is lightweight and runs on CPU; hardware requirements stem from the base model, not from the agent's construction.

BUILT ON

LLM

A Large Language Model (LLM) is a class of machine learning models based on the Transformer architecture, trained on large text datasets via autoregressive language modeling (next-token prediction). These models have billions of parameters and can generate coherent text, answer questions, write code, translate languages, and perform many other language-cognitive tasks without task-specific fine-tuning. The term covers models such as GPT, LLaMA, Gemini, Claude, and Mistral. Most modern LLMs are instruction-tuned (SFT + RLHF) after the pre-training phase.

GO TO CONCEPT
Tool-augmented LLM

Tool-augmented LLM is an architectural pattern in which a large language model is equipped with access to one or more external tools that it can invoke during inference by generating structured function-call or API-call outputs. The model learns when and how to call tools by producing special tokens or structured output (e.g., JSON function calls) that are intercepted by a host runtime, executed against the tool, and whose results are returned to the model as new context for continued generation. The canonical formalization appeared in the Toolformer paper (Schick et al., Meta AI, 2023), which demonstrated that LLMs can learn to self-supervise their own tool-use through API call annotation without requiring large labeled datasets. Toolformer showed that models trained this way can decide which tools to call, when, and with which arguments, and that tool use substantially improves performance on tasks requiring fresh information, arithmetic, multilingual lookup, and question answering. The pattern encompasses several mechanisms: (1) in-context tool specification, where tool interfaces are described in the system prompt or context (JSON Schema, OpenAPI, natural language); (2) function calling APIs, where the model produces structured output matched to a defined schema and the host dispatches the call; (3) ReAct-style interleaving, where the model alternates reasoning traces with tool-use observations; and (4) parallel tool calling, where the model emits multiple tool calls simultaneously to be executed concurrently. Key implementations include OpenAI function calling (GPT-4, June 2023), Anthropic tool use (Claude, 2023), Google Gemini function calling, and the Model Context Protocol (MCP, 2024) which standardizes tool server connectivity.

GO TO CONCEPT
CoT

Chain-of-Thought (CoT) Reasoning is a prompting technique introduced by Wei et al. (2022) in which a large language model is induced to generate a series of intermediate natural-language reasoning steps as part of its output, prior to producing a final answer. The technique was shown to significantly improve LLM performance on arithmetic, commonsense, and symbolic reasoning benchmarks where standard few-shot prompting yields flat or poor results. In the original formulation (few-shot CoT), a small number of exemplar question-answer pairs are included in the prompt, where each answer consists of a chain of thought followed by the final answer. The model learns from these demonstrations to produce its own reasoning chains. A subsequent zero-shot variant (Kojima et al., 2022) showed that appending the phrase 'Let's think step by step' to a question is sufficient to elicit reasoning chains from large models without any exemplars. CoT is an emergent property: empirical results in the originating paper show that reasoning ability via CoT prompting appears only in models above a certain parameter threshold (approximately 100B parameters for the models tested in 2022), with smaller models not benefiting or performing worse. This relationship has been revisited by subsequent work as smaller models have been fine-tuned on CoT data. Key extensions include Self-Consistency CoT (Wang et al., 2022), which samples multiple reasoning paths and selects the most frequent final answer; Tree of Thoughts (Yao et al., 2023), which frames reasoning as search over a tree of intermediate thoughts; and native reasoning models such as OpenAI o1 (2024) and DeepSeek-R1 (2025), which internalize extended reasoning through reinforcement learning on process reward signals rather than relying on prompting.

GO TO CONCEPT

EXTENDS

Agentic AI

Agentic AI denotes an architectural transition from single-turn, stateless generative models toward goal-directed systems capable of autonomous perception, planning, action, and adaptation through iterative control loops. An agentic system wraps a large language model in a runtime that gives the model access to tools (web search, code execution, APIs, file I/O), persistent memory, and feedback from prior steps. The model then decides dynamically which tools to call, in what order, and whether to loop or stop, rather than following a predefined code path. Two primary system types are commonly distinguished: (1) Workflows, in which LLMs and tools are orchestrated through predefined code paths, and (2) Agents, in which the LLM itself directs its process and tool usage dynamically. Both can be composed into multi-agent systems where specialized agents collaborate, with one acting as orchestrator and others as subagents. Key design patterns identified by Anthropic (2024) include prompt chaining, routing, parallelization, orchestrator-workers, and evaluator-optimizer loops. Andrew Ng's 2024 taxonomy describes four foundational patterns: Reflection, Tool Use, Planning, and Multi-Agent Collaboration. Formal frameworks model agentic control loops as Partially Observable Markov Decision Processes (POMDPs). The control loop is: perceive state β†’ reason/plan β†’ select action β†’ execute tool β†’ observe result β†’ update state β†’ repeat. Agentic systems introduce risks not present in single-turn models, including hallucination in action, prompt injection through observed content, infinite loops, reward hacking, and tool misuse.

GO TO CONCEPT

Connects

Tool-augmented LLM

Tool-augmented LLM is an architectural pattern in which a large language model is equipped with access to one or more external tools that it can invoke during inference by generating structured function-call or API-call outputs. The model learns when and how to call tools by producing special tokens or structured output (e.g., JSON function calls) that are intercepted by a host runtime, executed against the tool, and whose results are returned to the model as new context for continued generation. The canonical formalization appeared in the Toolformer paper (Schick et al., Meta AI, 2023), which demonstrated that LLMs can learn to self-supervise their own tool-use through API call annotation without requiring large labeled datasets. Toolformer showed that models trained this way can decide which tools to call, when, and with which arguments, and that tool use substantially improves performance on tasks requiring fresh information, arithmetic, multilingual lookup, and question answering. The pattern encompasses several mechanisms: (1) in-context tool specification, where tool interfaces are described in the system prompt or context (JSON Schema, OpenAPI, natural language); (2) function calling APIs, where the model produces structured output matched to a defined schema and the host dispatches the call; (3) ReAct-style interleaving, where the model alternates reasoning traces with tool-use observations; and (4) parallel tool calling, where the model emits multiple tool calls simultaneously to be executed concurrently. Key implementations include OpenAI function calling (GPT-4, June 2023), Anthropic tool use (Claude, 2023), Google Gemini function calling, and the Model Context Protocol (MCP, 2024) which standardizes tool server connectivity.

GO TO CONCEPT
CoT

Chain-of-Thought (CoT) Reasoning is a prompting technique introduced by Wei et al. (2022) in which a large language model is induced to generate a series of intermediate natural-language reasoning steps as part of its output, prior to producing a final answer. The technique was shown to significantly improve LLM performance on arithmetic, commonsense, and symbolic reasoning benchmarks where standard few-shot prompting yields flat or poor results. In the original formulation (few-shot CoT), a small number of exemplar question-answer pairs are included in the prompt, where each answer consists of a chain of thought followed by the final answer. The model learns from these demonstrations to produce its own reasoning chains. A subsequent zero-shot variant (Kojima et al., 2022) showed that appending the phrase 'Let's think step by step' to a question is sufficient to elicit reasoning chains from large models without any exemplars. CoT is an emergent property: empirical results in the originating paper show that reasoning ability via CoT prompting appears only in models above a certain parameter threshold (approximately 100B parameters for the models tested in 2022), with smaller models not benefiting or performing worse. This relationship has been revisited by subsequent work as smaller models have been fine-tuned on CoT data. Key extensions include Self-Consistency CoT (Wang et al., 2022), which samples multiple reasoning paths and selects the most frequent final answer; Tree of Thoughts (Yao et al., 2023), which frames reasoning as search over a tree of intermediate thoughts; and native reasoning models such as OpenAI o1 (2024) and DeepSeek-R1 (2025), which internalize extended reasoning through reinforcement learning on process reward signals rather than relying on prompting.

GO TO CONCEPT

Commonly used with

MAS

Multi-Agent Systems (MAS) are a paradigm in Distributed Artificial Intelligence in which multiple autonomous software entities β€” agents β€” interact within a shared environment to achieve individual or collective goals. Each agent perceives its environment through sensors or interfaces, reasons about its state, and acts through actuators or API calls. In the context of LLM-based MAS (emerging prominently from 2023 onward), agents are powered by large language models that provide the cognitive core (planning, reasoning, natural language communication), supplemented by memory modules, tool-use interfaces, and role-specific prompts. The system architecture defines how agents coordinate: coordination topologies include sequential pipelines, hierarchical orchestration (orchestrator-worker), parallel fan-out/fan-in, publish-subscribe messaging, and decentralized peer-to-peer communication. Core agent properties, as defined by Wooldridge and Jennings (1995), include autonomy, social ability, reactivity, and pro-activeness. In LLM-based systems, key components are: the agent (an LLM with a system prompt defining its role), a communication channel (natural language messages, structured function calls, or shared memory), an orchestrator or coordinator (managing task decomposition, routing, and state), tool-use interfaces (external APIs, code execution, web search), and a memory subsystem (short-term context, long-term vector storage). Prominent frameworks implementing LLM-based MAS include AutoGen (Microsoft, 2023), CAMEL (2023), MetaGPT (2023), CrewAI, and LangGraph.

GO TO CONCEPT
MCP

Model Context Protocol (MCP) is an open protocol developed by Anthropic and released in November 2024. It addresses the MΓ—N integration problem in AI systems: connecting M different LLM applications to N different external tools previously required MΓ—N bespoke connectors. MCP defines a standardized client-host-server architecture where hosts (LLM applications) manage one or more clients, each maintaining a stateful session with a specific server. Servers expose capabilities as three primitives: Resources (structured data for context), Prompts (templated instructions), and Tools (executable functions). Clients expose two primitives: Roots (filesystem entry points) and Sampling (server-initiated LLM completions). Communication is based on JSON-RPC 2.0. Capability negotiation occurs at session initialization. The protocol is transport-agnostic and has been implemented in Python, TypeScript, C#, and Java SDKs. In December 2025, Anthropic donated MCP governance to the Agentic AI Foundation (AAIF) under the Linux Foundation.

GO TO CONCEPT
RAG

Retrieval-Augmented Generation (RAG) was introduced by Lewis et al. (2020) as a general-purpose fine-tuning recipe combining pre-trained parametric memory (a seq2seq language model, specifically BART in the original paper) with non-parametric memory (a dense vector index of Wikipedia, accessed via Dense Passage Retrieval, DPR). In the original formulation, both the retriever and the generator are fine-tuned end-to-end: given an input query x, the retriever retrieves top-k documents z from the corpus, and the generator produces an output y conditioned on x and z. Two formulations were proposed: RAG-Sequence (the same retrieved documents condition the full output sequence) and RAG-Token (different documents may be used per generated token, marginalized during generation). In widespread contemporary usage (post-2022, with the growth of LLM applications), 'RAG' has expanded to describe a broader class of retrieve-then-generate pipelines, typically with a frozen LLM, a vector store containing pre-computed dense embeddings of document chunks, and a retrieval step that fetches top-k relevant chunks based on embedding similarity to the query. The retrieved chunks are appended to the prompt as context before the LLM generates a response. This non-trainable pipeline variant is technically distinct from the original Lewis et al. formulation but is the dominant practical interpretation of RAG as of 2023–2025. The canonical modern RAG pipeline consists of an offline indexing phase (document chunking, embedding computation, storage in a vector database) and an online query phase (query embedding, approximate nearest neighbor search, context-augmented generation). Key design decisions include: chunk size and overlap, embedding model choice, retrieval strategy (dense, sparse/BM25, or hybrid), number of retrieved documents k, and context integration method (prepend to prompt, cross-attention injection, or fusion-in-decoder). RAG addresses two fundamental limitations of parametric-only LLMs: the knowledge cutoff problem (inability to access post-training information) and hallucination (generation of factually incorrect content). However, RAG introduces its own failure modes, including retrieval of irrelevant or misleading context and the LLM's susceptibility to being distracted by retrieved content that contradicts its parametric knowledge.

GO TO CONCEPT
Reasoning model

A reasoning model (also: large reasoning model, LRM, reasoning language model, RLM) is a type of large language model that has been specifically post-trained to solve complex multi-step problems by explicitly generating intermediate reasoning steps before committing to a final response. Unlike standard LLMs that generate a direct response in a single forward pass, reasoning models allocate additional computation at inference time β€” a property known as test-time compute scaling β€” by producing a long internal chain of thought (CoT). The reasoning trace typically includes steps such as problem decomposition, hypothesis generation, self-verification, reflection, and correction of errors. The defining characteristics of reasoning models are: (1) post-training via large-scale reinforcement learning (RL) using reward signals based on final answer correctness (and sometimes intermediate step quality via process reward models); (2) the emergence of extended, often hidden, reasoning traces that precede the final answer; (3) a consistent empirical relationship between the length or computational budget allocated to the reasoning trace and final answer quality (test-time scaling law); (4) superior performance on verifiable tasks requiring multi-step logic, such as mathematics, competitive programming, and scientific reasoning. The term 'reasoning model' was introduced as a product category by OpenAI in September 2024 with the release of the o1-preview model. OpenAI described o1 as trained via a large-scale RL algorithm teaching the model to use chain of thought productively. The approach does not rely on explicit tree search algorithms; instead, implicit search emerges via RL-trained CoT generation. In January 2025, DeepSeek published the first detailed open technical description of this class of models in the DeepSeek-R1 paper (arXiv:2501.12948), demonstrating that reasoning capabilities can be incentivized via pure RL without supervised fine-tuning, using Group Relative Policy Optimization (GRPO) as the RL framework. Reasoning models typically employ the same base Transformer decoder architecture as standard LLMs, with the key difference residing entirely in the post-training pipeline: RL replaces or augments standard RLHF/SFT, and reward signals are grounded in verifiable outcomes. The resulting models generate substantially longer token sequences during inference (reasoning tokens), which are often hidden from end users but incur real compute costs. Performance consistently improves with both more training-time RL compute and more inference-time thinking budget.

GO TO CONCEPT
AaaS

Agents as a Service (AaaS) is a software delivery model in which a vendor provides the customer with an autonomous AI agent that performs concrete business tasks, instead of a traditional human-operated application. The term was publicly introduced on March 25, 2026, by Sierra co-founders Bret Taylor and Clay Bavor in a blog manifesto announcing their Ghostwriter agent as their own realization of this paradigm. Unlike Software as a Service (SaaS), where customers buy access to an interface (menus, form fields, tables) and perform the work themselves through clicks, in AaaS the customer defines a desired outcome in natural language, and the vendor delivers an agent that builds and improves production agents or performs the work directly. The defining property is full autonomy: the agent has access to data, tools, a sandboxed test environment, and the deployment pipeline, while the human acts as a supervisor approving changes. The key technical enabler is the agent harness β€” scaffolding of tools, memory, planning, and task context β€” combined with refactoring the platform into headless infrastructure that an agent can invoke programmatically rather than navigating through a UI. The work cycle includes analyzing interactions, identifying improvement opportunities, validating in a sandbox, and preparing for review β€” which Sierra calls an "agent assembly line." AaaS is tightly coupled with the Agentic AI paradigm (its technical foundation) and outcome-based billing models (its commercial superstructure).

GO TO CONCEPT
Building effective agents

Canonical definition of agent vs. workflow, and five compositional patterns for production agentic systems.

blogAnthropic
ReAct: Synergizing Reasoning and Acting in Language Models

The technical foundation of the reasoning-action loop with LLMs.

scientific articlearXiv (Yao et al.)
What's next for AI agentic workflows (Andrew Ng)

Four agent design patterns: Reflection, Tool use, Planning, Multi-Agent.

articleDeepLearning.AI