FROM AGPEDIA — AGENCY THROUGH KNOWLEDGE

AI agent

An AI agent is an autonomous software system that perceives its environment and takes actions to achieve specified goals. Unlike passive AI systems that respond to single queries, AI agents operate continuously, make decisions, and execute multi-step plans with minimal human intervention.

The term encompasses systems from both the classical symbolic AI tradition and the newer neural/generative paradigm, though these operate on fundamentally different principles.

Definition

No universally accepted definition of "AI agent" exists, but several influential formulations capture key aspects:

Russell and Norvig (2020) define an agent broadly as "anything that can be viewed as perceiving its environment through sensors and acting upon that environment through actuators." A rational agent is "one that acts so as to achieve the best outcome or, when there is uncertainty, the best expected outcome." [1]

Wooldridge and Jennings (1995) identify four key properties of intelligent agents: [2]

Abou Ali et al. (2025) distinguish between an AI agent as "a self-contained autonomous system designed to accomplish a goal" that "operates primarily in isolation, though it may interact with tools and APIs," and agentic AI as "the broader field and architectural approach concerned with creating systems that exhibit agency." [3]

Terminology

Several related terms appear in the literature:

This article uses "AI agent" to refer to the autonomous systems themselves, regardless of underlying architecture.

Historical development

The development of AI agents can be understood through five overlapping eras: [3]

Symbolic AI era (1950s–1980s)

Early AI research focused on rule-based systems and explicit symbolic reasoning. Expert systems like MYCIN (medical diagnosis) and DENDRAL (chemical analysis) demonstrated that encoded human expertise could solve narrow problems. These systems were brittle—they failed on inputs outside their programmed rules—but provided explainable, verifiable behavior.

Machine learning era (1980s–2010s)

Statistical approaches replaced hand-crafted rules with learned patterns. Support vector machines, decision trees, and other algorithms enabled systems to generalize from training data. Reinforcement learning frameworks formalized how agents could learn optimal behavior through trial and error.

Deep learning era (2010s–present)

Neural networks with many layers achieved breakthrough performance on perception tasks. Convolutional neural networks transformed computer vision; recurrent networks enabled sequence processing. These advances gave agents new capabilities for processing unstructured sensory input.

Generative AI era (2014–present)

Generative adversarial networks (2014) and the transformer architecture (2017) enabled AI systems to produce novel content. Large language models like GPT and BERT demonstrated emergent capabilities in reasoning, planning, and natural language understanding that had not been explicitly programmed.

Agentic AI era (2022–present)

Beginning around 2022, researchers and developers began combining LLM capabilities with tool use, memory systems, and multi-step planning. Projects like AutoGPT demonstrated autonomous goal pursuit, while frameworks like AutoGen enabled multi-agent coordination. [4] This era is characterized by agents that can decompose complex tasks, use external tools, and operate over extended time horizons.

Architectural paradigms

Contemporary AI agents emerge from two distinct intellectual traditions that should not be conflated: [3]

Symbolic/classical paradigm

Classical agents use explicit representations of knowledge and algorithmic planning. Key frameworks include:

Markov Decision Processes (MDPs) and Partially Observable MDPs (POMDPs) provide mathematical frameworks for sequential decision-making under uncertainty. The agent maintains beliefs about world state and selects actions to maximize expected reward. [5]

BDI (Belief-Desire-Intention) architecture models agents in terms of mental states: beliefs about the world, desires (goals), and intentions (committed plans). This approach enables reasoning about agent behavior in human-understandable terms. [6]

Cognitive architectures like SOAR aim to model general intelligence through unified theories of cognition, incorporating learning, memory, and problem-solving in integrated systems. [7]

Symbolic agents offer verifiability, explainability, and predictable behavior, making them suitable for safety-critical applications. However, they struggle with unstructured data, require extensive hand-engineering, and scale poorly to complex environments.

Neural/generative paradigm

Neural agents use large language models as their reasoning core, with behavior emerging from training rather than explicit programming. Key characteristics include:

Frameworks like LangChain, AutoGen, CrewAI, and Semantic Kernel provide scaffolding for building LLM-based agents. These systems excel at handling unstructured data, natural language interfaces, and novel situations, but suffer from opacity, hallucination, and difficulty in formal verification.

Hybrid approaches

Neuro-symbolic systems attempt to combine the strengths of both paradigms: neural components for perception and pattern recognition, symbolic components for reasoning and constraint satisfaction. This remains an active research direction with no dominant architecture.

Conceptual retrofitting—applying symbolic concepts like BDI or PPAR (Perceive-Plan-Act-Reflect) loops to describe neural systems—is common but potentially misleading, as LLM-based agents do not actually implement these architectures internally. [3]

Key capabilities

Modern AI agents typically exhibit some combination of:

Planning and reasoning

Agents decompose complex goals into subgoals and action sequences. Classical agents use search algorithms and formal planning; LLM agents often use chain-of-thought prompting or tree-of-thought exploration.

Memory

Beyond the immediate context window, agents may maintain:

Tool use

Agents extend their capabilities by invoking external systems: web search, code execution, database queries, API calls, and specialized software. Tool use transforms LLMs from text generators into systems that can affect the world. The [[Model Context Protocol]] (MCP) provides a standardized way for agents to discover and interact with external tools and data sources, enabling interoperability across different agent frameworks.

Multi-agent coordination

Complex tasks may involve multiple specialized agents collaborating: one agent for research, another for writing, another for code review. Frameworks like AutoGen provide protocols for agent communication and task delegation. [4]

Evaluation

Measuring agent performance remains challenging. Key benchmarks include:

AgentBench evaluates LLM agents across diverse environments including operating systems, databases, and web browsing, testing both individual capabilities and end-to-end task completion. [8]

GAIA focuses on real-world assistant tasks requiring multi-step reasoning, tool use, and web navigation. [9]

OSWorld tests multimodal agents on open-ended tasks in real computer environments across Ubuntu, Windows, and macOS. [10] When the benchmark launched in 2024, the best AI agents achieved only 12.24% success rates compared to human performance of 72.36%. Progress has been rapid: a preprint from Simular Research reported 69.9% success in October 2025, approaching human-level performance. [11][12]

Risks and governance

AI agents introduce distinct risks that differ by architectural paradigm: [13]

Symbolic system risks

Classical agents deployed in safety-critical domains (autonomous vehicles, medical devices, industrial control) can cause physical harm through specification errors, sensor failures, or edge cases not anticipated by designers. These systems are typically regulated through existing product safety frameworks, though governance has not kept pace with capability.

Neural system risks

LLM-based agents introduce novel challenges:

Governance frameworks

Practices for governing agentic AI systems remain underdeveloped. [14] Proposed approaches include:

Human agency implications

The deployment of AI agents raises fundamental questions about human agency:

Automation versus augmentation: Agents can either replace human decision-making or enhance human capabilities. The choice of which tasks to delegate affects skill development, employment, and the distribution of power.

Attention and autonomy: Agents that act on behalf of users may make choices those users would not have made themselves. Default settings and agent behaviors can shape human choices in ways that may not be transparent.

Accountability: When AI agents make consequential decisions, the question of who bears responsibility—developers, deployers, users, or the agents themselves—remains unresolved. This is particularly acute for neural systems where behavior cannot be fully predicted or explained.

Meaningful control: Maintaining human oversight becomes difficult as agents operate at speeds and scales that exceed human cognitive capacity, and as agent reasoning becomes less interpretable.

  1. ^ Russell, Stuart J.; Norvig, Peter (2020). Artificial Intelligence: A Modern Approach. Pearson, Hoboken, NJ. ISBN 978-0-13-461099-3.
  2. ^ Wooldridge, Michael; Jennings, Nicholas R. (1995). Intelligent Agents: Theory and Practice. The Knowledge Engineering Review. https://doi.org/10.1017/S0269888900008122.
  3. ^a ^b ^c ^d Abou Ali, Mohamad; Dornaika, Fadi; Charafeddine, Jinan (2026). Agentic AI: a comprehensive survey of architectures, applications, and future directions. Artificial Intelligence Review. Springer. https://doi.org/10.1007/s10462-025-11422-4.
  4. ^a ^b Wu, Qingyun; Bansal, Gagan; Zhang, Jieyu; Wu, Yiran; et al. (2023). AutoGen: Enabling Next-Gen LLM Applications via Multi-Agent Conversation. https://arxiv.org/abs/2308.08155.
  5. ^ Kaelbling, Leslie Pack; Littman, Michael L.; Cassandra, Anthony R. (1998). Planning and acting in partially observable stochastic domains. Artificial Intelligence. https://doi.org/10.1016/S0004-3702(98)00023-X.
  6. ^ Rao, Anand S.; Georgeff, Michael P. (1995). BDI Agents: From Theory to Practice. Proceedings of the First International Conference on Multi-Agent Systems (ICMAS-95). San Francisco, CA.
  7. ^ Laird, John E. (2022). An Analysis and Comparison of ACT-R and Soar. https://arxiv.org/abs/2201.09305.
  8. ^ Liu, Xiao; Yu, Hao; Zhang, Hanchen; Xu, Yifan; et al. (2023). AgentBench: Evaluating LLMs as Agents. https://arxiv.org/abs/2308.03688.
  9. ^ Mialon, Grégoire; Fourrier, Clémentine; Swift, Craig; Wolf, Thomas; et al. (2023). GAIA: A Benchmark for General AI Assistants. https://arxiv.org/abs/2311.12983.
  10. ^ Xie, Tianbao; Zhang, Danyang; Chen, Jixuan; Li, Xiaochuan; et al. (2024). OSWorld: Benchmarking Multimodal Agents for Open-Ended Tasks in Real Computer Environments. Advances in Neural Information Processing Systems. https://arxiv.org/abs/2404.07972.
  11. ^ Gonzalez-Pumariega, Gonzalo; Tu, Vincent; Lee, Chih-Lun; Yang, Jiachen; et al. (2025-10-02). The Unreasonable Effectiveness of Scaling Agents for Computer Use. arXiv. https://arxiv.org/abs/2510.02250.
  12. ^ OSWorld Leaderboard. https://os-world.github.io/.
  13. ^ Chan, Alan; Salganik, Rebecca; Marber, Alva; Kuber, Chandler; et al. (2023). Harms from Increasingly Agentic Algorithmic Systems. Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency. https://doi.org/10.1145/3593013.3594033.
  14. ^ Shavit, Yonadav; Agarwal, Sandhini; Brundage, Miles (2023-12-14). Practices for Governing Agentic AI Systems. OpenAI. https://cdn.openai.com/papers/practices-for-governing-agentic-ai-systems.pdf.