
Agentic RAG extends traditional Retrieval-Augmented Generation by giving the retrieval process itself to an autonomous agent that can iteratively search, evaluate, refine queries, and synthesize results across multiple sources. Unlike standard RAG which performs a single retrieval step before generation, Agentic RAG systems can: decide when additional retrieval is needed, reformulate queries based on partial results, cross-reference multiple sources for accuracy, and use tools beyond simple vector search (web browsing, API calls, database queries). This pattern is enabled by the new tool orchestration capabilities seen in OpenAI's Responses API computer environment, Anthropic's programmatic tool calling, and the Everything Claude Code framework's pre-engineered retrieval skills.
Why it matters
Standard RAG has a fundamental limitation: it performs one retrieval pass and hopes the results are sufficient. For simple factual queries this works, but for complex questions that require synthesizing information across documents, connecting disparate facts, or resolving contradictions between sources, single-pass retrieval frequently returns incomplete or misleading context. Agentic RAG addresses this by letting an autonomous agent control the entire retrieval loop — deciding what to search for, evaluating whether results are sufficient, and iterating until the answer is complete. This is particularly critical for enterprise knowledge bases, legal research, and scientific literature review where answer quality depends on thoroughness.

How it works
The agent operates in a retrieve-evaluate-refine loop. It starts with an initial query, retrieves candidate documents, then evaluates whether the results adequately address the question. If gaps remain, it reformulates the query — narrowing scope, using different terminology, or targeting a different retrieval backend entirely. The agent can access multiple tools beyond vector search: web browsing for current information, SQL queries for structured data, API calls to specialized databases, and even other agents for sub-questions. Tool orchestration via MCP or programmatic tool calling enables efficient chaining of these retrieval steps. The final synthesis step cross-references all gathered evidence, resolves contradictions, and cites sources with provenance.
Example
A legal research agent asked 'What are the compliance requirements for deploying AI in healthcare across EU member states?' illustrates the difference. Standard RAG might return a general EU AI Act summary. An Agentic RAG system first retrieves the EU AI Act text, identifies healthcare as a high-risk category, then searches for member-state implementation guidance, cross-references with recent enforcement actions, checks for sector-specific regulations in key markets (Germany, France, Netherlands), and queries medical device regulation databases. After five retrieval iterations, it synthesizes a comprehensive answer with jurisdiction-specific requirements and primary source citations — far beyond what a single retrieval pass could produce.