stack-profile:agentic-rag
Agentic RAG Stack (LlamaIndex, ChromaDB, LangChain, FastAPI, React) overview
A retrieval-augmented generation architecture where an AI agent dynamically decides what to retrieve, how to chunk and re-rank results, and when to perform multi-hop retrieval across heterogeneous data sources. LlamaIndex provides the data connectors, indexing pipelines, and query engines. ChromaDB (or Qdrant) serves as the vector store for embedding-based similarity search. LangChain handles prompt orchestration, tool integration, and output parsing. FastAPI exposes the RAG pipeline as an async API with streaming support. React powers the chat frontend with real-time token streaming. This stack is best for enterprise knowledge bases, legal document QA, and customer support copilots where static retrieval falls short and the agent must reason over retrieval strategy. The tradeoff is latency — agentic retrieval adds LLM calls per query compared to single-shot RAG.
Attributes
Outgoing edges
- domain:ml-ai·DomainML/AI
- domain:knowledge-management·DomainKnowledge Management
- library:llama-index·LibraryLlamaIndex
- tool:chromadb·ToolChroma
- framework:langchain·FrameworkLangChain
- framework:fastapi·FrameworkFastAPI
- framework:react·FrameworkReact
- language:python·LanguagePython
- language:typescript·LanguageTypeScript
- tool:docker·ToolDocker
- workflow:rag-pipeline-evaluation·WorkflowRAG Pipeline Evaluation
- workflow:prompt-engineering-iteration·WorkflowPrompt Engineering Iteration
- skill-area:retrieval-augmented-generation·SkillAreaRetrieval-Augmented Generation
- skill-area:rag-pipeline-engineering·SkillAreaRAG Pipeline Engineering
- skill-area:embedding-optimization·SkillAreaEmbedding Optimization
- skill-area:prompt-engineering·SkillAreaPrompt Engineering
- skill-area:context-management·SkillAreaLLM Context Management
- role:ml-engineer·RoleMachine Learning Engineer
- role:backend-engineer·RoleBackend Engineer
- role:fullstack-engineer·RoleFullstack Engineer