Overview
AgenticRAG is the complete RAG system that orchestrates all components into an intelligent pipeline. It’s called “agentic” because it makes smart decisions about:
- Query optimization: Automatically rewrites queries for better retrieval
- Retrieval strategy: Uses semantic, keyword, or hybrid search
- Result refinement: Re-ranks results for optimal relevance
- Answer generation: Synthesizes answers from retrieved context
Quick Start
Configuration
AgenticRAG uses a configuration-based API for clean, organized settings:Minimal Configuration
LLM Configuration
Retrieval Configuration
Complete Configuration
The Agentic Pipeline
When you query the system, AgenticRAG executes an intelligent pipeline:1
Query Rewriting (Optional)
Generates multiple query variations to improve retrieval coverage
2
Embedding
Converts query (and variations) into vector embeddings
3
Retrieval
Searches vector store using:
- Semantic search (default)
- Hybrid search (semantic + BM25)
top_k most relevant chunks4
Re-ranking (Optional)
Re-ranks retrieved chunks using:
- LLM-based scoring
- Cohere Rerank API
- Local cross-encoder models
rerank_top_k chunks5
Answer Generation
Uses LLM to generate answer based on re-ranked context
Operations
Index Documents
Query the System
Access Response Details
Get System Statistics
Features
Query Rewriting
Automatically generates query variations for better retrieval:Learn more
Explore query rewriting in detail
Hybrid Search
Combine semantic and keyword search:Learn more
Discover hybrid search capabilities
Re-ranking
Improve result quality with re-ranking:Learn more
Explore re-ranking strategies
Observability
Track and monitor your RAG pipeline:Learn more
Set up observability and monitoring
Best Practices
Configuration Management
Configuration Management
Organize configurations logically:
Start Simple
Start Simple
Begin with defaults, then optimize:
Monitor Performance
Monitor Performance
Use observability to understand behavior:
Metadata Strategy
Metadata Strategy
Use metadata for filtering and organization:
Common Patterns
Pattern 1: Document Q&A System
Pattern 2: Research Assistant
Pattern 3: Multi-lingual Support
Troubleshooting
Poor answer quality
Poor answer quality
Solutions:
- Enable query rewriting
- Increase
top_kfor more context - Enable re-ranking
- Try hybrid search
Slow queries
Slow queries
Solutions:
- Reduce
top_k - Disable query rewriting
- Use faster embedding model
- Optimize vector store index
High costs
High costs
Solutions:
- Use smaller LLM model
- Reduce
top_kandrerank_top_k - Disable query rewriting
- Cache results
