Build a Complete, Production-Grade RAG System and Evaluate It Rigorously
Project Requirements
1. Document Corpus Selection (Choose one):- Option A: Use provided corpus (technical documentation, 50 docs)
- Option B: Bring your own (company docs, research papers, etc.)
- Two-stage retrieval (first-pass + reranking)
- Hybrid search (BM25 + semantic)
- Appropriate chunking strategy with metadata
- Error handling and fallbacks
- Logging and observability
- Advanced pattern (GraphRAG, iterative, or agentic)
- Unstructured data handling (PDFs, images, tables)
- Query preprocessing/expansion
- Result caching
- Golden dataset (minimum 20 test cases)
- Retrieval metrics (Recall@5, Precision@5, NDCG@10) — prefer LlamaIndex evaluations
- Generation metrics (Faithfulness, Relevance) — prefer LlamaIndex evaluations
- Component-level analysis (retrieval vs. generation failures)
- Failure case analysis (identify and document 3 worst queries)
- Chunking strategy comparison (test 2+ strategies)
- Search strategy comparison (BM25, semantic, hybrid)
- Document optimization decisions with metrics
Module Exercises
Chunking Strategies
Experiment with chunking strategies:Pseudocode
Search Strategy Selection
Implement and compare all three approaches on the same dataset:Pseudocode
Reranking
Pseudocode
Unstructured Data
Process a mixed document corpus:Pseudocode