Problem
Many teams describe a retrieval-augmented generation system as if it were a single feature. In practice, the system behaves more like a chain of narrow dependencies: chunking, indexing, retrieval, reranking, context assembly, generation, and output validation.
What breaks in production
- Retrieval recall drops when indexing assumptions stop matching query behavior.
- Latency budgets get consumed by avoidable search and post-processing steps.
- Prompt changes mask underlying retrieval quality problems instead of fixing them.
- Teams optimize for demo success rather than observed production usefulness.
Practical design approach
Treat the system as a pipeline with explicit checkpoints.
- Define the user task and what constitutes a useful answer.
- Measure retrieval quality before discussing final answer quality.
- Track whether ranking, chunking, or source freshness is actually limiting the system.
- Keep each layer observable enough that the team can explain failures.
Tradeoffs
The best retrieval setup is not always the one with the most complex architecture. In regulated or operational environments, simpler systems with clearer evaluation boundaries often outperform more elaborate stacks because engineers can reason about them when something goes wrong.
Production lesson
RAG becomes credible when retrieval quality, latency, and operational risk are measured directly. If those signals are missing, the system is not production-ready even if the generated answers look impressive in a demo.