Tuesday Jul 29, 2025

RAG Is Not Solved – Your Evaluation Just Sucks

RAG Is Not Solved – Your Evaluation Just Sucks

Your RAG pipeline is passing benchmarks, but failing reality. In this episode, Yuval sits down with Niv from AI21 to expose why most RAG evaluation is fundamentally flawed. From overhyped retrieval scores to chunking strategies that collapse under real-world complexity, they break down why your system isn’t as good as you think — and how structured RAG solves problems that traditional pipelines simply can't. 

Bonus: what do Seinfeld trivia, World Cup stats, and your enterprise SharePoint have in common? (hint: your RAG pipeline chokes on all of them).

Key Topics:

  1. Why most RAG benchmarks reward the wrong thing (and hide real failures)
  2. The chunking trap: how bad segmentation sabotages good retrieval
  3. When LLMs ace the answer—but your pipeline still fails
  4. Structured RAG: pipeline that solves RAG problem over aggregative data (such as financial reports)
  5. Evaluation tips, tricks, and traps for AI builders


Comment (0)

No comments yet. Be the first to say something!

© 2025 AI21

Podcast Powered By Podbean

Version: 20241125