DiscoverVanishing GradientsEpisode 52: Why Most LLM Products Break at Retrieval (And How to Fix Them)
Episode 52: Why Most LLM Products Break at Retrieval (And How to Fix Them)

Episode 52: Why Most LLM Products Break at Retrieval (And How to Fix Them)

Update: 2025-07-021
Share

Description

Most LLM-powered features do not break at the model. They break at the context. So how do you retrieve the right information to get useful results, even under vague or messy user queries?



In this episode, we hear from Eric Ma, who leads data science research in the Data Science and AI group at Moderna. He shares what it takes to move beyond toy demos and ship LLM features that actually help people do their jobs.



We cover:

• How to align retrieval with user intent and why cosine similarity is not the answer

• How a dumb YAML-based system outperformed so-called smart retrieval pipelines

• Why vague queries like “what is this all about” expose real weaknesses in most systems

• When vibe checks are enough and when formal evaluation is worth the effort

• How retrieval workflows can evolve alongside your product and user needs



If you are building LLM-powered systems and care about how they work, not just whether they work, this one is for you.



LINKS





🎓 Learn more:





📺 Watch the video version on YouTube: YouTube link

Comments 
loading
In Channel
loading
00:00
00:00
1.0x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

Episode 52: Why Most LLM Products Break at Retrieval (And How to Fix Them)

Episode 52: Why Most LLM Products Break at Retrieval (And How to Fix Them)

Hugo Bowne-Anderson