DiscoverSuper Data Science: ML & AI Podcast with Jon Krohn915: How to Jailbreak LLMs (and How to Prevent It), with Michelle Yi
915: How to Jailbreak LLMs (and How to Prevent It), with Michelle Yi

915: How to Jailbreak LLMs (and How to Prevent It), with Michelle Yi

Update: 2025-08-19
Share

Description

Tech leader, investor, and Generationship cofounder Michelle Yi talks to Jon Krohn about finding ways to trust and secure AI systems, the methods that hackers use to jailbreak code, and what users can do to build their own trustworthy AI systems. Learn all about “red teaming” and how tech teams can handle other key technical terms like data poisoning, prompt stealing, jailbreaking and slop squatting. 




This episode is brought to you by Trainium2, the latest AI chip from AWS and by the Dell AI Factory with NVIDIA.




Additional materials: ⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠www.superdatascience.com/915⁠⁠⁠⁠⁠




Interested in sponsoring a SuperDataScience Podcast episode? Email natalie@superdatascience.com for sponsorship information.




In this episode you will learn:





    • (03:31 ) What “trustworthy AI” means     




    • (31:15 ) How to build trustworthy AI systems 




    • (46:55 ) About Michelle’s “sorry bench”  




    • (48:13 ) How LLMs help construct causal graphs  




    • (51:45 ) About Generationship 




Comments 
In Channel
loading
00:00
00:00
1.0x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

915: How to Jailbreak LLMs (and How to Prevent It), with Michelle Yi

915: How to Jailbreak LLMs (and How to Prevent It), with Michelle Yi

Jon Krohn