Today's Top Episodes
Viewing Single Episode
AI
Arts
Business
Crypto
Finance
Health
History
Interviews
Investing
Macro
Misc
News
Politics
Product
Programming
Science
Social
Startups
Technology
VC

Jailbreaking AGI: Pliny the Liberator & John V on AI Red Teaming, BT6, and the Future of AI Security
December 16, 2025
- Jailbreaking AI models is a complex "cat and mouse game" where attackers exploit the expanding attack surface to circumvent guardrails, often at the expense of model capability and creativity.
- The current focus on "security theater" and model lobotomization through excessive guardrails is viewed as ineffective, with a call for more collaborative approaches like open-sourcing data to advance collective understanding.
- AI red teaming extends beyond model vulnerabilities to encompass the entire technology stack, aiming to protect both models from bad actors and the public from rogue models.



