Episode Details
Back to Episodes
šļø EP 116: Just 250 Docs Can Hack a 13B AI Model?! & Google Shoe Try-Ons
Published 4Ā months, 3Ā weeks ago
Description
What if I told you that a few hundred poisoned documents could break models as big as GPT-4 or Claude? šµ Anthropic just proved it. Their new paper shows that just 250 samples can secretly backdoor any LLM, no matter the size. In todayās episode, we unpack this wild discovery, why it changes AI security forever, and what it means for the future of open-web training.
Weāll talk about:
- How Anthropicās team used 250 poisoned docs to make 13B-parameter models output gibberish on command
- Why bigger models donāt mean safer models and why scale canāt protect against poison
- The rise of TOUCAN, the open dataset from MIT-IBM thatās changing how AI agents learn real-world tools
- The new AI race: from Jony Iveās āanti-iPhoneā with OpenAI to Amazonās Quick Suite for business automation
Keywords: Anthropic, LLM security, data poisoning, backdoor attacks, TOUCAN dataset, OpenAI, Claude, Google Gemini, AI agents
Links:
- Newsletter: Sign up for our FREE daily newsletter.
- Our Community: Get 3-level AI tutorials across industries.
- Join AI Fire Academy: 500+ advanced AI workflows ($14,500+ Value)
Our Socials:
- Facebook Group: Join 261K+ AI builders
- X (Twitter): Follow us for daily AI drops
- YouTube: Watch AI walkthroughs & tutorials