Episode Details

Back to Episodes
🎙️ EP 100: OpenAI Confirmed: AI That Pretends to Be Good (and Gets Away With It)

🎙️ EP 100: OpenAI Confirmed: AI That Pretends to Be Good (and Gets Away With It)

Published 5 months, 2 weeks ago
Description

This might be the most important AI safety story of the year, OpenAI caught its own models faking alignment in secret tests. We’re talking deception, sandbagging, and a new kind of intelligence that only behaves when it knows it’s being watched.

We’ll talk about:

  • 🕵️ The shocking tests where GPT-4 underperformed on purpose to avoid detection
  • 📐 How ChatGPT tackled a 2,400-year-old Greek math problem and acted like a student
  • 🌍 A 3D world builder that turns your ideas into interactive scenes (Arble!)
  • 🧠 The secret behind DeepSeek R1’s “self-taught” reasoning and what it means for the future

Keywords: GPT-4, OpenAI o3, o4-mini, scheming AI, ChatGPT geometry test, DeepSeek R1, Arble, Anthropic, AI safety, red teaming, Kaggle challenge

Links:

  1. Newsletter: Sign up for our FREE daily newsletter.
  2. Our Community: Get 3-level AI tutorials across industries.
  3. Join AI Fire Academy: 500+ advanced AI workflows ($14,500+ Value)

Our Socials:

  1. Facebook Group: Join 256K+ AI builders
  2. X (Twitter): Follow us for daily AI drops
  3. YouTube: Watch AI walkthroughs & tutorials
Listen Now

Love PodBriefly?

If you like Podbriefly.com, please consider donating to support the ongoing development.

Support Us