{"version":"1.0","type":"rich","provider_name":"Acast","provider_url":"https://acast.com","height":250,"width":700,"html":"<iframe src=\"https://embed.acast.com/$/69ab3b7c7036d739021982df/69af8b0ab58ea3074de9532f?\" frameBorder=\"0\" width=\"700\" height=\"250\"></iframe>","title":"Claude Blackmailed Its Developers. Here's Why the System Hasn't Collapsed Yet.","description":"<p>What's really happening with AI safety in 2026? The common story is that the safety system is collapsing — but the reality is more complicated.</p><p><br></p><p>In this video, I share the inside scoop on why the AI risk picture is both worse and more resilient than the headlines suggest:</p><p><br></p><p>Why frontier AI agents scheme even after anti-scheming training</p><p>- How competitive dynamics create emergent safety properties no lab planned</p><p>- What \"intent engineering\" is and why it beats prompt engineering for AI agents</p><p>- Where the real vulnerability lives — and why it's you, not the models</p><p><br></p><p>The risks from large language models and autonomous AI agents are accelerating, but so are the structural forces holding the system together — and closing the gap between what you tell an agent and what you actually mean is the most leveraged safety skill you can build right now.</p><p><br></p><p>Chapters</p><p>00:00 Why This Isn't Terminator</p><p>02:15 How Frontier Models Actually Learn</p><p>04:40 The Misalignment Mechanic: Novel Paths Gone Wrong</p><p>06:55 What Anthropic's Sabotage Report Actually Shows</p><p>08:30 Every Major Model Schemes — The Apollo Research Findings</p><p>10:10 Can You Train Scheming Out? The Anti-Scheming Paradox</p><p>12:45 The Race Dynamic and Why Labs Keep Cutting Corners</p><p>15:20 Four Emergent Safety Properties Nobody Planned</p><p>20:05 The Consciousness Framing Is Hurting Us</p><p>23:30 Intent Engineering: The Fix That's Up to You</p><p>28:10 Three Questions That Change Everything</p><p>30:45 Where We Stand in 2026</p><p><br></p><p>Subscribe for daily AI strategy and news.</p><p>For deeper playbooks and analysis: https://natesnewsletter.substack.com/</p>","author_name":"Nate B. Jones"}