{"version":"1.0","type":"rich","provider_name":"Acast","provider_url":"https://acast.com","height":250,"width":700,"html":"<iframe src=\"https://embed.acast.com/$/64dcccacf78da20011671a85/688096dbfd9acfeba4c2696c?\" frameBorder=\"0\" width=\"700\" height=\"250\"></iframe>","title":"How We Lose Control of AI - and How to Stop It, with Kristian Rönn","thumbnail_width":200,"thumbnail_height":200,"thumbnail_url":"https://open-images.acast.com/shows/64dcccacf78da20011671a85/1753257351528-de714f3a-540c-4da7-b03d-b210797dbc41.jpeg?height=200","description":"<p>Everyone’s talking about ChatGPT—but what if the real threat from AI is happening elsewhere?</p><p><br></p><p>Kristian Rönn is the co-founder of Lucid, a company helping governments track the movement and usage of AI chips globally. He previously founded Normative, the carbon accounting platform that helped shape EU and UK climate disclosure rules—and began his career researching existential risk at Oxford’s Future of Humanity Institute.</p><p><br></p><p>He’s also the author of <em>The Darwinian Trap</em>, a bestselling book that explores how misaligned incentives—and short-term thinking—can push systems toward catastrophic outcomes. It’s a framework he now applies to AI, arguing that without global coordination, the very infrastructure powering this technology could spiral out of control.</p><p><br></p><p>In this conversation, he explains why AI’s biggest risks aren’t in the models—but in the chips, supply chains, and silent diffusion happening behind the scenes.</p><p><br></p><p>We explore:</p><p>• Why frontier models are a distraction from the real governance challenge</p><p>• The one global policy move governments must make before it’s too late</p><p>• What a chicken sandwich teaches us about AI’s hidden complexity</p><p>• And what carbon accounting taught Kristian about building systems that actually scale</p><p><br></p><p>If you care about who’s really in control of AI—and how we avoid losing the plot—this is essential listening.</p><p><br></p><p>Chapters</p><p>00:00 – Intro: Welcome to High Net Purpose</p><p>00:29 – Kristian Rönn: From Philosopher to AI Safety Pioneer</p><p>01:31 – Early Life: What Shaped Kristian’s Worldview</p><p>04:29 – The Impact of Peter Singer and Utilitarian Thinking</p><p>08:31 – Exploring Existential Risks and the Road to AI</p><p>13:20 – Building Normative: The Startup That Changed Carbon Accounting</p><p>21:29 – Why Kristian Left Climate Tech for AI Governance</p><p>23:22 – The Darwinian Trap: Why Good Incentives Lead to Bad Outcomes</p><p>25:36 – Darwinian Demons vs. Cooperative Systems</p><p>34:47 – The Challenge of Global AI Governance</p><p>35:23 – Centralization vs. Decentralization in AI Control</p><p>37:33 – The Chicken Sandwich Analogy: AI and Hidden Value Chains</p><p>39:28 – Can Decentralized Governance Actually Work?</p><p>40:49 – Enlightenment Thinking and Biological Drives</p><p>45:11 – AI Assurance, Risk Management, and Value Chain Complexity</p><p>48:27 – Who Should Govern AI? Security, Policy, and Global Standards</p><p>53:09 – What the Future of AI Regulation Could Look Like</p><p>01:05:18 – Final Thoughts: Purpose, Power, and What Comes Next</p>","author_name":"Islandbridge"}