{"version":"1.0","type":"rich","provider_name":"Acast","provider_url":"https://acast.com","height":250,"width":700,"html":"<iframe src=\"https://embed.acast.com/$/60518a52f69aa815d2dba41c/649c7d984a397000125bb01b?\" frameBorder=\"0\" width=\"700\" height=\"250\"></iframe>","title":"Talking Transparency With Meta’s Nick Clegg","thumbnail_width":200,"thumbnail_height":200,"thumbnail_url":"https://open-images.acast.com/shows/60518a52f69aa815d2dba41c/show-cover.png?height=200","description":"<p>How much transparency do big technology companies owe to their users? The question has become pointed in recent years as users, researchers, and politicians voice discontent about the absence of public information available about how platforms moderate and amplify content.&nbsp;</p><p>Today, Meta’s President of Global Affairs, Nick Clegg, <a href=\"https://about.fb.com/news/2023/06/how-ai-ranks-content-on-facebook-and-instagram/\" rel=\"noopener noreferrer\" target=\"_blank\">announced a new initiative</a> to provide more information about how the company’s ranking algorithms work on Facebook and Instagram. On this episode of <em>Arbiters of Truth</em>, <em>Lawfare</em>’s occasional series on the information ecosystem, <em>Lawfare</em> Senior Editors Quinta Jurecic and Alan Rozenshtein talked with Clegg about how Meta has approached transparency for both users and researchers. They also discussed <a href=\"https://nickclegg.medium.com/you-and-the-algorithm-it-takes-two-to-tango-7722b19aa1c2\" rel=\"noopener noreferrer\" target=\"_blank\">Clegg’s controversial 2021 essay</a> on how Meta’s algorithms interact with user preferences.</p><p><em>Meta provides support for </em>Lawfare<em>’s Digital Social Contract paper series. This podcast episode is not part of that series, and Meta does not have any editorial role in </em>Lawfare<em>.</em></p>","author_name":"The Lawfare Institute"}