{"version":"1.0","type":"rich","provider_name":"Acast","provider_url":"https://acast.com","height":250,"width":700,"html":"<iframe src=\"https://embed.acast.com/$/60518a52f69aa815d2dba41c/65ef604559f4740016fa9121?\" frameBorder=\"0\" width=\"700\" height=\"250\"></iframe>","title":"How Should Governments Use Deepfakes?","description":"<p>Progress in deepfake technology and artificial intelligence can make manipulated media hard to identify, making deepfakes an appealing tool for governments seeking to advance their national security objectives. But in a low-trust information environment, balancing the risks and rewards of a government-run deepfake campaign is trickier than it may seem.</p><p>To talk through how democracies should think about using deepfakes, <em>Lawfare</em>'s Fellow in Technology Policy and Law, Eugenia Lostri, was joined by Daniel Byman, Senior Fellow at the Center for Strategic &amp; International Studies and professor at Georgetown University; Daniel Linna, Director of Law and Technology Initiatives at Northwestern University; and V.S. Subrahmanian, the Walter P. Murphy Professor of Computer Science and Buffett Faculty Fellow at Northwestern University. They recently published a <a href=\"https://www.csis.org/analysis/government-use-deepfakes\" rel=\"noopener noreferrer\" target=\"_blank\">report</a> examining two critical points: the questions that a government agency should address before deploying a deepfake, and the governance mechanisms that should be in place to assess its risks and benefits.</p>","author_name":"The Lawfare Institute"}