"El rápido avance de la inteligencia artificial y las tecnologías de clonación de voz harán que saber cuándo un audio es real y cuándo es falso sea cada vez más complicado. Y con ello, la manipulación de audio utilizada para fines maliciosos, amenaza con poner en peligro la integridad de inf
...
ormaciones clave para el ciudadano [...] VerificAudio es la herramienta de Inteligencia Artificial de Prisa para combatir la desinformación en los contenidos sonoros, una amenaza en ascenso propiciada por el avance de las últimas tecnologías generativas. Estas tecnologías se encuentran a la orden del día, en especial, este año en el que 73 países -casi la mitad de la población mundial- celebrarán comicios cruciales para el panorama internacional. VerificAudio combina técnicas de verificación periodística y modelos de inteligencia artificial avanzados para detectar y analizar la fiabilidad de archivos sonoros con contenido informativo." (Introduction)
more
"This report examines political communication and media trust in the age of generative artificial intelligence systems (AI). Firstly, it provides a brief explainer of generative AI tools and techniques, looking separately at systems that generate text and those that generate or manipulate images, vi
...
deos and audio. By reference to real-world examples, the paper then surveys the ways in which generative AI systems have recently been used by political actors, distinguishing between three different use-cases: political campaigning, entertainment and disinformation campaigns. Building on this empirical analysis, the paper distils important insights for policymakers, which highlight the need to: refrain from falsely labelling content as AI-generated to avoid overstating the technical capabilities and persuasive power of those spreading disinformation; acknowledge the multimodality of threats posed by generative AI, in particular voice-generation; delimit fair-use cases of generative AI for political campaigning, given these technologies are already widely used for legitimate political communication purposes; raise awareness of how seemingly non-political uses of generative AI can be exploited for politics, in particular the creation of non-consensual intimate content. This is followed by an evaluation of emerging technical and policy solutions, namely the detection and labelling of deepfakes as well as the development of systems to certify content authenticity and provenance. The section concludes with a discussion of the emerging legal landscape, including the European Union’s AI Act." (Executive summary)
more
"This study contributes to research on digital inequalities in the context of artificial intelligence by examining user perceptions of deepfake technology. We focus on the stratification of deepfake knowledge and attitudes towards deepfakes as critical elements of technology access. Based on a surve
...
y of 1,421 German internet users, we analyse the role of sociodemographic variables, digital skills, and personal innovativeness in predicting deepfake knowledge. We then examine the role of deepfake knowledge in users’ assessments of risks and potentials associated with the technology. Our results point to a generally low level of knowledge and a strong focus on risks in internet users’ perceptions of deepfakes. We find that age, gender and educational attainment predict knowledge about deepfakes. Digital skills, personal innovativeness, and social media use also positively relate to deepfake knowledge. This knowledge, in turn, is shown to play a role in users’ positive attitude towards the technology. While age plays only a minor role, female gender strongly relates to low knowledge and negative attitudes towards deepfakes. We thus find evidence of a sizeable gender divide in user access to the novel deepfake technology." (Abstract)
more
"The potential to weaponize deepfakes is growing at an alarming rate. The study aimed to explore how education can help youth develop resilience to malicious deepfakes and the ability to counter disinformation, regardless of context. Sixteen youth between the ages of 18 and 24 participated in a 9-h,
...
cutting-edge, experiential, and reflective learning experience on deepfakes and disinformation informed by personal construct theory (PCT). Participants experienced the creation of deepfakes and assessed their ability to counter disinformation. They delved into their own construct systems and reflected on the genesis of their vulnerabilities. They moved from being unfamiliar with the deepfake phenomenon to becoming empowered digital citizens, motivated to develop their skills in assessing the validity of online information and resisting manipulation regardless of its source. The study provides recommendations for more targeted education about deepfakes and disinformation for youth. Educators, curriculum developers, and policymakers can use these findings to ensure that a well-equipped generation of digital citizens protects society from the growing disinformation plague. With this proof of concept, the next step is to bring this approach to a larger number of youth and contribute to the fight against malicious deepfakes, while developing strategies to integrate PCT-informed learning experiences into education." (Abstract)
more
"Visuals are expected to amplify the impact of disinformation, but it is rarely specified how, and what exactly distinguishes them from text. This article is one of the first to treat visual disinformation as its own type of falsehood, arguing that it differs from textual disinformation in its produ
...
ction, processing and effects. We suggest that visual disinformation is determined by varying levels of modal richness and manipulative sophistication. Because manipulated visuals are processed differently on a psychological level, they have unique effects on citizens’ behaviours and attitudes." (Abstract)
more
"In particular we recommend to strengthen collaboration (platforms should adopt a collaborative approach involving various stakeholders, including governments, civil society organisations, and fact-checkers, to counter the spread and impact of disinformation. This can include sharing information, be
...
st practices, and resources to develop effective strategies); enhance transparency (platforms should prioritise transparency by providing clear and comprehensive information on their policies, algorithms, and content moderation processes. Users should have a better understanding of how their data is used, and how algorithms work to prevent the amplification of false and misleading narratives); implement effective content moderation (platforms need to allocate sufficient resources to effectively monitor and moderate harmful content. This includes investing in advanced AI systems and human moderation teams to detect and remove disinformation in a timely manner. Transparent and consistent guidelines should be in place to ensure fairness and accountability in content moderation decisions); promote fact-based information (platforms should prioritise the promotion of fact-based information from reliable sources. This can be done by partnering with credible news organizations and fact-checkers to provide accurate information and combat false narratives. Advertising promoting climate change denial or other forms of misinformation should be prevented); improve the access to data for researchers (platforms should make efforts to provide access to data for independent researchers to evaluate the effectiveness of their policies and initiatives in countering disinformation. This will enable better analysis and understanding of the impact of disinformation and the effectiveness of countermeasures); comply with regulatory frameworks (platforms should fully comply with regulatory frameworks, such as the Digital Services Act (DSA) or other relevant International, EU and National laws and regulations, that provide for obligations on addressing disinformation and mitigating associated risks, the Code of Practice on Disinformation that aims to commit signatories to a range of actions to counter disinformation. These actions include providing transparency reports on political advertising, restricting advertising placements on disinformation websites, disrupting advertising revenue for purveyors of disinformation, and enabling user feedback and fact-checking mechanisms. In this framework, compliance should not be limited to large platforms but extended, with adjustments, to smaller platforms to ensure a comprehensive approach)." (Recommendations, page 6)
more
"Current developments in technology, such as facial recognition, have already disproportionately affected people of color, especially people of African descent. The rise of DeepFakes and other forms of Fake News online has brought a host of new impacts and potential obstacles to the way that Black c
...
ommunities communicate. With a focus on the emergence of DeepFakes, and AI Synthetic Media, contributors have explored a range of themes and topics, including but not limited to: How do AI and digital algorithms impact people of color? How does Social Media shape Black women's perception of their body? How vulnerable are young Africans to social media generated fake news? Contributions have examined how Black virtual, in person and digital communication is affected by the current onslaught of misinformation, manipulated images and videos, and changing social media landscape." (Publisher description)
more
"Deception and media manipulation have always been an integral part of wartime propaganda. But never before has it been so easy to create high-quality fabrications of images as well as sound and video recordings. The tendency to react emotionally to these media opens up a whole new possibility for a
...
buse by their creators. A call to surrender by President Volodymyr Zelensky, which was immediately exposed as a deepfake, is the first attempt to use the new technology in an armed conflict. The quality of such fabrications is improving, detecting them is becoming increasingly complex and there is no end in sight to these developments. Banning deepfakes would be futile. It is therefore time to look at current and potential applications and possible counter-strategies." (Page 1)
more
"Day in, day out, we journalists receive images from combat zones in Ukraine, without knowing whether they are authentic or have been manipulated. It is our job to find out, conducting analysis that takes both time and basic knowledge of image forensics. There are many testing tools and platforms fo
...
r this. But many of those working in journalism are unsure about what exactly these tools look at and how to interpret the findings they produce [...] Image forensics and source analysis cannot convict war criminals – but they can be used to research initial indications of when and where a crime was committed. And it is vital that people are warned against the error of relying solely on a reverse image search to detect manipulation of videos and photos. This paper examines these problem areas, as well as the urgent need to incorporate these research methods into journalistic training at all levels." (Abstract)
more
"In this article, a theoretical model on the detection of deepfakes by ordinary citizens is introduced. The authors conducted three studies in which deepfakes with political content were presented. The deepfakes showed UK’s Prime Minister Boris Johnson or Barack Obama. In the deepfake videos, the
...
two politicians said things they had never said in real life. The authors expected that people who regularly and automatically reflect on information they see (analytic thinking) are more likely to identify deepfakes correctly than people who tend to be less reflective, more intuitive. The authors further expected that interest in politics is positively related to detecting political deepfakes. Indeed, the higher participants’ scores on analytic thinking (Studies 1-2) and political interest (Study 1), the better participants identified the deepfakes. Moreover, people with high analytic thinking and political interest were better at identifying a fake news article to be inaccurate (whether or not a warranting deepfake video was presented, Study 3). It is discussed how researchers, everyday people, and whole societies can deal with deepfakes." (Lay summary)
more
"Any technology that can be used to generate false or misleading content, from photocopiers and Photoshop software to deep fakes, can be weaponised. This paper argues that policymakers face a narrowing window of opportunity to minimise the consequences of weaponised deep fakes. Any response must inc
...
lude measures across three lines of effort: 1. investment in and deployment of deep fake detection technologies; 2. changing online behaviour, including via policy measures that empower digital audiences to critically engage with content and that bolster trusted communication channels; 3. creation and enforcement of digital authentication standards." (Page 4)
more
"Was macht Desinformation im deutschsprachigen Internet aus? Wie wirkt Desinformation? Wie kann sie mithilfe technischer Mittel erkannt werden? Was kann und könnte mit regulatorischen und rechtlichen Maßnahmen gegen Desinformation getan werden? Aus den Erkenntnissen von Journalistik, Medienpsychol
...
ogie, Informatik und Recht werden Handlungsempfehlungen an die relevanten Adressaten hergeleitet: An den Gesetzgeber, Presserat, Medienschaffende, Betreiber von Social Networks, Einrichtungen der Forschungsförderung und nicht zuletzt Mediennutzende. Dieser Band endet nicht bei der Analyse, sondern zeigt auf, wie die Verbreitung von Desinformationen über das Internet wirkungsvoll eingedämmt werden kann." (Verlagsbeschreibung)
more
"This article explores the challenges and opportunities presented by advances in artificial intelligence (AI) in the context of information operations. The article first examines the ways in which AI can be used to counter disinformation online. It then dives into some of the limitations of AI solut
...
ions and threats associated with AI techniques, namely user profiling, micro-targeting, and deep fakes. Finally, the paper reviews a number of solutions that could help address the spread of AI-powered disinformation and improve the online environment. The article recognises that in the fight against disinformation, there is no single fix. The next wave of disinformation calls first and foremost for societal resilience." (Abstract)
more
"Our research revealed that the deepfake phenomenon is growing rapidly online, with the number of deepfake videos almost doubling over the last seven months to 14,678. This increase is supported by the growing commodification of tools and services that lower the barrier for non-experts to create dee
...
pfakes. Perhaps unsurprisingly, we observed a significant contribution to the creation and use of synthetic media tools from web users in China and South Korea, despite the totality of our sources coming from the English-speaking Internet. Another key trend we identified is the prominence of non-consensual deepfake pornography, which accounted for 96% of the total deepfake videos online. We also found that the top four websites dedicated to deepfake pornography received more than 134 million views on videos targeting hundreds of female celebrities worldwide. This significant viewership demonstrates a market for websites creating and hosting deepfake pornography, a trend that will continue to grow unless decisive action is taken. Deepfakes are also making a significant impact on the political sphere. Two landmark cases from Gabon and Malaysia that received minimal Western media coverage saw deepfakes linked to an alleged government cover-up and a political smear campaign. One of these cases was related to an attempted military coup, while the other continues to threaten a highprofile politician with imprisonment. Seen together, these examples are possibly the most powerful indications of how deepfakes are already destabilizing political processes. Without defensive countermeasures, the integrity of democracies around the world are at risk."
more