Main Topic: Increasing use of AI in manipulative information campaigns online.
Key Points:
1. Mandiant has observed the use of AI-generated content in politically-motivated online influence campaigns since 2019.
2. Generative AI models make it easier to create convincing fake videos, images, text, and code, posing a threat.
3. While the impact of these campaigns has been limited so far, AI's role in digital intrusions is expected to grow in the future.
### Summary
The rise of generative artificial intelligence (AI) is making it difficult for the public to differentiate between real and fake content, raising concerns about deceptive fake political content in the upcoming 2024 presidential race. However, the Content Authenticity Initiative is working on a digital standard to restore trust in online content.
### Facts
- Generative AI is capable of producing hyper-realistic fake content, including text, images, audio, and video.
- Tools using AI have been used to create deceptive political content, such as images of President Joe Biden in a Republican Party ad and a fabricated voice of former President Donald Trump endorsing Florida Gov. Ron DeSantis' White House bid.
- The Content Authenticity Initiative, a coalition of companies, is developing a digital standard to restore trust in online content.
- Truepic, a company involved in the initiative, uses camera technology to add verified content provenance information to images, helping to verify their authenticity.
- The initiative aims to display "content credentials" that provide information about the history of a piece of content, including how it was captured and edited.
- The hope is for widespread adoption of the standard by creators to differentiate authentic content from manipulated content.
- Adobe is having conversations with social media platforms to implement the new content credentials, but no platforms have joined the initiative yet.
- Experts are concerned that generative AI could further erode trust in information ecosystems and potentially impact democratic processes, highlighting the importance of industry-wide change.
- Regulators and lawmakers are engaging in conversations and discussions about addressing the challenges posed by AI-generated fake content.
AI Algorithms Battle Russian Disinformation Campaigns on Social Media
A mysterious individual known as Nea Paw has developed an AI-powered project called CounterCloud to combat mass-produced AI disinformation. In response to tweets from Russian media outlets and the Chinese embassy that criticized the US, CounterCloud produced tweets, articles, and even journalists and news sites that were entirely generated by AI algorithms. Paw believes that the project highlights the danger of easily accessible generative AI tools being used for state-backed propaganda. While some argue that educating users about manipulative AI-generated content or equipping browsers with AI-detection tools could mitigate the issue, Paw believes that these solutions are not effective or elegant. Disinformation researchers have long warned about the potential of AI language models being used for personalized propaganda campaigns and influencing social media users. Evidence of AI-powered disinformation campaigns has already emerged, with academic researchers uncovering a botnet powered by AI language model ChatGPT. Legitimate political campaigns, such as the Republican National Committee, have also utilized AI-generated content, including fake images. AI-generated text can still be fairly generic, but with human finesse, it becomes highly effective and difficult to detect using automated filters. OpenAI has expressed concern about its technology being utilized to create tailored automated disinformation at a large scale, and while it has updated its policies to restrict political usage, it remains a challenge to block the generation of such material effectively. As AI tools become increasingly accessible, society must become aware of their presence in politics and protect against their misuse.
AI on social media platforms, both as a tool for manipulation and for detection, is seen as a potential threat to voter sentiment in the upcoming US presidential elections, with China-affiliated actors leveraging AI-generated visual media to emphasize politically divisive topics, while companies like Accrete AI are employing AI to detect and predict disinformation threats in real-time.
Google has announced that starting in November, political advertisements featuring synthetic content, such as images generated by artificial intelligence, must disclose this prominently, as concerns rise about the potential for AI-generated election misinformation that social media platforms may struggle to handle.
AI-generated content is becoming increasingly prevalent in political campaigns and poses a significant threat to democratic processes as it can be used to spread misinformation and disinformation to manipulate voters.
AI-generated images have the potential to create alternative history and misinformation, raising concerns about their impact on elections and people's ability to discern truth from manipulated visuals.
An image of Senator Cory Booker wearing pink shorts at the Capitol is not real, as it was generated with AI and shared as satire in response to the Senate's dress code changes.
AI-altered images of celebrities are being used to promote products without their consent, raising concerns about the misuse of artificial intelligence and the need for regulations to protect individuals from unauthorized AI-generated content.
The rise of false and misleading information on social media, exacerbated by advances in artificial intelligence, has created an authenticity crisis that is eroding trust in traditional news outlets and deepening social and political divisions.