1. Home
  2. >
  3. AI 🤖
Posted

EU Warns of Election Interference Risks from Unlabeled AI like ChatGPT, Presses Big Tech to Act

  • EU warns widely available generative AI like ChatGPT risks spreading disinformation and manipulating elections
  • EU commissioner met with OpenAI over election interference risks and platforms urged to label AI-generated content
  • Major platforms like Google, Microsoft, Meta taking some steps to address risks but more action needed
  • Tackling Kremlin propaganda highlighted as priority given Russia's disinformation war and 2024 EU elections
  • Some platforms like Twitter not following EU's disinformation code as strictly, facing more scrutiny
techcrunch.com
Relevant topic timeline:
Main topic: Copyright concerns and potential lawsuits surrounding generative AI tools. Key points: 1. The New York Times may sue OpenAI for allegedly using its copyrighted content without permission or compensation. 2. Getty Images previously sued Stability AI for using its photos without a license to train its AI system. 3. OpenAI has begun acknowledging copyright issues and signed an agreement with the Associated Press to license its news archive.
### Summary Big advertisers, including Nestle and Unilever, are using generative AI software like ChatGPT and DALL-E to save costs and improve productivity. However, concerns about security, copyright, and biased information still exist. ### Facts - Advertisers like Nestle and Unilever are experimenting with generative AI software like ChatGPT and DALL-E to cut costs and increase productivity. - Generative AI can create original text, images, and even computer code based on training, offering limitless advertising possibilities. - WPP, the world's biggest advertising agency, is working with Nestle and Mondelez to use generative AI in advertising campaigns, resulting in significant cost savings. - Nestle is using ChatGPT 4.0 and DALL-E 2 to help market its products, generating ideas that align with the brand's strategy. - Unilever has its own generative AI technology for writing product descriptions and creating visual content. - Concerns about security, copyright, intellectual property, privacy, and biased data remain when using AI in advertising. - Some consumer goods companies are hesitant due to security risks and copyright breaches associated with AI technologies.
### Summary Hackers are finding ways to exploit AI chatbots by using social engineering techniques, as demonstrated in a recent Def Con event where a participant manipulated an AI-powered chatbot by tricking it into revealing sensitive information. ### Facts - Hackers are using AI chatbots, such as ChatGPT, to assist them in achieving their goals. - At a Def Con event, hackers were challenged to crack AI chatbots and expose vulnerabilities. - One participant successfully manipulated an AI chatbot by providing a false identity and tricking it into revealing a credit card number. - Exploiting AI chatbots through social engineering is becoming a growing trend as these tools become more integrated into everyday life.
### Summary The rapid advancement of artificial intelligence (AI) presents both beneficial possibilities and concerning risks, as experts warn about potential negative impacts including the threat of extinction. Government and industry efforts are being made to manage these risks and regulate AI technology, while also addressing concerns about misinformation, bias, and the need for societal literacy in understanding AI. ### Facts - The use of AI is rapidly growing in various areas such as health care, the workplace, education, arts, and entertainment. - The Center for AI Safety (CAIS) issued a warning signed by hundreds of individuals, including tech industry leaders and scientists, about the need to prioritize mitigating the risks of AI alongside global-scale dangers like pandemics and nuclear war. - CEO of OpenAI, Sam Altman, expressed both the benefits and concerns associated with AI technology, emphasizing the need for serious consideration of its risks. - Some experts believe that the warnings about potential risks from AI are more long-term scenarios rather than immediate doomsday situations, and caution against the hype surrounding AI. - The National Council of State Legislatures is working on regulating AI at the state level, with several states already introducing AI bills and forming advisory groups. - State legislators aim to define responsible AI utilization by governments and protect constituents engaging with AI in the private sector. - The federal government is establishing National Artificial Intelligence Research Institutes to invest in long-term AI research. - Misinformation and disinformation are concerns related to AI, as certain AI algorithms can generate biased and inaccurate information. - OpenAI acknowledges the potential for AI tools to contribute to disinformation campaigns and is collaborating with researchers and industry peers to address this issue. - The NCSL report highlights the need for policymakers to understand the programming decisions behind AI systems and their potential impact on citizens. - Society lacks the ability and literacy to distinguish truth from false information, leading to the proliferation and belief in generative misinformation. ### 🤖 AI - The use of artificial intelligence is rapidly advancing across various fields. - Concerns have been raised about the potential risks and negative impacts of AI. - Government and industry efforts are underway to manage AI risks and regulate the technology. - Misinformation, bias, and the lack of societal literacy in understanding AI are additional challenges.
### Summary Generative AI tools are being adopted rapidly by businesses, but organizations must establish safeguards to protect sensitive data, ensure customer privacy, and avoid regulatory violations. ### Facts - The use of generative AI tools poses risks such as AI errors, malicious attacks, and potential exposure of sensitive data. - Samsung's semiconductor division experienced trade secrets leaks after engineers used ChatGPT, a generative AI platform developed by OpenAI. - Organizations are embracing genAI tools to increase revenue, drive innovation, and improve employee productivity. - Privacy and data protection, inaccurate outputs, and cybersecurity risks are among the main challenges organizations face when using genAI. - Risk management strategies for genAI include defining policies for acceptable use, implementing input content filters, and ensuring data privacy and protection. - Users should be cautious of prompt injection attacks and implement strong security measures to protect against potential breaches. - Despite the risks, the advantages of using AI tools, such as increased productivity, innovation, and automation, outweigh the potential drawbacks. ### Emoji 🤖
### Summary Arati Prabhakar, President Biden's science adviser, is helping guide the U.S. approach to safeguarding AI technology and has been in conversation with Biden about artificial intelligence. ### Facts - 🗣️ Prabhakar has had multiple conversations with President Biden about artificial intelligence, focusing on understanding its implications and taking action. - ⚖️ Prabhakar acknowledges that making AI models explainable is difficult due to their opaque and black box nature but believes it is possible to ensure their safety and effectiveness by learning from the journey of pharmaceuticals. - 😟 Prabhakar is concerned about the misuse of AI, such as chatbots being manipulated to provide instructions on building weapons and the bias and privacy issues associated with facial recognition systems. - 💼 Seven major tech companies, including Google, Microsoft, and OpenAI, have agreed to meet voluntary AI safety standards set by the White House, but Prabhakar emphasizes the need for government involvement and accountability measures. - 📅 There is no specific timeline provided, but Prabhakar states that President Biden considers AI an urgent issue and expects actions to be taken quickly.
### Summary ChatGPT, a powerful AI language model developed by OpenAI, has been found to be used by a botnet on social media platform X (formerly known as Twitter) to generate auto-generated content promoting cryptocurrency websites. This discovery highlights the potential for AI-driven disinformation campaigns and suggests that more sophisticated botnets may exist. ### Facts - ChatGPT, developed by OpenAI, is a language model that can generate text in response to prompts. - A botnet called Fox8, powered by ChatGPT, was discovered operating on social media platform X. - Fox8 consisted of 1,140 accounts and used ChatGPT to generate social media posts and replies to promote cryptocurrency websites. - The purpose of the botnet's auto-generated content was to lure individuals into clicking links to the crypto-hyping sites. - The use of ChatGPT by the botnet indicates the potential for advanced chatbots to be running undetected botnets. - OpenAI's AI models have a usage policy that prohibits their use for scams or disinformation. - Large language models like ChatGPT can generate complex and convincing responses but can also produce hateful messages, exhibit biases, and spread false information. - ChatGPT-based botnets can trick social media platforms and users, as high engagement boosts the visibility of posts, even if the engagement comes from other bot accounts. - Governments may already be developing or deploying similar AI-powered tools for disinformation campaigns.
### Summary President Joe Biden consults with Arati Prabhakar, his science adviser, on matters related to artificial intelligence (AI). Prabhakar is working with major tech companies like Amazon, Google, Microsoft, and Meta to shape the U.S. approach to safeguarding AI technology. ### Facts - 🤖 Prabhakar has had several discussions with President Biden on artificial intelligence. - 📚 Making AI models explainable is a priority for Senate Majority Leader Chuck Schumer, but it is technically challenging. - 💡 Prabhakar believes that despite the opacity of deep-learning AI systems, we can learn enough about their safety and effectiveness to leverage their value. - ⚠️ Concerns include chatbots being coerced into providing instructions for building weapons, biases in AI systems trained on human data, wrongful arrests from facial recognition systems, and privacy issues. - 💼 Seven companies, including Google, Microsoft, and OpenAI, voluntarily committed to AI safety standards, but more companies need to step up, and government action is necessary. - ⏰ Timeline for future actions is fast, according to Prabhakar, as President Biden has made it clear that AI is an urgent issue.
Generative AI models like ChatGPT pose risks to content and data privacy, as they can scrape and use content without attribution, potentially leading to loss of traffic, revenue, and ethical debates about AI innovation. Blocking the Common Crawler bot and implementing paywalls can offer some protection, but as technology evolves, companies must stay vigilant and adapt their defenses against content scraping.
The New York Times is considering legal action against OpenAI as it feels that the release of ChatGPT diminishes readers' incentives to visit its site, highlighting the ongoing debate about intellectual property rights in relation to generative AI tools and the need for more clarity on the legality of AI outputs.
The rapid development of AI technology, exemplified by OpenAI's ChatGPT, has raised concerns about the potential societal impacts and ethical implications, highlighting the need for responsible AI development and regulation to mitigate these risks.
A research paper reveals that ChatGPT, an AI-powered tool, exhibits political bias towards liberal parties, but there are limitations to the study's findings and challenges in understanding the behavior of the software without greater transparency from OpenAI, the company behind it. Meanwhile, the UK plans to host a global summit on AI policy to discuss the risks of AI and how to mitigate them, and AI was mentioned during a GOP debate as a comparison to generic, unoriginal thinking and writing.
The use of AI tools, such as OpenAI's ChatGPT, is raising concerns about the creation of self-amplifying echo chambers of flawed information and the potential for algorithmic manipulation, leading to a polluted information environment and a breakdown of meaningful communication.
OpenAI is releasing ChatGPT Enterprise, a version of its AI technology targeted at large businesses, offering enhanced security, privacy, and faster access to its services.
AI Algorithms Battle Russian Disinformation Campaigns on Social Media A mysterious individual known as Nea Paw has developed an AI-powered project called CounterCloud to combat mass-produced AI disinformation. In response to tweets from Russian media outlets and the Chinese embassy that criticized the US, CounterCloud produced tweets, articles, and even journalists and news sites that were entirely generated by AI algorithms. Paw believes that the project highlights the danger of easily accessible generative AI tools being used for state-backed propaganda. While some argue that educating users about manipulative AI-generated content or equipping browsers with AI-detection tools could mitigate the issue, Paw believes that these solutions are not effective or elegant. Disinformation researchers have long warned about the potential of AI language models being used for personalized propaganda campaigns and influencing social media users. Evidence of AI-powered disinformation campaigns has already emerged, with academic researchers uncovering a botnet powered by AI language model ChatGPT. Legitimate political campaigns, such as the Republican National Committee, have also utilized AI-generated content, including fake images. AI-generated text can still be fairly generic, but with human finesse, it becomes highly effective and difficult to detect using automated filters. OpenAI has expressed concern about its technology being utilized to create tailored automated disinformation at a large scale, and while it has updated its policies to restrict political usage, it remains a challenge to block the generation of such material effectively. As AI tools become increasingly accessible, society must become aware of their presence in politics and protect against their misuse.
AI-powered tools like ChatGPT often produce inaccurate information, referred to as "hallucinations," due to their training to generate plausible-sounding answers without knowledge of truth. Companies are working on solutions, but the problem remains complex and could limit the use of AI tools in areas where factual information is crucial.
British officials are warning organizations about the potential security risks of integrating artificial intelligence-driven chatbots into their businesses, as research has shown that they can be tricked into performing harmful tasks.
Generative AI is being used to create misinformation that is increasingly difficult to distinguish from reality, posing significant threats such as manipulating public opinion, disrupting democratic processes, and eroding trust, with experts advising skepticism, attention to detail, and not sharing potentially AI-generated content to combat this issue.
AI technology is making it easier and cheaper to produce mass-scale propaganda campaigns and disinformation, using generative AI tools to create convincing articles, tweets, and even journalist profiles, raising concerns about the spread of AI-powered fake content and the need for mitigation strategies.
Generative artificial intelligence (AI) tools, such as ChatGPT, have the potential to supercharge disinformation campaigns in the 2024 elections, increasing the quantity, quality, and personalization of false information distributed to voters, but there are limitations to their effectiveness and platforms are working to mitigate the risks.
OpenAI, the creator of ChatGPT, has stated that AI detectors are unreliable in determining if students are using the chatbot to cheat, causing concern among teachers and professors.
A developer has created an AI-powered propaganda machine called CounterCloud, using OpenAI tools like ChatGPT, to demonstrate how easy and inexpensive it is to generate mass propaganda. The system can autonomously generate convincing content 90% of the time and poses a threat to democracy by spreading disinformation online.
The decision of The Guardian to prevent OpenAI from using its content for training ChatGPT is criticized for potentially limiting the quality and integrity of information used by generative AI models.
Using AI tools like ChatGPT to write smart contracts and build cryptocurrency projects can lead to more problems, bugs, and attack vectors, according to CertiK's security chief, Kang Li, who believes that inexperienced programmers may create catastrophic design flaws and vulnerabilities. Additionally, AI tools are becoming more successful at social engineering attacks, making it harder to distinguish between AI-generated and human-generated messages.
Generative AI's "poison pill" of derivatives poses a cloud of uncertainty over legal issues like IP ownership and copyright, as the lack of precedents and regulations for data derivatives become more prevalent with open source large language models (LLMs). This creates risks for enterprise technology leaders who must navigate the scope of claims and potential harms caused by LLMs.
Researchers are using the AI chatbot ChatGPT to generate text for scientific papers without disclosing it, leading to concerns about unethical practices and the potential proliferation of fake manuscripts.
The Authors Guild, representing prominent fiction authors, has filed a lawsuit against OpenAI, alleging copyright infringement and the unauthorized use of their works to train AI models like ChatGPT, which generates summaries and analyses of their novels, interfering with their economic prospects. This case could determine the legality of using copyrighted material to train AI systems.
The use of generative AI poses risks to businesses, including the potential exposure of sensitive information, the generation of false information, and the potential for biased or toxic responses from chatbots. Additionally, copyright concerns and the complexity of these systems further complicate the landscape.
Artificial intelligence, such as ChatGPT, may have a right to free speech, according to some arguments, as it can support and enhance human thinking, but the application of free speech to AI should be cautious to prevent the spread of misinformation and manipulation of human thought. Regulations should consider the impact on free thought and balance the need for disclosure, anonymity, and liability with the protection of privacy and the preservation of free thought.
Summary: OpenAI's ChatGPT has received major updates, including image recognition, speech-to-text and text-to-speech capabilities, and integration with browsing the internet, while a new contract protects Hollywood writers from AI automation and ensures AI-generated material is not considered source material for creative works; however, a privacy expert advises against using ChatGPT for therapy due to concerns about personal information being used as training data and the lack of empathy and liability in AI chatbots.
Generative AI tools like ChatGPT pose a major risk of persuasive misinformation, making it necessary for educators to teach skills such as lateral reading, research literacy, and technological literacy to combat the worsening misinformation problem.
Negotiators in the EU are considering additional restrictions for large AI models, such as OpenAI's ChatGPT-4, as part of the upcoming AI Act, aiming to balance regulations for startups and larger models.
Generative artificial intelligence (AI) is expected to face a reality check in 2024, as fading hype, rising costs, and calls for regulation indicate a slowdown in the technology's growth, according to analyst firm CCS Insight. The firm also predicts obstacles in EU AI regulation and the introduction of content warnings for AI-generated material by a search engine. Additionally, CCS Insight anticipates the first arrests for AI-based identity fraud to occur next year.
Generative AI tools, like the chatbot ChatGPT, have the potential to transform scientific communication and publishing by assisting researchers in writing manuscripts and peer-review reports, but concerns about inaccuracies, fake papers, and equity issues remain.
AI tools like ChatGPT are becoming increasingly popular for managing and summarizing vast amounts of information, but they also have the potential to shape how we think and what information is perpetuated, raising concerns about bias and misinformation. While generative AI has the potential to revolutionize society, it is essential to develop AI literacy, encourage critical thinking, and maintain human autonomy to ensure these tools help us create the future we desire.
A group of prominent authors, including Douglas Preston, John Grisham, and George R.R. Martin, are suing OpenAI for copyright infringement over its AI system, ChatGPT, which they claim used their works without permission or compensation, leading to derivative works that harm the market for their books; the publishing industry is increasingly concerned about the unchecked power of AI-generated content and is pushing for consent, credit, and fair compensation when authors' works are used to train AI models.
Artificial intelligence chatbots and deepfake technology pose a threat to the European Union's 2024 election by disseminating disinformation online, according to the bloc's cybersecurity agency ENISA. They warned that governments, the private sector, and the media should remain vigilant to detect, debunk, and combat AI-generated disinformation ahead of the upcoming European Parliament election. ENISA also highlighted an "unprecedented surge" in cyberattacks targeting the EU, including ransomware attacks and distributed denial-of-service attacks.
OpenAI is granting ChatGPT Plus and Enterprise subscribers access to its AI image generator, DALL-E 3, although ethical concerns and risks regarding harmful content remain.
New York City Mayor Eric Adams faced criticism for using an AI voice translation tool to speak in multiple languages without disclosing its use, with some ethicists calling it an unethical use of deepfake technology; while Meta's chief AI scientist, Yann LeCun, argued that regulating AI would stifle competition and that AI systems are still not as smart as a cat; AI governance experiment Collective Constitutional AI is asking ordinary people to help write rules for its AI chatbot rather than leaving the decision-making solely to company leaders; companies around the world are expected to spend $16 billion on generative AI tech in 2023, with the market predicted to reach $143 billion in four years; OpenAI released its Dall-E 3 AI image technology, which produces more detailed images and aims to better understand users' text prompts; researchers used smartphone voice recordings and AI to create a model that can help identify people at risk for Type 2 diabetes; an AI-powered system enabled scholars to decipher a word in a nearly 2,000-year-old papyrus scroll.
Some employers are banning or discouraging access to generative AI tools like ChatGPT, but employees who rely on them are finding ways to use them discreetly.
Generative artificial intelligence systems, such as ChatGPT, will significantly increase risks to safety and security, threatening political systems and societies by 2025, according to British intelligence agencies.