1. Home
  2. >
  3. AI 🤖
Posted

US Attorneys General call for national law to combat AI CSAM

The National Association of Attorneys General has called on Congress to pass legislation that prohibits the use of AI to generate child sex abuse images, citing the need to protect children from exploitation and the potential harm caused by AI-generated content.

theregister.com
Relevant topic timeline:
AI-generated child pornography: A controversial solution or a Pandora's Box? The emergence of generative AI models that can produce realistic fake images of child sexual abuse has sparked concern and debate among regulators and child safety advocates. On one hand, there is fear that this technology may exacerbate an already abhorrent practice. On the other hand, some experts argue that AI-generated child pornography could offer a less harmful alternative to the existing market for such explicit content. They believe that pedophilia is rooted in biology and that finding a way to redirect pedophilic urges without involving real children could be beneficial. While psychiatrists strive for a cure, utilizing AI-generated imagery as a temporary solution to replace the demand for real child pornography may have its merits. Currently, law enforcement comb through countless images in their efforts to identify victims, and the introduction of AI-generated images further complicates their task. Additionally, these images often exploit the likenesses of real people, further perpetuating abuse of a different nature. However, AI technology could also play a role in helping distinguish between real and simulated content, aiding law enforcement in targeting actual cases of child sexual abuse. There are differing opinions on whether satisfying pedophilic urges through AI-generated child pornography can actually prevent harm in the long run. Some argue that exposure to such content might reinforce and legitimize these attractions, potentially leading to more severe offenses. Others suggest that AI-generated images could serve as an outlet for pedophiles who do not wish to harm children, allowing them to find sexual catharsis without real-world implications. By providing a controlled environment for these individuals, AI-generated images could potentially help curb their behavior and encourage them to seek therapeutic treatment. Concerns about the normalization of child pornography and the potential gateway effect are addressed by experts. They argue that individuals without pedophilic tendencies are unlikely to be enticed by AI-generated child pornography, and the scientific research indicates that the act of viewing alone does not necessarily lead to hands-on offenses. Moreover, redirecting potential viewers to AI-generated images could reduce the circulation of real images, offering some protection to victims. While the idea of utilizing AI-generated child pornography as a form of harm reduction may be difficult to accept, it parallels the philosophy behind other public health policies aimed at minimizing damage. However, it is crucial to differentiate between controlled psychiatric settings and uncontrolled proliferation on the web. Integrating AI-generated images into therapy and treatment plans, tailored to each individual's needs, could offer a way to diminish risks and prioritize the safety of both victims and potential offenders.
AI-generated inventions need to be allowed patent protection to encourage innovation and maximize social benefits, as current laws hinder progress in biomedicine; jurisdictions around the world have differing approaches to patenting AI-generated inventions, and the US falls behind in this area, highlighting the need for legislative action.
The Alliance of Motion Picture and Television Producers has proposed guidelines for the usage of artificial intelligence (AI) and data transparency in the entertainment industry, stating that AI-created material cannot be considered literary or intellectually protected, and ensuring that credit, rights, and compensation for AI-generated scripts are given to the original human writer or reworker.
The U.S. is falling behind in regulating artificial intelligence (AI), while Europe has passed the world's first comprehensive AI law; President Joe Biden recently met with industry leaders to discuss the need for AI regulation and companies pledged to develop safeguards for AI-generated content and prioritize user privacy.
Major media organizations are calling for new laws to protect their content from being used by AI tools without permission, expressing concerns over unauthorized scraping and the potential for AI to produce false or biased information.
Salesforce has released an AI Acceptable Use Policy that outlines the restrictions on the use of its generative AI products, including prohibiting their use for weapons development, adult content, profiling based on protected characteristics, medical or legal advice, and more. The policy emphasizes the need for responsible innovation and sets clear ethical guidelines for the use of AI.
A federal judge in the US rejected an attempt to copyright an artwork created by an AI, ruling that copyright law only protects works of human creation. However, the judge also acknowledged that as AI becomes more involved in the creation process, challenging questions about human input and authorship will arise.
A Washington D.C. judge has ruled that AI-generated art should not be awarded copyright protections since no humans played a central role in its creation, establishing a precedent that art should require human authorship; YouTube has partnered with Universal Music Group to launch an AI music incubator to protect artists from unauthorized use of their content; Meta has introduced an automated translator that works for multiple languages, but concerns have been raised regarding the impact it may have on individuals who wish to learn multiple languages; major studios are hiring "AI specialists" amidst a writers' strike, potentially leading to a future of automated entertainment that may not meet audience expectations.
Generative AI has revolutionized various sectors by producing novel content, but it also raises concerns around biases, intellectual property rights, and security risks. Debates on copyrightability and ownership of AI-generated content need to be resolved, and existing laws should be modified to address the risks associated with generative AI.
Artificial intelligence (AI) tools can put human rights at risk, as highlighted by researchers from Amnesty International on the Me, Myself, and AI podcast, who discuss scenarios in which AI is used to track activists and make automated decisions that can lead to discrimination and inequality, emphasizing the need for human intervention and changes in public policy to address these issues.
AI-generated videos are targeting children online, raising concerns about their safety, while there are also worries about AI causing job losses and becoming oppressive bosses; however, AI has the potential to protect critical infrastructure and extend human life.
The United States Copyright Office has issued a notice of inquiry seeking public comment on copyright and artificial intelligence (AI), specifically on issues related to the content AI produces and how it should be treated when it imitates or mimics human artists.
The use of AI in the entertainment industry, such as body scans and generative AI systems, raises concerns about workers' rights, intellectual property, and the potential for broader use of AI in other industries, infringing on human connection and privacy.
The top prosecutors in all 50 states are urging Congress to establish an expert commission to study and legislate against the use of artificial intelligence to exploit children through pornography.
The infiltration of artificial intelligence into children's lives is causing anxiety and sparking fears about the perversion of children's culture, as AI tools create unsettling and twisted representations of childhood innocence. This trend continues a long history of cultural anxieties about dangerous interactions between children and technology, with films like M3GAN and Frankenstein depicting the dangers of AI. While there is a need to address children's use and understanding of AI, it is important not to succumb to moral panics and instead focus on promoting responsible AI use and protecting children's rights.
Attorneys general from all 50 states and four territories are urging Congress to establish an expert commission to study the potential exploitation of children through generative AI and to expand laws against child sexual abuse material (CSAM) to cover AI-generated materials.
Australia's eSafety Commissioner has introduced an industry code that requires tech giants like Google and Microsoft to eliminate child abuse material from their search results and prevent generative AI from producing deepfake versions of such material.
Australia's internet regulator has drafted a new code that requires search engines like Google and Bing to prevent the sharing of child sexual abuse material created by artificial intelligence, and also prohibits the AI functions of search engines from producing deepfake content.
State attorneys general, including Oklahoma's Attorney General Gentner Drummond, are urging Congress to address the consequences of artificial intelligence on child pornography, expressing concern that AI-powered tools are making prosecution more challenging and creating new opportunities for abuse.
Congressman Clay Higgins (R-LA) plans to introduce legislation prohibiting the use of artificial intelligence (AI) by the federal government for law enforcement purposes, in response to the Internal Revenue Service's recently announced AI-driven tax enforcement initiative.
Paedophiles are using open source AI models to create child sexual abuse material, according to the Internet Watch Foundation, raising concerns about the potential for realistic and widespread illegal content.
Adobe has joined other companies in committing to safe AI development and has proposed a federal anti-impersonation law that would allow creators to seek damages from individuals using AI to impersonate them or their style for commercial purposes, which would make the impersonator, not the tool's vendor, the target of legal action.
A surge in AI-generated child sexual abuse material (CSAM) circulating online has been observed by the Internet Watch Foundation (IWF), raising concerns about the ability to identify and protect real children in need. Efforts are being made by law enforcement and policymakers to address the growing issue of deepfake content created using generative AI platforms, including the introduction of legislation in the US to prevent the use of deceptive AI in elections.
The Department of Homeland Security (DHS) has released new guidelines for the use of artificial intelligence (AI), including a policy that prohibits the collection and dissemination of data used in AI activities and a requirement for thorough testing of facial recognition technologies to ensure there is no unintended bias.
California lawmakers have introduced a bill that would protect actors and artists from being replaced by digital clones created using artificial intelligence technology in response to concerns about the potential impact on entertainment jobs.
The Subcommittee on Cybersecurity, Information Technology, and Government Innovation discussed the federal government's use of artificial intelligence (AI) and emphasized the need for responsible governance, oversight, and accountability to mitigate risks and protect civil liberties and privacy rights.
Adversaries and criminal groups are exploiting artificial intelligence (AI) technology to carry out malicious activities, according to FBI Director Christopher Wray, who warned that while AI can automate tasks for law-abiding citizens, it also enables the creation of deepfakes and malicious code, posing a threat to US citizens. The FBI is working to identify and track those misusing AI, but is cautious about using it themselves. Other US security agencies, however, are already utilizing AI to combat various threats, while concerns about China's use of AI for misinformation and propaganda are growing.
The US Copyright Office has ruled for the third time that AI-generated art cannot be copyrighted, raising questions about whether AI-generated art is categorically excluded from copyright protection or if human creators should be listed as the image's creator. The office's position, which is based on existing copyright doctrine, has been criticized for being unscalable and a potential quagmire, as it fails to consider the creative choices made by AI systems similar to those made by human photographers.