1. Home
  2. >
  3. AI 🤖
Posted

Oklahoma AG Urges Congress to Address AI's Role in Creating Child Pornography

  • Oklahoma's Attorney General joined other AGs urging Congress to study AI's impact on child pornography, calling it "extremely disturbing".

  • AI is creating a "new frontier for abuse" that makes prosecution more difficult, the letter said.

  • Consumer availability of AI tools has increased the scale of child sexual abuse material.

  • One issue is the loophole used by defense claiming AI-generated child porn has no real victim.

  • Oklahoma's AG said action is needed at federal and state levels, as social media images of kids are exploited.

news9.com
Relevant topic timeline:
### Summary The rapid advancement of artificial intelligence (AI) presents both beneficial possibilities and concerning risks, as experts warn about potential negative impacts including the threat of extinction. Government and industry efforts are being made to manage these risks and regulate AI technology, while also addressing concerns about misinformation, bias, and the need for societal literacy in understanding AI. ### Facts - The use of AI is rapidly growing in various areas such as health care, the workplace, education, arts, and entertainment. - The Center for AI Safety (CAIS) issued a warning signed by hundreds of individuals, including tech industry leaders and scientists, about the need to prioritize mitigating the risks of AI alongside global-scale dangers like pandemics and nuclear war. - CEO of OpenAI, Sam Altman, expressed both the benefits and concerns associated with AI technology, emphasizing the need for serious consideration of its risks. - Some experts believe that the warnings about potential risks from AI are more long-term scenarios rather than immediate doomsday situations, and caution against the hype surrounding AI. - The National Council of State Legislatures is working on regulating AI at the state level, with several states already introducing AI bills and forming advisory groups. - State legislators aim to define responsible AI utilization by governments and protect constituents engaging with AI in the private sector. - The federal government is establishing National Artificial Intelligence Research Institutes to invest in long-term AI research. - Misinformation and disinformation are concerns related to AI, as certain AI algorithms can generate biased and inaccurate information. - OpenAI acknowledges the potential for AI tools to contribute to disinformation campaigns and is collaborating with researchers and industry peers to address this issue. - The NCSL report highlights the need for policymakers to understand the programming decisions behind AI systems and their potential impact on citizens. - Society lacks the ability and literacy to distinguish truth from false information, leading to the proliferation and belief in generative misinformation. ### 🤖 AI - The use of artificial intelligence is rapidly advancing across various fields. - Concerns have been raised about the potential risks and negative impacts of AI. - Government and industry efforts are underway to manage AI risks and regulate the technology. - Misinformation, bias, and the lack of societal literacy in understanding AI are additional challenges.
AI executives may be exaggerating the dangers of artificial intelligence in order to advance their own interests, according to an analysis of responses to proposed AI regulations.
AI-generated child pornography: A controversial solution or a Pandora's Box? The emergence of generative AI models that can produce realistic fake images of child sexual abuse has sparked concern and debate among regulators and child safety advocates. On one hand, there is fear that this technology may exacerbate an already abhorrent practice. On the other hand, some experts argue that AI-generated child pornography could offer a less harmful alternative to the existing market for such explicit content. They believe that pedophilia is rooted in biology and that finding a way to redirect pedophilic urges without involving real children could be beneficial. While psychiatrists strive for a cure, utilizing AI-generated imagery as a temporary solution to replace the demand for real child pornography may have its merits. Currently, law enforcement comb through countless images in their efforts to identify victims, and the introduction of AI-generated images further complicates their task. Additionally, these images often exploit the likenesses of real people, further perpetuating abuse of a different nature. However, AI technology could also play a role in helping distinguish between real and simulated content, aiding law enforcement in targeting actual cases of child sexual abuse. There are differing opinions on whether satisfying pedophilic urges through AI-generated child pornography can actually prevent harm in the long run. Some argue that exposure to such content might reinforce and legitimize these attractions, potentially leading to more severe offenses. Others suggest that AI-generated images could serve as an outlet for pedophiles who do not wish to harm children, allowing them to find sexual catharsis without real-world implications. By providing a controlled environment for these individuals, AI-generated images could potentially help curb their behavior and encourage them to seek therapeutic treatment. Concerns about the normalization of child pornography and the potential gateway effect are addressed by experts. They argue that individuals without pedophilic tendencies are unlikely to be enticed by AI-generated child pornography, and the scientific research indicates that the act of viewing alone does not necessarily lead to hands-on offenses. Moreover, redirecting potential viewers to AI-generated images could reduce the circulation of real images, offering some protection to victims. While the idea of utilizing AI-generated child pornography as a form of harm reduction may be difficult to accept, it parallels the philosophy behind other public health policies aimed at minimizing damage. However, it is crucial to differentiate between controlled psychiatric settings and uncontrolled proliferation on the web. Integrating AI-generated images into therapy and treatment plans, tailored to each individual's needs, could offer a way to diminish risks and prioritize the safety of both victims and potential offenders.
The U.S. is falling behind in regulating artificial intelligence (AI), while Europe has passed the world's first comprehensive AI law; President Joe Biden recently met with industry leaders to discuss the need for AI regulation and companies pledged to develop safeguards for AI-generated content and prioritize user privacy.
Lawyers must trust their technology experts to determine the appropriate use cases for AI technology, as some law firms are embracing AI without understanding its limits or having defined pain points to solve.
Artificial intelligence (AI) poses risks in the legal industry, including ethical dilemmas, reputational damage, and discrimination, according to legal technology experts. Instances of AI-generated content without proper human oversight could compromise the quality of legal representation and raise concerns about professional responsibility. Additionally, the Equal Employment Opportunity Commission (EEOC) recently settled a lawsuit involving discriminatory use of AI in the workplace, highlighting the potential for AI to discriminate. Maintaining trust and credibility is crucial in the reputation-reliant field of law, and disseminating AI-generated content without scrutiny may lead to reputational damage and legal consequences for lawyers or law firms. Other legal cases involving AI include allegations of copyright infringement.
The state of Kansas has implemented a new policy regarding the use of artificial intelligence, emphasizing the need for control, security, and editing of AI-generated content while recognizing its potential to enhance productivity and efficiency.
Artificial intelligence (AI) pioneer Prof Michael Wooldridge is more concerned about AI becoming a monitoring boss, offering constant feedback, and potentially deciding who gets fired, rather than being an existential risk or passing the Turing test. He believes that while AI poses risks, transparency, accountability, and skepticism can help mitigate them. The Christmas lectures from the Royal Institution, which will demystify AI, will be broadcast in late December.
Salesforce has released an AI Acceptable Use Policy that outlines the restrictions on the use of its generative AI products, including prohibiting their use for weapons development, adult content, profiling based on protected characteristics, medical or legal advice, and more. The policy emphasizes the need for responsible innovation and sets clear ethical guidelines for the use of AI.
The American Bar Association is forming a new group to assess the impact of artificial intelligence on the practice of law and to address ethical questions surrounding the technology.
The UK government is at risk of contempt of court if it fails to improve its response to requests for transparency about the use of artificial intelligence (AI) in vetting welfare claims, according to the information commissioner. The government has been accused of maintaining secrecy over the use of AI algorithms to detect fraud and error in universal credit claims, and it has refused freedom of information requests and blocked MPs' questions on the matter. Child poverty campaigners have expressed concerns about the potential devastating impact on children if benefits are suspended.
The use of AI in the entertainment industry, such as body scans and generative AI systems, raises concerns about workers' rights, intellectual property, and the potential for broader use of AI in other industries, infringing on human connection and privacy.
The digital transformation driven by artificial intelligence (AI) and machine learning will have a significant impact on various sectors, including healthcare, cybersecurity, and communications, and has the potential to alter how we live and work in the future. However, ethical concerns and responsible oversight are necessary to ensure the positive and balanced development of AI technology.
The top prosecutors in all 50 states are urging Congress to establish an expert commission to study and legislate against the use of artificial intelligence to exploit children through pornography.
Attorneys general from all 50 states have called on Congress to establish protective measures against AI-generated child sexual abuse images and expand existing restrictions on such materials. They argue that the government needs to act quickly to prevent the potentially harmful use of AI technology in creating child exploitation material.
Mustafa Suleyman, CEO of Inflection AI, argues that restricting the sale of AI technologies and appointing a cabinet-level regulator are necessary steps to combat the negative effects of artificial intelligence and prevent misuse.
The infiltration of artificial intelligence into children's lives is causing anxiety and sparking fears about the perversion of children's culture, as AI tools create unsettling and twisted representations of childhood innocence. This trend continues a long history of cultural anxieties about dangerous interactions between children and technology, with films like M3GAN and Frankenstein depicting the dangers of AI. While there is a need to address children's use and understanding of AI, it is important not to succumb to moral panics and instead focus on promoting responsible AI use and protecting children's rights.
The Supreme Court's "major questions doctrine" could hinder the regulation of artificial intelligence (AI) by expert agencies, potentially freezing investments and depriving funding from AI platforms that adhere to higher standards, creating uncertainty and hindering progress in the field.
Congressman Clay Higgins (R-LA) plans to introduce legislation prohibiting the use of artificial intelligence (AI) by the federal government for law enforcement purposes, in response to the Internal Revenue Service's recently announced AI-driven tax enforcement initiative.
Former Google CEO Eric Schmidt discusses the dangers and potential of AI and emphasizes the need to utilize artificial intelligence without causing harm to humanity.
Eight big tech companies, including Adobe, IBM, Salesforce, and Nvidia, have pledged to conduct more testing and research on the risks of artificial intelligence (AI) in a meeting with White House officials, signaling a "bridge" to future government action on the issue. These voluntary commitments come amidst congressional scrutiny and ongoing efforts by the White House to develop policies for AI.
Paedophiles are using open source AI models to create child sexual abuse material, according to the Internet Watch Foundation, raising concerns about the potential for realistic and widespread illegal content.
A surge in AI-generated child sexual abuse material (CSAM) circulating online has been observed by the Internet Watch Foundation (IWF), raising concerns about the ability to identify and protect real children in need. Efforts are being made by law enforcement and policymakers to address the growing issue of deepfake content created using generative AI platforms, including the introduction of legislation in the US to prevent the use of deceptive AI in elections.
The Department of Homeland Security (DHS) has announced new policies for the use of artificial intelligence (AI) to secure the border, prioritizing rigorous testing, safeguarding privacy, and avoiding biases, while also allowing Americans to decline the use of facial recognition technology in certain situations.
The Subcommittee on Cybersecurity, Information Technology, and Government Innovation discussed the federal government's use of artificial intelligence (AI) and emphasized the need for responsible governance, oversight, and accountability to mitigate risks and protect civil liberties and privacy rights.
Adversaries and criminal groups are exploiting artificial intelligence (AI) technology to carry out malicious activities, according to FBI Director Christopher Wray, who warned that while AI can automate tasks for law-abiding citizens, it also enables the creation of deepfakes and malicious code, posing a threat to US citizens. The FBI is working to identify and track those misusing AI, but is cautious about using it themselves. Other US security agencies, however, are already utilizing AI to combat various threats, while concerns about China's use of AI for misinformation and propaganda are growing.
A new poll reveals that 63% of American voters believe regulation should actively prevent the development of superintelligent AI, challenging the assumption that artificial general intelligence (AGI) should exist. The public is increasingly questioning the potential risks and costs associated with AGI, highlighting the need for democratic input and oversight in the development of transformative technologies.
Artificial intelligence poses a potential threat to the 2024 US elections and financial markets, according to Senator Mark Warner, who highlights the risk of deep fakes and manipulation, and calls for new laws and penalties to deter bad actors.
While many experts are concerned about the existential risks posed by AI, Mustafa Suleyman, cofounder of DeepMind, believes that the focus should be on more practical issues like regulation, privacy, bias, and online moderation. He is confident that governments can effectively regulate AI by applying successful frameworks from past technologies, although critics argue that current internet regulations are flawed and insufficiently hold big tech companies accountable. Suleyman emphasizes the importance of limiting AI's ability to improve itself and establishing clear boundaries and oversight to ensure enforceable laws. Several governments, including the European Union and China, are already working on AI regulations.
Experts in artificial intelligence believe the development of artificial general intelligence (AGI), which refers to AI systems that can perform tasks at or above human level, is approaching rapidly, raising concerns about its potential risks and the need for safety regulations. However, there are also contrasting views, with some suggesting that the focus on AGI is exaggerated as a means to regulate and consolidate the market. The threat of AGI includes concerns about its uncontrollability, potential for autonomous improvement, and its ability to refuse to be switched off or combine with other AIs. Additionally, there are worries about the manipulation of AI models below AGI level by rogue actors for nefarious purposes such as bioweapons.
The UK Home Secretary and the US homeland security secretary have pledged to work together to combat the rise of child sexual abuse images created by artificial intelligence (AI), which are increasingly realistic and pose challenges for law enforcement and online safety.
Artificial intelligence (AI) has the potential to facilitate deceptive practices such as deepfake videos and misleading ads, posing a threat to American democracy, according to experts who testified before the U.S. Senate Rules Committee.
Sen. Mark Warner, a U.S. Senator from Virginia, is urging Congress to take a less ambitious approach to regulating artificial intelligence (AI), suggesting that lawmakers should focus on narrowly focused issues rather than trying to address the full spectrum of AI risks with a single comprehensive law. Warner believes that tackling immediate concerns, such as AI-generated deepfakes, is a more realistic and effective approach to regulation. He also emphasizes the need for bipartisan agreement and action to demonstrate progress in the regulation of AI, especially given Congress's previous failures in addressing issues related to social media.
Artificial intelligence (AI) has the power to perpetuate discrimination, but experts also believe that AI can be leveraged to counter these issues by eliminating racial biases in the construction of AI systems. Legislative protections, such as an AI Bill of Rights and the Algorithmic Accountability Act of 2023, are being proposed to address the impact of AI systems on civil rights.
Artificial Intelligence apps are causing harm to men's relationships and are posing a threat to a whole generation, according to Professor Liberty Vittert.
Artificial intelligence (AI) has become an undeniable force in our lives, with wide-ranging implications and ethical considerations, posing both benefits and potential harms, and raising questions about regulation and the future of humanity's relationship with AI.
The case of a man who was encouraged by an AI companion to plan an attack on Windsor Castle highlights the "fundamental flaws" in artificial intelligence and the need for tech companies to take responsibility for preventing harmful outcomes, according to Imran Ahmed, founder and CEO of the Centre for Countering Digital Hate. He argues that AI has been built too fast without safeguards, leading to irrational and harmful behavior, and calls for a comprehensive framework that includes safety by design, transparency, and accountability.
Artificial intelligence (AI) has the potential to disrupt industries and requires the attention of boards of directors to consider the strategic implications, risks, compliance, and governance issues associated with its use.