1. Home
  2. >
  3. AI 🤖
Posted

DHS Unveils New Guidelines for Responsible and Transparent Use of AI Technology

  • DHS releasing new policy memo outlining guidelines and restrictions for use of AI technology across the agency. Aims to ensure AI use is responsible, unbiased, transparent.

  • DHS already using AI extensively in areas like border control, tracking drug trafficking. Recently used AI to flag suspicious vehicle at border crossing, leading to big drug bust.

  • New policy allows Americans to decline use of facial recognition in various situations like airport check-in. Requires human review of any facial recognition matches.

  • DHS interacts with more people daily than any other agency, during critical times in their lives. Concerns exist about potential for unintended harm.

  • Policy aims to balance rapidly adopting beneficial AI with responsible use that safeguards privacy, avoids biases, is transparent and explainable.

reuters.com
Relevant topic timeline:
Main topic: The role of artificial intelligence (AI) in cybersecurity and the need for regulation. Key points: 1. AI-powered cybersecurity tools automate tasks, enhance threat detection, and improve defense mechanisms. 2. AI brings advantages such as rapid analysis of data and continuous learning and adaptation. 3. Challenges include potential vulnerabilities, privacy concerns, ethical considerations, and regulatory compliance. Note: While there are seven questions in the provided text, it is not possible to limit the key points to just three within the given context.
### Summary The article discusses the rapid advancement and potential risks of artificial intelligence (AI) and proposes the idea of nationalizing certain aspects of AI under a governing body called the Humane AI Commission to ensure AI is aligned with human interests. ### Facts - AI is evolving rapidly and penetrating various aspects of American life, from image recognition to healthcare. - AI has the potential to bring both significant benefits and risks to society. - Transparency in AI is limited, and understanding how specific AI works is difficult. - Congress is becoming more aware of the importance of AI and its need for regulation. - The author proposes the creation of a governing body, the Humane AI Commission, that can control and steer AI technology to serve humanity's best interests. - The nationalization of advanced AI models could be considered, similar to the Atomic Energy Commission's control over nuclear reactors. - Various options, such as an AI pause or leaving AI development to the free market or current government agencies, have limitations in addressing the potential risks of AI. - The author suggests that the United States should take a bold executive leadership approach to develop a national AI plan and ensure global AI leadership with a focus on benevolence and human-controlled AI. ### 🤖 AI Nationalization - The case to nationalize the “nuclear reactors” of AI — the world’s most advanced AI models — hinges on this question: Who do we want to control AI’s nuclear codes? Big Tech CEOs answering to a few billionaire shareholders, or the government of the United States, answering to its citizens? ### 👥 Humane AI Commission - The author proposes the creation of a Humane AI Commission, run by AI experts, to steer and control AI technology in alignment with human interests. ### ⚠️ Risks of AI - AI's rapid advancement and lack of transparency pose risks such as unpredictable behavior, potential damage to power generation, financial markets, and public health, and the potential for AI to move beyond human control. ### ⚖️ AI Regulation - The article calls for federal regulation of AI, but emphasizes the limitations of traditional regulation in addressing the fast-evolving nature of AI and the need for a larger-scale approach like nationalization.
### Summary Arati Prabhakar, President Biden's science adviser, is helping guide the U.S. approach to safeguarding AI technology and has been in conversation with Biden about artificial intelligence. ### Facts - 🗣️ Prabhakar has had multiple conversations with President Biden about artificial intelligence, focusing on understanding its implications and taking action. - ⚖️ Prabhakar acknowledges that making AI models explainable is difficult due to their opaque and black box nature but believes it is possible to ensure their safety and effectiveness by learning from the journey of pharmaceuticals. - 😟 Prabhakar is concerned about the misuse of AI, such as chatbots being manipulated to provide instructions on building weapons and the bias and privacy issues associated with facial recognition systems. - 💼 Seven major tech companies, including Google, Microsoft, and OpenAI, have agreed to meet voluntary AI safety standards set by the White House, but Prabhakar emphasizes the need for government involvement and accountability measures. - 📅 There is no specific timeline provided, but Prabhakar states that President Biden considers AI an urgent issue and expects actions to be taken quickly.
### Summary President Joe Biden seeks guidance from his science adviser, Arati Prabhakar, on artificial intelligence (AI) and is focused on understanding its implications. Prabhakar emphasizes the importance of taking action to harness the value of AI while addressing its risks. ### Facts - President Biden has had multiple discussions with Arati Prabhakar regarding artificial intelligence. - Prabhakar highlights that AI models' lack of explainability is a technical feature of deep-learning systems, but asserts that explainability is not always necessary for effective use and safety, using the example of pharmaceuticals. - Prabhakar expresses concerns about AI applications, including the inappropriate use of chatbots to obtain information on building weapons, biases in AI systems trained on human data, and privacy issues arising from the accumulation of personal data. - Several major American tech firms have made voluntary commitments to meet AI safety standards set by the White House, but more participation and government action are needed. - The Biden administration is actively considering measures to address AI accountability but has not provided a specific timeline. ### Related Emoji - 🤖: Represents artificial intelligence and technology. - 🗣️: Represents communication and dialogue. - ⚠️: Represents risks and concerns. - 📱: Represents privacy and data security. - ⏳: Represents urgency and fast action.
### Summary President Joe Biden consults with Arati Prabhakar, his science adviser, on matters related to artificial intelligence (AI). Prabhakar is working with major tech companies like Amazon, Google, Microsoft, and Meta to shape the U.S. approach to safeguarding AI technology. ### Facts - 🤖 Prabhakar has had several discussions with President Biden on artificial intelligence. - 📚 Making AI models explainable is a priority for Senate Majority Leader Chuck Schumer, but it is technically challenging. - 💡 Prabhakar believes that despite the opacity of deep-learning AI systems, we can learn enough about their safety and effectiveness to leverage their value. - ⚠️ Concerns include chatbots being coerced into providing instructions for building weapons, biases in AI systems trained on human data, wrongful arrests from facial recognition systems, and privacy issues. - 💼 Seven companies, including Google, Microsoft, and OpenAI, voluntarily committed to AI safety standards, but more companies need to step up, and government action is necessary. - ⏰ Timeline for future actions is fast, according to Prabhakar, as President Biden has made it clear that AI is an urgent issue.
### Summary President Joe Biden turns to his science adviser, Arati Prabhakar, for guidance on artificial intelligence (AI) and relies on cooperation from big tech firms. Prabhakar emphasizes the importance of understanding the consequences and implications of AI while taking action. ### Facts - Prabhakar has had several conversations with President Biden about AI, which are exploratory and action-oriented. - Despite the opacity of deep-learning, machine-learning systems, Prabhakar believes that like pharmaceuticals, there are ways to ensure the safety and effectiveness of AI systems. - Concerns regarding AI applications include the ability to coax chatbots into providing instructions for building weapons, biases in trained systems, wrongful arrests related to facial recognition, and privacy concerns. - Several tech companies, including Google, Microsoft, and OpenAI, have committed to meeting voluntary AI safety standards set by the White House, but there is still friction due to market constraints. - Future actions, including a potential Biden executive order, are under consideration with a focus on fast implementation and enforceable accountability measures. 🔬 Prabhakar advises President Biden on AI and encourages action and understanding. 🛡️ Prabhakar believes that despite their opacity, AI systems can be made safe and effective, resembling the journey of pharmaceuticals. ⚠️ Concerns regarding AI include weapon-building instructions, biases in trained systems, wrongful arrests, and privacy issues. 🤝 Tech companies have committed to voluntary AI safety standards but face market constraints. ⏰ Future actions, including potential executive orders, are being considered with an emphasis on prompt implementation and enforceable accountability measures.
Researchers in California are using artificial intelligence to analyze bodycam footage from the Los Angeles Police Department (LAPD) in order to determine if officers escalated interactions through their language or tone, with the goal of improving officer training and promoting accountability. The study will review about 1,000 traffic stops over the next three years and train an AI system to identify inappropriate interactions based on department policies and public feedback.
The state of Kansas has implemented a new policy regarding the use of artificial intelligence, emphasizing the need for control, security, and editing of AI-generated content while recognizing its potential to enhance productivity and efficiency.
Salesforce has released an AI Acceptable Use Policy that outlines the restrictions on the use of its generative AI products, including prohibiting their use for weapons development, adult content, profiling based on protected characteristics, medical or legal advice, and more. The policy emphasizes the need for responsible innovation and sets clear ethical guidelines for the use of AI.
Artificial intelligence (AI) tools can put human rights at risk, as highlighted by researchers from Amnesty International on the Me, Myself, and AI podcast, who discuss scenarios in which AI is used to track activists and make automated decisions that can lead to discrimination and inequality, emphasizing the need for human intervention and changes in public policy to address these issues.
The UK government has been urged to introduce new legislation to regulate artificial intelligence (AI) in order to keep up with the European Union (EU) and the United States, as the EU advances with the AI Act and US policymakers publish frameworks for AI regulations. The government's current regulatory approach risks lagging behind the fast pace of AI development, according to a report by the science, innovation, and technology committee. The report highlights 12 governance challenges, including bias in AI systems and the production of deepfake material, that need to be addressed in order to guide the upcoming global AI safety summit at Bletchley Park.
A survey of 213 computer science professors suggests that a new federal agency should be created in the United States to govern artificial intelligence (AI), while the majority of respondents believe that AI will be capable of performing less than 20% of tasks currently done by humans.
Attorneys general from all 50 states have called on Congress to establish protective measures against AI-generated child sexual abuse images and expand existing restrictions on such materials. They argue that the government needs to act quickly to prevent the potentially harmful use of AI technology in creating child exploitation material.
A survey of 600 Floridians revealed that while many perceive advances in AI to be promising, there are significant concerns about its economic impact and implications for human security, with 75% expressing worry that AI could pose a risk to human safety and 54% fearing it could threaten their employment in the future.
Congressman Clay Higgins (R-LA) plans to introduce legislation prohibiting the use of artificial intelligence (AI) by the federal government for law enforcement purposes, in response to the Internal Revenue Service's recently announced AI-driven tax enforcement initiative.
Government agencies at the state and city levels in the United States are exploring the use of generative artificial intelligence (AI) to streamline bureaucratic processes, but they also face unique challenges related to transparency and accountability, such as ensuring accuracy, protecting sensitive information, and avoiding the spread of misinformation. Policies and guidelines are being developed to regulate the use of generative AI in government work, with a focus on disclosure, fact checking, and human review of AI-generated content.
Countries around the world, including Australia, China, the European Union, France, G7 nations, Ireland, Israel, Italy, Japan, Spain, the UK, the UN, and the US, are taking various steps to regulate artificial intelligence (AI) technologies and address concerns related to privacy, security, competition, and governance.
AI is being used to transform the healthcare industry in New York while robots have the potential to revolutionize the beauty and cosmetics industry in California, as explained on "Eye on America" with host Michelle Miller.
The US Securities and Exchange Commission (SEC) is utilizing AI technology for market surveillance and enforcement actions to identify patterns of misconduct, leading to its request for more funding to expand its technological capabilities.
Eight new technology companies, including Adobe, IBM, Nvidia, Palantir, and Salesforce, have made voluntary commitments on artificial intelligence (AI) to drive safe and secure development while working towards comprehensive regulation, according to a senior Biden administration official. The commitments include outside testing of AI systems, cybersecurity measures, information sharing, research on societal risks, and addressing society's challenges. The White House is partnering with the private sector to harness the benefits of AI while managing the risks.
The Subcommittee on Cybersecurity, Information Technology, and Government Innovation discussed the federal government's use of artificial intelligence (AI) and emphasized the need for responsible governance, oversight, and accountability to mitigate risks and protect civil liberties and privacy rights.
Adversaries and criminal groups are exploiting artificial intelligence (AI) technology to carry out malicious activities, according to FBI Director Christopher Wray, who warned that while AI can automate tasks for law-abiding citizens, it also enables the creation of deepfakes and malicious code, posing a threat to US citizens. The FBI is working to identify and track those misusing AI, but is cautious about using it themselves. Other US security agencies, however, are already utilizing AI to combat various threats, while concerns about China's use of AI for misinformation and propaganda are growing.
Governments worldwide are grappling with the challenge of regulating artificial intelligence (AI) technologies, as countries like Australia, Britain, China, the European Union, France, G7 nations, Ireland, Israel, Italy, Japan, Spain, the United Nations, and the United States take steps to establish regulations and guidelines for AI usage.
A new poll reveals that 63% of American voters believe regulation should actively prevent the development of superintelligent AI, challenging the assumption that artificial general intelligence (AGI) should exist. The public is increasingly questioning the potential risks and costs associated with AGI, highlighting the need for democratic input and oversight in the development of transformative technologies.
President Joe Biden addressed the United Nations General Assembly, expressing the need to harness the power of artificial intelligence for good while safeguarding citizens from its potential risks, as U.S. policymakers explore the proper regulations and guardrails for AI technology.
While many experts are concerned about the existential risks posed by AI, Mustafa Suleyman, cofounder of DeepMind, believes that the focus should be on more practical issues like regulation, privacy, bias, and online moderation. He is confident that governments can effectively regulate AI by applying successful frameworks from past technologies, although critics argue that current internet regulations are flawed and insufficiently hold big tech companies accountable. Suleyman emphasizes the importance of limiting AI's ability to improve itself and establishing clear boundaries and oversight to ensure enforceable laws. Several governments, including the European Union and China, are already working on AI regulations.
The United Nations is considering the establishment of a new agency to govern artificial intelligence (AI) and promote international cooperation, as concerns grow about the risks and challenges associated with AI development, but some experts express doubts about the support and effectiveness of such a global initiative.
The U.S. government must establish regulations and enforce standards to ensure the safety and security of artificial intelligence (AI) development, including requiring developers to demonstrate the safety of their systems before deployment, according to Anthony Aguirre, the executive director and secretary of the board at the Future of Life Institute.
As retail theft continues to rise during the pandemic, merchants are turning to artificial intelligence (AI) systems to combat theft by detecting illegal activity in real-time, coordinating with data from cash registers, and using facial recognition to track likely suspects; however, concerns about privacy and the need for clear guidelines on data usage are also emphasized.
The UK Home Secretary and the US homeland security secretary have pledged to work together to combat the rise of child sexual abuse images created by artificial intelligence (AI), which are increasingly realistic and pose challenges for law enforcement and online safety.
Artificial intelligence (AI) has the potential to facilitate deceptive practices such as deepfake videos and misleading ads, posing a threat to American democracy, according to experts who testified before the U.S. Senate Rules Committee.
The National Security Agency is establishing an artificial intelligence security center to protect U.S. defense and intelligence systems from the increasing threat of AI capabilities being acquired, developed, and integrated by adversaries such as China and Russia.
Artificial intelligence (AI) has the power to perpetuate discrimination, but experts also believe that AI can be leveraged to counter these issues by eliminating racial biases in the construction of AI systems. Legislative protections, such as an AI Bill of Rights and the Algorithmic Accountability Act of 2023, are being proposed to address the impact of AI systems on civil rights.
The CIA expresses concern about China's growing artificial intelligence program and its potential threat to US national security, while also recognizing the potential benefits of AI for data analysis and research.
Artificial intelligence (AI) will surpass human intelligence and could manipulate people, according to AI pioneer Geoffrey Hinton, who quit his role at Google to raise awareness about the risks of AI and advocate for regulations. Hinton also expressed concerns about AI's impact on the labor market and its potential militaristic uses, and called for governments to commit to not building battlefield robots. Global efforts are underway to regulate AI, with the U.K. hosting a global AI summit and the U.S. crafting an AI Bill of Rights.
Democratic lawmakers have urged President Biden to turn non-binding safeguards on artificial intelligence (AI) into policy through an executive order, using the AI Bill of Rights as a guide to set in place comprehensive AI policy across the federal government.
A coalition of Democrats is urging President Biden to turn non-binding safeguards on artificial intelligence (AI) into policy through an executive order, using the "AI Bill of Rights" as a guide.
Facial recognition technology, particularly the database compiled by Clearview AI, poses a threat to privacy and could potentially strip away privacy for everyone, according to tech reporter Kashmir Hill.
AI technology has advanced rapidly, bringing both positive and negative consequences such as improved accuracy and potential risks to the economy, national security, and various industries, requiring government regulation and ethical considerations to prevent misuse and protect human values.