1. Home
  2. >
  3. AI 🤖
Posted

Biden Warns UN of AI's 'Enormous Potential and Enormous Peril' Amid Efforts to Set Safeguards Without Stifling Innovation

  • Biden said at UN that AI holds "enormous potential and enormous peril" and countries must work together to ensure it's "used as tools of opportunity, not as weapons of oppression."

  • His comments come as US policymakers try to learn more about AI to set proper safeguards without stifling innovation, amid competition with China.

  • Senate Majority Leader Schumer hosted tech CEOs to discuss regulating AI; said government needs to play a role but must not move too fast like EU.

  • Multiple US agencies have asserted ability to rein in AI abuses with existing powers, like NIST's risk framework and voluntary commitments from AI companies.

  • Lawmakers differ on best body to regulate AI and how light a touch to take; Schumer said timeline should be in "months" not "days or weeks, but nor should it be years."

cnbc.com
Relevant topic timeline:
In this episode of the "Have a Nice Future" podcast, Gideon Lichfield and Lauren Goode interview Mustafa Suleyman, the co-founder of DeepMind and InflectionAI. The main topic of discussion is Suleyman's new book, "The Coming Wave," which examines the potential impact of AI and other technologies on society and governance. Key points discussed include Suleyman's concern that AI proliferation could undermine nation-states and increase inequality, the potential for AI to help lift people out of poverty, and the need for better AI assessment tools.
### Summary The article discusses the rapid advancement and potential risks of artificial intelligence (AI) and proposes the idea of nationalizing certain aspects of AI under a governing body called the Humane AI Commission to ensure AI is aligned with human interests. ### Facts - AI is evolving rapidly and penetrating various aspects of American life, from image recognition to healthcare. - AI has the potential to bring both significant benefits and risks to society. - Transparency in AI is limited, and understanding how specific AI works is difficult. - Congress is becoming more aware of the importance of AI and its need for regulation. - The author proposes the creation of a governing body, the Humane AI Commission, that can control and steer AI technology to serve humanity's best interests. - The nationalization of advanced AI models could be considered, similar to the Atomic Energy Commission's control over nuclear reactors. - Various options, such as an AI pause or leaving AI development to the free market or current government agencies, have limitations in addressing the potential risks of AI. - The author suggests that the United States should take a bold executive leadership approach to develop a national AI plan and ensure global AI leadership with a focus on benevolence and human-controlled AI. ### 🤖 AI Nationalization - The case to nationalize the “nuclear reactors” of AI — the world’s most advanced AI models — hinges on this question: Who do we want to control AI’s nuclear codes? Big Tech CEOs answering to a few billionaire shareholders, or the government of the United States, answering to its citizens? ### 👥 Humane AI Commission - The author proposes the creation of a Humane AI Commission, run by AI experts, to steer and control AI technology in alignment with human interests. ### ⚠️ Risks of AI - AI's rapid advancement and lack of transparency pose risks such as unpredictable behavior, potential damage to power generation, financial markets, and public health, and the potential for AI to move beyond human control. ### ⚖️ AI Regulation - The article calls for federal regulation of AI, but emphasizes the limitations of traditional regulation in addressing the fast-evolving nature of AI and the need for a larger-scale approach like nationalization.
### Summary Arati Prabhakar, President Biden's science adviser, is helping guide the U.S. approach to safeguarding AI technology and has been in conversation with Biden about artificial intelligence. ### Facts - 🗣️ Prabhakar has had multiple conversations with President Biden about artificial intelligence, focusing on understanding its implications and taking action. - ⚖️ Prabhakar acknowledges that making AI models explainable is difficult due to their opaque and black box nature but believes it is possible to ensure their safety and effectiveness by learning from the journey of pharmaceuticals. - 😟 Prabhakar is concerned about the misuse of AI, such as chatbots being manipulated to provide instructions on building weapons and the bias and privacy issues associated with facial recognition systems. - 💼 Seven major tech companies, including Google, Microsoft, and OpenAI, have agreed to meet voluntary AI safety standards set by the White House, but Prabhakar emphasizes the need for government involvement and accountability measures. - 📅 There is no specific timeline provided, but Prabhakar states that President Biden considers AI an urgent issue and expects actions to be taken quickly.
### Summary President Joe Biden seeks guidance from his science adviser, Arati Prabhakar, on artificial intelligence (AI) and is focused on understanding its implications. Prabhakar emphasizes the importance of taking action to harness the value of AI while addressing its risks. ### Facts - President Biden has had multiple discussions with Arati Prabhakar regarding artificial intelligence. - Prabhakar highlights that AI models' lack of explainability is a technical feature of deep-learning systems, but asserts that explainability is not always necessary for effective use and safety, using the example of pharmaceuticals. - Prabhakar expresses concerns about AI applications, including the inappropriate use of chatbots to obtain information on building weapons, biases in AI systems trained on human data, and privacy issues arising from the accumulation of personal data. - Several major American tech firms have made voluntary commitments to meet AI safety standards set by the White House, but more participation and government action are needed. - The Biden administration is actively considering measures to address AI accountability but has not provided a specific timeline. ### Related Emoji - 🤖: Represents artificial intelligence and technology. - 🗣️: Represents communication and dialogue. - ⚠️: Represents risks and concerns. - 📱: Represents privacy and data security. - ⏳: Represents urgency and fast action.
### Summary President Joe Biden consults with Arati Prabhakar, his science adviser, on matters related to artificial intelligence (AI). Prabhakar is working with major tech companies like Amazon, Google, Microsoft, and Meta to shape the U.S. approach to safeguarding AI technology. ### Facts - 🤖 Prabhakar has had several discussions with President Biden on artificial intelligence. - 📚 Making AI models explainable is a priority for Senate Majority Leader Chuck Schumer, but it is technically challenging. - 💡 Prabhakar believes that despite the opacity of deep-learning AI systems, we can learn enough about their safety and effectiveness to leverage their value. - ⚠️ Concerns include chatbots being coerced into providing instructions for building weapons, biases in AI systems trained on human data, wrongful arrests from facial recognition systems, and privacy issues. - 💼 Seven companies, including Google, Microsoft, and OpenAI, voluntarily committed to AI safety standards, but more companies need to step up, and government action is necessary. - ⏰ Timeline for future actions is fast, according to Prabhakar, as President Biden has made it clear that AI is an urgent issue.
### Summary President Joe Biden turns to his science adviser, Arati Prabhakar, for guidance on artificial intelligence (AI) and relies on cooperation from big tech firms. Prabhakar emphasizes the importance of understanding the consequences and implications of AI while taking action. ### Facts - Prabhakar has had several conversations with President Biden about AI, which are exploratory and action-oriented. - Despite the opacity of deep-learning, machine-learning systems, Prabhakar believes that like pharmaceuticals, there are ways to ensure the safety and effectiveness of AI systems. - Concerns regarding AI applications include the ability to coax chatbots into providing instructions for building weapons, biases in trained systems, wrongful arrests related to facial recognition, and privacy concerns. - Several tech companies, including Google, Microsoft, and OpenAI, have committed to meeting voluntary AI safety standards set by the White House, but there is still friction due to market constraints. - Future actions, including a potential Biden executive order, are under consideration with a focus on fast implementation and enforceable accountability measures. 🔬 Prabhakar advises President Biden on AI and encourages action and understanding. 🛡️ Prabhakar believes that despite their opacity, AI systems can be made safe and effective, resembling the journey of pharmaceuticals. ⚠️ Concerns regarding AI include weapon-building instructions, biases in trained systems, wrongful arrests, and privacy issues. 🤝 Tech companies have committed to voluntary AI safety standards but face market constraints. ⏰ Future actions, including potential executive orders, are being considered with an emphasis on prompt implementation and enforceable accountability measures.
President Joe Biden relies on his science adviser Arati Prabhakar to guide the US approach to safeguarding AI technology, with cooperation from tech giants like Amazon, Google, Microsoft and Meta. Prabhakar discusses the need for understanding the implications and consequences of AI, the challenge of making AI models explainable, concerns about biases and privacy, and the importance of voluntary commitments from tech companies along with government actions.
The U.S. is falling behind in regulating artificial intelligence (AI), while Europe has passed the world's first comprehensive AI law; President Joe Biden recently met with industry leaders to discuss the need for AI regulation and companies pledged to develop safeguards for AI-generated content and prioritize user privacy.
Artificial intelligence should be controlled by humans to prevent its weaponization and ensure safety measures are in place, according to Microsoft's president Brad Smith. He stressed the need for regulations and laws to govern AI, comparing it to other technologies that have required safety breaks and human oversight. Additionally, Smith emphasized that AI is a tool to assist humans, not to replace them, and that it can help individuals think more efficiently.
Artificial intelligence (AI) tools can put human rights at risk, as highlighted by researchers from Amnesty International on the Me, Myself, and AI podcast, who discuss scenarios in which AI is used to track activists and make automated decisions that can lead to discrimination and inequality, emphasizing the need for human intervention and changes in public policy to address these issues.
The UK government has been urged to introduce new legislation to regulate artificial intelligence (AI) in order to keep up with the European Union (EU) and the United States, as the EU advances with the AI Act and US policymakers publish frameworks for AI regulations. The government's current regulatory approach risks lagging behind the fast pace of AI development, according to a report by the science, innovation, and technology committee. The report highlights 12 governance challenges, including bias in AI systems and the production of deepfake material, that need to be addressed in order to guide the upcoming global AI safety summit at Bletchley Park.
Several tech giants in the US, including Alphabet, Microsoft, Meta Platforms, and Amazon, have pledged to collaborate with the Biden administration to address the risks associated with artificial intelligence, focusing on safety, security, and trust in AI development.
A survey of 213 computer science professors suggests that a new federal agency should be created in the United States to govern artificial intelligence (AI), while the majority of respondents believe that AI will be capable of performing less than 20% of tasks currently done by humans.
The rivalry between the US and China over artificial intelligence (AI) is intensifying as both countries compete for dominance in the emerging field, but experts suggest that cooperation on certain issues is necessary to prevent conflicts and ensure global governance of AI. While tensions remain high and trust is lacking, potential areas of cooperation include AI safety and regulations. However, failure to cooperate could increase the risk of armed conflict and hinder the exploration and governance of AI.
The G20 member nations have pledged to use artificial intelligence (AI) in a responsible manner, addressing concerns such as data protection, biases, human oversight, and ethics, while also planning for the future of cryptocurrencies and central bank digital currencies (CBDCs).
Eight big tech companies, including Adobe, IBM, Salesforce, and Nvidia, have pledged to conduct more testing and research on the risks of artificial intelligence (AI) in a meeting with White House officials, signaling a "bridge" to future government action on the issue. These voluntary commitments come amidst congressional scrutiny and ongoing efforts by the White House to develop policies for AI.
The United Nations is urging the international community to confront the potential risks and benefits of Artificial Intelligence, which has the power to transform the world.
The Subcommittee on Cybersecurity, Information Technology, and Government Innovation discussed the federal government's use of artificial intelligence (AI) and emphasized the need for responsible governance, oversight, and accountability to mitigate risks and protect civil liberties and privacy rights.
Representatives from several countries and companies announced commitments to harness the power of artificial intelligence (AI) to advance progress in achieving the United Nations' Sustainable Development Goals (SDGs) during a ministerial side event at the United Nations' 78th Session High Level Week. These commitments focused on using AI to address issues related to health, education, food security, energy, and climate action, with an emphasis on inclusive and responsible governance of AI.
Governments worldwide are grappling with the challenge of regulating artificial intelligence (AI) technologies, as countries like Australia, Britain, China, the European Union, France, G7 nations, Ireland, Israel, Italy, Japan, Spain, the United Nations, and the United States take steps to establish regulations and guidelines for AI usage.
A new poll reveals that 63% of American voters believe regulation should actively prevent the development of superintelligent AI, challenging the assumption that artificial general intelligence (AGI) should exist. The public is increasingly questioning the potential risks and costs associated with AGI, highlighting the need for democratic input and oversight in the development of transformative technologies.
The United States must prioritize global leadership in artificial intelligence (AI) and win the platform competition with China in order to protect national security, democracy, and economic prosperity, according to Ylli Bajraktari, the president and CEO of the Special Competitive Studies Project and former Pentagon official.
While many experts are concerned about the existential risks posed by AI, Mustafa Suleyman, cofounder of DeepMind, believes that the focus should be on more practical issues like regulation, privacy, bias, and online moderation. He is confident that governments can effectively regulate AI by applying successful frameworks from past technologies, although critics argue that current internet regulations are flawed and insufficiently hold big tech companies accountable. Suleyman emphasizes the importance of limiting AI's ability to improve itself and establishing clear boundaries and oversight to ensure enforceable laws. Several governments, including the European Union and China, are already working on AI regulations.
Pennsylvania Governor Josh Shapiro signed an executive order establishing standards and a governance framework for the use of artificial intelligence (AI) by state agencies, as well as creating a Generative AI Governing Board and outlining core values to govern AI use. The order aims to responsibly integrate AI into government operations and enhance employee job functions.
The UK's deputy prime minister, Oliver Dowden, will use a speech at the UN general assembly to warn that artificial intelligence is developing too fast for regulation, and will call on other countries to collaborate in creating an international regulatory system to address the potential threats posed by AI technology.
The U.S. government must establish regulations and enforce standards to ensure the safety and security of artificial intelligence (AI) development, including requiring developers to demonstrate the safety of their systems before deployment, according to Anthony Aguirre, the executive director and secretary of the board at the Future of Life Institute.
Israeli Prime Minister Benjamin Netanyahu warned of the potential dangers of artificial intelligence (AI) and called for responsible and ethical development of AI during his speech at the United Nations General Assembly, emphasizing that nations must work together to prevent the perils of AI and ensure it brings more freedom and benefits humanity.
The United Nations aims to bring inclusiveness, legitimacy, and authority to the regulation of artificial intelligence, leveraging its experience with managing the impact of various technologies and creating compliance pressure for commitments made by governments, according to Amandeep Gill, the organization's top tech-policy official. Despite the challenges of building consensus and engaging stakeholders, the U.N. seeks to promote diverse and inclusive innovation to ensure equal opportunities and prevent concentration of economic power. Gill also emphasizes the potential of AI in accelerating progress towards the Sustainable Development Goals but expresses concerns about potential misuse and concentration of power.
The leaked information about a possible executive order by U.S. President Joe Biden on artificial intelligence is causing concern in the bitcoin and crypto industry, as it could have spillover effects on the market.
The White House plans to introduce an executive order on artificial intelligence in the coming weeks, as President Biden aims for responsible AI innovation and collaboration with international partners.
Sen. Mark Warner, a U.S. Senator from Virginia, is urging Congress to take a less ambitious approach to regulating artificial intelligence (AI), suggesting that lawmakers should focus on narrowly focused issues rather than trying to address the full spectrum of AI risks with a single comprehensive law. Warner believes that tackling immediate concerns, such as AI-generated deepfakes, is a more realistic and effective approach to regulation. He also emphasizes the need for bipartisan agreement and action to demonstrate progress in the regulation of AI, especially given Congress's previous failures in addressing issues related to social media.
Artificial intelligence (AI) will have a significant impact on geopolitics and globalization, driving a new globalization but also posing risks that the world is not yet ready for, according to political scientist Ian Bremmer. Global leaders and policymakers are now catching up and discussing the implications of AI, but a greater understanding of the technology is needed for effective regulation. Bremmer suggests international cooperation, such as a United Nations-driven process, to establish global oversight and prevent the U.S. versus China competition in AI development.
China's use of artificial intelligence (AI) for surveillance and oppression should deter the United States from collaborating with China on AI development and instead focus on asserting itself in international standards-setting bodies, open sourcing AI technologies, and promoting explainable AI to ensure transparency and uphold democratic values.
Advisers to UK Chancellor Rishi Sunak are working on a statement to be used in a communique at the AI safety summit next month, although they are unlikely to reach an agreement on establishing a new international organisation to oversee AI. The summit will focus on the risks of AI models, debate national security agencies' scrutiny of dangerous versions of the technology, and discuss international cooperation on AI that poses a threat to human life.
Artificial intelligence (AI) will surpass human intelligence and could manipulate people, according to AI pioneer Geoffrey Hinton, who quit his role at Google to raise awareness about the risks of AI and advocate for regulations. Hinton also expressed concerns about AI's impact on the labor market and its potential militaristic uses, and called for governments to commit to not building battlefield robots. Global efforts are underway to regulate AI, with the U.K. hosting a global AI summit and the U.S. crafting an AI Bill of Rights.
Democratic lawmakers have urged President Biden to turn non-binding safeguards on artificial intelligence (AI) into policy through an executive order, using the AI Bill of Rights as a guide to set in place comprehensive AI policy across the federal government.
A coalition of Democrats is urging President Biden to turn non-binding safeguards on artificial intelligence (AI) into policy through an executive order, using the "AI Bill of Rights" as a guide.
President Biden's executive order on artificial intelligence is expected to use the federal government's purchasing power to influence American AI standards, tighten industry guidelines, require cloud computing companies to monitor users developing powerful AI systems, and boost AI talent recruitment and domestic training.
The United Kingdom will host an international summit on artificial intelligence safety in November 2023, focusing on the potential existential threat of AI and establishing the country as a mediator in technology post-Brexit. British Prime Minister Rishi Sunak, along with Vice President Kamala Harris and other distinguished guests, aims to initiate a global conversation on AI regulation and address concerns about its misuse.