### Summary
The article discusses the rapid advancement and potential risks of artificial intelligence (AI) and proposes the idea of nationalizing certain aspects of AI under a governing body called the Humane AI Commission to ensure AI is aligned with human interests.
### Facts
- AI is evolving rapidly and penetrating various aspects of American life, from image recognition to healthcare.
- AI has the potential to bring both significant benefits and risks to society.
- Transparency in AI is limited, and understanding how specific AI works is difficult.
- Congress is becoming more aware of the importance of AI and its need for regulation.
- The author proposes the creation of a governing body, the Humane AI Commission, that can control and steer AI technology to serve humanity's best interests.
- The nationalization of advanced AI models could be considered, similar to the Atomic Energy Commission's control over nuclear reactors.
- Various options, such as an AI pause or leaving AI development to the free market or current government agencies, have limitations in addressing the potential risks of AI.
- The author suggests that the United States should take a bold executive leadership approach to develop a national AI plan and ensure global AI leadership with a focus on benevolence and human-controlled AI.
### 🤖 AI Nationalization - The case to nationalize the “nuclear reactors” of AI — the world’s most advanced AI models — hinges on this question: Who do we want to control AI’s nuclear codes? Big Tech CEOs answering to a few billionaire shareholders, or the government of the United States, answering to its citizens?
### 👥 Humane AI Commission - The author proposes the creation of a Humane AI Commission, run by AI experts, to steer and control AI technology in alignment with human interests.
### ⚠️ Risks of AI - AI's rapid advancement and lack of transparency pose risks such as unpredictable behavior, potential damage to power generation, financial markets, and public health, and the potential for AI to move beyond human control.
### ⚖️ AI Regulation - The article calls for federal regulation of AI, but emphasizes the limitations of traditional regulation in addressing the fast-evolving nature of AI and the need for a larger-scale approach like nationalization.
A new poll conducted by the AI Policy Institute reveals that 72 percent of American voters want to slow down the development of AI, signaling a divergence between elite opinion and public opinion on the technology. Additionally, the poll shows that 82 percent of American voters do not trust AI companies to self-regulate. To address these concerns, the AI Now Institute has proposed a framework called "Zero Trust AI Governance," which calls for lawmakers to vigorously enforce existing laws, establish bold and easily administrable rules, and place the burden of proof on companies to demonstrate the safety of their AI systems.
The U.S. is falling behind in regulating artificial intelligence (AI), while Europe has passed the world's first comprehensive AI law; President Joe Biden recently met with industry leaders to discuss the need for AI regulation and companies pledged to develop safeguards for AI-generated content and prioritize user privacy.
The GZERO World podcast episode discusses the explosive growth and potential risks of generative AI, as well as the proposed 5 principles for effective AI governance.
Artificial intelligence (AI) is likely to subtract jobs without producing new ones, with evidence suggesting that jobs will disappear rather than be replaced, according to experts, and regulation should only be considered once AI is controllable.
Artificial intelligence expert Michael Wooldridge is not worried about the growth of AI, but is concerned about the potential for AI to become a controlling and invasive boss that monitors employees' every move. He emphasizes the immediate and concrete existential concerns in the world, such as the escalation of conflict in Ukraine, as more important things to worry about.
AI developments in Eastern Europe have the potential to boost economic growth and address issues such as hate speech, healthcare, agriculture, and waste management, providing a "great equalizer" for the region's historically disadvantaged areas.
The UK government has been urged to introduce new legislation to regulate artificial intelligence (AI) in order to keep up with the European Union (EU) and the United States, as the EU advances with the AI Act and US policymakers publish frameworks for AI regulations. The government's current regulatory approach risks lagging behind the fast pace of AI development, according to a report by the science, innovation, and technology committee. The report highlights 12 governance challenges, including bias in AI systems and the production of deepfake material, that need to be addressed in order to guide the upcoming global AI safety summit at Bletchley Park.
The rise of AI and other emerging technologies will lead to a significant redistribution of power, giving individuals and organizations unprecedented capabilities and disrupting established power structures.
The rise of artificial intelligence (AI) is a hot trend in 2023, with the potential to add trillions to the global economy by 2030, and billionaire investors are buying into AI stocks like Nvidia, Meta Platforms, Okta, and Microsoft.
The rapid advancement of AI technology poses significant challenges for democratic societies, including the need for nuanced debates, public engagement, and ethical considerations in regulating AI to mitigate unintended consequences.
AI has the potential to transform numerous industries, including medicine, law, art, retail, film, tech, education, and agriculture, by automating tasks, improving productivity, and enhancing decision-making, while still relying on the unique human abilities of empathy, creativity, and intuition. The impact of AI will be felt differently in each industry and will require professionals to adapt and develop new skills to work effectively with AI systems.
A survey of 213 computer science professors suggests that a new federal agency should be created in the United States to govern artificial intelligence (AI), while the majority of respondents believe that AI will be capable of performing less than 20% of tasks currently done by humans.
Countries around the world, including Australia, China, the European Union, France, G7 nations, Ireland, Israel, Italy, Japan, Spain, the UK, the UN, and the US, are taking various steps to regulate artificial intelligence (AI) technologies and address concerns related to privacy, security, competition, and governance.
Artificial intelligence (AI) is poised to be the biggest technological shift of our lifetimes, and companies like Nvidia, Amazon, Alphabet, Microsoft, and Tesla are well-positioned to capitalize on this AI revolution.
Eight new technology companies, including Adobe, IBM, Nvidia, Palantir, and Salesforce, have made voluntary commitments on artificial intelligence (AI) to drive safe and secure development while working towards comprehensive regulation, according to a senior Biden administration official. The commitments include outside testing of AI systems, cybersecurity measures, information sharing, research on societal risks, and addressing society's challenges. The White House is partnering with the private sector to harness the benefits of AI while managing the risks.
The United Nations is urging the international community to confront the potential risks and benefits of Artificial Intelligence, which has the power to transform the world.
The book "The Age of AI: And Our Human Future" by Henry Kissinger, Eric Schmidt, and Daniel Huttenlocher explores the transformational impact of AI on human society and the need for humans to shape its development and use with their values.
Leading economist Daron Acemoglu argues that the prevailing optimism about artificial intelligence (AI) and its potential to benefit society is flawed, as history has shown that technological progress often fails to improve the lives of most people; he warns of a future two-tier system with a small elite benefiting from AI while the majority experience lower wages and less meaningful jobs, emphasizing the need for societal action to ensure shared prosperity.
A bipartisan group of senators is expected to introduce legislation to create a government agency to regulate AI and require AI models to obtain a license before deployment, a move that some leading technology companies have supported; however, critics argue that licensing regimes and a new AI regulator could hinder innovation and concentrate power among existing players, similar to the undesirable economic consequences seen in Europe.
Governments worldwide are grappling with the challenge of regulating artificial intelligence (AI) technologies, as countries like Australia, Britain, China, the European Union, France, G7 nations, Ireland, Israel, Italy, Japan, Spain, the United Nations, and the United States take steps to establish regulations and guidelines for AI usage.
A new poll reveals that 63% of American voters believe regulation should actively prevent the development of superintelligent AI, challenging the assumption that artificial general intelligence (AGI) should exist. The public is increasingly questioning the potential risks and costs associated with AGI, highlighting the need for democratic input and oversight in the development of transformative technologies.
While many experts are concerned about the existential risks posed by AI, Mustafa Suleyman, cofounder of DeepMind, believes that the focus should be on more practical issues like regulation, privacy, bias, and online moderation. He is confident that governments can effectively regulate AI by applying successful frameworks from past technologies, although critics argue that current internet regulations are flawed and insufficiently hold big tech companies accountable. Suleyman emphasizes the importance of limiting AI's ability to improve itself and establishing clear boundaries and oversight to ensure enforceable laws. Several governments, including the European Union and China, are already working on AI regulations.
AI adoption is rapidly increasing, but it is crucial for businesses to establish governance and ethical usage policies to prevent potential harm and job loss, while utilizing AI to automate tasks, augment human work, enable change management, make data-driven decisions, prioritize employee training, and establish responsible AI governance.
Deputy Prime Minister Oliver Dowden will warn the UN that artificial intelligence (AI) poses a threat to world order unless governments take action, with fears that the rapid pace of AI development could lead to job losses, misinformation, and discrimination without proper regulations in place. Dowden will call for global regulation and emphasize the importance of making rules in parallel with AI development rather than retroactively. Despite the need for regulation, experts note the complexity of reaching a quick international agreement, with meaningful input needed from smaller countries, marginalized communities, and ethnic minorities. The UK aims to take the lead in AI regulation, but there are concerns that without swift action, the European Union's AI Act could become the global standard instead.
AI is dramatically reshaping industries and driving productivity, but businesses that lag behind in adaptation risk falling behind and becoming obsolete. Job displacement may occur, but history suggests that new roles will emerge. The responsibility lies with us to guide AI's evolution responsibly and ensure its transformative power benefits all of society.
Israeli Prime Minister Benjamin Netanyahu warned of the potential dangers of artificial intelligence (AI) and called for responsible and ethical development of AI during his speech at the United Nations General Assembly, emphasizing that nations must work together to prevent the perils of AI and ensure it brings more freedom and benefits humanity.
There is a need for more policy balance in discussions about artificial intelligence (AI) to focus on the potential for good and how to ensure societal benefit, as AI has the potential to advance education, national security, and economic success, while also providing new economic opportunities and augmenting human capabilities.
The United Nations aims to bring inclusiveness, legitimacy, and authority to the regulation of artificial intelligence, leveraging its experience with managing the impact of various technologies and creating compliance pressure for commitments made by governments, according to Amandeep Gill, the organization's top tech-policy official. Despite the challenges of building consensus and engaging stakeholders, the U.N. seeks to promote diverse and inclusive innovation to ensure equal opportunities and prevent concentration of economic power. Gill also emphasizes the potential of AI in accelerating progress towards the Sustainable Development Goals but expresses concerns about potential misuse and concentration of power.
The spread of information technology, including advancements in AI, is pushing for governments to adapt to digitalization in order to make informed and efficient decisions, as it becomes increasingly clear that the world is moving towards a digital future.
Artificial intelligence (AI) will have a significant impact on geopolitics and globalization, driving a new globalization but also posing risks that the world is not yet ready for, according to political scientist Ian Bremmer. Global leaders and policymakers are now catching up and discussing the implications of AI, but a greater understanding of the technology is needed for effective regulation. Bremmer suggests international cooperation, such as a United Nations-driven process, to establish global oversight and prevent the U.S. versus China competition in AI development.
AI has the potential to augment human work and create shared prosperity, but without proper implementation and worker power, it can lead to job replacement, economic inequality, and concentrated political power.
AI will surpass the industrial revolution, according to a leading AI executive.
Artificial intelligence (AI) has become an undeniable force in our lives, with wide-ranging implications and ethical considerations, posing both benefits and potential harms, and raising questions about regulation and the future of humanity's relationship with AI.
As the 2023 election campaign in New Zealand nears its end, the rise of Artificial Intelligence (AI) and its potential impact on the economy, politics, and society is being largely overlooked by politicians, despite growing concerns from AI experts and the public. The use of AI raises concerns about job displacement, increased misinformation, biased outcomes, and data sovereignty issues, highlighting the need for stronger regulation and investment in AI research that benefits all New Zealanders.
Artificial intelligence (AI) has the potential to disrupt the creative industry, with concerns raised about AI-generated models, music, and other creative works competing with human artists, leading to calls for regulation and new solutions to protect creators.
The birth of the PC, Internet, and now mainstream artificial intelligence (AI) has ushered us into uncharted territories, requiring collaboration, shared principles, security, and sustainability to unlock AI's true value ethically and for the benefit of all.
China's use of artificial intelligence (AI) for surveillance and oppression should deter the United States from collaborating with China on AI development and instead focus on asserting itself in international standards-setting bodies, open sourcing AI technologies, and promoting explainable AI to ensure transparency and uphold democratic values.
Ukraine's Ministry of Digital Transformation has unveiled a regulatory roadmap for artificial intelligence (AI), aiming to help local companies prepare for adopting a law similar to the EU's AI Act and educate citizens on protecting themselves from AI risks. The roadmap follows a bottom-up approach, providing tools for businesses to prepare for future requirements before implementing any laws.
Artificial intelligence (AI) will surpass human intelligence and could manipulate people, according to AI pioneer Geoffrey Hinton, who quit his role at Google to raise awareness about the risks of AI and advocate for regulations. Hinton also expressed concerns about AI's impact on the labor market and its potential militaristic uses, and called for governments to commit to not building battlefield robots. Global efforts are underway to regulate AI, with the U.K. hosting a global AI summit and the U.S. crafting an AI Bill of Rights.
A new study warns that the artificial intelligence (AI) industry could consume as much energy as a country the size of the Netherlands by 2027, but its environmental impact could be less than feared if growth slows down.