A new poll conducted by the AI Policy Institute reveals that 72 percent of American voters want to slow down the development of AI, signaling a divergence between elite opinion and public opinion on the technology. Additionally, the poll shows that 82 percent of American voters do not trust AI companies to self-regulate. To address these concerns, the AI Now Institute has proposed a framework called "Zero Trust AI Governance," which calls for lawmakers to vigorously enforce existing laws, establish bold and easily administrable rules, and place the burden of proof on companies to demonstrate the safety of their AI systems.
Congress should prioritize maintaining bipartisan commitment to AI, generating global AI guardrails, and seeking out local perspectives in order to develop effective and responsible AI policies.
Wisconsin has established a task force to study the impact of artificial intelligence on the state's workforce, following a trend among other states. The task force, comprised of government leaders, educational institutions, and representatives from various sectors, aims to gather information and create an action plan to understand and adapt to the transformations brought about by AI.
The U.S. is falling behind in regulating artificial intelligence (AI), while Europe has passed the world's first comprehensive AI law; President Joe Biden recently met with industry leaders to discuss the need for AI regulation and companies pledged to develop safeguards for AI-generated content and prioritize user privacy.
Artificial intelligence (AI) has the potential to deliver significant productivity gains, but its current adoption may further consolidate the dominance of Big Tech companies, raising concerns among antitrust authorities.
As calls for regulation of artificial intelligence (A.I.) grow, history suggests that implementing comprehensive federal regulation of advanced A.I. systems in the U.S. will likely be a slow process, given Congress's historical patterns of responding to revolutionary technologies.
China's People's Liberation Army aims to be a leader in generative artificial intelligence for military applications, but faces challenges including data limitations, political restrictions, and a need for trust in the technology. Despite these hurdles, China is at a similar level or even ahead of the US in some areas of AI development and views AI as a crucial component of its national strategy.
A recent poll conducted by Pew Research Center shows that 52% of Americans are more concerned than excited about the use of artificial intelligence (AI) in their daily lives, marking an increase from the previous year; however, there are areas where they believe AI could have a positive impact, such as in online product and service searches, self-driving vehicles, healthcare, and finding accurate information online.
Artificial intelligence should be controlled by humans to prevent its weaponization and ensure safety measures are in place, according to Microsoft's president Brad Smith. He stressed the need for regulations and laws to govern AI, comparing it to other technologies that have required safety breaks and human oversight. Additionally, Smith emphasized that AI is a tool to assist humans, not to replace them, and that it can help individuals think more efficiently.
The UK government has been urged to introduce new legislation to regulate artificial intelligence (AI) in order to keep up with the European Union (EU) and the United States, as the EU advances with the AI Act and US policymakers publish frameworks for AI regulations. The government's current regulatory approach risks lagging behind the fast pace of AI development, according to a report by the science, innovation, and technology committee. The report highlights 12 governance challenges, including bias in AI systems and the production of deepfake material, that need to be addressed in order to guide the upcoming global AI safety summit at Bletchley Park.
Artificial intelligence regulation varies across countries, with Brazil focusing on user rights and risk assessments, China emphasizing "true and accurate" content generation, the EU categorizing AI into three risk levels, Israel promoting responsible innovation and self-regulation, Italy allocating funds for worker support, Japan adopting a wait-and-see approach, and the UAE prioritizing AI development and integration.
The United States Copyright Office has launched a study on artificial intelligence (AI) and copyright law, seeking public input on various policy issues and exploring topics such as AI training, copyright liability, and authorship. Other U.S. government agencies, including the SEC, USPTO, and DHS, have also initiated inquiries and public forums on AI, highlighting its impact on innovation, governance, and public policy.
A survey of 600 Floridians revealed that while many perceive advances in AI to be promising, there are significant concerns about its economic impact and implications for human security, with 75% expressing worry that AI could pose a risk to human safety and 54% fearing it could threaten their employment in the future.
The rivalry between the US and China over artificial intelligence (AI) is intensifying as both countries compete for dominance in the emerging field, but experts suggest that cooperation on certain issues is necessary to prevent conflicts and ensure global governance of AI. While tensions remain high and trust is lacking, potential areas of cooperation include AI safety and regulations. However, failure to cooperate could increase the risk of armed conflict and hinder the exploration and governance of AI.
Congressman Clay Higgins (R-LA) plans to introduce legislation prohibiting the use of artificial intelligence (AI) by the federal government for law enforcement purposes, in response to the Internal Revenue Service's recently announced AI-driven tax enforcement initiative.
The G20 member nations have pledged to use artificial intelligence (AI) in a responsible manner, addressing concerns such as data protection, biases, human oversight, and ethics, while also planning for the future of cryptocurrencies and central bank digital currencies (CBDCs).
Countries around the world, including Australia, China, the European Union, France, G7 nations, Ireland, Israel, Italy, Japan, Spain, the UK, the UN, and the US, are taking various steps to regulate artificial intelligence (AI) technologies and address concerns related to privacy, security, competition, and governance.
AI is being used to transform the healthcare industry in New York while robots have the potential to revolutionize the beauty and cosmetics industry in California, as explained on "Eye on America" with host Michelle Miller.
Eight big tech companies, including Adobe, IBM, Salesforce, and Nvidia, have pledged to conduct more testing and research on the risks of artificial intelligence (AI) in a meeting with White House officials, signaling a "bridge" to future government action on the issue. These voluntary commitments come amidst congressional scrutiny and ongoing efforts by the White House to develop policies for AI.
Senators Richard Blumenthal and Josh Hawley are holding a hearing to discuss legislation on regulating artificial intelligence (AI), with a focus on protecting against potential dangers posed by AI and improving transparency and public trust in AI companies. The bipartisan legislation framework includes creating an independent oversight body, clarifying legal liability for AI harms, and requiring companies to disclose when users are interacting with AI models or systems. The hearing comes ahead of a major AI Insight Forum, where top tech executives will provide insights to all 100 senators.
The US Securities and Exchange Commission (SEC) is utilizing artificial intelligence (AI) technologies to monitor the financial sector for fraud and manipulation, according to SEC Chair Gary Gensler.
AI has the potential to fundamentally change governments and society, with AI-powered companies and individuals usurping traditional institutions and creating a new world order, warns economist Samuel Hammond. Traditional governments may struggle to regulate AI and keep pace with its advancements, potentially leading to a loss of global power for these governments.
China's targeted and iterative approach to regulating artificial intelligence (AI) could provide valuable lessons for the United States, despite ideological differences, as the U.S. Congress grapples with comprehensive AI legislation covering various issues like national security, job impact, and democratic values. Learning from China's regulatory structure and process can help U.S. policymakers respond more effectively to the challenges posed by AI.
California Senator Scott Wiener is introducing a bill to regulate artificial intelligence (AI) in the state, aiming to establish transparency requirements, legal liability, and security measures for advanced AI systems. The bill also proposes setting up a state research cloud called "CalCompute" to support AI development outside of big industry.
Tesla CEO Elon Musk suggests the need for government regulation of artificial intelligence, even proposing the creation of a Department of AI, during a gathering of tech CEOs in Washington. Senate Majority Leader Chuck Schumer and other attendees also expressed the view that government should play a role in regulating AI. The options for regulation range from a standalone department to leveraging existing agencies, but the debate is expected to continue in the coming months.
The U.S. Department of Homeland Security is set to announce new limits on its use of artificial intelligence (AI) technology, aiming to ensure responsible and effective use while safeguarding privacy, civil rights, and civil liberties. The agency plans to adopt AI in various missions, including border control and supply chain security, but acknowledges the potential for unintended harm and the need for transparency. The new policy will allow Americans to decline the use of facial recognition technology and require manual review of AI-generated facial recognition matches for accuracy.
Spain has established Europe's first artificial intelligence (AI) policy task force, the Spanish Agency for the Supervision of Artificial Intelligence (AESIA), to determine laws and provide a framework for the development and implementation of AI technology in the country. Many governments are uncertain about how to regulate AI, balancing its potential benefits with fears of abuse and misuse.
The United Nations is urging the international community to confront the potential risks and benefits of Artificial Intelligence, which has the power to transform the world.
A bipartisan group of senators is expected to introduce legislation to create a government agency to regulate AI and require AI models to obtain a license before deployment, a move that some leading technology companies have supported; however, critics argue that licensing regimes and a new AI regulator could hinder innovation and concentrate power among existing players, similar to the undesirable economic consequences seen in Europe.
The Subcommittee on Cybersecurity, Information Technology, and Government Innovation discussed the federal government's use of artificial intelligence (AI) and emphasized the need for responsible governance, oversight, and accountability to mitigate risks and protect civil liberties and privacy rights.
Artificial intelligence (AI) requires leadership from business executives and a dedicated and diverse AI team to ensure effective implementation and governance, with roles focusing on ethics, legal, security, and training data quality becoming increasingly important.
Adversaries and criminal groups are exploiting artificial intelligence (AI) technology to carry out malicious activities, according to FBI Director Christopher Wray, who warned that while AI can automate tasks for law-abiding citizens, it also enables the creation of deepfakes and malicious code, posing a threat to US citizens. The FBI is working to identify and track those misusing AI, but is cautious about using it themselves. Other US security agencies, however, are already utilizing AI to combat various threats, while concerns about China's use of AI for misinformation and propaganda are growing.
President Biden has called for the governance of artificial intelligence to ensure it is used as a tool of opportunity and not as a weapon of oppression, emphasizing the need for international collaboration and regulation in this area.
Salesforce CEO Marc Benioff argues that the government needs to step up and regulate artificial intelligence before it becomes too powerful, citing the failures in regulating social media companies.
Governments worldwide are grappling with the challenge of regulating artificial intelligence (AI) technologies, as countries like Australia, Britain, China, the European Union, France, G7 nations, Ireland, Israel, Italy, Japan, Spain, the United Nations, and the United States take steps to establish regulations and guidelines for AI usage.
A new poll reveals that 63% of American voters believe regulation should actively prevent the development of superintelligent AI, challenging the assumption that artificial general intelligence (AGI) should exist. The public is increasingly questioning the potential risks and costs associated with AGI, highlighting the need for democratic input and oversight in the development of transformative technologies.
Artificial intelligence (AI) can be used to improve lives and address global challenges, such as poverty, hunger, and climate change, according to US Secretary of State Antony Blinken, who emphasized the need to use AI to achieve the Sustainable Development Goals (SDGs) in a speech at the New York Public Library. He highlighted the potential benefits of AI in various areas, including weather forecasting, agriculture, disease control, and clean energy, while acknowledging the risks and hazards associated with AI. The United States is committed to supporting AI innovation and governance, working with partners to develop international frameworks and involving a wide range of voices in the discussion. A new $15 million commitment has been made to help governments leverage AI for the SDGs.
The United States must prioritize global leadership in artificial intelligence (AI) and win the platform competition with China in order to protect national security, democracy, and economic prosperity, according to Ylli Bajraktari, the president and CEO of the Special Competitive Studies Project and former Pentagon official.
While many experts are concerned about the existential risks posed by AI, Mustafa Suleyman, cofounder of DeepMind, believes that the focus should be on more practical issues like regulation, privacy, bias, and online moderation. He is confident that governments can effectively regulate AI by applying successful frameworks from past technologies, although critics argue that current internet regulations are flawed and insufficiently hold big tech companies accountable. Suleyman emphasizes the importance of limiting AI's ability to improve itself and establishing clear boundaries and oversight to ensure enforceable laws. Several governments, including the European Union and China, are already working on AI regulations.
Pennsylvania state government is preparing to use artificial intelligence in its operations and is taking steps to understand and regulate its impact, including the formation of an AI governing board and the development of training programs for state employees.
The use of artificial intelligence for deceptive purposes should be a top priority for the Federal Trade Commission, according to three commissioner nominees at a recent confirmation hearing.
The U.S. government must establish regulations and enforce standards to ensure the safety and security of artificial intelligence (AI) development, including requiring developers to demonstrate the safety of their systems before deployment, according to Anthony Aguirre, the executive director and secretary of the board at the Future of Life Institute.