Discussions on regulation of artificial intelligence in the United States have included topics such as the timeliness of regulating AI, the nature of the federal regulatory framework to govern and promote AI, including what agency should lead, the regulatory and governing powers of that agency, and how to update regulations in the face of rapidly changing technology, as well as the roles of state governments and courts.[1]
As early as 2016, the Obama administration had begun to focus on the risks and regulations for artificial intelligence. In a report titled Preparing For the Future of Artificial Intelligence,[2] the National Science and Technology Council set a precedent to allow researchers to continue to develop new AI technologies with few restrictions. It is stated within the report that "the approach to regulation of AI-enabled products to protect public safety should be informed by assessment of the aspects of risk....".[3] These risks would be the principal reason to create any form of regulation, granted that any existing regulation would not apply to AI technology.
The first main report was the National Strategic Research and Development Plan for Artificial Intelligence.[4] On August 13, 2018, Section 1051 of the Fiscal Year 2019 John S. McCain National Defense Authorization Act (P.L. 115-232) established the National Security Commission on Artificial Intelligence "to consider the methods and means necessary to advance the development of artificial intelligence, machine learning, and associated technologies to comprehensively address the national security and defense needs of the United States."[5] Steering on regulating security-related AI is provided by the National Security Commission on Artificial Intelligence.[6] The Artificial Intelligence Initiative Act (S.1558) is a proposed bill that would establish a federal initiative designed to accelerate research and development on AI for, inter alia, the economic and national security of the United States.[7] [8]
On January 7, 2019, following an Executive Order on Maintaining American Leadership in Artificial Intelligence,[9] the White House's Office of Science and Technology Policy released a draft Guidance for Regulation of Artificial Intelligence Applications,[10] which includes ten principles for United States agencies when deciding whether and how to regulate AI.[11] In response, the National Institute of Standards and Technology has released a position paper,[12] and the Defense Innovation Board has issued recommendations on the ethical use of AI.[13] A year later, the administration called for comments on regulation in another draft of its Guidance for Regulation of Artificial Intelligence Applications.[14]
Other specific agencies working on the regulation of AI include the Food and Drug Administration,[15] which has created pathways to regulate the incorporation of AI in medical imaging.[16] National Science and Technology Council also published the National Artificial Intelligence Research and Development Strategic Plan,[17] which received public scrutiny and recommendations to further improve it towards enabling Trustworthy AI.[18]
In March 2021, the National Security Commission on Artificial Intelligence released their final report.[19] In the report, they stated that "Advances in AI, including the mastery of more general AI capabilities along one or more dimensions, will likely provide new capabilities and applications. Some of these advances could lead to inflection points or leaps in capabilities. Such advances may also introduce new concerns and risks and the need for new policies, recommendations, and technical advances to assure that systems are aligned with goals and values, including safety, robustness and trustworthiness. The US should monitor advances in AI and make necessary investments in technology and give attention to policy so as to ensure that AI systems and their uses align with our goals and values."
In June 2022, Senators Rob Portman and Gary Peters introduced the Global Catastrophic Risk Mitigation Act. The bipartisan bill "would also help counter the risk of artificial intelligence... from being abused in ways that may pose a catastrophic risk".[20] [21] On October 4, 2022, President Joe Biden unveiled a new AI Bill of Rights,[22] which outlines five protections Americans should have in the AI age: 1. Safe and Effective Systems, 2. Algorithmic Discrimination Protection, 3.Data Privacy, 4. Notice and Explanation, and 5. Human Alternatives, Consideration, and Fallback. The Bill was introduced in October 2021 by the Office of Science and Technology Policy (OSTP), a US government department that advises the president on science and technology.[23]
In July 2023, the Biden–Harris Administration secured voluntary commitments from seven companies – Amazon, Anthropic, Google, Inflection, Meta, Microsoft, and OpenAI – to manage the risks associated with AI. The companies committed to ensure AI products undergo both internal and external security testing before public release; to share information on the management of AI risks with the industry, governments, civil society, and academia; to prioritize cybersecurity and protect proprietary AI system components; to develop mechanisms to inform users when content is AI-generated, such as watermarking; to publicly report on their AI systems' capabilities, limitations, and areas of use; to prioritize research on societal risks posed by AI, including bias, discrimination, and privacy concerns; and to develop AI systems to address societal challenges, ranging from cancer prevention to climate change mitigation. In September 2023, eight additional companies – Adobe, Cohere, IBM, Nvidia, Palantir, Salesforce, Scale AI, and Stability AI – subscribed to these voluntary commitments.[24] [25]
The Biden administration, in October 2023 signaled that they would release an executive order leveraging the federal government's purchasing power to shape AI regulations, hinting at a proactive governmental stance in regulating AI technologies.[26] On October 30, 2023, President Biden released this Executive Order on Safe, Secure, and Trustworthy Artificial Intelligence. The Executive Order addresses a variety of issues, such as focusing on standards for critical infrastructure, AI-enhanced cybersecurity, and federally funded biological synthesis projects.[27]
The Executive Order provides the authority to various agencies and departments of the US government, including the Energy and Defense departments, to apply existing consumer protection laws to AI development.[28]
The Executive Order builds on the Administration’s earlier agreements with AI companies to instate new initiatives to "red-team" or stress-test AI dual-use foundation models, especially those that have the potential to pose security risks, with data and results shared with the federal government.
The Executive Order also recognizes AI's social challenges, and calls for companies building AI dual-use foundation models to be wary of these societal problems. For example, the Executive Order states that AI should not “worsen job quality”, and should not “cause labor-force disruptions”. Additionally, Biden’s Executive Order mandates that AI must “advance equity and civil rights”, and cannot disadvantage marginalized groups.[29] It also called for foundation models to include "watermarks" to help the public discern between human and AI-generated content, which has raised controversy and criticism from deepfake detection researchers.[30]
In January 2023, the New York City Bias Audit Law (Local Law 144[31]) was enacted by the NYC Council in November 2021. Originally due to come into effect on 1 January 2023, the enforcement date for Local Law 144 has been pushed back due to the high volume of comments received during the public hearing on the Department of Consumer and Worker Protection's (DCWP) proposed rules to clarify the requirements of the legislation. It eventually became effective on July 5, 2023.[32] From this date, the companies that are operating and hiring in New York City are prohibited from using automated tools to hire candidates or promote employees, unless the tools have been independently audited for bias.
On March 21, 2024, the State of Tennessee enacted legislation called the ELVIS Act, aimed specifically at audio deepfakes, and voice cloning.[33] This legislation was the first enacted legislation in the nation aimed at regulating AI simulation of image, voice and likeness.[34] The bill passed unanimously in the Tennessee House of Representatives and Senate.[35] This legislation's success was hoped by its supporters to inspire similar actions in other states, contributing to a unified approach to copyright and privacy in the digital age, and to reinforce the importance of safeguarding artists' rights against unauthorized use of their voices and likenesses.[36] [37]
In February 2024, Senator Scott Wiener introduced the Safe and Secure Innovation for Frontier Artificial Intelligence Models Act to the California legislature. The bill has the goal of reducing catastrophic risks by mandating safety tests for the most powerful AI models. If passed, the bill will also establish a publicly-funded cloud computing cluster in California.[38]
In 2016, Joy Buolamwini, AI researcher at Massachusetts Institute of Technology, shared her personal experiences with discrimination in facial recognition software at a TED Talk conference.[39] Facial recognition software is vastly understood to be inaccurate in its identification of darker-skinned peoples, which matters especially in the context of policing, the criminal justice system, healthcare system, and employment sectors. [40]
In 2022, the PEW Research Center's study of Americans revealed that only 18% of respondents are more excited than they are concerned about AI.[41] Biases in AI algorithms and methods that lead to discrimination are causes for concern among many activist organizations and academic institutions. Recommendations include increasing diversity among creators of AI algorithms and addressing existing systemic bias in current legislation and AI development practices.[42]