The United States will launch a AI safety institute to evaluate known and emerging risks of so-called “frontier” artificial intelligence models.
“I will almost certainly be calling on many of you in the audience who are in academia and industry to be part of this consortium,” secretary of commerce Gina Raimondo said in a speech to the AI Safety Summit in Britain.
“We can’t do it alone, the private sector must step up.”
Raimondo added that she would also commit to the US institute establishing a formal partnership with the United Kingdom Safety Institute.
The new effort will be under the National Institute of Standards and Technology (NIST) and lead the US government’s efforts on AI safety, especially for reviewing advanced AI models.
The institute “will facilitate the development of standards for safety, security, and testing of AI models, develop standards for authenticating AI-generated content, and provide testing environments for researchers to evaluate emerging AI risks and address known impacts,” the department said.
President Joe Biden on Monday signed an artificial intelligence executive order, requiring developers of AI systems that pose risks to US national security, the economy, public health or safety to share the results of safety tests with the US government, in line with the Defense Production Act, before they are released to the public.
The order also directs agencies to set standards for that testing and address related chemical, biological, radiological, nuclear, and cyber security risks.