Skip to content
Tech News
← Back to articles

Spooked by Mythos, Trump suddenly realized AI safety testing might be good

read original get AI Safety Testing Kit → more articles
Why This Matters

This shift in policy highlights a growing recognition within the tech industry and government of the importance of AI safety and regulation, especially as advanced models pose increasing security and ethical risks. It signals a potential move towards more structured oversight of frontier AI systems, balancing innovation with safety concerns for consumers and national security. The collaboration between government agencies and leading AI firms underscores the need for rigorous testing to prevent misuse and ensure responsible AI development.

Key Takeaways

This week, the Trump administration backpedaled and signed agreements with Google DeepMind, Microsoft, and xAI to run government safety checks on the firms’ frontier AI models before and after their release.

Previously, Donald Trump had stubbornly cast aside the Biden-era policy, dismissing the need for voluntary safety checks as overregulation blocking unbridled innovation. Soon after taking office, he took the extra step of rebranding the US AI Safety Institute to the Center for AI Standards and Innovation (CAISI), removing “safety” from the name in a pointed jab at Joe Biden.

But after Anthropic announced that it would be too risky to release its latest Claude Mythos model—fearing that bad actors might exploit its advanced cybersecurity capabilities—Trump’s suddenly concerned about AI safety. According to White House National Economic Council Director Kevin Hassett, Trump may soon issue an executive order mandating government testing of advanced AI systems prior to release, Fortune reported.

In CAISI’s press release, the center acknowledges that the voluntary agreements signed by Google, Microsoft, and xAI “build on” Biden’s policy. Celebrating the new partnerships, CAISI Director, Chris Fall, did not mention Mythos but promised that the “expanded industry collaborations” would help CAISI scale its work “in the public interest at a critical moment.”

“Independent, rigorous measurement science is essential to understanding frontier AI and its national security implications,” Fall said.

To date, CAISI said it has completed about 40 evaluations, including those of frontier models that have yet to be released. When conducting tests, CAISI frequently gains access to models with “reduced or removed safeguards,” which CAISI said allowed them to more “thoroughly evaluate national security-related capabilities and risks.”

Through the evaluations, the government will also gain a better understanding of model capabilities, CAISI claimed. And to ensure that evaluators understand top national security concerns as they emerge across government, a “group of interagency experts” has formed a task force “focused on AI national security concerns,” CAISI said.