U.S. expands pre-release AI safety testing program to Google, Microsoft and xAI models
- Both agree
- Early access lets federal scientists evaluate unreleased frontier AI models for capabilities and security risks before public launch, extending an existing arrangement to more major companies and shifting oversight earlier than post-release scrutiny alone would allow.
- They split on
- Whether the story is about building public safeguards around advanced AI before deployment, or about keeping security testing voluntary rather than letting it become a more formal government review process.
The Facts
- Google DeepMind, Microsoft and xAI agreed to provide the U.S. government early access to unreleased AI models for review before public release.
- The reviews will be conducted through the Center for AI Standards and Innovation (CAISI) within the U.S. Department of Commerce.
- Under the agreements, CAISI will conduct pre-deployment evaluations and targeted research to assess model capabilities and security risks.
- The new agreements expand an existing voluntary testing arrangement that already included OpenAI and Anthropic.
- A central purpose of the testing is to examine national security-related risks from advanced AI systems, including cybersecurity threats.
- The program matters because it gives federal scientists a way to assess frontier AI systems before launch, rather than only after the tools are already public.
- The arrangements are voluntary agreements with the companies, not a mandatory government approval system for AI models.
- What remains unresolved is whether the Trump administration will move beyond these voluntary agreements to create a more formal government review process for AI models.
Context
What will the government actually do with these AI models?
CAISI says it will perform pre-deployment evaluations and targeted research on unreleased models to better understand their capabilities and security risks before launch POLITICO,Reuters,CNN International.
Which companies were already participating before this expansion?
OpenAI and Anthropic were already voluntarily working with CAISI to test unreleased models for vulnerabilities, and the new deals extend that program to Google DeepMind, Microsoft and xAI Reuters,BBC,U.S. News & World R….
Why is the U.S. focusing on these reviews now?
Sources say officials are focused on demonstrable risks from advanced AI, especially cybersecurity and other national security dangers, and some reports tie the urgency to concerns raised by Anthropic's newly unveiled Mythos model Reuters,Al Jazeera Online,CNN International.
View all 100 sources
Wire services (6)
Independent coverage (50)
About these frames
See this differently than someone you know would? Two ways to keep it going.
The dial works on any URL — paste an article you read elsewhere this week.