spoonai
TOPCAISINISTAI Safety

CAISI Signs Pre-Deployment AI Safety Deals with Google DeepMind, Microsoft, and xAI

On May 5, NIST's CAISI announced pre-deployment evaluation agreements with Google DeepMind, Microsoft, and xAI. OpenAI and Anthropic renegotiated their 2024 deals to align with the Trump AI Action Plan — bringing all five major U.S. frontier labs into voluntary government pre-release testing.

·8분 소요·NISTNIST
공유
CAISI frontier AI pre-deployment evaluations — Google, Microsoft, xAI sign on; OpenAI, Anthropic renegotiate
Source: NIST

All Five U.S. Frontier Labs Now Have a Government Stamp

Here's the deal: on May 5, the Commerce Department's CAISI (housed within NIST) announced that Google DeepMind, Microsoft, and xAI signed pre-deployment evaluation agreements covering cyber, biosecurity, and chemical-weapons risks. Same day, OpenAI and Anthropic renegotiated their August 2024 deals to align with the Trump administration's AI Action Plan. Net result: every major U.S. frontier lab now participates in voluntary pre-release government testing. This is a meaningful shift — Trump's previously light-touch posture on AI has tilted toward "government inspects first." Anthropic's Mythos preview surfacing thousands of high-severity vulnerabilities autonomously in April was the trigger.

The Players — CAISI, Five Frontier Labs, the White House

CAISI was set up under NIST in 2024, originally as the "U.S. AI Safety Institute," then rebranded under the Trump administration to "Center for AI Standards and Innovation" — dropping "safety" for "standards and innovation." Has completed 40+ model evaluations. Focus areas: cybersecurity, biosecurity, chemical weapons (CBRN). Distinctive practice — receives models with safeguards reduced or fully removed, to model worst-case scenarios. Findings flow to TRAINS Taskforce (DOD, CIA, NSA, DOE, etc.).

Google DeepMind signed for the first time. CEO Demis Hassabis published an open letter on AI safety governance in April; this agreement is the follow-on. The implication: Gemini 3 launch could face a 30-90 day pre-release evaluation window.

Microsoft is interesting because it's primarily known as an OpenAI redistributor via Azure, but explicitly included its own Phi family and forthcoming proprietary frontier models in scope. That's a structural acknowledgement that Microsoft is moving from "OpenAI dependence" toward "internal frontier models."

xAI's signing reflects the Musk-Trump alignment. With xAI Grok 4/5 in evaluation scope, plus the same-week Anthropic-SpaceX compute deal, Musk is now central to both the AI infrastructure and policy axes simultaneously.

OpenAI and Anthropic re-papered their August 2024 MOUs. Key changes: voluntary submission → mandatory pre-notification, public results → confidential by default, fully government-funded → some company cost-sharing.

NIST's release states CAISI will conduct pre-deployment evaluations and targeted research to better assess frontier AI capabilities and advance AI security.

The Mechanics — 5 Labs, 3 Domains, Mandatory Pre-Notification

Lab Agreement Date Evaluation Domains Notes
OpenAI 2024-08 → 2026-05 renewal Cyber·Bio·Chem Some cost-sharing
Anthropic 2024-08 → 2026-05 renewal Cyber·Bio·Chem Deepest cyber assessment
Google DeepMind 2026-05-05 new Cyber·Bio·Chem Applies to Gemini 3
Microsoft 2026-05-05 new Cyber·Bio·Chem Includes proprietary frontier models
xAI 2026-05-05 new Cyber·Bio·Chem Grok 4/5 in scope

Cyber means autonomous vulnerability discovery, exploit writing, and network penetration. Bio/chem means evaluating capability to design pathogens or chemical weapon precursors. Critically, evaluations run with "safeguards reduced or removed" — measuring how far the underlying model can go without alignment guardrails.

Workflow: company submits 30-60 days pre-launch → CAISI runs 7-9 benchmarks (mix of public and confidential) → results flow to TRAINS Taskforce → if national-security signals trigger, launch could be blocked or modified → company has 30 days to respond. Results stay confidential, but "evaluation completed" status is disclosed.

The biggest change is the "mandatory" element. The 2024 MOUs were voluntary submission; the 2026 renewals require pre-notification. That puts hard floors under launch timelines for GPT-6, Claude Opus 5, and Gemini 3.

Who Wins — Government, Labs, Allies

U.S. government wins twice. First, frontier capability visibility — see what GPT-6, Claude Opus 6, Gemini 3 can really do before launch, then adjust defense posture, intelligence collection, and export controls. Second, international leverage — having all five U.S. labs in a government inspection regime gives the U.S. a reference model in negotiating AI governance with the U.K., E.U., Japan, etc.

Frontier labs get regulatory clarity. Knowing exactly which domains evaluate, and what triggers a launch block, sharpens R&D investment priorities. Labs also get implicit competitive protection — only U.S. labs are on the inside, raising barriers for foreign entrants in U.S. government procurement. Costs: 30-90 day launch delays plus shared evaluation expenses.

Allies (U.K., E.U., Japan, Australia) signal: "U.S. models can be imported safely." U.K. AISI has shared evaluation results with U.S. CAISI since 2024; this expansion deepens that pipeline.

China and Russia get two messages: U.S. government visibility into model capability sharpens military application risk, and U.S. is locking down its own labs — implying export controls and tech restrictions will tighten further. Both messages accelerate Chinese frontier model ramps (DeepSeek V4, Qwen 4, MiniMax).

Past Parallels — Wins and Losses

FDA pre-market clinical trial mandate (1962): post-thalidomide, U.S. mandated FDA review pre-launch. Over 50 years, U.S. pharma cemented global #1 — pre-market review didn't kill competitiveness, it strengthened it. Pro-evaluation analogy.

NPT + IAEA inspections (1968-present): nuclear states accept inspections in exchange for peaceful-use rights and tech-sharing privileges. U.S./Russia/U.K./France/China cooperated, and nonproliferation worked. AI governance could follow a "5 powers + inspection" structure.

Internet self-regulation (1996-2018): Section 230 left platforms self-regulated, and disinformation, harassment, and child safety problems metastasized. AI cannot be self-regulated without similar outcomes — the rationale for mandatory evaluation.

GDPR Phase 1 ramp (2018-2020): E.U. tightened data regulation but the first two years were ambiguous, costly, and disruptive. CAISI's first 1-2 years could see similar friction.

Counter-Plays — China, E.U., U.K.

China builds parallel evaluation. CAC operates pre-launch model registration since 2024, but it's content-censorship rather than capability evaluation. Reports suggest a Chinese capability evaluation center is under consideration — following the U.S. CAISI pattern while building separate Chinese standards.

The E.U. is in AI Act implementation, with broader "general-purpose AI model" coverage than CAISI's frontier focus. But E.U. evaluation infrastructure trails CAISI by 6-12 months. Short term, the U.S. is setting the global AI governance standard.

The U.K. AISI has been sharing results with CAISI since 2024, and its data pool just got bigger. U.K. AISI's own evaluation capacity is ~30-40% of U.S. capacity, so de facto reliance on U.S. evaluations will continue.

Canada, Australia, Japan, and Korea receive subsets via 5-Eyes/AUKUS/QUAD channels. Korea may announce an AI Safety Evaluation Institute by end-2026; Japan may use NEDO as host.

What Changes — Devs, Founders, Investors, End Users

Devs: launches of GPT-6, Claude Opus 6, Gemini 3 likely slip 30-90 days for evaluation. AI alignment and safety engineering hiring picks up — getting strong CBRN-domain evaluation scores requires more alignment R&D.

Founders: AI application startups in regulated industries (finance, healthcare, legal) get pulled toward CAISI-evaluated models, and Chinese models effectively get walled out of U.S. federal procurement. New "AI safety engineering as a service" startups likely fundraise across 2026.

Investors: AI safety/alignment/evaluation is now a real category. Frontier labs (OpenAI, Anthropic, Google) face slight launch-delay drag on revenue ramp but stronger moats against new entrants — a wash to slight positive.

End users: model trust improves (only government-evaluated models reach market), but trade-off is slower release cadence.

Stakes

  • Wins: Howard Lutnick (Commerce) — all five labs in government evaluation; CAISI/NIST — institutional expansion; allied nations — leveraged U.S. evaluations.
  • Loses: Five U.S. labs — launch delays + cost-sharing; Chinese labs (DeepSeek, Alibaba, MiniMax) — sharper U.S. market barriers; E.U. AI Act — losing global standard-setting leverage to U.S.
  • Watching: Korea/Japan governments — own evaluation infrastructure timing; UN/OECD — global AI governance frameworks; academia (Bengio, Hinton) — judging if mandatory evaluation actually improves safety.

The Skeptics — "Pre-Evaluation = Censorship and Protectionism"

Free-market voices like Marc Andreessen (a16z) frame mandatory pre-evaluation as government censorship plus de facto protectionism — only U.S. Big 5 labs in scope, raising barriers for entrants like Reflection or Mistral and entrenching a cartel. Confidentiality of results compounds opacity.

Skeptics like Yann LeCun (Meta AI Chief) argue current LLMs don't actually pose meaningful CBRN risks, making evaluation more political performance than real safety work. Confidential results also block academic verification.

Two skeptic lines: (1) mandatory evaluation = entry barrier + Big 5 cartel, (2) current capability levels make CBRN evaluations theatrical. Both converge on "this is protectionism dressed as safety."

TL;DR

  • CAISI (NIST) signed pre-deployment evaluation deals with Google, Microsoft, xAI on May 5; OpenAI and Anthropic renegotiated — all five U.S. frontier labs in scope.
  • Cyber, bio, and chemical-weapons evaluation domains; mandatory pre-notification, confidential results, partial cost-sharing.
  • GPT-6, Claude Opus 6, Gemini 3 launch timelines could shift 30-90 days; alignment/safety hiring increases.

References

관련 기사

무료 뉴스레터

AI 트렌드를 앞서가세요

매일 아침, 엄선된 AI 뉴스를 받아보세요. 스팸 없음. 언제든 구독 취소.

매일 30개+ 소스 분석 · 한국어/영어 이중 언어광고 없음 · 1-클릭 해지