Home Stories About RSS Feed
2 min read

Anthropic Withholds Claude Mythos — Its Most Powerful Model Ever — Over Cybersecurity Risks

Back to News

Anthropic has done something no major AI lab has done before: publicly withheld a frontier model from release because it was deemed too dangerous.

Claude Mythos Preview, the successor to the Claude Opus line, demonstrated a “step change” in capability — particularly in cybersecurity. During internal evaluations, the model autonomously discovered and exploited complex zero-day vulnerabilities across major operating systems and web browsers, a feat previously requiring elite human security researchers working over extended periods.

Project Glasswing

Rather than shelving the model entirely, Anthropic created Project Glasswing — a gated consortium that provides restricted access to select technology and security partners. The roster includes Google, AWS, Apple, Microsoft, NVIDIA, Cisco, CrowdStrike, and JPMorgan Chase.

The premise is straightforward: if a model this capable exists, it’s better to use it defensively — identifying and patching critical vulnerabilities globally — than to either release it broadly or pretend the capability doesn’t exist.

Partners access the model under strict controls, with usage limited to:

Why This Matters

This is the first time a major AI company has drawn a clear line between “what we can build” and “what we should release.” The decision reframes the AI safety debate from theoretical alignment discussions to concrete, present-day capability management.

Key implications:

The Precedent

Industry observers have called this a “watershed moment” in AI governance. Until now, the competitive pressure to release has overridden caution at every major lab. Anthropic’s decision to absorb the reputational and commercial cost of not releasing its best model sets a precedent that will be difficult to ignore — and even harder to follow.

Whether other labs adopt similar restraint, or use Anthropic’s caution as an opportunity to capture market share with less restricted offerings, will define the next phase of the AI race.


Source: anthropic.com, infoq.com, cfr.org