8 April 2026

7 min. Reading Time

Anthropic has developed an AI model, Claude Mythos, capable of identifying vulnerabilities within critical infrastructure-bugs embedded in code for 27 years that slipped past millions of automated tests. Crucially, the model was not trained specifically as a security tool. It remains fundamentally a coding model, with these security capabilities emerging as an unintended side effect. Rather than releasing it publicly, Anthropic is granting defenders a strategic head start. What this means for the cloud industry.

Key Takeaways

  • Claude Mythos achieves 93.9 percent on SWE-bench Verified – compared to 80.8 percent for Opus 4.6 (Anthropic, April 2026).
  • On the CyberGym benchmark for Vulnerability Detection, Mythos scores 83.1 percent. These security capabilities are a byproduct of superior code competence.
  • The model identified a 27-year-old bug in OpenBSD and a 16-year-old bug in FFmpeg that 5 million automated tests had overlooked.
  • Project Glasswing grants controlled access to over 40 organizations (AWS, Google, Microsoft, Apple, NVIDIA, CrowdStrike) – a Defender-First approach.
  • $100 million in usage credits and $4 million for open-source security committed. All findings will be published within 90 days.

A Coding Model Becomes the Most Powerful Security Scanner

Claude Mythos wasn’t trained as a hacker. Anthropic optimized the model to understand and write code better than any other available system. On SWE-bench Verified-the industry benchmark for bug-fixing capabilities-Mythos achieves 93.9 percent. Opus 4.6, the most powerful public model to date, scores 80.8 percent.

The leap from 80 to 94 percent sounds like optimization. In practice, it is a category shift. At these levels, the model solves nearly every real-world software task you give it.

The real news lies in the side effect: whoever understands code at this level also understands where code breaks. On the CyberGym benchmark for Vulnerability Detection, Mythos reaches 83.1 percent-compared to 66.6 percent for Opus. Anthropic compares this to a locksmith skilled enough to open every lock without ever having been a burglar.

93.9 %
SWE-bench Verified (Mythos)

83.1 %
CyberGym Vuln Detection

27 Years
Oldest Found Bug

What Mythos Discovered in Practice

Three findings reveal the magnitude.

In OpenBSD, Mythos uncovered a vulnerability in the SACK (Selective Acknowledgment) implementation of the TCP stack-a mechanism for handling packet loss. It was 27 years old. The flaw enables a remote Denial of Service attack against any OpenBSD server. OpenBSD is regarded as one of the most secure operating systems. Its code is regularly audited by experienced security researchers. Nevertheless, the bug remained undetected.

In FFmpeg-the multimedia library providing video and audio processing for practically every platform-Mythos found a bug in the H.264 codec. It was 16 years old. 5 million automated tests had traversed the affected area without triggering the vulnerability. FFmpeg processes the majority of video traffic on the internet. FFmpeg confirmed and patched the bug following the report.

Additionally, Mythos identified several privilege escalation vulnerabilities in the Linux kernel, including one in the DRR (Deficit Round Robin) scheduler-an algorithm for distributing network bandwidth. An unprivileged user could gain full root access via this path.

The Market Shift: Code Competence Becomes Security Competence

This is the point where this announcement stops being an Anthropic press release and starts affecting the cloud industry.

If code competence automatically generates security competence, then every next generation of coding models-whether from Anthropic, OpenAI, Google, or Open Source-will develop similar capabilities. This is not a feature that gets switched on. It is an emergent property that grows with code quality.

For cloud infrastructure, this means three things.

First, major providers will reinforce their security scans with this technology. AWS, Azure, and GCP are Glasswing partners. Patches for previously unknown infrastructure bugs will arrive in the coming months. Cloud teams benefit automatically.

Second, the bar for your own code rises. If an AI model finds 27-year-old bugs that manual audits missed, “we conducted a penetration test” is no longer sufficient as proof of security. CI/CD pipelines will have to integrate AI-supported security scans as standard gates.

Third, a new category of security tools is emerging. The market for AI-supported vulnerability detection will fundamentally change within 12 to 24 months. Not because Mythos is a specialized security tool, but because it shows that every good coding model automatically becomes a good security model too.

“Mythos is very powerful and should feel terrifying. I am proud of our approach to responsible deployment.”
– Boris Cherny, Head of the Claude Code Team at Anthropic (April 2026)

Project Glasswing: Controlled Distribution to Defenders

Anthropic has opted against a public release. Instead, Mythos operates under the Project Glasswing banner within a controlled environment. The partner roster: AWS, Apple, Google, Microsoft, NVIDIA, Cisco, CrowdStrike, JPMorgan Chase, Palo Alto Networks, Broadcom, the Linux Foundation, and more than 30 additional organizations.

The rationale: If a model identifies vulnerabilities in the infrastructure powering the internet, the operators of that infrastructure must receive the findings first. Three concrete commitments: All findings will be shared publicly within 90 days. $100 million USD in usage credits are available to partners. $4 million USD flow directly to open-source security communities.

Patches are already rolling out. FFmpeg has confirmed and patched the 16-year-old bug. OpenBSD and Linux kernel fixes are currently in deployment.

What Cloud Teams Should Do Now

In the short term, there is little to do. Organizations with automatic updates enabled are largely protected. The Glasswing patches are delivered through the standard update channels of cloud providers and operating system distributions.

Over the medium term, engineering teams should evaluate AI-assisted code analysis within their tool stacks. This is not because Mythos will be available tomorrow, but because the next generation of coding assistants will incorporate similar capabilities. Those who prepare their CI/CD pipeline today for AI-assisted security scans gain an advantage.

In the long run, the balance of power is shifting. Defenders are acquiring tools that were previously available only to elite attackers. However, the same technology becomes available to attackers once comparable open-source models emerge. Security remains an arms race. The difference: This time, defenders have the upper hand.

Frequently Asked Questions

What is Claude Mythos, and how does it differ from Claude Opus?

Claude Mythos is Anthropic’s next model generation, achieving 93.9 percent on SWE-bench Verified – compared to 80.8 percent for Opus. Security capabilities aren’t a trained feature but rather a byproduct of superior code competence.

Why isn’t Anthropic releasing the model publicly?

A model capable of finding vulnerabilities in critical infrastructure and combining them into exploit chains would be an attack tool in the wrong hands. Through Project Glasswing, defenders get access first so bugs can be patched before becoming public knowledge.

Do smaller companies also benefit from Project Glasswing?

Indirectly, yes. The patches roll out via regular updates from major platforms. Anyone using AWS, Azure, or GCP benefits automatically. Currently, direct access is limited to over 40 partner organizations.

Will other AI labs develop similar models?

The Mythos results indicate that security capabilities are a byproduct of superior code competence. Every subsequent frontier coding model will develop similar capabilities. Whether OpenAI, Google, and Meta choose the same controlled approach remains to be seen.

What concrete steps should DevOps teams take now?

Keep all systems up to date – the Glasswing patches arrive via regular updates. Evaluate AI-driven security scans for integration into the CI/CD pipeline. Adjust your internal security benchmarks: If AI can find 27-year-old bugs, an annual penetration test is no longer sufficient.

More from the MBF Media Network

Source Cover Image: AI-generated mood image (FLUX.2) – no product depiction

Also available in

A magazine by Evernine Media GmbH