Anthropic's AI Model, Claude Mythos Preview, Raises Global Cybersecurity Concerns

On April 7, 2026, Anthropic announced that its latest general-purpose large language model, Claude Mythos Preview, had demonstrated unprecedented—and unintended—capabilities in cybersecurity. The artificial intelligence system autonomously identified and exploited software vulnerabilities at an unprecedented rate, igniting widespread concern among governments, the public, and the IT sector about AI's potential to undermine global cybersecurity.

Anthropic deemed the model too risky for immediate public release, citing a moral responsibility to disclose these vulnerabilities. Instead, the company granted exclusive access to tech giants for testing under Project Glasswing, a controlled evaluation process.

Mythos' Capabilities: A Reflection of Systemic Fragility

As a cybersecurity researcher, I view Mythos' capabilities as impressive but not revolutionary. The model does not introduce a new threat; rather, it mirrors existing vulnerabilities and human behaviors in modern systems. During controlled evaluations, engineers with minimal security experience used Mythos to scan thousands of software codebases for vulnerabilities. The model autonomously conducted multistep, complex attacks that typically take human experts weeks or months to develop.

Key Findings from Mythos' Evaluations

Mythos demonstrated striking capabilities, including:

  • Discovering 271 vulnerabilities in Mozilla’s Firefox, with exploits developed for 181 of them.
  • Identifying thousands of zero-day vulnerabilities in major operating systems, web browsers, and applications—flaws that remain unpatched and exploitable immediately.
  • Uncovering a 27-year-old dormant security flaw in OpenBSD, a security-focused operating system, and a 16-year-old bug in FFmpeg, a video/audio processing tool. Some of these flaws allow unauthenticated users to gain control of host machines.
  • Enabling relatively inexperienced engineers to complete attacks overnight—from vulnerability identification to exploitation—a process that typically takes human experts weeks.

In an evaluation by the AI Security Institute, Mythos successfully took over a simulated corporate network in three out of 10 attempts, marking the first time an AI model achieved this feat. National Security Agency officials testing Mythos praised the tool’s speed and efficiency in finding software vulnerabilities, according to a news report.

Anthropic's Response: Restricting Public Access

Anthropic’s announcement of Mythos and its cybersecurity implications drew significant media attention. The company’s decision to restrict public access to the model sparked debates about AI governance, ethical responsibilities, and the balance between innovation and security. By granting exclusive access to tech giants under Project Glasswing, Anthropic aims to mitigate risks while further evaluating the model’s capabilities.

The results of Mythos’ evaluations highlight both its potential and the systemic vulnerabilities in modern software. While Mythos does not represent a radical departure in cybersecurity threats, its ability to autonomously identify and exploit flaws at scale underscores the urgent need for improved security practices and AI governance.