Anthropic’s “Too Dangerous” AI Was Accessed by Guessing the URL

Anthropic’s “Too Dangerous” AI Was Accessed by Guessing the URL

Boing Boing
Boing BoingApr 23, 2026

Key Takeaways

  • Anthropic's Mythos claim disproved; public Claude Opus found bug
  • Bug discovered via simple URL guess, not secret access
  • Highlights public models' capability in security research
  • Undermines trust in "dangerous AI" marketing narrative

Pulse Analysis

Anthropic’s recent publicity around Mythos—a restricted AI it warned could be "too dangerous"—has taken an unexpected turn. A 244‑page system card revealed that the celebrated Linux kernel flaw was actually identified by Claude Opus 4.6, a model anyone can reach by merely guessing a URL. The discovery was made public by researcher Devansh, who demonstrated that the vulnerability was not the product of a secret, high‑risk model but rather of a widely accessible service. This nuance strips away the mystique surrounding Mythos and forces a reassessment of Anthropic’s marketing narrative.

The episode shines a spotlight on a broader debate in AI safety: the line between proprietary, gated models and openly available ones. While companies often argue that restricting access mitigates misuse, the Claude Opus case shows that public models can already perform sophisticated tasks such as vulnerability discovery. Security researchers can leverage these tools without special permissions, suggesting that openness may accelerate defensive research rather than exacerbate threats. At the same time, the incident raises questions about how firms quantify and communicate risk, especially when hype can outpace technical reality.

For investors and regulators, the takeaway is clear: claims of exclusive, dangerous capabilities need rigorous verification. Anthropic’s credibility has taken a hit, which could influence partnership decisions and funding allocations in the competitive generative‑AI market. Moreover, the incident may prompt policymakers to scrutinize how AI firms label and manage “high‑risk” models, potentially leading to clearer guidelines on transparency and responsible disclosure. As the industry matures, balancing innovation with honest risk communication will be essential for sustained growth and public trust.

Anthropic’s “too dangerous” AI was accessed by guessing the URL

Comments

Want to join the conversation?