Anthropic just announced Project Glasswing, an AI model that has apparently found security vulnerabilities in every major operating system and web browser. That is not a small claim. The model, called Claude Mythos Preview, was not specifically built for cybersecurity but turns out to be exceptionally good at it anyway, thanks to its reasoning and agentic coding capabilities. Anthropic is keeping it locked to a tight circle of "defensive security" partners including Nvidia, Google, AWS, Apple, and Microsoft, precisely because the same tool that finds vulnerabilities could absolutely be weaponised to exploit them.
Here is the part that deserves your attention. This is not a niche security product for a niche security team. This is a general purpose model that stumbled into being genuinely dangerous in the best possible way. That tells you something important about where AI capability development is heading: the gap between "useful assistant" and "nation state level threat analyst" is getting uncomfortably thin.
The restricted rollout is a deliberate choice, and a revealing one. Anthropic is essentially admitting that releasing this publicly would be like handing a lockpick set to everyone on the street and hoping most of them are locksmiths. The fact that major tech companies are lining up as partners suggests this is not a PR stunt. Organisations are genuinely scrambling to get ahead of AI powered attacks before the offensive side of this equation becomes democratised.