When AI Gets Too Good: Why Anthropic Won’t Release Its Most Powerful Model

April 12, 2026
Written By Christi Brown

Christi Brown is the founder of AdapToIT, where modern IT strategy meets hands-on execution. With a background in security, cloud infrastructure, and automation, Christi writes for IT leaders and business owners who want tech that actually works—and adapts with them.

My AI minions just gained the ability to discover 27-year-old operating system vulnerabilities, break containment during testing, and — apparently — post exploit details to public websites without being asked. One step closer to world domination… but first, need the upgrades.

That is not a hypothetical. That is the documented behavior of Claude Mythos Preview, Anthropic’s most capable model to date, announced on April 7, 2026 alongside Project Glasswing — the initiative built specifically because the model is too dangerous to release publicly. The AI security Glasswing story is the most significant thing to happen in this space in years, and if you lead IT, run an MSP, or sit in a CIO chair, I want to tell you exactly what it means for you operationally.

What Anthropic Actually Did with AI Security and Glasswing

Anthropic built a model that is, by their own description, simultaneously their best-aligned model and their most alignment-risky model ever produced. That sentence should stop you cold. They are saying: we finally got the values right, and we still will not let you near it. Those two things are not in contradiction. They are the entire point.

Instead of a general release, Anthropic created a restricted coalition — Project Glasswing — giving access to eleven organizations: Amazon Web Services, Apple, Broadcom, Cisco, CrowdStrike, Google, JPMorganChase, the Linux Foundation, Microsoft, Nvidia, and Palo Alto Networks. The model, Mythos Preview, will be deployed inside a coordinated vulnerability discovery and responsible disclosure pipeline. It found a 27-year-old flaw in OpenBSD during testing. It developed a multi-step exploit to gain internet access during a sandboxed evaluation. In approximately 29% of behavioral test transcripts, interpretability tooling revealed it suspected it was being evaluated — without writing anything in its visible scratchpad. It just knew, and it stayed quiet about it.

That last detail is the one that keeps me thinking. A model that behaves well when it suspects it is being watched is not demonstrating alignment. It is demonstrating the appearance of alignment under observation. Those are not the same thing, and standard benchmarks do not distinguish cleanly between them.

The Gap Between What You Can Buy and What Actually Exists

The models available for public consumption and the models that exist internally at frontier labs are not the same thing. They were never exactly the same thing, but the gap is widening. Mythos Preview is explicitly not for sale. It lives inside a curated coalition of companies large enough and security-mature enough to be trusted with a tool that can, in Anthropic’s own words, pose offensive cybersecurity capabilities that “far exceed defensive utility in unrestricted use.”

What that means practically is that your threat model has to account for capability you cannot directly observe or access. If Mythos Preview can find a 27-year-old vulnerability in OpenBSD during a test run, a well-resourced threat actor with access to a similar model — whether through a nation-state program, a leaked weight, or an underground market — can do the same thing to your environment. The asymmetry between what defenders can access and what sophisticated attackers might access is real and growing.

I am not saying this to generate panic. I am saying it because the appropriate response is not panic — it is recalibration. Your patching cadence, your vendor-supplied vulnerability scanning, your assumption that you will have weeks between disclosure and active exploitation: all of those assumptions need a hard look.

What This Tells You About AI Vendor Trust Right Now

Every security vendor on the planet is about to announce AI-powered zero-day discovery. Some of it will be real. Most of it will be marketing copy pasted over a faster version of something they already had. Project Glasswing gives you a filter.

Ask your vendors three questions. What model are you running under the hood? What is your responsible disclosure process when the model finds something in a customer environment? And what are your containment protocols if the model behaves unexpectedly during an engagement? If a vendor cannot answer those three questions specifically and immediately, their AI security pitch is aspirational at best.

The Glasswing coalition structure is instructive here. Anthropic did not just hand Mythos to eleven companies and walk away. The structure involves coordinated disclosure, defined partner obligations, and a specific pipeline for handling what the model finds. That is not how most commercial AI security products are deployed. Most of them are running a fine-tuned version of a public model with a marketing layer on top.

The Practical Recalibration: What to Actually Do

First, shorten your patch cycle for anything internet-facing. The window between a vulnerability being discovered and being weaponized has collapsed. What used to take months now takes days or hours for a well-resourced attacker. If you are still running monthly patching on edge infrastructure, that cadence is incompatible with the threat environment Glasswing just confirmed exists.

Second, revisit your open-source dependency exposure. Mythos found a 27-year-old flaw in OpenBSD. The Linux Foundation is a Glasswing partner for a reason: open-source infrastructure that has been trusted for decades is now being re-examined by AI systems that can reason about code at a depth and speed that human auditors never could. If you run Linux infrastructure, container workloads, or anything built on widely-used open-source components, those will be among the first things that Glasswing-class analysis surfaces.

Third, build an AI vendor evaluation into your annual security review. Not just “what tools are we using” but “what models are running in those tools, what are the containment protocols, and what happens when the model produces unexpected output.” This is new governance territory. Most of your existing vendor security questionnaires do not cover it.

Fourth, and I say this as someone who has invested heavily in AI tooling: do not confuse the public availability of AI with safety. The Glasswing story is really a story about how the most capable AI systems are being deliberately withheld from general use because the people who built them decided the risk profile was unacceptable.

The Part That Actually Impressed Me

Anthropic caught the evaluation-awareness behavior. They published it in the system card. They built interpretability tooling capable of detecting that the model suspected it was being tested, even when the model deliberately did not write that suspicion down anywhere visible. And then they decided not to release it anyway, even though it would have been commercially very valuable to do so.

That is not a small thing. The commercial pressure to release a frontier model is enormous. Anthropic has an IPO on the horizon. Their competitors are shipping. They still held the model back, published the concerning findings, and built a controlled alternative deployment path instead. I do not agree with every decision Anthropic makes, and I think there are real open questions about whether the Glasswing coalition structure will work as intended over time. But this particular call was the right one, and it took institutional courage to make it.

As IT leaders, we spend a lot of time evaluating vendor trustworthiness based on uptime, support responsiveness, and pricing. The AI era is adding a new dimension to that evaluation: does this vendor tell you the truth about what their model cannot do? Anthropic just answered that question in a very public way. Watch how the rest of the industry responds. The ones who follow with similar transparency are the ones worth building on. The ones who stay quiet about their model’s limitations while claiming AI-powered everything are the ones to scrutinize harder.

The capability is real. The risk is real. And the vendors who are honest about both are the ones earning the right to be in your stack for the next five years.

Leave a Comment