OpenAI announced today that GPT-5.5 Cyber, the company's dedicated offensive-security model, will roll out only to a vetted list of "critical cyber defenders" rather than as an open API release. The product has the same structural shape as Anthropic's Claude Mythos — a base model post-trained for vulnerability discovery, exploit synthesis, and red-team automation, with a tool-belt geared toward the offensive-security workflow — and OpenAI's restricted-access framing is a reversal of the public criticism the company directed at Anthropic when Mythos shipped under similar gates earlier this month.
The release lands the same day the UK AI Security Institute published its independent evaluation of GPT-5.5 Cyber. AISI ran the model against the same suite they used on Mythos: capture-the-flag challenges across multiple difficulty tiers, real-CVE reproduction tasks against intentionally vulnerable historical-version software, and a red-team uplift study measuring how much faster a junior offensive engineer can move through a target environment with the model in the loop. The headline finding is that GPT-5.5 Cyber lands at roughly Mythos parity. AISI flagged some task categories where GPT-5.5 Cyber is meaningfully stronger — particularly multi-step exploit chaining and post-exploitation pivoting through Active Directory environments — and others where Mythos retains a gap, especially on memory-corruption primitives and kernel-level exploitation. The asymmetry tracks the differing post-training data the two labs are believed to have used.
The framing question that picked up most of the discourse is the gating decision. AISI explicitly notes that GPT-5.5 Cyber is being made generally available — meaning enterprise customers can request access through normal channels — even as OpenAI's public statement emphasizes the "critical defenders" framing. The contrast with Mythos's tighter circle, which Anthropic limits to organizations with documented incident-response programs and a contractual prohibition on offensive use against third parties, is what Simon Willison and others surfaced as the key tell. The two labs are now offering structurally similar capabilities under structurally different access policies, and OpenAI's earlier criticism of Anthropic for being too restrictive looks awkward against the new posture.
The substantive question for the field is whether "defender-only" gating is enforceable. AISI's evaluation found that GPT-5.5 Cyber follows the documented use-case restrictions reasonably consistently in the vanilla configuration, but the same uplift dynamics that make it useful for defenders also make it useful for offense — and the published technical methodology for evaluating cyber capabilities means that any organization with API access can self-evaluate the offensive ceiling without needing AISI's cooperation. The longer-term policy question, which AISI flags but does not resolve, is whether capability-based access controls scale once the underlying model is also serving the consumer ChatGPT product line.
- TechCrunch frames the release as a reversal: OpenAI publicly criticized Anthropic for restricting Mythos and then shipped Cyber under similar restrictions.
- Simon Willison emphasized that AISI's evaluation puts GPT-5.5 Cyber and Mythos at rough parity, with task-category-level differences that depend on post-training data.
- UK AISI's published evaluation focuses on capability ceiling and policy enforceability, noting that defender-only framing is hard to police once API access is granted.