GPT-5.5 System Card: What OpenAI Is Telling Us

GPT-5.5 System Card: What OpenAI Is Telling Us

OpenAI dropped the GPT-5.5 system card on April 23, 2026, and if you’re the kind of person who reads these things carefully — not just the press release version — there’s quite a bit worth unpacking. System cards are easy to dismiss as corporate box-ticking. This one shouldn’t be. It outlines where GPT-5.5 sits on OpenAI’s internal risk spectrum, how the model performed on a battery of safety and capability evaluations, and what guardrails are actually in place before it reaches your API calls or ChatGPT sessions. That’s real information, and the details tell a story that goes beyond the usual launch hype.

How We Got Here: The GPT-5 Family Is Getting Complicated

Cast your mind back to early 2025. OpenAI shipped GPT-4o, o3, and GPT-4.5 in fairly quick succession, and the version numbering started feeling like airline boarding zones — technically meaningful, practically confusing. GPT-5 arrived as a consolidation moment: one flagship model that folded reasoning and multimodal capability into a single deployment. The reception was strong. Enterprise uptake accelerated. Then came the sub-variants.

GPT-5.4 was the first notable fork. OpenAI used it as the base for GPT-5.4-Cyber, a purpose-tuned model aimed squarely at defensive security work. That told you something important: OpenAI is no longer treating each model release as a monolithic product. They’re shipping a platform with variants, each carrying its own system card and its own evaluated risk profile. GPT-5.5 is the next step up that trunk line — not a specialty fork, but a general-purpose capability upgrade that still needs its own safety documentation because the underlying model has meaningfully changed.

Why does that matter? Because each jump in general capability also tends to expand what the model can be coaxed into doing, both helpfully and otherwise. A system card isn’t marketing. It’s OpenAI saying, on the record, “here’s what we tested, here’s what we found, here’s what we decided to do about it.”

What the GPT-5.5 System Card Actually Contains

The official GPT-5.5 system card follows the structure OpenAI has been refining since GPT-4’s card in 2023, but it’s noticeably more detailed in a few areas that reflect where the AI safety conversation has moved.

Safety Evaluations and Risk Classification

OpenAI uses a tiered preparedness framework — think of it as a color-coded risk dial with levels from low to critical. GPT-5.5 is classified at medium risk across most evaluated threat categories, which is consistent with where GPT-5 landed. The categories tested include:

  • CBRN uplift — whether the model provides meaningful assistance to someone attempting to develop chemical, biological, radiological, or nuclear weapons. GPT-5.5 showed no significant uplift beyond what’s available through standard internet searches, according to OpenAI’s red-team findings.
  • Cyberoffense — the model’s ability to assist in writing functional exploits or novel malware. Scored medium, with enhanced refusal behaviors trained in specifically for this capability category.
  • Persuasion and influence operations — how effectively the model can generate targeted disinformation or manipulative political content at scale. This category received extended attention in the card, reflecting post-election scrutiny across the industry.
  • Autonomy and self-replication — evaluated in the context of agentic deployments, where GPT-5.5 might be running multi-step tasks with tool access. Low risk classification here, with noted behavioral constraints around acquiring unintended resources.

The red-teaming section is worth reading in full. OpenAI worked with external evaluators — including teams from academic institutions and independent security researchers — not just internal staff. That’s a meaningful distinction. Self-reported safety findings from a lab deploying the model commercially have obvious credibility limits. Third-party adversarial testing doesn’t eliminate that tension, but it at least adds a layer of accountability.

Capability Benchmarks: Where GPT-5.5 Lands

On raw capability, GPT-5.5 shows measurable improvements over GPT-5 on several standard evaluation suites. The card cites gains on MMLU-Pro, GPQA (Graduate-Level Google-Proof Q&A), and coding benchmarks including HumanEval and SWE-bench Verified. The gains aren’t enormous — this isn’t a generational leap — but they’re consistent across domains rather than concentrated in one area.

What’s more interesting than the numbers is the reasoning quality under adversarial prompting. OpenAI’s internal evals show GPT-5.5 is less susceptible to certain jailbreak patterns that worked on GPT-5, particularly multi-turn social engineering attempts where the model is gradually walked toward a harmful output. That’s a targeted behavioral improvement, not just a capability scaling artifact.

For context: Anthropic’s Claude 4 series has been making similar claims about adversarial robustness, and Google’s Gemini 2.5 Pro has posted competitive numbers on GPQA and reasoning tasks. The frontier is genuinely crowded right now. GPT-5.5 doesn’t obviously lap the competition on benchmarks — OpenAI’s advantage, if there is one, is in deployment infrastructure, the ChatGPT distribution channel, and the agent tooling they’ve been building out aggressively. Speaking of which, if you’re running GPT-5.5 in agentic pipelines, the latency and caching improvements OpenAI rolled out earlier this year matter a lot for real-world performance, not just benchmark scores.

Deployment Decisions and Access Tiers

GPT-5.5 is available via the API and in ChatGPT, with access gated by tier in familiar ways. API customers on the paid tier get access immediately. Free ChatGPT users get a rate-limited version. Enterprise and Teams accounts get priority throughput. Nothing shocking there.

What’s worth flagging is the system card’s treatment of agentic deployment constraints. OpenAI is explicit that certain tool-use combinations trigger additional runtime monitoring — specifically when GPT-5.5 has simultaneous access to web browsing, code execution, and external API calls. This is a direct response to concerns about multi-agent systems compounding errors or being manipulated through prompt injection in retrieved content. If you’re building workspace automation agents on top of ChatGPT, this monitoring layer is something to understand, not just accept as background infrastructure.

What This Actually Means for Developers and Enterprises

Here’s the thing: a system card is partly a safety document and partly a negotiation with the developer community about what the model will and won’t do. The refusal behaviors, the tool-use constraints, the monitoring triggers — these directly affect what you can build.

For most enterprise use cases, GPT-5.5 is going to behave very similarly to GPT-5. The capability improvements are real but incremental. The safety improvements are meaningful but not disruptive to legitimate applications. If you were happy with GPT-5 in production, GPT-5.5 is a drop-in upgrade that handles edge cases better and performs slightly stronger on complex reasoning tasks.

For security and research applications, the picture is more nuanced. The enhanced cyberoffense refusals might occasionally trip on legitimate security research prompts — that’s been a consistent complaint from penetration testers and CTF participants with previous models. The system card acknowledges this tension but doesn’t fully resolve it. OpenAI points to its safety and usage policies for guidance on research use cases, which isn’t a satisfying answer if you’re a security professional who needs the model to actually help with offensive security work in a controlled context.

For AI researchers tracking how labs are handling transparency, the GPT-5.5 system card is a step forward. It’s more specific than many prior releases. But the fundamental asymmetry hasn’t changed: OpenAI controls the evaluation methodology, chooses which results to publish, and decides what constitutes an acceptable risk level. Independent verification of these claims is still essentially impossible. The METR (Model Evaluation and Threat Research) organization and similar groups are working on closing that gap, but we’re not there yet.

Key Takeaways

  • GPT-5.5 is classified as medium risk across CBRN, cyberoffense, and influence operations categories — consistent with GPT-5, not an escalation.
  • Capability improvements are real but incremental — gains on MMLU-Pro, GPQA, and coding benchmarks, with notably better resistance to adversarial multi-turn jailbreaks.
  • Agentic deployments get additional scrutiny — runtime monitoring kicks in for multi-tool combinations, relevant for anyone building complex automation pipelines.
  • Third-party red-teaming was used — more credible than pure internal evaluation, though independent verification of claims is still limited.
  • API access is live for paid tiers — enterprise and Teams customers get priority throughput, free users get rate-limited access.
  • Security researchers face familiar friction — enhanced cyberoffense refusals may affect legitimate use cases; OpenAI’s policy guidance here remains vague.

Frequently Asked Questions

What is the GPT-5.5 system card and why does it matter?

A system card is a formal document where OpenAI discloses the safety evaluations, capability benchmarks, and deployment constraints for a new model before or at release. It matters because it’s the closest thing to a public accountability record for how the model was tested and what risks the lab decided were acceptable — details that directly affect what developers and enterprises can build on top of it.

How does GPT-5.5 compare to Claude 4 and Gemini 2.5?

On standard benchmarks, GPT-5.5 is competitive but not clearly dominant — Anthropic’s Claude 4 and Google’s Gemini 2.5 Pro post similar numbers on reasoning and knowledge tasks. OpenAI’s real differentiation is deployment scale, ChatGPT’s distribution reach, and the agent infrastructure it’s been building out. The benchmark race matters less than it used to; integration and reliability in production are where the competition is actually playing out.

Is GPT-5.5 available right now?

Yes, as of April 23, 2026. Paid API customers and ChatGPT Enterprise and Teams users have full access. Free ChatGPT users can access a rate-limited version. Pricing follows the same tier structure as GPT-5, with no announced changes to per-token costs at launch.

What should enterprises pay attention to in the system card?

The agentic deployment section is the most practically relevant part for enterprise builders — specifically the runtime monitoring triggers for multi-tool agent configurations. If you’re running automated workflows where GPT-5.5 has simultaneous access to browsing, code execution, and external APIs, understand that additional monitoring is active and factor that into your latency and compliance planning.

OpenAI is clearly treating the GPT-5.x series as a sustained deployment platform rather than a stepping stone to GPT-6, and that changes the calculus for anyone building long-term on their APIs. The more granular system cards, the specialty forks like GPT-5.4-Cyber, the tightening integration with agent infrastructure — it all points toward a company that expects enterprises to go deep rather than just experiment. Whether the transparency in these documents is sufficient for the stakes involved is a question the industry hasn’t finished answering.