OpenAI Launches GPT-5.4-Cyber and Scales Access for Cyber to Thousands of Defenders
GPT-5.4-Cyber lowers refusal thresholds and adds reverse engineering, following Anthropic’s Project Glasswing cybersecurity model launch last week.
OpenAI on April 14, 2026, announced GPT-5.4-Cyber, a fine-tuned GPT-5.4 variant designed to reduce refusal friction for legitimate cybersecurity tasks and enable capabilities that are absent in the standard model.
OpenAI describes it as “cyber-permissive,” lowering refusal boundaries for real security work while adding advanced defenses, including binary reverse engineering.
The launch comes exactly one week after Anthropic introduced Project Glasswing and its restricted Claude Mythos model to about 40 organizations for defensive cybersecurity use, signaling the rising AI cybersecurity competition.
What Binary Reverse Engineering Means for Security Teams
The standout capability in GPT-5.4-Cyber is binary reverse engineering. As XDA Developers confirmed, it enables security professionals to analyze compiled software, the machine code of a program, for malware, vulnerabilities, and weaknesses without source code access.
In practice, teams investigating suspicious applications or embedded firmware can input compiled programs into GPT-5.4-Cyber and receive analysis of behavior, structure, and potential attack surfaces without original developer cooperation or source code availability.
The company also notes that GPT-5.4-Cyber is available to a limited, vetted group of security testers for vulnerability identification.
Standard ChatGPT models often decline dual-use cybersecurity queries, which can be used for defense and misuse, creating friction for legitimate researchers analyzing real threats. GPT-5.4-Cyber’s cyber-permissive design directly addresses this friction for verified defenders.
How the Trusted Access for Cyber Program Works
The deployment mechanism behind GPT-5.4-Cyber is the Trusted Access for Cyber program, launched in February 2026 alongside a $10 million cybersecurity grant initiative.
As Reuters confirmed, OpenAI is now scaling it from an invitation-only structure to thousands of verified individual defenders and hundreds of security teams protecting critical software infrastructure.
As OpenAI confirmed, the program uses a tiered verification system. Lower tiers provide access to standard models with reduced friction for security tasks, while the highest tier unlocks GPT-5.4-Cyber with binary reverse engineering and a more permissive setup.
Foud emphasized that cybersecurity is a team effort where every organization should be empowered to secure its systems. “No one should be in the business of picking winners and losers when it comes to cybersecurity,” he added.
This structure ties GPT-5.4-Cyber access directly to verification level and trusted cybersecurity roles.
How Codex Security Establishes the Baseline
Alongside the GPT-5.4-Cyber launch, OpenAI highlighted its Codex Security product, which launched in private beta six months ago, a key part of its growing OpenAI Superapp plans.
Reports note that it has helped fix over 3,000 critical and high-severity vulnerabilities, plus additional lower-severity issues across 1,000 open-source projects through Codex for Open Source.
As XDA Developers reported, Codex Security continuously monitors codebases, validates issues, and proposes fixes, creating an always-on AI-assisted security layer instead of periodic manual audits.
The Competitive Context and What OpenAI’s Benchmarks Signal
OpenAI’s cybersecurity benchmark progression shows rapid advancement in its artificial intelligence capability.
As SiliconAngle confirmed, capture-the-flag benchmark performance rose from 27% on GPT-5 in August 2025 to 76% on GPT-5.1-Codex-Max in November 2025, a 49-point increase in three months.
OpenAI also stated it is planning future evaluations as though each new model could reach “High” cybersecurity capability under its Preparedness Framework, marking a stricter internal safety stance for models still in development.
The GPT-5.4-Cyber rollout is broader than Anthropic’s Mythos Preview access, which is limited to around 40 organizations via 12 core partners. OpenAI is instead targeting thousands of individual users and hundreds of teams through automated verification.
OpenAI said the goal is to make advanced defensive tools widely available while preventing misuse, framing identity verification and tiered access as a scalable alternative to Anthropic’s more tightly gated approach.



