News

GPT-5.4 Leaked Twice in Codex Repo PRs - Here Is What We Know

Two pull requests in OpenAI's public Codex GitHub repo referenced GPT-5.4 before being scrubbed - one adding full-resolution vision support, the other a fast mode toggle. Seven force pushes and a deleted employee screenshot confirm this was not intentional.

GPT-5.4 Leaked Twice in Codex Repo PRs - Here Is What We Know

OpenAI has accidentally revealed GPT-5.4 - twice - through pull requests in its public Codex GitHub repository. Both references were quickly scrubbed via force pushes and edits, but not before screenshots circulated and the community noticed.

This isn't speculation. The PRs are public. The force push history is auditable. And an OpenAI employee's now-deleted screenshot showed GPT-5.4 in the Codex app's model selector.

TL;DR

  • PR #13050 (Feb 27) added full-resolution vision support with a minimum model version set to (5, 4) - changed to (5, 3) after seven force pushes in five hours
  • PR #13212 (Mar 2) added a /fast slash command originally described as "toggle Fast mode for GPT-5.4" - scrubbed within three hours
  • An OpenAI employee accidentally posted a screenshot showing GPT-5.4 in the Codex app model selector, then deleted it
  • Prediction markets put GPT-5.4 release before April 2026 at 55%, before June at 74%
  • GPT-5.3-Codex shipped just three weeks ago - if GPT-5.4 drops soon, OpenAI will have shipped five GPT-5.x models in seven months
  • OpenAI has not commented; their only response has been quietly cleaning up the code

The First Leak - Full-Resolution Vision (Feb 27)

On February 27, OpenAI engineer Curtis "Fjord" Hawthorne (fjord-oai) opened PR #13050 in the Codex repo, titled "Add under-development original-resolution view_image support." The feature adds a flag (view_image_original_resolution) that preserves original PNG, JPEG, and WebP image bytes and sends detail: "original" to the Responses API instead of resizing and compressing them.

The problem: the code set the minimum model version constant to (5, 4) - meaning this feature requires GPT-5.4 or newer. X user Lisan al Gaib (@scaling01) screenshotted it before the cleanup.

What followed was a frantic scrub. The PR received seven force pushes between 19:30 and 22:44 UTC on February 27 - five hours of damage control. The current version of the code now reads MIN_ORIGINAL_RESOLUTION_MODEL_VERSION: (u32, u32) = (5, 3), changed from the original (5, 4).

Seven force pushes to change a two-digit version number. That's not a typo fix.

The Second Leak - Fast Mode Toggle (Mar 2)

Five days later it happened again. On March 2 at 04:07 UTC, OpenAI engineer pash-openai opened PR #13212, adding a fast mode toggle to Codex. The PR introduces a ServiceTier enum with Standard and Fast variants, a /fast slash command in the terminal UI, and service_tier=priority on API requests when enabled.

The original version reportedly contained a direct function call using gpt-5.4 as the model argument and a slash command explicitly described as "toggle Fast mode for GPT-5.4." By 07:14 UTC - about three hours later - the GPT-5.4 references were gone.

The Employee Screenshot

Adding to the evidence: users on X shared screenshots showing GPT-5.4 appearing in the Codex desktop app's model selector dropdown. More damningly, OpenAI employee Tibo (confirmed as Codex contributor tibo-openai via PR #13181) accidentally posted a screenshot on X that included GPT-5.4 in the model picker. The post was deleted.

What GPT-5.4 Appears to Include

Based on the two PRs and surrounding evidence, here is what we can confirm about GPT-5.4 features:

FeatureSourceStatus
Full-resolution visionPR #13050 - original-res image passthrough with detail: "original"Confirmed in code (feature flag)
Fast mode / priority tierPR #13212 - /fast slash command, service_tier=priorityConfirmed in code
Model exists internallyEmployee screenshot, model selector dropdownScreenshots circulated, post deleted

What we cannot confirm: rumors of a 2 million token context window and persistent memory. No evidence for either appeared in the PRs or any credible source. The Codex repo does have several memory-related PRs from late February, but none reference GPT-5.4 specifically or use any branded feature name.

The Pace Is Relentless

The GPT-5 family has shipped at a pace that makes the GPT-4 era look glacial:

ModelDateGap
GPT-5August 7, 2025-
GPT-5-CodexSeptember 15, 20255 weeks
GPT-5.1November 12, 20258 weeks
GPT-5.2December 11, 20254 weeks
GPT-5.3-CodexFebruary 5, 20268 weeks
GPT-5.4 (leaked)Not released3+ weeks since 5.3

Five major model versions in seven months. Prediction markets on Manifold give GPT-5.4 a 55% chance of shipping before April 2026 and 74% before June. Some traders note this would require "unprecedented acceleration" since GPT-5.3's general-purpose variant (non-Codex) has not even shipped yet.

The competitive pressure is obvious. Claude Opus 4.6 launched with agent teams and a 1M context window. Anthropic's Claude Code dominates the coding market with 54% share. DeepSeek V4 is training on Huawei hardware outside the NVIDIA ecosystem entirely. OpenAI cannot afford to slow down.

What This Means

The leak pattern is notable. OpenAI's Codex repo is public by design - it is their open-source terminal coding agent. But having engineers accidentally commit internal model references to a public repo, twice in five days, suggests the development pace is outrunning internal processes. Force-pushing seven times to clean up a version number isn't how a well-coordinated release pipeline looks.

The features themselves are gradual but meaningful. Full-resolution vision support means Codex could work with detailed diagrams, high-DPI screenshots, and architectural drawings without lossy compression destroying detail. A fast mode toggle with priority service tiers suggests OpenAI is building tiered inference - paying more for lower latency - which aligns with their pricing structure for Codex subscriptions.

OpenAI hasn't commented on any of this. Their response has been limited to code cleanup and a deleted employee post. Given the cadence of the GPT-5 family so far, we'll probably not have to wait long for the official announcement.

Sources:

GPT-5.4 Leaked Twice in Codex Repo PRs - Here Is What We Know
About the author Senior AI Editor & Investigative Journalist

Elena is a technology journalist with over eight years of experience covering artificial intelligence, machine learning, and the startup ecosystem.