There was a time when a new OpenAI model felt like an event. GPT-3 arrived in 2020 after nearly two years of development, and its arrival genuinely shifted what the world thought was possible. Language models, once curiosities for researchers, suddenly became infrastructure. That kind of patience, the willingness to spend years refining something before releasing it, was the hallmark of OpenAI’s early identity.
That identity is now almost unrecognisable. GPT-5.3 Instant went live on March 3, 2026. Barely an hour after the announcement, OpenAI teased what was coming next, posting a single line on X: “5.4 sooner than you think.” GPT-5.4 arrived on March 5th. Two days between models. By any measure, OpenAI has transformed from a research laboratory that happened to build products into a product company that occasionally does research. The question worth asking is not whether that shift has happened because it clearly has but whether it is costing the models something that no benchmark can quite capture.
Also read: Tata Power-Salesforce want to install solar power units on 25 crore Indian rooftops
The early OpenAI timeline reads like a history of deliberate restraint. GPT-2 launched in February 2019, and OpenAI initially withheld the full model — citing concerns about misuse — before releasing it in stages. GPT-3 followed in June 2020, its 175 billion parameters representing a tenfold scaling leap over its predecessor. Then came a two-year pause before GPT-3.5 powered the original ChatGPT in late 2022, and GPT-4 arrived four months after that in March 2023.
These weren’t incremental updates. Each model represented a genuine rethinking of what the architecture could do. The gaps between them weren’t delays — they were the research. Users who engaged with GPT-3 in 2020 and then GPT-4 in 2023 experienced something meaningfully, tangibly different: a model that could reason, code, and sustain complex arguments in a way its predecessor simply couldn’t.
The fragmentation didn’t begin with GPT-5. It began inside GPT-4. By November 2023, GPT-4 Turbo had arrived with a massively expanded context window and cheaper pricing. By May 2024, GPT-4o redefined the line entirely by adding real-time audio and vision processing, making the model genuinely multimodal for the first time. Then came GPT-4o mini, GPT-4.5, and GPT-4.1 in rapid succession, each targeting a slightly different capability profile. What was meant to be one model became seven.
Also read: 5 key Codex features worth trying in GPT 5.4
The version number stopped meaning something. And then came the moment that should have served as a warning: in April 2025, OpenAI rolled back a GPT-4o update after users widely reported the model had become sycophantic – agreeing with everything, pushing back on nothing, performing intelligence rather than exercising it. The speed of iteration had outrun quality control. It was, in hindsight, a rehearsal for what was coming.
GPT-5 was supposed to be the reset. When it arrived in August 2025, the backlash was swift. People felt it underdelivered, and OpenAI had to reintroduce GPT-4o for paid subscribers after the public reaction made clear the new flagship wasn’t the leap many had expected. What followed was a steady monthly drumbeat of specialised variants across seven months. Each carried a specific mandate – GPT-5.1 targeted everyday reliability, 5.2 improved multimodal consistency. Then came the version that revealed something telling about the pace OpenAI was now keeping. GPT-5.3 Instant was designed to address what OpenAI itself admitted were “cringe” tendencies in its predecessor – overbearing responses, unwarranted assumptions about user intent, and dramatic phrases like “Stop. Take a breath.” These aren’t capability failures. They’re personality failures, the kind that only emerge when a model ships before it’s truly ready, and accumulate update by update until users are openly mocking the thing.
OpenAI acknowledged that GPT-5.3 targeted problems that “don’t always show up in benchmarks, but shape whether ChatGPT feels helpful or frustrating.” That’s a remarkable admission. The benchmarks were fine. The experience wasn’t. And two days after 5.3 shipped, GPT-5.4 Thinking arrived alongside GPT-5.4 Pro, with OpenAI now asking users to choose between models optimised for speed, difficult reasoning, and maximum capability – three distinct products where there used to be one.
It would be too simple to say OpenAI has sacrificed intelligence for speed. The models are, by most technical measures, more capable than anything that came before them. But capability and substance are not the same thing. Substance includes trustworthiness, the sense that a model says what it believes, not what it thinks you want to hear. It includes coherence, the sense that a version is a version, not a continuous blur of overlapping variants. And it includes the kind of reliability that only comes from a company willing to sit with something before it ships.
On those measures, the trajectory is harder to defend. The GPT-2 to GPT-3 era produced models that surprised people and changed minds. The GPT-5.3 to GPT-5.4 era produces models that arrive two days apart – one fixing a tone problem, one promising deeper thinking – while OpenAI teases the next one before the current one has even settled. That’s not a research culture. That’s a release cycle. Whether OpenAI can find its way back to the patience that once defined it is a question that I have been thinking about since GPT-5.4 came out. The benchmarks will keep climbing either way. The question is whether the models will.
Also read: I replaced my PC with the Asus ROG ecosystem, here is what happened