OpenAI fires back at Google with GPT-5.2 after ‘code red’ memo

Sam Altman, chief executive officer of OpenAI

OpenAI launched its latest frontier model, GPT-5.2, on Thursday amid increasing competition from Google, pitching it as its most advanced model yet and one designed for developers and everyday professional use.

OpenAI’s GPT-5.2 comes to ChatGPT paid users and developers via the API in three flavors: Instant, a speed-optimized model for routine queries like information retrieval, writing and translation; Thinking that excels in complex structured work such as coding, analyzing long documents, mathematics and planning; and Pro, the top model that aims to deliver maximum accuracy and reliability for difficult problems.

“We designed 5.2 to unlock even more economic value for people,” Fidji Simo, OpenAI’s chief product officer, said Thursday during a briefing with reporters. “It’s better for creating spreadsheets, building presentations, writing code, perceiving images, understanding long context, using tools, and then connecting complex multi-step projects.”

GPT-5.2 lands in the middle of an arms race with Google’s Gemini 3, which tops LMArena’s rankings across most benchmarks (apart from encoding – which Anthropic’s Claude Opus-4.5 still has locked).

Earlier this month, The Information reported that CEO Sam Altman issued an internal “code red” memo to staff amid ChatGPT traffic declines and concerns that it is losing consumer market share to Google. The code red called for a shift in priorities, including ending commitments like introducing ads and instead focusing on creating a better ChatGPT experience.

GPT-5.2 is OpenAI’s push to regain leadership, though some employees reportedly asked for the model release to be pushed back to give the company more time to improve it. And despite indications that OpenAI would focus its attention on consumer cases by adding more personalization and customization to ChatGPT, the launch of GPT-5.2 appears to increase the company’s capabilities.

The company is specifically targeting developers and the tools ecosystem with the goal of becoming the standard foundation for building AI-powered applications. Earlier this week, OpenAI released new data showing that enterprise use of its AI tools has increased dramatically over the past year.

Techcrunch event

San Francisco
|
13.-15. October 2026

This comes as Gemini 3 has been tightly integrated into Google’s product and cloud ecosystem for multimodal and agentic workflows. Google this week launched managed MCP servers that make its Google and cloud services like Maps and BigQuery easier for agents to connect to. (MCPs are the connectors between AI systems and data and tools.)

OpenAI says GPT-5.2 sets new benchmark scores in coding, math, science, vision, long-context reasoning, and tooling, which the company claims can lead to “more reliable agent workflows, production-quality code, and complex systems that operate across large contexts and real-world data.”

These features put it in direct competition with Gemini 3’s Deep Think mode, which has been touted as an important reasoning advance aimed at mathematics, logic and science. On OpenAI’s own benchmark chart, GPT-5.2 edges out Thinking Gemini 3 and Anthropic’s Claude Opus 4.5 in nearly every reasoning test listed, from real-world software engineering tasks (SWE-Bench Pro) and doctoral-level science knowledge (GPQA Diamond) to abstract reasoning and pattern discovery (ARC-AGI).

Research leader Adain Clark said stronger maths results aren’t just about solving equations. Mathematical reasoning, he explained, is a proxy for whether a model can follow multistep logic, keep numbers consistent over time, and avoid subtle errors that can worsen over time.

“These are all attributes that really matter across a wide variety of workloads,” Clark said. “Things like financial modeling, forecasting, doing an analysis of data.”

During the briefing, OpenAI product manager Max Schwarzer said that GPT-5.2 “makes significant improvements to code generation and debugging” and can step through complex math and logic. Coding startups like Windsurf and CharlieCode, he added, report “state-of-the-art agent coding performance” and measurable gains on complex multi-step workflows.

In addition to coding, Schwarzer said GPT-5.2 Thinking responses contain 38% fewer errors than its predecessor, making the model more reliable for everyday decision-making, research and writing.

GPT-5.2 appears to be less of a reinvention and more of a consolidation of OpenAI’s last two upgrades. GPT-5, which dropped in August, was a reset that laid the foundation for a unified system with a router to switch the model between a fast default model and a deeper “Thinking” mode. November’s GPT-5.1 focused on making the system warmer, more conversational, and better suited for agent and coding tasks. The latest model, GPT-5.2, seems to ramp up all these advances, making it a more reliable foundation for production use.

For OpenAI, the stakes have never been higher. The company has committed $1.4 trillion worth of AI infrastructure build-out over the next few years to support its growth — commitments it made when it still had a first-mover advantage among AI companies. But now that Google, which lagged behind at first, is pushing forward, that bet may be what drives Altman’s ‘code red.’

OpenAI’s renewed focus on reasoning models is also a risky flex. The systems behind its Thinking and Deep Research modes are more expensive to run than standard chatbots because they chew through more computing. By doubling down on that kind of model with GPT-5.2, OpenAI can create a vicious cycle: spend more on compute to win the rankings, then spend even more to keep these high-cost models running at scale.

OpenAI is reportedly already using more computers than it had previously allowed. As TechCrunch reported recently, most of OpenAI’s front-end costs — the money it spends on computers to run a trained AI model — are paid in cash rather than through cloud credits, suggesting that the company’s computing costs have grown beyond what partnerships and credits can subsidize.

For all its focus on reasoning, one thing missing from today’s launch is a new image generator. Altman reportedly said in his red memo that image generation would be a key priority going forward, especially after Google’s Nano Banana (the nickname for Google’s Gemini 2.5 Flash Image model) had a viral moment after its release in August.

Last month, Google launched Nano Banana Pro (AKA Gemini 3 Pro Image), an upgraded version with even better text rendering, world knowledge and a spooky, truly unedited feel to its images. It also integrates better across Google products, as shown over the past week as it shows up in tools and workflows like Google Labs Mixboard for automatic presentation generation.

OpenAI reportedly plans to release another new model in January with better visuals, improved speed and better personality, though the company did not confirm those plans on Thursday.

OpenAI also said Thursday it is rolling out new safeguards around mental health use and age verification for teenagers, but didn’t spend much of the launch pitching those changes.