Back to stories
Models

OpenAI Releases GPT-5.5 'Spud', Tops Artificial Analysis Intelligence Index

Michael Ouroumis3 min read
OpenAI Releases GPT-5.5 'Spud', Tops Artificial Analysis Intelligence Index

OpenAI released GPT-5.5 on Thursday, April 23, 2026, its most capable frontier model to date. The release — internally codenamed 'Spud' — comes exactly one month after pretraining wrapped on March 24 and ends weeks of leak-driven speculation after an unreleased OpenAI model was caught live-testing on public APIs earlier in the month.

GPT-5.5 is rolling out immediately to paid ChatGPT and Codex subscribers on the Plus, Pro, Business and Enterprise tiers. API access is coming shortly, pending the integration of additional cybersecurity guardrails that OpenAI says it wants in place before opening the model up to developer traffic.

A new top of the leaderboard

GPT-5.5 debuts at the top of the Artificial Analysis Intelligence Index with a score of 60 — three points ahead of Claude Opus 4.7 and Gemini 3.1 Pro Preview, which had been tied at 57. The lead is narrow on paper but meaningful competitively: Anthropic has held or shared the pole position across much of the last two quarters, and the index is one of the closer-watched composite measures in the industry.

OpenAI's pitch, however, is about agentic capability rather than trivia-style benchmarks. On Artificial Analysis's Coding Agent Index — which measures end-to-end performance on real coding tasks — GPT-5.5 posts state-of-the-art numbers at roughly half the cost of competing frontier coding models. The company is framing the model as one that can take a messy, multi-part request and 'plan, use tools, check its work, navigate through ambiguity, and keep going' with minimal user prompting.

Benchmark highlights

OpenAI published a cluster of headline results alongside the launch:

Speed has historically been the trade-off as frontier models get bigger, but OpenAI says GPT-5.5 matches GPT-5.4's per-token latency in real-world serving while operating at a substantially higher level of capability.

Why it matters

GPT-5.5 raises the bar on three fronts at once: computer use, autonomous coding, and knowledge work across tools. Those are precisely the axes that matter for the 'agent' products enterprises are now rolling out — Codex-style coding agents, browsing agents, and workspace assistants that have to stay on task across dozens of steps. A coding-agent SOTA at half the cost of the competition is likely to pressure Anthropic and Google on pricing as much as on capability.

It also resets expectations for what a point-release cadence means at OpenAI. GPT-5 launched less than a year ago; since then the company has shipped 5.2, 5.3, 5.4 and now 5.5, with Codex variants layered on top. The platform story — models plus Codex plus native computer use — is clearly the product now, and GPT-5.5 is the clearest expression of it yet.

Learn AI for Free — FreeAcademy.ai

Take "AI Essentials: Understanding AI in 2026" — a free course with certificate to master the skills behind this story.

More in Models

xAI Launches Grok Voice Think Fast 1.0, Tops τ-Voice Bench and Powers Starlink Support
Models

xAI Launches Grok Voice Think Fast 1.0, Tops τ-Voice Bench and Powers Starlink Support

xAI's new voice model scored 67.3% on the τ-voice Bench — well ahead of Gemini 3.1 Flash Live and GPT Realtime — and is now powering Starlink's phone sales and support with a 70% autonomous resolution rate.

2 days ago2 min read
Tencent Drops Hy3 Preview: 295B Open-Source MoE Model Kicks DeepSeek Out of Yuanbao
Models

Tencent Drops Hy3 Preview: 295B Open-Source MoE Model Kicks DeepSeek Out of Yuanbao

Tencent has open-sourced Hy3 Preview, a 295B/21B-activated mixture-of-experts model built in under three months. The Yuanbao chatbot is switching its primary engine from DeepSeek to the new in-house model.

4 days ago2 min read
DeepSeek V4 Preview Lands: 1.6T-Parameter Open Model With 1M Context, Flash Pricing at $0.14/M
Models

DeepSeek V4 Preview Lands: 1.6T-Parameter Open Model With 1M Context, Flash Pricing at $0.14/M

DeepSeek on April 24 released preview versions of V4-Pro and V4-Flash, an open-weight MoE family with a 1M-token context window and pricing that undercuts Western frontier labs.

4 days ago2 min read