Llm March 6, 2026

OpenAI GPT-5.4 adds a 1M-token context window and Tool Search API

OpenAI’s GPT-5.4 release is aimed at teams building real systems, not chatbot demos. The main additions are easy to spot: a 1 million token context window, a new Tool Search mechanism in the API, and three model variants with different jobs. There’s ...

OpenAI GPT-5.4 adds a 1M-token context window and Tool Search API

OpenAI’s GPT-5.4 pushes harder on agents, long context, and tool-heavy workflows

OpenAI’s GPT-5.4 release is aimed at teams building real systems, not chatbot demos.

The main additions are easy to spot: a 1 million token context window, a new Tool Search mechanism in the API, and three model variants with different jobs. There’s the base gpt-5.4, a faster gpt-5.4-pro, and gpt-5.4-thinking for heavier reasoning work. OpenAI also says GPT-5.4 uses fewer tokens than GPT-5.2 to finish the same tasks. That matters more than another foggy claim about “intelligence.”

Taken together, the release points in one direction: agents, long-running sessions, messy enterprise software, and work that spans documents, browser actions, spreadsheets, and internal systems.

What stands out first

The 1M-token context window is the flashy number. Tool Search may matter more for developers.

A lot of production agent systems are ugly in the same way. Teams stuff tool definitions into the prompt because the model needs to know what functions exist. Once you have dozens of tools, prompts get bloated, latency rises, costs creep up, and tool routing gets sloppier because the model has to scan a wall of schemas before it can do anything useful.

GPT-5.4 changes that by letting the model look up tool specs on demand. Instead of preloading every function schema into context, you keep a catalog and let the model retrieve the right one when it needs it. If this works well outside demos, it removes one of the most annoying pieces of agent plumbing.

That’s a meaningful API improvement.

The three-model split is sensible

OpenAI’s naming is straightforward:

  • gpt-5.4 for general-purpose use
  • gpt-5.4-pro for lower latency and higher throughput
  • gpt-5.4-thinking for deeper reasoning and chain-of-thought control

That split matches how teams deploy models now. One model usually doesn’t do every job well enough.

If you’re building customer-facing workflows or batch automation, Pro is probably the first one to test. Throughput and response time often matter more than squeezing out a few extra reasoning points. For legal review, financial modeling, scientific writing, or multi-step software planning, Thinking is the one worth a closer look.

OpenAI is turning the speed-versus-deliberation trade-off into a product choice instead of pretending one endpoint covers both cleanly.

A 1M-token context changes some architecture, not all of it

A million tokens will change some systems. It won’t wipe out retrieval-augmented generation.

You can keep much larger working sets inside the model’s active context: whole project histories, long threads, multi-file repos, discovery docs, planning notes, tool outputs, and style constraints. For software teams, that means fewer brittle chunk boundaries during codebase-wide refactors. For internal knowledge work, it means less prompt gymnastics to stitch fragmented context back together.

Useful, yes. Magic, no.

Long context doesn’t remove the need for document structure, ranking, and careful prompt design. If you dump a million tokens into a session without organization, the model still has to figure out what matters. Quality still depends on how information is laid out and refreshed. Garbage scales just fine.

There’s also latency. Even with sparse attention, segmented memory, caching, and the usual long-context tricks, huge prompts are still expensive. In plenty of applications, smart retrieval will stay cheaper and faster than brute-forcing the whole corpus into context every time.

So yes, 1M tokens opens up new patterns. Teams still have to design systems properly.

The benchmark story points to computer-use agents

OpenAI says GPT-5.4 sets records on OSWorld-Verified and WebArena Verified, both aimed at software use across desktop and web environments. It also reports 83% on GDPval, its knowledge-work eval, plus a leading score on Mercor’s APEX-Agents, which focuses on professional tasks in areas like law and finance.

Those are more useful signals than the usual benchmark chest-thumping because they line up better with actual enterprise work.

A lot of AI spending right now is chasing one specific question: can the model complete real tasks across tools without falling apart halfway through? Open a browser, search the web, inspect a spreadsheet, update a system, summarize findings, produce a deliverable, keep state, recover from small errors. That’s where agent projects usually fail. Not on trivia. On tedious, multi-step workflow execution.

GPT-5.4’s reported gains suggest OpenAI is getting better in that category. Mercor CEO Brendan Foody’s comment that it handles “slide decks, financial models, and legal analysis” faster and at lower cost than other frontier models fits the pattern. Those are long-horizon deliverables with plenty of chances to drift off course.

Benchmarks are still benchmarks. Expect degradation once your own auth flows, naming conventions, tool failures, and half-documented APIs show up. Still, these results are more relevant than another leaderboard built around academic QA.

Token efficiency matters

OpenAI says GPT-5.4 solves equivalent tasks with fewer tokens than GPT-5.2. That’s a meaningful claim.

Most production teams care about cost per completed job. If a model reaches the right answer with fewer generated tokens, or wastes less time narrating itself while calling tools more cleanly, the economics improve fast at scale.

That shows up in a few places:

  • multi-turn support agents
  • long-running research or analysis sessions
  • batch document processing
  • tool-heavy orchestration where the model should decide, call, and move on

Token efficiency also changes how teams behave. If each run is cheaper and shorter, engineers iterate more aggressively. Features that looked marginal at one price point start to make sense. That’s usually how platform shifts show up in real teams: a pile of small “we can finally ship this” decisions.

Tool Search is probably the part developers will copy

There’s a good chance Tool Search spreads quickly across the ecosystem.

The old pattern was crude and common: inject every tool schema into the system prompt and hope the model picks the right one. That works until the tool catalog gets large, versioned, or domain-specific. Then you’re paying context tax on tools that never get used.

OpenAI’s approach is closer to a function registry with retrieval. The model infers intent, looks up matching tool definitions, pulls in the relevant schema, then executes. That means cleaner prompt state, lower token overhead, and better scaling for organizations with sprawling internal APIs.

If you build agent infrastructure, this idea will feel familiar. The industry has been moving the same way elsewhere: stop stuffing raw metadata into prompts and start indexing it properly.

A decent implementation probably needs:

  • concise tool descriptions
  • strict input schemas
  • versioning
  • cost or latency hints
  • permission boundaries
  • strong validation on outputs

Without that, you get a nicer registry and the same old failures.

The safety angle is narrower, but still relevant

OpenAI also introduced a new chain-of-thought safety evaluation, and says the Thinking variant is less prone to deceptive reasoning traces. It argues that chain-of-thought monitoring still has value as a safety control.

That matters for teams that inspect reasoning during sensitive workflows, especially in regulated environments or internal review pipelines. If you’re using reasoning traces to audit why the model chose a path, you want some confidence that the visible process hasn’t drifted too far from the actual one.

This area still deserves skepticism.

Reasoning traces aren’t ground truth. They’re evidence. Sometimes useful evidence, sometimes polished rationalization. If you depend on them, pair them with tool logs, cited sources, deterministic checks, and post-processing rules. A readable thought process helps. An auditable execution trail helps more.

What engineering teams should do with this

If you’re evaluating GPT-5.4 for production, the obvious mistake is to treat all three variants as a drop-in upgrade and stop there.

A better plan:

  • test Pro on agent and batch workloads where latency and cost dominate
  • test Thinking on high-stakes, multi-step tasks where reasoning quality changes outcomes
  • redesign your tool layer around a registry instead of prompt stuffing
  • revisit whether some RAG-heavy flows can be simplified with larger session memory
  • keep retrieval for ranking, freshness, and cost control
  • measure completed-task cost, not just token pricing

The broader shift is clear enough. Frontier model competition is becoming less about who writes the prettiest paragraph and more about who can sit inside a software stack, keep state across long sessions, call the right tools, and finish work with fewer retries.

That’s a better direction. It’s also less forgiving. Once models are embedded in real workflows, benchmarks matter less than failure recovery, observability, access control, and whether your infra team can tell what the system is doing when it breaks.

GPT-5.4 looks strong on the parts builders care about. The open question is whether OpenAI’s mix of long context, on-demand tool lookup, and stronger agent performance holds up once it hits production mess. That’s where this release gets judged.

Keep going from here

Useful next reads and implementation paths

If this topic connects to a real workflow, these links give you the service path, a proof point, and related articles worth reading next.

Relevant service
AI model evaluation and implementation

Compare models against real workflow needs before wiring them into production systems.

Related proof
Internal docs RAG assistant

How model-backed retrieval reduced internal document search time by 62%.

Related article
OpenAI's GPT-4.1 API models add 1M-token context with lower latency

OpenAI has released a new API-only model family: GPT-4.1, GPT-4.1 mini, and GPT-4.1 nano. The headline numbers are straightforward: up to 1 million tokens of context across the lineup, better coding performance than GPT-4o, lower latency, and much lo...

Related article
OpenAI launches GPT-5.2 with Instant, Thinking, and Pro for production AI

OpenAI has launched GPT-5.2, and the important part is the product shape. The release comes in three profiles: Instant, Thinking, and Pro. The pitch is aimed squarely at teams putting AI into production. A fast mode for cheaper, everyday work. A deep...

Related article
OpenAI's GPT-5 pricing puts real pressure on model margins

OpenAI priced GPT-5 low enough to force a serious conversation about margins across the model market. The headline numbers: - $1.25 per 1 million input tokens - $10 per 1 million output tokens - $0.125 per 1 million cached input tokens That matters r...