OpenAI Releases GPT-5.4 Mini and Nano for High-Volume AI Applications

AI News Hub Editorial
Senior AI Reporter
March 18, 2026
OpenAI Releases GPT-5.4 Mini and Nano for High-Volume AI Applications

OpenAI has released two new smaller models, GPT-5.4 mini and GPT-5.4 nano, aimed at delivering faster performance and lower costs for high-volume AI workloads. The models are available now, with GPT-5.4 mini accessible across the API, Codex and ChatGPT, and GPT-5.4 nano offered through the API for use cases where speed and cost are priorities.

GPT-5.4 mini is positioned as a step up from GPT-5 mini, with improvements in coding, reasoning, multimodal understanding and tool use, while running more than twice as fast. OpenAI said the model approaches the performance of the larger GPT-5.4 system on benchmarks such as SWE-Bench Pro and OSWorld-Verified, highlighting a focus on balancing capability with responsiveness.

GPT-5.4 nano is designed as the smallest and least expensive option in the lineup. The company recommends it for tasks like classification, data extraction, ranking and lightweight coding support, where lower latency and cost are more important than maximum capability.

The models target applications where speed directly affects user experience, including coding assistants, subagent workflows and systems that process images or screenshots in real time. OpenAI said these use cases benefit from models that can respond quickly, reliably use tools and still handle complex tasks when needed.

In coding environments, GPT-5.4 mini is intended for rapid iteration workflows such as debugging, navigating codebases and generating front-end components. The model delivers higher pass rates than its predecessor at similar latency, while running faster, according to the company’s benchmarks.

OpenAI also emphasized the role of smaller models in multi-model systems. In setups like Codex, larger models can handle planning and decision-making, while smaller models like GPT-5.4 mini execute narrower tasks in parallel. This approach allows developers to manage cost and performance by assigning work based on complexity.

The models also support multimodal inputs and computer-use tasks. GPT-5.4 mini can interpret screenshots of user interfaces and complete actions based on them, with performance close to the larger GPT-5.4 model in certain evaluations.

Pricing reflects the focus on efficiency. GPT-5.4 mini costs $0.75 per million input tokens and $4.50 per million output tokens, while GPT-5.4 nano is priced at $0.20 per million input tokens and $1.25 per million output tokens.

By expanding its lineup with smaller models, OpenAI is targeting workloads where latency, cost and scalability are key considerations, particularly in applications that require frequent, real-time interactions rather than maximum model size.

This analysis is based on reporting from openai.

Image courtesy of openai.

This article was generated with AI assistance and reviewed for accuracy and quality.

Last updated: March 18, 2026

About this article: This article was generated with AI assistance and reviewed by our editorial team to ensure it follows our editorial standards for accuracy and independence. We maintain strict fact-checking protocols and cite all sources.

Word count: 414Reading time: 0 minutes

AI Tools for this Article

📧 Stay Updated

Get the latest AI news delivered to your inbox every morning.

Browse All Articles
Share this article:
Next Article