India, March 18 -- OpenAI has introduced GPT-5.4 mini and nano, positioning them as optimised models for high-volume, latency-sensitive AI workloads. While the release builds on the capabilities of GPT-5.4, the focus shifts towards efficiency, cost control, and scalable deployment across enterprise environments.

GPT-5.4 mini delivers improvements over GPT-5 mini across coding, reasoning, multimodal understanding, and tool use while running more than twice as fast. On select benchmarks such as SWE-Bench Pro and OSWorld-Verified, it approaches the performance of the larger GPT-5.4 model.

GPT-5.4 nano, positioned as the smallest and most cost-efficient variant, is designed for simpler, high-frequency tasks, including classification, data e...