OpenAI
OpenAI

GPT-5.4

Flagship reasoning LLM with 1M context, native computer use, and high factual accuracy

Text to TextImage to Text

GPT-5.4 Overview

GPT-5.4 is OpenAI's flagship large language model, featuring a 1 million token context window, native computer use, and a 33% reduction in factual errors over GPT-5.2. It integrates coding capabilities from GPT-5.3-Codex, is 47% more token-efficient, and supports configurable reasoning effort for complex professional tasks.

From $2.51/ tokens
Input tokens$2.5 / 1M
Output token$15.00 / 1M

Commercial use

How to Use GPT-5.4

Overview

ChatGPT 5.4 is a multimodal AI model designed for advanced reasoning, content generation, and interactive workflows. It supports text, image, and structured input, enabling high-quality outputs across a wide range of use cases including writing, coding, analysis, and creative tasks.

The model is optimized for accuracy, contextual understanding, and production-grade performance, making it suitable for both conversational and structured applications.

How it Works

Prompt Interpretation

ChatGPT 5.4 interprets prompts with strong contextual awareness and reasoning capabilities. It can understand complex instructions, maintain coherence over long inputs, and adapt to different tones, formats, and task requirements.

Clear and structured prompts improve reliability and controllability of outputs.

Multimodal Generation

The model supports multiple input and output types, including text and images. It can generate detailed written content, assist with coding tasks, analyze data, and respond to visual inputs when provided.

It is designed to handle both open-ended creative tasks and precise, instruction-based workflows.

Reasoning and Task Execution

ChatGPT 5.4 applies advanced reasoning to solve problems, follow multi-step instructions, and generate structured outputs. It can perform transformations, summarization, analysis, and planning while maintaining logical consistency.

This makes it suitable for complex workflows requiring both understanding and execution.

Key Features

  • Advanced Reasoning

    Handles complex instructions, multi-step tasks, and logical problem solving.

  • Multimodal Capabilities

    Supports text and image inputs for more flexible workflows.

  • High-Quality Text Generation

    Produces coherent, structured, and context-aware content.

  • Code Generation and Analysis

    Assists with programming, debugging, and technical explanations.

  • Instruction Following

    Executes detailed and structured prompts with strong accuracy.

  • Consistent Output Quality

    Maintains reliability and coherence across long interactions.

How to Use

  1. Provide a clear prompt describing the task or objective.
  2. Include any necessary context, constraints, or formatting instructions.
  3. Run the generation.
  4. Refine the prompt to improve accuracy, tone, or structure if needed.

For best results, use specific instructions and break down complex tasks into clear steps.

Example usage:

“Write a detailed product launch plan for a new AI-powered photo editing app, including target audience, positioning, marketing channels, timeline, and key metrics, in a structured and professional format.”

Documentation

You can find full usage details, parameters, and examples here: https://runware.ai/docs/models/openai-gpt-5-4

More models from OpenAI

GPT Image 2 is a general-purpose GPT Image family model for text-to-image generation and image editing. Its strengths include strong prompt adherence, readable embedded text, detailed edits, photorealistic rendering, and structured visual outputs such as posters, packaging, product comps, diagrams, and other layout-sensitive images.

GPT-5.4 Nano is the smallest and fastest variant of GPT-5.4, designed for high-throughput, low-latency tasks such as classification, data extraction, ranking, and lightweight automation. It prioritizes speed and cost efficiency for simple, high-volume workloads and is available exclusively via the API.

GPT-5.4 Mini is a compact, efficient variant of GPT-5.4 designed for coding assistants, subagent orchestration, and multimodal applications requiring faster responsiveness. It supports a 400K token context window and retains native computer use and configurable reasoning effort at a lower cost than the flagship model.

GPT-5.4 Pro

Coming Soon

GPT-5.4 Pro is the high-performance variant of GPT-5.4, optimized for enterprise-grade professional tasks. It offers deeper reasoning, enhanced accuracy, and extended compute for complex multi-step workflows including document creation, spreadsheet analysis, and autonomous agent orchestration. It shares the 1 million token context window and native computer use capabilities of the standard GPT-5.4.

GPT Image 1.5 is OpenAI’s newest flagship image model powering the latest ChatGPT Images. It delivers significantly faster image generation with stronger instruction following, more precise edits that preserve original details, more believable transformations, and improved rendering of dense or small text. It is suited for practical creative workflows, detailed design tasks, and production use cases.

Sora 2 is OpenAI’s flagship generative model for video and audio. It accepts text prompts and generates visually rich clips with synchronized dialogue and sound. It improves physical realism and scene control. It also supports editing and extension of existing video inputs.

Sora 2 Pro is the higher quality Sora 2 variant for precision video work. It supports text prompts and image inputs. It outputs synchronized video with sound, higher resolution frames, and stronger temporal consistency. Ideal for production clips and demanding pipelines.

GPT Image 1 Mini is a lighter variant of OpenAI's GPT Image 1 model. It offers faster generation at a lower cost while retaining core capabilities including text-to-image generation, image editing, and text rendering. It is suited for high-volume workflows, rapid prototyping, and cost-sensitive applications where the full GPT Image 1 model may be excessive.

GPT Image 1 is OpenAI’s native GPT 4o image model. It creates detailed visuals from text prompts. It supports diverse styles and precise layouts. It can edit existing images with masks. It renders readable text in scenes. It suits design tools and production workflows.

DALL·E 3 converts natural language prompts into detailed images with strong caption fidelity. It improves handling of complex instructions and visual details. It integrates with ChatGPT and the OpenAI API for programmatic image creation and workflow automation.

DALL·E 2 is OpenAI’s diffusion based text to image model. It generates high quality images from prompts. It supports inpainting for local edits and outpainting for extended canvases. Developers use it through an API for creative tools, design workflows, and content pipelines.

OpenAI CLIP ViT-L/14 is a contrastive vision-language model that embeds images and text into a shared representation space. It enables tasks like zero-shot image classification, semantic search, and similarity scoring by computing aligned feature vectors for images and texts.