Anthropic
Anthropic

Claude Haiku 4.5

Fast, cost-efficient multimodal language model for low-latency agents and scaled reasoning workloads

Text to TextImage to Text

Claude Haiku 4.5 Overview

Claude Haiku 4.5 is Anthropic's fastest and most cost-efficient Claude model. It is built for latency-sensitive applications, high-volume agents, sub-agent orchestration, coding assistance, and budget-conscious deployments that still need strong reasoning and multimodal understanding.

Token based
Input tokens / 1M$1.00
Output tokens / 1M$5.00

Commercial use

How to Use Claude Haiku 4.5

Overview

Claude Haiku 4.5 is a fast, cost-efficient multimodal language model for low-latency applications, scaled agent workloads, and budget-sensitive deployments.

It is best suited to teams that need strong coding and reasoning performance at higher throughput and lower cost than larger frontier models.

Strengths

Low Latency and High Throughput

Claude Haiku 4.5 is built for fast responses and scaled deployments. It is a strong fit for interactive systems, background workers, and high-volume AI products.

Cost-Efficient Agent Work

The model works well for sub-agent orchestration, parallel task execution, and workloads where many model calls need to remain affordable.

Strong Coding Performance for the Size

Claude Haiku 4.5 is a capable coding model despite its smaller footprint, making it useful for coding assistants, rapid prototyping, and development workflows that prioritize speed.

Practical Multimodal Understanding

The model supports image understanding in addition to text, which helps it handle visual context in workflows that need screenshot interpretation or image-grounded reasoning.

Good Fit for Real-Time Products

Haiku 4.5 is well suited to chat assistants, customer support systems, real-time analysis tools, and other product experiences where responsiveness matters as much as raw intelligence.

Capabilities

Text-to-Text

Claude Haiku 4.5 supports general-purpose text generation, structured responses, coding assistance, reasoning, and agent execution workflows.

Image-to-Text

The model accepts image inputs for multimodal tasks such as screenshot understanding, visual reasoning, and image-grounded assistance.

Input and Output

  • AIR ID: anthropic:[email protected]
  • Input: text and image inputs
  • Output: text responses

Best Fit

  • Real-time chat and support systems
  • High-volume AI agents
  • Coding sub-agents and rapid prototyping
  • Budget-conscious multimodal products
  • Parallelized reasoning and automation workloads

More models from Anthropic

Claude Opus 4.7

Api Only

Claude Opus 4.7 is Anthropic's highest-capability generally available Claude model. It is designed for demanding coding, agent orchestration, multimodal reasoning, and high-stakes professional workflows, with stronger instruction following, better high-resolution vision, adaptive thinking, and a 1M-token context window.

Claude Sonnet 4.6

Api Only

Claude Sonnet 4.6 is Anthropic's most capable Sonnet model, built for daily production use across coding, agent workflows, long-context reasoning, computer use, and professional knowledge work. It supports adaptive and extended thinking, strong instruction following, high-volume automation, and a 1M-token context window in beta.