TL;DR
Anthropic's Claude Haiku 4.5 delivers Sonnet 4-level coding performance at one-third the cost and twice the speed. Here is what developers need to know.
Five months ago, Claude Sonnet 4 was state-of-the-art. Now Claude Haiku 4.5 matches its coding performance at one-third the cost and more than twice the speed.
That is not marketing spin. On SWE-bench Verified - the benchmark that measures performance on real-world GitHub issues - Haiku 4.5 sits right alongside models that were considered frontier just months earlier. Anthropic released it on October 15, 2025, and it immediately changed the math on which model to use for what.
| Metric | Haiku 4.5 | Sonnet 4 | Delta |
|---|---|---|---|
| SWE-bench Verified | Near-Sonnet 4 | Frontier (at release) | Comparable |
| Speed | 2x+ faster | Baseline | Major improvement |
| Cost (input) | $1/M tokens | $3/M tokens | 3x cheaper |
| Cost (output) | $5/M tokens | $15/M tokens | 3x cheaper |
| Computer use | Surpasses Sonnet 4 | Strong | Haiku wins |
The pricing is $1 per million input tokens and $5 per million output tokens. For context, that means a typical coding session with 50K input tokens and 10K output tokens costs about $0.10. Run that same session on Sonnet 4.5 and you are paying significantly more.
Sub-agent orchestration. This is the killer use case. Sonnet 4.5 breaks down a complex problem into a multi-step plan, then dispatches a team of Haiku 4.5 instances to execute subtasks in parallel. You get frontier-level planning with fast, cheap execution. Claude Code uses this pattern heavily - Haiku 4.5 runs as the sub-agent model by default.
# In Claude Code, Haiku 4.5 powers sub-agents automatically
# The main agent (Sonnet/Opus) orchestrates, Haiku executes
claude "Refactor the auth module and update all tests"
# -> Opus plans the refactor
# -> Multiple Haiku 4.5 sub-agents execute file changes in parallel
Real-time applications. Chat assistants, customer service agents, pair programming tools - anything where latency matters. Haiku 4.5 responds fast enough that the AI feels instant rather than sluggish.
Computer use. Surprisingly, Haiku 4.5 surpasses Sonnet 4 on computer use tasks. If you are building desktop automation, the small model is actually the better choice.
High-volume batch processing. At 3x cheaper than Sonnet, running Haiku 4.5 on thousands of files, PRs, or code reviews becomes economically viable in ways that frontier models are not.
Get the weekly deep dive
Tutorials on Claude Code, AI agents, and dev tools - delivered free every week.
Through the API, just swap the model name:
import anthropic
client = anthropic.Anthropic()
response = client.messages.create(
model="claude-haiku-4-5",
max_tokens=1024,
messages=[
{"role": "user", "content": "Review this function for bugs..."}
]
)
In Claude Code, Haiku 4.5 is already integrated as the default sub-agent model. You do not need to configure anything - it handles the fast, parallel execution tasks while the primary model (Opus or Sonnet) handles planning and complex reasoning.
On claude.ai, Haiku 4.5 is available in the model selector for all users, including free tier.
The model lineup has a clear hierarchy now:
The practical pattern most teams settle on: Opus or Sonnet as the orchestrator, Haiku 4.5 as the executor. Planning happens once at the top. Execution happens many times in parallel at the bottom. This gives you the best of both worlds.
Augment reported that Haiku 4.5 achieves 90% of Sonnet 4.5's performance in their agentic coding evaluation. Warp called it "a leap forward for agentic coding, particularly for sub-agent orchestration." Vercel noted that "just six months ago, this level of performance would have been state-of-the-art on our internal benchmarks."
The consensus is the same from every direction: the speed-intelligence tradeoff that used to define small models is disappearing. Haiku 4.5 is not a compromise. It is a genuinely capable model that happens to be fast and cheap.
Haiku 4.5 represents a pattern in AI development. Today's frontier becomes tomorrow's commodity. The model that was cutting-edge in May is the small, cheap option by October. This compression benefits developers enormously - the capabilities you are building against keep getting cheaper to run.
For teams building on Claude, the practical takeaway is straightforward: audit your model usage. Anything running on Sonnet 4 that does not require frontier reasoning can likely drop to Haiku 4.5 with no quality loss and 3x cost savings.
Yes. It matches Sonnet 4 on SWE-bench Verified, which tests real-world GitHub issue resolution. For most coding tasks - code review, bug fixes, test generation, refactoring - Haiku 4.5 delivers results that are indistinguishable from what the larger models produce.
Haiku 4.5 outperforms both on coding benchmarks while maintaining competitive pricing. Its particular strength is agentic workflows - multi-step tasks where the model needs to use tools, navigate codebases, and make sequential decisions.
You can, but you will hit its limits on complex architectural reasoning and novel problem-solving. The recommended pattern is to use it alongside a larger model - let Sonnet or Opus handle the hard thinking, and Haiku handles the execution.
Haiku 4.5 supports a 200K token context window, same as the larger Claude models. This means it can process entire codebases, long documents, and extended conversation histories without truncation.
Yes. Full tool use, function calling, computer use, and all Claude API features are supported. There are no capability restrictions compared to larger models - only differences in reasoning depth on complex tasks.
Technical content at the intersection of AI and development. Building with AI agents, Claude Code, and modern dev tools - then showing you exactly how it works.
Anthropic's AI. Opus 4.6 for hard problems, Sonnet 4.6 for speed, Haiku 4.5 for cost. 200K context window. Best coding m...
View ToolAnthropic's agentic coding CLI. Runs in your terminal, edits files autonomously, spawns sub-agents, and maintains memory...
View Tool
New tutorials, open-source projects, and deep dives on coding agents - delivered weekly.
Anthropic's Python SDK for building production agent systems. Tool use, guardrails, agent handoffs, and orchestration. R...
Configure Claude Code for maximum productivity -- CLAUDE.md, sub-agents, MCP servers, and autonomous workflows.
AI AgentsInstall Claude Code, configure your first project, and start shipping code with AI in under 5 minutes.
Getting StartedInstall Ollama, pull your first model, and run AI locally for coding, chat, and automation - with zero cloud dependency.
Getting Started
Anthropic releases Claude Haiku 4.5, a faster, more cost-effective model for coding tasks, rivaling previous models and competitors. While maintaining similar performance to Claude Sonnet 4,...

In this video, we dive into Anthropic's newly launched Cowork, a user-friendly extension of Claude Code designed to streamline work for both developers and non-developers. This discussion includes an

In this video, we dive into Anthrop's latest release, Claude Opus 4.5, touted as the best model for coding agents and computer use. We review the blog post and significant announcements, such...

Claude Opus 4.6 vs GPT-5.3 for real TypeScript work. Benchmarks, pricing, context windows, and practical differences.

A developer's comparison of OpenAI and Anthropic ecosystems - models, coding tools, APIs, pricing, and which to choose f...

Anthropic's Sonnet 4.6 narrows the gap to Opus on agentic tasks, leads computer use benchmarks, and ships with a beta mi...