Anthropic launches faster, cheaper AI model

Anthropic, the San Francisco-based AI firm, has announced the launch of Claude Haiku 4.5, its latest compact model engineered to deliver near-frontier performance with reduced computational demands and operating costs. This release marks a material acceleration of the industry’s pivot toward models optimized for speed, affordability, and real-time enterprise integration.

Valued at an estimated $183 billion, Anthropic describes Haiku 4.5 as its most rapid and energy-frugal model to date. It has been specifically designed to handle low-latency applications necessary for enterprise operations, including sophisticated coding assistance, high-volume customer support systems, and continuous data-stream monitoring. Despite its position as the smallest offering within the Claude family, Haiku 4.5 has demonstrated coding proficiency benchmarks comparable to Claude Sonnet 4, the company’s flagship mid-tier model from the preceding year. This technical parity represents a substantial leap in efficiency for a system categorized as lightweight.

Anthropic maintains a clear, tiered model architecture: the Haiku models serve as fast, response-optimized systems; Sonnet models occupy the middle ground, handling complex reasoning and analysis; and Opus models represent the company’s most advanced, frontier-grade intelligence. The introduction of Haiku 4.5 comes closely on the heels of Claude Sonnet 4.5 and Opus 4.1. In fact, Anthropic has priced the new Haiku 4.5 model at approximately one-third the cost of its mid-tier Sonnet counterparts. “Anyone can chat with Claude using Haiku 4.5 on Claude.ai, available on web, iOS, and Android, Pricing for Haiku 4.5 on the Claude Developer Platform starts at $1 per million input tokens and $5 per million output tokens, with up to 90% cost savings with prompt caching and 50% cost savings with the Message Batches API,” the AI firm announced in a statement.

This pricing strategy seems to be intended to aggressively capture market share among businesses that require scalable AI solutions without incurring prohibitive infrastructure expenditure, making capabilities accessible for high-volume, cost-sensitive use cases. The model is immediately available via the Claude web application, its API, Amazon Bedrock, and Google Cloud’s Vertex AI, ensuring broad accessibility across major cloud platforms.

According to the firm, in benchmark evaluations, Haiku 4.5 achieved a 73% score on SWE-Bench Verified, a critical test of an AI’s ability to autonomously resolve real-world software engineering tasks within live codebases. This places its coding capability within a narrow margin of significantly larger systems like Claude Sonnet 4.5 and OpenAI’s GPT-5. The model also delivered strong results in complex “computer-use” scenarios, which assess an AI’s capacity to execute scripts, navigate software interfaces, and manage external tools.

According to internal Anthropic metrics, Haiku 4.5 processes input more than twice as fast as Sonnet 4.5, achieving a latency profile that facilitates near-instantaneous responses in interactive and distributed settings. This architecture allows a larger, higher-reasoning model to decompose a complex problem into subtasks, which are then executed simultaneously by numerous rapid-response Haiku instances, embodying a transition from monolithic to distributed cognition. The launch of a highly efficient model directly addresses the rising concerns over the substantial financial and environmental costs associated with large-scale generative AI. Estimates suggest that a single inference request from a frontier-grade model can require over 6,000 joules of energy. By contrast, a smaller system like Haiku 4.5, which is likely operating within the eight-billion-parameter range, may consume closer to 100 joules per query.

The Tech Portal is published by Blue Box Media Private Limited. Our investors have no influence over our reporting. Read our full Ownership and Funding Disclosure →