Pricing comparison

Claude Haiku 4.5 vs Gemini 3.1 Pro (preview)

Per-token pricing, full-workload cost ladders, and monthly volume projections. Numbers sourced directly from each provider's rate card.

Anthropic

Claude Haiku 4.5

Input
$1.00 / 1M
Output
$5.00 / 1M
Cached input
-
Context
200K
Max output
64K

Google

Gemini 3.1 Pro (preview)

Input
$2.00 / 1M
Output
$12.00 / 1M
Cached input
$0.200 / 1M
Context
1M
Max output
-

Cost per request

Four common workload shapes, input tokens and a 1:2 output ratio (a standard chat/completion pattern). Long-context surcharges apply automatically where the provider charges them.

ScenarioTokens (in / out)Claude Haiku 4.5Gemini 3.1 Pro (preview)Winner
Short prompt100 / 200$0.0011$0.0026Claude Haiku 4.5
Typical request1,000 / 2,000$0.0110$0.0260Claude Haiku 4.5
Long document10,000 / 5,000$0.0350$0.0800Claude Haiku 4.5
Large prompt100,000 / 10,000$0.1500$0.3200Claude Haiku 4.5

Monthly bill at scale

Projected monthly cost at typical request volume, assuming the "typical request" shape above (1k in, 2k out).

TrafficReq / monthClaude Haiku 4.5Gemini 3.1 Pro (preview)Delta
Small SaaS1,000$11.00$26.00Claude Haiku 4.5 -$15.00
Growing product10,000$110.00$260.00Claude Haiku 4.5 -$150.00
Heavy usage100,000$1,100$2,600Claude Haiku 4.5 -$1,500

Which should you use?

For the typical chat-shape request (~1k input, 2k output), Claude Haiku 4.5 comes out 136% cheaper. If you're picking one as the default, that's usually the right choice on cost alone.

Claude Haiku 4.5 wins on both sides of the bill - cheaper input and cheaper output. If the quality gap is in your favour, there's no cost argument for the other side.

Context window differs: Gemini 3.1 Pro (preview) holds 1M of input vs 200K on the other side. If you regularly push past the smaller ceiling, the comparison ends there.

Heads-up: Gemini 3.1 Pro (preview) applies a long-context surcharge above 200K input tokens ($4.00 input / $18.00 output per 1M). Workloads that push past that threshold pay roughly 2x the numbers above.

Live cost calculator

Type in any token counts - both prices update instantly. Uses base input/output rates (no cache discount, no long-context tier).

Claude Haiku 4.5

$0.0035

per request

Gemini 3.1 Pro (preview)

$0.0080

per request

Claude Haiku 4.5 is $0.0045 cheaper per request (56.3% less).

Try both in the estimator →

Drop your actual prompt in, tokens are counted with the provider's own tokenizer, and the dollar number matches what lands on your invoice.

Frequently asked

Which is cheaper, Claude Haiku 4.5 or Gemini 3.1 Pro (preview)?
On a typical 1,000-input / 2,000-output request, Claude Haiku 4.5 costs ~$0.0110 vs ~$0.0260 on Gemini 3.1 Pro (preview). Input or output rates can flip the answer for very lopsided workloads - see the cost ladder above.
What's the difference in per-token pricing?
Claude Haiku 4.5 charges $1.00 per 1M input tokens and $5.00 per 1M output tokens. Gemini 3.1 Pro (preview) charges $2.00 / $12.00 per 1M.
Which has the bigger context window?
Gemini 3.1 Pro (preview) is larger (1M) vs 200K on the other.
Is there a cached-input discount on either?
Claude Haiku 4.5 does not publish a cached-input rate. Gemini 3.1 Pro (preview) caches at $0.200 per 1M (90% off). Workloads with repeated static prefixes see the biggest savings.
Does Gemini 3.1 Pro (preview) have a long-context surcharge?
Yes. Above 200K input tokens, Gemini 3.1 Pro (preview) bills at $4.00 input / $18.00 output per 1M instead of the standard rate.
How fresh is this comparison?
Claude Haiku 4.5 was re-verified on 2026-04-06 and Gemini 3.1 Pro (preview) on 2026-04-06 against each provider's published rate card. Calcis re-checks every row on a rolling schedule and re-deploys when a provider changes pricing.

Claude Haiku 4.5 verified 2026-04-06 · Gemini 3.1 Pro (preview) verified 2026-04-06. Rate cards at Anthropic and Google.