Comparison hub

LLM API pricing comparison: every model, ranked

25+ models across OpenAI, Anthropic, and Google. Sorted by real cost, not marketing.

Section 1

Rankings by category

Four top-10 leaderboards covering the workloads that drive most API bills. Every number is computed from the pricing dataset at render time — retune the rates in lib/pricing.tsand this page updates on the next build.

Cheapest per request

1,000 input / 500 output tokens

RankModelProviderCost / requestInput / 1MOutput / 1M
1GPT-5 nanoOpenAI$0.000250$0.05$0.40
2Gemini 2.5 Flash-LiteGoogle$0.000300$0.10$0.40
3GPT-4o miniOpenAI$0.000450$0.15$0.60
4GPT-5.4 nanoOpenAI$0.000825$0.20$1.25
5Gemini 3.1 Flash-Lite (preview)Google$0.001000$0.25$1.50
6GPT-4.1 miniOpenAI$0.001200$0.40$1.60
7GPT-5 miniOpenAI$0.001250$0.25$2.00
8Gemini 2.5 FlashGoogle$0.001550$0.30$2.50
9Gemini 3 Flash (preview)Google$0.002000$0.50$3.00
10GPT-5.4 miniOpenAI$0.003000$0.75$4.50

Cheapest for long context

100,000 input / 10,000 output tokens

RankModelProviderCost / requestInput / 1MContext
1GPT-5 nanoOpenAI$0.009000$0.05400k
2Gemini 2.5 Flash-LiteGoogle$0.0140$0.101M
3GPT-4o miniOpenAI$0.0210$0.15128k
4GPT-5.4 nanoOpenAI$0.0325$0.20400k
5Gemini 3.1 Flash-Lite (preview)Google$0.0400$0.251M
6GPT-5 miniOpenAI$0.0450$0.25400k
7Gemini 2.5 FlashGoogle$0.0550$0.301M
8GPT-4.1 miniOpenAI$0.0560$0.401.0M
9Gemini 3 Flash (preview)Google$0.0800$0.501M
10GPT-5.4 miniOpenAI$0.1200$0.75400k

Cheapest at high volume

1,000,000 requests / month at 1k input + 500 output each

RankModelProviderCost / requestMonthly total
1GPT-5 nanoOpenAI$0.000250$250.00
2Gemini 2.5 Flash-LiteGoogle$0.000300$300.00
3GPT-4o miniOpenAI$0.000450$450.00
4GPT-5.4 nanoOpenAI$0.000825$825.00
5Gemini 3.1 Flash-Lite (preview)Google$0.001000$1,000
6GPT-4.1 miniOpenAI$0.001200$1,200
7GPT-5 miniOpenAI$0.001250$1,250
8Gemini 2.5 FlashGoogle$0.001550$1,550
9Gemini 3 Flash (preview)Google$0.002000$2,000
10GPT-5.4 miniOpenAI$0.003000$3,000

Largest context window

For RAG, long-document, and large-codebase workloads

RankModelProviderContextMax output
1Gemini 2.5 ProGoogle2M
2GPT-5.4OpenAI1.1M128k
3GPT-4.1OpenAI1.0M33k
4GPT-4.1 miniOpenAI1.0M33k
5Claude Opus 4.7Anthropic1M128k
6Claude Opus 4.6Anthropic1M128k
7Claude Sonnet 4.6Anthropic1M64k
8Gemini 3.1 Pro (preview)Google1M
9Gemini 3 Flash (preview)Google1M
10Gemini 3.1 Flash-Lite (preview)Google1M

Section 2

Full comparison matrix

Every tracked model, sortable by any column. Filter by provider, minimum context window, or maximum cost per request; search by model name or id. Costs are computed at render time fromlib/pricing.ts.

Provider
Showing 25 of 25 models
Compare
GPT-5 nanogpt-5-nanoOpenAI$0.05$0.40400k128k$0.000250$0.002500Details →
Gemini 2.5 Flash-Litegemini-2.5-flash-liteGoogle$0.10$0.401M$0.000300$0.003000Details →
GPT-4o minigpt-4o-miniOpenAI$0.15$0.60128k16k$0.000450$0.004500Details →
GPT-5.4 nanogpt-5.4-nanoOpenAI$0.20$1.25400k128k$0.000825$0.008250Details →
Gemini 3.1 Flash-Lite (preview)gemini-3.1-flash-lite-previewGoogle$0.25$1.501M$0.001000$0.0100Details →
GPT-4.1 minigpt-4.1-miniOpenAI$0.40$1.601.0M33k$0.001200$0.0120Details →
GPT-5 minigpt-5-miniOpenAI$0.25$2.00400k128k$0.001250$0.0125Details →
Gemini 2.5 Flashgemini-2.5-flashGoogle$0.30$2.501M$0.001550$0.0155Details →
Gemini 3 Flash (preview)gemini-3-flash-previewGoogle$0.50$3.001M$0.002000$0.0200Details →
GPT-5.4 minigpt-5.4-miniOpenAI$0.75$4.50400k128k$0.003000$0.0300Details →
o4-minio4-miniOpenAI$1.10$4.40200k100k$0.003300$0.0330Details →
Claude Haiku 4.5claude-haiku-4-5Anthropic$1.00$5.00200k64k$0.003500$0.0350Details →
GPT-4.1gpt-4.1OpenAI$2.00$8.001.0M33k$0.006000$0.0600Details →
o3o3OpenAI$2.00$8.00200k100k$0.006000$0.0600Details →
Gemini 2.5 Progemini-2.5-proGoogle$1.25$10.002M$0.006250$0.0625Details →
GPT-5gpt-5OpenAI$1.25$10.00400k128k$0.006250$0.0625Details →
GPT-4ogpt-4oOpenAI$2.50$10.00128k16k$0.007500$0.0750Details →
Gemini 3.1 Pro (preview)gemini-3.1-pro-previewGoogle$2.00$12.001M$0.008000$0.0800Details →
GPT-5.4gpt-5.4OpenAI$2.50$15.001.1M128k$0.0100$0.1000Details →
Claude Sonnet 4.6claude-sonnet-4-6Anthropic$3.00$15.001M64k$0.0105$0.1050Details →
Claude Sonnet 4.5claude-sonnet-4-5Anthropic$3.00$15.00200k64k$0.0105$0.1050Details →
Claude Opus 4.6claude-opus-4-6Anthropic$5.00$25.001M128k$0.0175$0.1750Details →
Claude Opus 4.5claude-opus-4-5Anthropic$5.00$25.00200k64k$0.0175$0.1750Details →
Claude Opus 4.7claude-opus-4-7Anthropic$5.00$25.001M128k$0.0201$0.2012Details →
Claude Opus 4.1claude-opus-4-1Anthropic$15.00$75.00200k32k$0.0525$0.5250Details →

Section 3

Cost scenarios

Same workload, every model. The bar is proportional to the most expensive model in each scenario so you can eyeball the 10×-and-more spread without doing arithmetic.

Simple chatbot

Lightweight conversational UX: 500 in / 200 out tokens, 10,000 requests/month.

RankModelProviderCost / reqMonthlyRelative
1GPT-5 nanoOpenAI$0.000105$1.05
2Gemini 2.5 Flash-LiteGoogle$0.000130$1.30
3GPT-4o miniOpenAI$0.000195$1.95
4GPT-5.4 nanoOpenAI$0.000350$3.50
5Gemini 3.1 Flash-Lite (preview)Google$0.000425$4.25
6GPT-4.1 miniOpenAI$0.000520$5.20
7GPT-5 miniOpenAI$0.000525$5.25
8Gemini 2.5 FlashGoogle$0.000650$6.50
9Gemini 3 Flash (preview)Google$0.000850$8.50
10GPT-5.4 miniOpenAI$0.001275$12.75
11o4-miniOpenAI$0.001430$14.30
12Claude Haiku 4.5Anthropic$0.001500$15.00
13GPT-4.1OpenAI$0.002600$26.00
14o3OpenAI$0.002600$26.00
15Gemini 2.5 ProGoogle$0.002625$26.25
16GPT-5OpenAI$0.002625$26.25
17GPT-4oOpenAI$0.003250$32.50
18Gemini 3.1 Pro (preview)Google$0.003400$34.00
19GPT-5.4OpenAI$0.004250$42.50
20Claude Sonnet 4.6Anthropic$0.004500$45.00
21Claude Sonnet 4.5Anthropic$0.004500$45.00
22Claude Opus 4.6Anthropic$0.007500$75.00
23Claude Opus 4.5Anthropic$0.007500$75.00
24Claude Opus 4.7Anthropic$0.008625$86.25
25Claude Opus 4.1Anthropic$0.0225$225.00

Coding agent

Iterative code edits: 5,000 in / 2,000 out tokens, 1,000 requests/month.

RankModelProviderCost / reqMonthlyRelative
1GPT-5 nanoOpenAI$0.001050$1.05
2Gemini 2.5 Flash-LiteGoogle$0.001300$1.30
3GPT-4o miniOpenAI$0.001950$1.95
4GPT-5.4 nanoOpenAI$0.003500$3.50
5Gemini 3.1 Flash-Lite (preview)Google$0.004250$4.25
6GPT-4.1 miniOpenAI$0.005200$5.20
7GPT-5 miniOpenAI$0.005250$5.25
8Gemini 2.5 FlashGoogle$0.006500$6.50
9Gemini 3 Flash (preview)Google$0.008500$8.50
10GPT-5.4 miniOpenAI$0.0128$12.75
11o4-miniOpenAI$0.0143$14.30
12Claude Haiku 4.5Anthropic$0.0150$15.00
13GPT-4.1OpenAI$0.0260$26.00
14o3OpenAI$0.0260$26.00
15Gemini 2.5 ProGoogle$0.0263$26.25
16GPT-5OpenAI$0.0263$26.25
17GPT-4oOpenAI$0.0325$32.50
18Gemini 3.1 Pro (preview)Google$0.0340$34.00
19GPT-5.4OpenAI$0.0425$42.50
20Claude Sonnet 4.6Anthropic$0.0450$45.00
21Claude Sonnet 4.5Anthropic$0.0450$45.00
22Claude Opus 4.6Anthropic$0.0750$75.00
23Claude Opus 4.5Anthropic$0.0750$75.00
24Claude Opus 4.7Anthropic$0.0862$86.25
25Claude Opus 4.1Anthropic$0.2250$225.00

RAG pipeline

Retrieval-augmented answers: 50,000 in / 1,000 out tokens, 5,000 requests/month.

RankModelProviderCost / reqMonthlyRelative
1GPT-5 nanoOpenAI$0.002900$14.50
2Gemini 2.5 Flash-LiteGoogle$0.005400$27.00
3GPT-4o miniOpenAI$0.008100$40.50
4GPT-5.4 nanoOpenAI$0.0113$56.25
5Gemini 3.1 Flash-Lite (preview)Google$0.0140$70.00
6GPT-5 miniOpenAI$0.0145$72.50
7Gemini 2.5 FlashGoogle$0.0175$87.50
8GPT-4.1 miniOpenAI$0.0216$108.00
9Gemini 3 Flash (preview)Google$0.0280$140.00
10GPT-5.4 miniOpenAI$0.0420$210.00
11Claude Haiku 4.5Anthropic$0.0550$275.00
12o4-miniOpenAI$0.0594$297.00
13Gemini 2.5 ProGoogle$0.0725$362.50
14GPT-5OpenAI$0.0725$362.50
15GPT-4.1OpenAI$0.1080$540.00
16o3OpenAI$0.1080$540.00
17Gemini 3.1 Pro (preview)Google$0.1120$560.00
18GPT-4oOpenAI$0.1350$675.00
19GPT-5.4OpenAI$0.1400$700.00
20Claude Sonnet 4.6Anthropic$0.1650$825.00
21Claude Sonnet 4.5Anthropic$0.1650$825.00
22Claude Opus 4.6Anthropic$0.2750$1,375
23Claude Opus 4.5Anthropic$0.2750$1,375
24Claude Opus 4.7Anthropic$0.3162$1,581
25Claude Opus 4.1Anthropic$0.8250$4,125

Document processing

Long-document summarisation: 100,000 in / 5,000 out tokens, 500 requests/month.

RankModelProviderCost / reqMonthlyRelative
1GPT-5 nanoOpenAI$0.007000$3.50
2Gemini 2.5 Flash-LiteGoogle$0.0120$6.00
3GPT-4o miniOpenAI$0.0180$9.00
4GPT-5.4 nanoOpenAI$0.0263$13.13
5Gemini 3.1 Flash-Lite (preview)Google$0.0325$16.25
6GPT-5 miniOpenAI$0.0350$17.50
7Gemini 2.5 FlashGoogle$0.0425$21.25
8GPT-4.1 miniOpenAI$0.0480$24.00
9Gemini 3 Flash (preview)Google$0.0650$32.50
10GPT-5.4 miniOpenAI$0.0975$48.75
11Claude Haiku 4.5Anthropic$0.1250$62.50
12o4-miniOpenAI$0.1320$66.00
13Gemini 2.5 ProGoogle$0.1750$87.50
14GPT-5OpenAI$0.1750$87.50
15GPT-4.1OpenAI$0.2400$120.00
16o3OpenAI$0.2400$120.00
17Gemini 3.1 Pro (preview)Google$0.2600$130.00
18GPT-4oOpenAI$0.3000$150.00
19GPT-5.4OpenAI$0.3250$162.50
20Claude Sonnet 4.6Anthropic$0.3750$187.50
21Claude Sonnet 4.5Anthropic$0.3750$187.50
22Claude Opus 4.6Anthropic$0.6250$312.50
23Claude Opus 4.5Anthropic$0.6250$312.50
24Claude Opus 4.7Anthropic$0.7188$359.38
25Claude Opus 4.1Anthropic$1.88$937.50

Section 4

Every pair, one click away

300 head-to-head pages, generated programmatically from the pricing dataset. Every link resolves to a dedicated comparison page with side-by-side rates, scenario cost ladders, and volume projections.

Anthropic vs OpenAI

84 pairs

OpenAI vs Google

72 pairs

Anthropic vs Google

42 pairs

Groupings reflect lib/pricing.ts providers: Anthropic · OpenAI · Google.

Section 5

Frequently asked questions

Every answer is composed from the current pricing dataset at render time — both the prose above and the FAQPage JSON-LD embedded for search engines.

Which LLM API is the cheapest?
Based on current pricing, GPT-5 nano is the cheapest model at $0.000250 per request for a standard workload of 1,000 input / 500 output tokens. If you care most about input-token rate at scale, GPT-5 nano offers the best per-token price at $0.05 per million input tokens.
How much does it cost to run an LLM chatbot?
A chatbot handling 10,000 messages per month (500 input / 200 output tokens each) costs between $1.05 on GPT-5 nano and $225.00 on Claude Opus 4.1. Use the comparison table above to pick the cheapest model that meets your quality bar.
Is GPT-5.4 cheaper than Claude Opus 4.7?
At a standard workload (1,000 input / 500 output tokens), GPT-5.4 costs $0.0100 per request while Claude Opus 4.7 costs $0.0201. GPT-5.4 is cheaper by $0.0101 per request — roughly 50% less.
What is the cheapest LLM for coding?
For coding tasks at 5,000 input / 2,000 output tokens, GPT-5 nano offers the best cost at $0.001050 per request. Budget-conscious teams can drop to Gemini 2.5 Flash-Lite at $0.001300 per request if quality permits — typically fine for boilerplate edits, riskier for architectural reasoning.
How do LLM API prices compare in 2026?
LLM API prices dropped roughly 80% from 2024 to 2026. Frontier models now cost $0.05–$15.00 per million input tokens, while budget models start at $0.05 per million. See the full comparison table above for every currently-tracked model.

Estimate the cost of your actual prompt

Paste a prompt, pick a model, and see the exact cost before you send anything.