LLM Pricing — Cost Per Token for Every Model

How much does an LLM cost per token? This page breaks down LLM pricing for every model across OpenAI, Anthropic, and Google — from the cheapest budget models to flagship reasoning engines. All prices updated for 2026.

LLM Cost Per Token

Every LLM API charges per token. The table below shows the exact cost per single token for every model. For breakpoint models, the lower rate (under the context threshold) is shown. Use our LLM cost calculator for custom estimates, or see the full pricing comparison with per-million rates.

ModelProviderInput / tokenOutput / tokenInput / 1MPricing
gpt-5-nanoOpenAI$0.00000005$0.00000040$0.05Flat
gemini-2.0-flash-liteGoogle Gemini$0.00000007$0.00000030$0.07Flat
gemini-2.0-flashGoogle Gemini$0.00000010$0.00000040$0.10Flat
gemini-2.5-flash-liteGoogle Gemini$0.00000010$0.00000040$0.10Flat
gpt-4.1-nanoOpenAI$0.00000010$0.00000040$0.10Flat
gpt-4o-miniOpenAI$0.00000015$0.00000060$0.15Flat
claude-haiku-3Anthropic Claude$0.00000025$0.00000125$0.25Flat
gpt-5-miniOpenAI$0.00000025$0.00000200$0.25Flat
gemini-2.5-flashGoogle Gemini$0.00000030$0.00000250$0.30Flat
gpt-4.1-miniOpenAI$0.00000040$0.00000160$0.40Flat
gemini-3-flashGoogle Gemini$0.00000050$0.00000300$0.50Flat
gemini-2.5-flash-preview-ttsGoogle Gemini$0.00000050$0.00001000$0.50Flat
claude-haiku-3-5Anthropic Claude$0.00000080$0.00000400$0.80Flat
gemini-2.5-pro-preview-ttsGoogle Gemini$0.00000100$0.00002000$1.00Flat
claude-haiku-4-5Anthropic Claude$0.00000100$0.00000500$1.00Flat
o1-miniOpenAI$0.00000110$0.00000440$1.10Flat
o3-miniOpenAI$0.00000110$0.00000440$1.10Flat
o4-miniOpenAI$0.00000110$0.00000440$1.10Flat
gemini-2.5-proGoogle Gemini$0.00000125$0.00001000$1.25Breakpoint (200K)
gemini-2.5-computer-useGoogle Gemini$0.00000125$0.00001000$1.25Breakpoint (200K)
gpt-5OpenAI$0.00000125$0.00001000$1.25Flat
gpt-5-codexOpenAI$0.00000125$0.00001000$1.25Flat
gpt-5.1OpenAI$0.00000125$0.00001000$1.25Flat
gpt-5.1-codexOpenAI$0.00000125$0.00001000$1.25Flat
gpt-5.1-codex-maxOpenAI$0.00000125$0.00001000$1.25Flat
gpt-5.2OpenAI$0.00000175$0.00001400$1.75Flat
gpt-5.2-codexOpenAI$0.00000175$0.00001400$1.75Flat
gpt-5.3-codexOpenAI$0.00000175$0.00001400$1.75Flat
gemini-3-pro-previewGoogle Gemini$0.00000200$0.00001200$2.00Breakpoint (200K)
gemini-3.1-pro-previewGoogle Gemini$0.00000200$0.00001200$2.00Breakpoint (200K)
gpt-4.1OpenAI$0.00000200$0.00000800$2.00Flat
o3OpenAI$0.00000200$0.00000800$2.00Flat
o4-mini-deep-researchOpenAI$0.00000200$0.00000800$2.00Flat
gpt-4oOpenAI$0.00000250$0.00001000$2.50Flat
gpt-5.4OpenAI$0.00000250$0.00001500$2.50Breakpoint (272K)
claude-3-7-sonnetAnthropic Claude$0.00000300$0.00001500$3.00Flat
claude-sonnet-4-0Anthropic Claude$0.00000300$0.00001500$3.00Breakpoint (200K)
claude-sonnet-4-5Anthropic Claude$0.00000300$0.00001500$3.00Breakpoint (200K)
claude-opus-4-5Anthropic Claude$0.00000500$0.00002500$5.00Flat
claude-opus-4-6Anthropic Claude$0.00000500$0.00002500$5.00Breakpoint (200K)
o3-deep-researchOpenAI$0.00001000$0.00004000$10.00Flat
claude-opus-3Anthropic Claude$0.00001500$0.00007500$15.00Flat
claude-opus-4-0Anthropic Claude$0.00001500$0.00007500$15.00Flat
claude-opus-4-1Anthropic Claude$0.00001500$0.00007500$15.00Flat
gpt-5-proOpenAI$0.00001500$0.0001200$15.00Flat
o1OpenAI$0.00001500$0.00006000$15.00Flat
o3-proOpenAI$0.00002000$0.00008000$20.00Flat
gpt-5.2-proOpenAI$0.00002100$0.0001680$21.00Flat
gpt-5.4-proOpenAI$0.00003000$0.0001800$30.00Breakpoint (272K)
o1-proOpenAI$0.0001500$0.0006000$150.00Flat

How LLM Token Pricing Works

LLM APIs charge per token — a unit of text roughly equal to 4 characters or 0.75 words in English. Every API call has two cost components:

  • Input tokens (your prompt) — processed in parallel, relatively cheap.
  • Output tokens (the model's response) — generated sequentially, typically 2-8x more expensive than input.

Some models use breakpoint pricing, where rates increase above a context length threshold (typically 200K tokens). This means a long conversation or document analysis costs more per token than a short prompt. Models with breakpoint pricing are marked in the table above.

Multimodal models (image and audio) have separate rates for each modality. Image output tokens can cost 10-60x more than text output on the same model. See individual provider pages for multimodal pricing details.

LLM Pricing by Provider

Frequently Asked Questions

How much does an LLM cost per token?

LLM costs range from $0.00000005 per input token (GPT-5 Nano) to $0.000150 per input token (o1-pro). Most mid-range models cost $0.000001-0.000003 per input token. Output tokens typically cost 2-8x more than input tokens.

What is the cheapest LLM?

The cheapest LLMs in 2026 are GPT-5 Nano ($0.05/1M input tokens), Gemini 2.0 Flash Lite ($0.075/1M input), and GPT-4.1 Nano and Gemini 2.0 Flash (both $0.10/1M input). These are ideal for high-volume, low-complexity tasks.

Why are output tokens more expensive than input tokens?

Output tokens require the model to generate text autoregressively — each token depends on all previous tokens. This is more computationally expensive than processing input tokens in parallel. Output tokens typically cost 2-8x more than input tokens.

How do I calculate my LLM API cost?

Multiply your input token count by the input rate, and your output token count by the output rate. For example, 1,000 input tokens + 500 output tokens on GPT-5 ($1.25/1M input, $10/1M output) costs: (1000 × $0.00000125) + (500 × $0.00001) = $0.00625. Use our LLM cost calculator for instant estimates.

Automate Your LLM Cost Estimation

Get programmatic access to real-time LLM pricing with our API. Or try the cost calculator and pricing comparison for quick estimates.

Get Started Free