Text/LLM Models

Text & LLM Model Pricing 2026 - Complete Comparison

Compare pricing for all major Large Language Models including GPT-5, Claude, Gemini, and more. Find the best model for your budget and use case.

36
Models
7
Providers
$0.100
Cheapest Input/1M
$75.00
Max Output/1M
OpenAI(8 models)
ModelInput / 1MOutput / 1MContextNotes
gpt-5.2$1.75$14.00128KOpenAI's most capable GPT-5 model with enhanced re
gpt-5.1$1.25$10.00128KBalanced GPT-5 variant with excellent cost-perform
gpt-5-mini$0.300$1.00128KEfficient GPT-5 variant for cost-sensitive applica
gpt-4.1$2.00$8.001MGPT-4 with 1M context window, ideal for long docum
gpt-4.1-mini$0.400$1.601MCost-effective GPT-4 with extended context window.
gpt-4.1-nano$0.100$0.4001MUltra-efficient GPT-4 for high-volume applications
gpt-4o$2.50$10.00128KGPT-4 Omni with native multimodal capabilities.
gpt-4o-mini$0.150$0.600128KFast, affordable GPT-4o for everyday tasks.
Anthropic(6 models)
ModelInput / 1MOutput / 1MContextNotes
claude-opus-4.5$5.00$25.00200KAnthropic's most capable model with extended think
claude-opus-4.1$15.00$75.00200KPrevious Opus generation with proven reliability.
claude-sonnet-4.5$3.00$15.00200KExcellent balance of intelligence, speed, and cost
claude-sonnet-4$3.00$15.00200KReliable Sonnet with strong coding capabilities.
claude-haiku-4.5$1.00$5.00200KFast, efficient model for high-throughput tasks.
claude-haiku-3.5$0.800$4.00200KCost-effective option for simple tasks.
Google(6 models)
ModelInput / 1MOutput / 1MContextNotes
gemini-3.0-pro$2.00$12.002MGoogle's flagship model with 2M context window.
gemini-3.0-flash$0.500$3.001MFast Gemini 3.0 variant with excellent throughput.
gemini-2.5-pro$1.25$10.002MProven Gemini with exceptional long-context handli
gemini-2.5-flash$0.300$2.501MFast, affordable Gemini for everyday use.
gemini-2.5-flash-lite$0.100$0.4001MUltra-efficient for high-volume applications.
gemini-2.0-flash$0.100$0.4001MLegacy Flash model, still highly capable.
xAI(6 models)
ModelInput / 1MOutput / 1MContextNotes
grok-4$3.00$15.00256KxAI's flagship model with real-time knowledge.
grok-4.1-fast$0.200$0.500256KUltra-fast Grok for real-time applications.
grok-4-fast$0.200$0.500256KFast Grok-4 variant for quick responses.
grok-3$3.00$15.00128KReliable Grok-3 with strong general capabilities.
grok-3-mini$0.300$0.500128KEfficient Grok-3 for cost-sensitive workloads.
grok-2-vision$2.00$10.0032KVision
Perplexity(2 models)
ModelInput / 1MOutput / 1MContextNotes
sonar-pro$3.00$15.00-Search + Citations
sonar$1.00$1.00-Basic Search
Mistral(6 models)
ModelInput / 1MOutput / 1MContextNotes
mistral-large-3$2.00$6.00128KMoE 675B
mistral-medium-3$1.00$3.00128KBalanced Mistral model for most use cases.
mistral-small-3$0.200$0.600128KEfficient small model with strong performance.
ministral-14b$0.150$0.150128K14B parameter model for edge deployments.
ministral-8b$0.100$0.100128KUltra-efficient 8B model for high-volume tasks.
devstral-2$0.500$1.50-Coding
DeepSeek(2 models)
ModelInput / 1MOutput / 1MContextNotes
deepseek-v3.2$0.270$1.10128KDeepSeek's latest MoE model, highly cost-effective
deepseek-chat$0.140$0.280128KUltra-affordable chat model for general use.

Track Your Text/LLM Models Costs

Burnwise monitors every API call. See exactly where your budget goes and get recommendations to optimize costs.

Start Free Trial