Value ranking
Best value on Output Speed
Median sustained output speed in tokens per second on the model's first-party API for medium-length prompts. Higher is faster.
“Value” is normalized benchmark score (0–100 for this leaderboard cohort) divided by input price per million tokens. Higher means more capability per dollar on this axis only — always sanity-check latency, context length, and your real workload.
- 1Gemini 1.5 FlashGoogle1333.33100.0 / $0.08/M
- 2Gemini 2.0 FlashGoogle1000.00100.0 / $0.10/M
- 3DeepSeek V3DeepSeek370.37100.0 / $0.27/M
- 4Gemini 2.5 FlashGoogle289.4786.8 / $0.30/M
- 5Llama 4 ScoutMeta248.5644.7 / $0.18/M
- 6DeepSeek R1DeepSeek181.82100.0 / $0.55/M
- 7Llama 4 MaverickMeta175.4447.4 / $0.27/M
- 8Claude 3.5 HaikuAnthropic125.00100.0 / $0.80/M
- 9GPT-4o miniOpenAI122.8018.4 / $0.15/M
- 10Qwen3 235BAlibaba (Qwen)92.1018.4 / $0.20/M
- 11Gemini 1.5 ProGoogle80.00100.0 / $1.25/M
- 12o4-miniOpenAI52.6357.9 / $1.10/M
- 13o3-miniOpenAI52.6357.9 / $1.10/M
- 14Grok 2xAI50.00100.0 / $2.00/M
- 15Llama 3.3 70B InstructMeta35.8931.6 / $0.88/M
- 16Gemini 2.5 ProGoogle35.7944.7 / $1.25/M
- 17o1-miniOpenAI33.33100.0 / $3.00/M
- 18Claude 3.5 SonnetAnthropic33.33100.0 / $3.00/M
- 19Claude 3.7 SonnetAnthropic33.33100.0 / $3.00/M
- 20GPT-5OpenAI25.2631.6 / $1.25/M
- 21GPT-4oOpenAI22.1055.3 / $2.50/M
- 22Qwen2.5 72B InstructAlibaba (Qwen)14.6213.2 / $0.90/M
- 23Claude 3 OpusAnthropic6.67100.0 / $15.00/M
- 24Grok 3xAI4.3913.2 / $3.00/M
- 25Claude Sonnet 4 (Thinking)Anthropic3.5110.5 / $3.00/M
AI Model Analyzer does not recommend specific vendors; rankings are derived from public data only.