Value ranking
Best value on Time to First Token
Median time from request to first output chunk in milliseconds on the model's first-party API for medium-length prompts. Lower is snappier; reasoning models are penalised here because they think before talking.
“Value” is normalized benchmark score (0–100 for this leaderboard cohort) divided by input price per million tokens. Higher means more capability per dollar on this axis only — always sanity-check latency, context length, and your real workload.
- 1Gemini 1.5 FlashGoogle1332.67100.0 / $0.08/M
- 2Gemini 2.0 FlashGoogle999.50100.0 / $0.10/M
- 3GPT-4o miniOpenAI666.67100.0 / $0.15/M
- 4Llama 4 ScoutMeta555.0699.9 / $0.18/M
- 5Qwen3 235BAlibaba (Qwen)495.2599.0 / $0.20/M
- 6DeepSeek V3DeepSeek370.19100.0 / $0.27/M
- 7Llama 4 MaverickMeta369.3799.7 / $0.27/M
- 8Gemini 2.5 FlashGoogle333.1399.9 / $0.30/M
- 9DeepSeek R1DeepSeek181.73100.0 / $0.55/M
- 10Claude 3.5 HaikuAnthropic124.94100.0 / $0.80/M
- 11Llama 3.1 70B InstructMeta113.4899.9 / $0.88/M
- 12Llama 3.3 70B InstructMeta113.4199.8 / $0.88/M
- 13Qwen2.5 72B InstructAlibaba (Qwen)109.8998.9 / $0.90/M
- 14o3-miniOpenAI81.8590.0 / $1.10/M
- 15Gemini 1.5 ProGoogle79.96100.0 / $1.25/M
- 16o4-miniOpenAI66.9573.7 / $1.10/M
- 17Gemini 2.5 ProGoogle54.1167.6 / $1.25/M
- 18Grok 2xAI49.98100.0 / $2.00/M
- 19Mistral Large 2Mistral49.9499.9 / $2.00/M
- 20GPT-4oOpenAI39.9599.9 / $2.50/M
- 21Grok 3xAI33.32100.0 / $3.00/M
- 22o1-miniOpenAI33.32100.0 / $3.00/M
- 23Claude 3.5 SonnetAnthropic33.32100.0 / $3.00/M
- 24Claude 3.7 SonnetAnthropic33.32100.0 / $3.00/M
- 25Claude Sonnet 4Anthropic33.0999.3 / $3.00/M
AI Model Analyzer does not recommend specific vendors; rankings are derived from public data only.