Value ranking
Best value on Chatbot Arena Elo
Crowdsourced pairwise human preference rankings of LLM responses. Higher Elo means more frequently preferred by users.
“Value” is normalized benchmark score (0–100 for this leaderboard cohort) divided by input price per million tokens. Higher means more capability per dollar on this axis only — always sanity-check latency, context length, and your real workload.
- 1GPT-5 nanoOpenAI944.0047.2 / $0.05/M
- 2Gemini 2.0 FlashGoogle573.7057.4 / $0.10/M
- 3Gemini 1.5 FlashGoogle496.8037.3 / $0.08/M
- 4Qwen3 235B (Thinking)Alibaba (Qwen)376.0075.2 / $0.20/M
- 5Qwen3 235BAlibaba (Qwen)304.3060.9 / $0.20/M
- 6Gemini 3 FlashGoogle303.5791.1 / $0.30/M
- 7DeepSeek V3 (Thinking)DeepSeek289.5678.2 / $0.27/M
- 8DeepSeek V3DeepSeek289.3378.1 / $0.27/M
- 9Gemini 2.5 FlashGoogle254.9776.5 / $0.30/M
- 10GPT-5 miniOpenAI252.9663.2 / $0.25/M
- 11GPT-4o miniOpenAI247.9337.2 / $0.15/M
- 12Llama 4 ScoutMeta197.0035.5 / $0.18/M
- 13GLM-4.6Zhipu AI (GLM)166.4883.2 / $0.50/M
- 14GLM-4.7Zhipu AI (GLM)163.8481.9 / $0.50/M
- 15DeepSeek R1DeepSeek144.4479.4 / $0.55/M
- 16Llama 4 MaverickMeta139.0737.5 / $0.27/M
- 17Kimi K2Moonshot (Kimi)125.7275.4 / $0.60/M
- 18Gemini 3 ProGoogle77.8497.3 / $1.25/M
- 19Gemini 2.5 ProGoogle71.2789.1 / $1.25/M
- 20Claude Haiku 4.5Anthropic67.5767.6 / $1.00/M
- 21GPT-5.1OpenAI66.8183.5 / $1.25/M
- 22GPT-5.5OpenAI61.9592.9 / $1.50/M
- 23GPT-5.2OpenAI61.2076.5 / $1.25/M
- 24GPT-5.4OpenAI61.1591.7 / $1.50/M
- 25GPT-5OpenAI58.1472.7 / $1.25/M
AI Model Analyzer does not recommend specific vendors; rankings are derived from public data only.