About

Intelligence ScoreEqual-weight average of exactly 3 benchmarks, chosen for high ceilings, contamination resistance, and real-world predictive value. Each is normalized to 0–1 using fixed goalposts, then averaged and scaled to 0–100. All 3 are required — models missing any benchmark show “—” instead of a score.

ReasoningGPQA Diamond
Expert ReasoningHumanity's Last Exam
CodingLiveCodeBench

Other benchmarks (SWE-Bench, AIME, MMLU-Pro, MMMU-Pro, Chatbot Arena Elo) are shown in model details but excluded from the composite. Speed and cost reflect best available provider.

Blended CostCost is shown as a weighted average of input and output token prices: (3 × input + 1 × output) ÷ 4. This reflects typical usage where prompts are longer than completions. Where a model is available from multiple providers, the lowest blended cost is used.

SourcesArtificial Analysis, SWE-bench, MathArena, Chatbot Arena, and provider documentation.

Built withNext.js, visx, Tailwind CSS, and Vercel.

Created by@davidhariri

Model Browser

Compare large language models side by side. Filter GPT-4o, Claude, Gemini, Llama, Mistral, DeepSeek, and Qwen models by intelligence benchmarks like GPQA Diamond, HLE, and LiveCodeBench. Sort by API pricing, tokens per second, and overall score. Find the best LLM for your use case.

0
0
Minimum Intelligence
50
$50 /1M
Maximum Blended Cost
No possible models
Intelligenceby

All Models

ModelIntelligenceCostSpeed
Gemini 3.1 Pro~81$4.50130tok/s
GPT-5.4~78$5.6387tok/s
Gemini 3 Pro76$4.50142tok/s
Gemini 3 Flash75$1.13215tok/s
GPT-5.274$4.8188tok/s
Claude Opus 4.672$10.0073tok/s
GPT-5.169$3.44136tok/s
Claude Opus 4.569$10.0088tok/s
Kimi K2.569$1.07345tok/s
Qwen3.5 397B68$1.3574tok/s
GLM-4.768$1.00136tok/s
GPT-567$3.4498tok/s
Grok 466$6.0036tok/s
Claude Sonnet 4.665$6.0064tok/s
DeepSeek V3.265$0.32219tok/s
GPT-5 mini63$0.6975tok/s
o362$3.50141tok/s
Gemini 2.5 Pro62$3.44159tok/s
Grok 4.1 Fast62$0.28150tok/s
GPT-OSS 120B61$0.262951tok/s
o4-mini60$1.93134tok/s
Claude Sonnet 4.557$6.00103tok/s
DeepSeek R157$2.36306tok/s
MiniMax M2.557$0.52273tok/s
DeepSeek V3.155$0.84347tok/s
GLM-555$1.55266tok/s
Gemini 2.5 Flash52$0.85282tok/s
Magistral Medium51$2.7529tok/s
GPT-5 nano49$0.14131tok/s
o148$26.25174tok/s
Claude Haiku 4.537$2.00109tok/s
Mistral Large 337$0.75147tok/s
GPT-4.136$3.50104tok/s
GPT-4.1 mini36$0.7078tok/s
Llama 4 Maverick34$0.30434tok/s
DeepSeek V328$0.52120tok/s
GPT-4.1 nano24$0.18142tok/s
Command A24$4.3849tok/s
GPT-4o23$4.38170tok/s
GPT-4o mini17$0.2654tok/s
Command R+9$4.3859tok/s