Gemma 4
AI Model Comparison · April 2026

How does
Gemma 4
stack up?

Side-by-side benchmarks, capability radar charts, and use-case rankings across Gemma 4, GPT-4o, Claude 3.5 Sonnet, Llama 4, Qwen 3.5, and Mistral Small 4.

Live data April 2026 6 models 8 benchmarks Apache 2.0 open
🏆
Gemma 4 31B — #3 Open Model Globally
Ranked #3 on Arena.ai leaderboard with an ELO of 1441. Outperforms models 20× its size. The most parameter-efficient open reasoning engine available — and it's free, Apache 2.0.
AIME 2026: 89.2% MMLU Pro: 85.2% Codeforces ELO: 2150 Arena #3 global Apache 2.0
Models
select to filter
Gemma 4 31B
GPT-4o
Claude 3.5 Sonnet
Llama 4 Maverick
Qwen 3.5 27B
Mistral Small 4
Capability Radar

Normalised scores across six capability dimensions. Larger area = stronger overall profile. Toggle models above to add / remove layers.

Model Developer Params (Total) Active Params Context Architecture License Vision Audio Thinking Fn Call Min VRAM Arena ELO MMLU Pro