AIBenchmarks
RankingsCompareGemini 1.5 Pro vs Claude 3.5 Sonnet

Gemini 1.5 ProvsClaude 3.5 Sonnet

Google

Gemini 1.5 Pro

Google's 1M-token multimodal model

Arena ELO
1261
Context
1000K
Speed
150 t/s
Input/1M
$1.25
Anthropic

Claude 3.5 Sonnet

🏆 Overall Winner

Anthropic's most intelligent model

Arena ELO
1298
Context
200K
Speed
85 t/s
Input/1M
$3

Capability Radar

Category performance across 6 domains

Benchmark Scores

MMLU · HumanEval · MATH · GSM8K · GPQA · BBH

Gemini 1.5 Pro

Pros

1M token context window
Fastest response times
Best price-to-performance

Cons

Slightly lower reasoning vs GPT-4o/Claude
Less consistent instruction following

Claude 3.5 Sonnet

Pros

Best coding (SWE-bench leader)
200K context window
Exceptional instruction following

Cons

Slower than GPT-4o
No native audio capabilities

🏆 Our Verdict

Based on overall benchmark averages, Claude 3.5 Sonnet has the edge with an average score of 84.6% across all benchmarks. However, the best choice depends on your use case — Gemini 1.5 Pro excels in 1m token context window, while Claude 3.5 Sonnet stands out for best coding (swe-bench leader).

More Comparisons

Gemini 1.5 Pro vs GPT-4oClaude 3.5 Sonnet vs GPT-4oGemini 1.5 Pro vs Llama 3.1 405BClaude 3.5 Sonnet vs Llama 3.1 405BGemini 1.5 Pro vs Grok 2Claude 3.5 Sonnet vs Grok 2Gemini 1.5 Pro vs Mistral Large 2Claude 3.5 Sonnet vs Mistral Large 2