Claude Opus 4.6 vs Gemini 3.1 Pro: Specs & Benchmark Comparison

CharacteristicClaude Opus 4.6Gemini 3.1 Pro
CompanyAnthropicGoogle
Release DateFebruary 4, 2026February 19, 2026
Parameters
MultimodalYesYes
Context (input)1.0M1.0M
Context (output)128K66K
Input Price / 1M$5.00$2.50
Output Price / 1M$25.00$15.00
Average Score0.80.8
Benchmarks
CharXiv-R0.70.9
ARC-AGI v20.70.8
Humanity's Last Exam0.50.5
GPQA0.90.9
SWE-Bench Verified0.80.8

Visual Benchmark Comparison

Claude Opus 4.6
Gemini 3.1 Pro
CharXiv-R0.7 vs 0.9
0.7
0.9
ARC-AGI v20.7 vs 0.8
0.7
0.8
Humanity's Last Exam0.5 vs 0.5
0.5
0.5
GPQA0.9 vs 0.9
0.9
0.9
SWE-Bench Verified0.8 vs 0.8
0.8
0.8

Verdict

Gemini 3.1 Pro leads in 3 out of 5 comparison categories.

Overall Performance

Both models show comparable average scores: Claude Opus 4.6 — 0.8, Gemini 3.1 Pro — 0.8.

Programming

On SWE-Bench, both models are nearly equal: Claude Opus 4.6 — 0.8, Gemini 3.1 Pro — 0.8.

API Cost

Gemini 3.1 Pro is 1.7x cheaper: input $2.50/1M vs $5.00/1M tokens.

Context Window

Gemini 3.1 Pro supports a larger context: 1M vs 1M tokens.

Recency

Gemini 3.1 Pro is newer: released 2/19/2026 vs 2/4/2026.

More About These Models

Related Comparisons

The Claude Opus 4.6 and Gemini 3.1 Pro comparison is updated for 2026. Data includes benchmark results, API pricing, context window size and other specifications. For more detailed information, visit the Claude Opus 4.6 or Gemini 3.1 Pro page. See also the complete list of AI model comparisons.