Compare workspace
Curated / indexable

Claude Sonnet 4.6 vs Gemini 3.1 Pro Preview

Claude Sonnet 4.6 and Gemini 3.1 Pro Preview are both strong shortlist candidates, but they usually appeal to different buyers once context posture and ecosystem preference are explicit.

Overall: Google: Gemini 3.1 Pro PreviewAnthropic: Claude Sonnet 4.6: Long-context research / MultimodalGoogle: Gemini 3.1 Pro Preview: Long-context research / Agent workflows
Verdict

Go with Gemini 3.1 Pro Preview when Google alignment and aggressive input pricing are high on the checklist. Go with Claude Sonnet 4.6 when you want Anthropic’s premium production profile and a direct alternative to the Google path.

Biggest tradeoff

This pair tends to pivot on stack and pricing more than on a clean universal quality gap. Gemini can look friendlier at the front of the cost curve, while Claude often earns consideration through its Anthropic positioning and workflow familiarity.

Quick Decision Cards

Winner cards before the full matrix

These cards call out the most useful early distinctions without hiding the fact that different public fields may point to different winners.

Best reasoning
Google: Gemini 3.1 Pro Preview
57

Highest reasoning score from the currently public benchmark fields.

Best coding
Google: Gemini 3.1 Pro Preview
56

Best coding posture from AA Coding Index, LiveCodeBench, or SWE Bench when present.

Lowest input cost
Google: Gemini 3.1 Pro Preview
$2.00

Lowest currently published input-token price.

Largest context
Google: Gemini 3.1 Pro Preview
1049K

Largest resolved context window from the public detail dataset.

Use-Case Framing

Which buyer questions this page is built to answer

Best for buyers deciding between Google and Anthropic before they ever get to an OpenAI branch.

Best when long-context analysis, price posture, and platform preference all influence the shortlist.

Best for teams that want an indexable alternative to the usual OpenAI-centered comparison pages.

Full Matrix

Every public compare field grouped by job to be done

Missing values stay visible as N/A, and softly tinted cells mark the leading value in each comparable row so the matrix scans faster.

Overview

Decision-first fields that summarize fit before the deeper benchmark matrix.

Field
Anthropic: Claude Sonnet 4.6
anthropic
Google: Gemini 3.1 Pro Preview
google
Creatoranthropicgoogle
Overall profileSelective fitSelective fit
Best forLong-context research / MultimodalLong-context research / Agent workflows
Vision supportYesYes
New in 2026YesYes

Intelligence / Reasoning

Broad reasoning quality, knowledge depth, and flagship benchmark posture.

Field
Anthropic: Claude Sonnet 4.6
anthropic
Google: Gemini 3.1 Pro Preview
google
Reasoning labelSituationalSituational
Intelligence score4457
Intelligence Index44.457.2
AA Intelligence Index44.457.2
MMLU ProN/AN/A
GPQA79.9%94.1%
HLE13.2%44.7%
Arena ELON/AN/A

Coding

Signals that matter for code generation, refactors, debugging, and software tasks.

Field
Anthropic: Claude Sonnet 4.6
anthropic
Google: Gemini 3.1 Pro Preview
google
Coding score4656
AA Coding Index46.455.5
LiveCodeBenchN/AN/A
LiveBenchN/AN/A
SWE BenchN/AN/A
SciCode46.9%58.9%

Math

Published math-oriented signals, including both summary indexes and narrower benchmark cuts.

Field
Anthropic: Claude Sonnet 4.6
anthropic
Google: Gemini 3.1 Pro Preview
google
Math scoreN/AN/A
AA Math IndexN/AN/A
Math 500N/AN/A
AIMEN/AN/A
AIME 25N/AN/A

Agent / Tool Use

Signals that better reflect tool loops, long-running tasks, and agent-style workflows.

Field
Anthropic: Claude Sonnet 4.6
anthropic
Google: Gemini 3.1 Pro Preview
google
Agent score5675
IFBench41.2%77.1%
TAU279.5%95.6%
TerminalBench Hard46.2%53.8%
LCR57.7%72.7%

Latency / Speed

Interactive responsiveness and throughput signals from the public detail dataset.

Field
Anthropic: Claude Sonnet 4.6
anthropic
Google: Gemini 3.1 Pro Preview
google
Latency tierBalancedHeavy
Speed labelSituationalLimited
Speed score5740
Tokens per second50113
TTFT1.13s22.16s
AA Tokens per second53115
AA TTFT0.97s20.66s
First answer token0.97s20.66s

Pricing

Published token pricing plus the lower-level OpenRouter and Artificial Analysis cost fields.

Field
Anthropic: Claude Sonnet 4.6
anthropic
Google: Gemini 3.1 Pro Preview
google
Price tierMid-rangeMid-range
Price labelCompetitiveCompetitive
Price score6262
Input price$3.00$2.00
Output price$15.00$12.00
AA input price$3.00$2.00
AA output price$15.00$12.00
AA blended 3:1$6.00$4.50
OR prompt price$3.0000$2.0000
OR completion price$15.0000$12.0000
OR request priceN/AN/A
OR image priceN/A$0.0000
OR audio priceN/A$0.0000
OR web search price$0.0100N/A
OR cache read price$0.0000$0.0000
OR cache write price$0.0000$0.0000
OR internal reasoning priceN/A$0.0000

Context

Window size and completion limits relevant to long-context tasks and workspace planning.

Field
Anthropic: Claude Sonnet 4.6
anthropic
Google: Gemini 3.1 Pro Preview
google
Context tierLargeLarge
Context labelAbove averageAbove average
Context score100100
Primary context window1000K Tokens1049K Tokens
OpenRouter context length1000K Tokens1049K Tokens
Top provider context1000K Tokens1049K Tokens
Max completion tokens12800065536

Modality / Vision

Modalities stay visible near the decision surface so multimodal support is easy to compare.

Field
Anthropic: Claude Sonnet 4.6
anthropic
Google: Gemini 3.1 Pro Preview
google
Vision supportYesYes
Modalitiestext, image->text, imagetext, image, file, audio, video->text, video
OpenRouter modalitytext+image->texttext+image+file+audio+video->text
OR input modalitiestext, imageaudio, file, image, text, video
OR output modalitiestexttext

Provider Internals

Lower-signal provider fields kept below the fold

Next step

Keep exploring from the curated hub or widen the shortlist in the leaderboard

Curated pages handle editorial intent. The leaderboard handles discovery. Custom compare URLs stay available for working sessions without being promoted as canonical landing pages.