Anthropic (14 models) vs Google (29 models) - compared across composite scores, pricing, capabilities, and context windows.
| Capability | Anthropic | Leader | |
|---|---|---|---|
Vision | 14/14 | 25/29 | |
Reasoning | 12/14 | 17/29 | |
Function Calling | 14/14 | 19/29 | |
JSON Mode | 8/14 | 26/29 | |
Web Search | 13/14 | 16/29 | |
Streaming | 14/14 | 27/29 | |
Image Output | 0/14 | 4/29 |
| Metric | Anthropic | |
|---|---|---|
| Cheapest Input (per 1M tokens) | $0.250 Claude 3 Haiku | $0.040 Gemma 3 4B |
| Cheapest Output (per 1M tokens) | $1.25 | $0.080 |
| Most Expensive Input (per 1M tokens) | $30.00 Claude Opus 4.6 (Fast) | $2.00 Gemini 3.1 Pro Preview Custom Tools |
| Most Expensive Output (per 1M tokens) | $150.00 | $12.00 |
| Free Models | 0 | 4 |
| Max Context Window | 1.0M | 1.0M |
| Model | Score | Input $/M | Output $/M |
|---|---|---|---|
| Claude Opus 4.6 (Fast) | 90 | $30.00 | $150.00 |
| Claude Opus 4.6 | 90 | $5.00 | $25.00 |
| Claude Sonnet 4.6 | 85 | $3.00 | $15.00 |
| Claude Opus 4.5 | 85 | $5.00 | $25.00 |
| Claude Sonnet 4.5 | 82 | $3.00 | $15.00 |
| Claude Opus 4 | 82 | $15.00 | $75.00 |
| Claude Opus 4.7 | 79 | $5.00 | $25.00 |
| Claude Opus 4.1 | 75 | $15.00 | $75.00 |
| Claude 3.7 Sonnet (thinking) | 75 | $3.00 | $15.00 |
| Claude Sonnet 4 | 74 | $3.00 | $15.00 |
| Claude 3.7 Sonnet | 73 | $3.00 | $15.00 |
| Claude Haiku 4.5 | 70 | $1.00 | $5.00 |
| Claude 3.5 Haiku | 58 | $0.800 | $4.00 |
| Claude 3 Haiku | 50 | $0.250 | $1.25 |
| Model | Score | Input $/M | Output $/M |
|---|---|---|---|
| Gemini 3 Flash Preview | 88 | $0.500 | $3.00 |
| Gemini 2.5 Pro | 84 | $1.25 | $10.00 |
| Gemini 2.5 Pro Preview 06-05 | 84 | $1.25 | $10.00 |
| Gemini 2.5 Pro Preview 05-06 | 84 | $1.25 | $10.00 |
| Gemini 3.1 Pro Preview Custom Tools | 81 | $2.00 | $12.00 |
| Gemini 3.1 Pro Preview | 81 | $2.00 | $12.00 |
| Gemma 4 31B (free) | 81 | Free | Free |
| Gemma 4 31B | 81 | $0.130 | $0.380 |
| Gemini 3.1 Flash Lite Preview | 80 | $0.250 | $1.50 |
| Gemini 2.5 Flash Lite Preview 09-2025 | 79 | $0.100 | $0.400 |
| Gemini 2.5 Flash Lite | 79 | $0.100 | $0.400 |
| Gemini 2.5 Flash | 79 | $0.300 | $2.50 |
| Gemma 2 27B | 77 | $0.650 | $0.650 |
| Gemma 4 26B A4B (free) | 73 | Free | Free |
| Gemma 4 26B A4B | 73 | $0.060 | $0.330 |
| Gemini 2.0 Flash | 72 | $0.100 | $0.400 |
| Gemini 2.0 Flash Lite | 59 | $0.075 | $0.300 |
| Lyria 3 Pro Preview | 40 | Free | Free |
| Lyria 3 Clip Preview | 40 | Free | Free |
| Gemma 3n 4B | 40 | $0.060 | $0.120 |
Compare any two AI providers side-by-side.
Google's portfolio strategy prioritizes accessibility and specialization: 9 models are completely free and 15 are open source, compared to Anthropic's zero in both categories. Their pricing spans from $0.040 to $12.00 per 1M tokens (300x range) versus Anthropic's narrower $1.25 to $150.00 range, suggesting Google targets everything from hobbyists to enterprise while Anthropic focuses exclusively on premium performance.
Anthropic has equipped 12 of 13 models with web search capabilities while Google provides zero web search across all 34 models, likely reflecting different architectural philosophies. This counterintuitive gap suggests Anthropic views real-time information access as core to assistant functionality, while Google may prefer keeping search separate from their LLM offerings to maintain clearer product boundaries.
Anthropic provides function calling on 100% of models (13/13) while Google covers only 47% (16/34), indicating Anthropic's all-in commitment to production API use cases. Combined with Anthropic's higher average score (55 vs 45), this suggests Anthropic optimizes for developers building sophisticated applications, while Google's broader portfolio includes many models unsuitable for complex integrations.
Despite equal context limits, the providers diverge sharply on vision economics: Anthropic covers 100% of models (13/13) with vision but starts at $1.25/M tokens, while Google provides vision on 79% of models (27/34) with options as low as $0.040/M tokens. For high-volume vision tasks, Google's 31x cheaper entry point likely outweighs Anthropic's 6-point performance advantage unless quality is absolutely critical.
The 10-point average score gap (Anthropic 55 vs Google 45) and reasoning capability coverage (Anthropic 85% vs Google 47%) justify the premium for production workloads where reliability matters. Anthropic's narrower 13-model portfolio also means less evaluation overhead compared to Google's 34 options, making it attractive for teams that value curation over choice.