by ·
Gemma 4 26B A4B IT is an instruction-tuned Mixture-of-Experts (MoE) model from Google DeepMind. Despite 25.2B total parameters, only 3.8B activate per token during inference — delivering near-31B quality at a fraction of the compute cost. Supports multimodal input including text, images, and video (up to 60s at 1fps). Features a 256K token context window, native function calling, configurable thinking/reasoning mode, and structured output support. Released under Apache 2.0.
| Signal | Strength | Weight | Impact |
|---|---|---|---|
| Benchmarksjust now | 72 | 30% | +21.6 |
| Capabilitiesjust now | 83 | 20% | +16.7 |
| Recencyjust now | 100 | 15% | +15.0 |
| Pricingjust now | 100 | 15% | +14.9 |
| Output Capacityjust now | 90 | 10% | +9.0 |
| Context Windowjust now | 86 | 10% | +8.6 |
View this model against the provider’s recent shipping cadence.
Gemma 4 26B A4B Current
coding
Gemma 4 31B
coding
Lyria 3 Pro Preview
coding
Lyria 3 Clip Preview
coding
Gemini 3.1 Flash Lite Preview
coding
Nano Banana 2 (Gemini 3.1 Flash Image Preview)
image generation
Gemini 3.1 Pro Preview Custom Tools
coding
Gemini 3.1 Pro Preview
coding
Community and practitioner feedback adds real-world signal on top of benchmarks and pricing.
Share your experience with Gemma 4 26B A4B and help the community make better decisions.
Pricing, benchmarks, and reliability come from different data surfaces, so they refresh on different cadences. The timestamps above show the latest verification point we have for each one.
Cost Estimator
You save $38.88/month vs category average
From verified sources.