by
Reka Flash 3 is a general-purpose, instruction-tuned large language model with 21 billion parameters, developed by Reka. It excels at general chat, coding tasks, instruction-following, and function calling. Featuring a 32K context length and optimized through reinforcement learning (RLOO), it provides competitive performance comparable to proprietary models within a smaller parameter footprint. Ideal for low-latency, local, or on-device deployments, Reka Flash 3 is compact, supports efficient quantization (down to 11GB at 4-bit precision), and employs explicit reasoning tags ("<reasoning>") to indicate its internal thought process. Reka Flash 3 is primarily an English model with limited multilingual understanding capabilities. The model weights are released under the Apache 2.0 license.
| Signal | Strength | Weight | Impact |
|---|---|---|---|
| Output Capacityjust now | 80 | 15% | +12.0 |
| Context Windowjust now | 76 | 15% | +11.5 |
| Capabilitiesjust now | 33 | 30% | +10.0 |
| Recencyjust now | 63 | 15% | +9.4 |
| Pricingjust now | 0 | 25% | +0.1 |
View this model against the provider’s recent shipping cadence.
Community and practitioner feedback adds real-world signal on top of benchmarks and pricing.
Share your experience with Reka Flash 3 and help the community make better decisions.
Pricing, benchmarks, and reliability come from different data surfaces, so they refresh on different cadences. The timestamps above show the latest verification point we have for each one.
Cost Estimator
You save $39.68/month vs category average