Which AI models are the most consistent over time? This report analyzes rank changes, state classifications, and sparkline volatility across 300 tracked models to produce a stability score from 0 to 100.
Rock Solid
55
Consistent
69
Variable
35
Volatile
141
Top 20 models with the highest stability scores. These models maintain consistent rankings with minimal volatility.
| # | Model | Score | Stability | 24h | 7d |
|---|---|---|---|---|---|
| 1 | GPT-5.4OpenAI | 94.0 | 100 | -1 | -1 |
| 2 | o3 Deep ResearchOpenAI | 91.5 | 100 | -1 | -1 |
| 3 | GPT-5OpenAI | 90.0 | 100 | 0 | +1 |
| 4 | o3 ProOpenAI | 87.6 | 100 | 0 | +1 |
| 5 | Qwen Plus 0728Alibaba | 76.8 | 100 | 0 | +1 |
| 6 | GPT Audio MiniOpenAI | 68.4 | 100 | -1 | -1 |
| 7 | SonarPerplexity | 53.5 | 100 | 0 | -1 |
| 8 | LFM2-8B-A1BLiquid AI | 53.2 | 100 | -1 | -1 |
| 9 | Olmo 2 32B InstructAllen AI | 44.3 | 100 | -1 | 0 |
| 10 | GPT-4 Turbo (older v1106)OpenAI | 42.7 | 100 | 0 | -1 |
| 11 | GPT-4OpenAI | 39.0 | 100 | 0 | +1 |
| 12 | Llama 3.2 3B InstructMeta | 35.8 | 100 | -1 | -1 |
| 13 | Llama 3.2 3B Instruct (free)Meta | 35.1 | 100 | 0 | -1 |
| 14 | GPT-3.5 Turbo InstructOpenAI | 32.2 | 100 | 0 | +2 |
| 15 | WizardLM-2 8x22BMicrosoft | 32.0 | 100 | +1 | 0 |
| 16 | Gemma 2 9BGoogle | 30.1 | 100 | -1 | 0 |
| 17 | Mistral 7B Instruct v0.1Mistral AI | 19.8 | 100 | 0 | 0 |
| 18 | Grok 4.1 FastxAI | 86.9 | 100 | 0 | -3 |
| 19 | GPT-5.4 MiniOpenAI | 93.3 | 99 | -1 | +1 |
| 20 | Llemma 7beleutherai | 47.3 | 98 | 0 | +3 |
Bottom 20 models with the lowest stability scores. These models show significant ranking fluctuations or inconsistent states.
| # | Model | Score | Stability | 24h | 7d |
|---|---|---|---|---|---|
| 1 | Ministral 3 8B 2512Mistral AI | 73.5 | 35 | -18 | +17 |
| 2 | Ministral 3 14B 2512Mistral AI | 73.5 | 35 | -20 | +16 |
| 3 | Devstral 2 2512Mistral AI | 67.7 | 35 | -10 | +14 |
| 4 | Composer 2Cursor | 76.4 | 35 | +10 | +15 |
| 5 | Mistral Small CreativeMistral AI | 59.0 | 35 | -14 | +7 |
| 6 | Mistral Small 3.2 24BMistral AI | 67.2 | 35 | +15 | +9 |
| 7 | Mistral Medium 3.1Mistral AI | 70.2 | 35 | -10 | +9 |
| 8 | Olmo 3.1 32B InstructAllen AI | 64.9 | 35 | +8 | +22 |
| 9 | MiMo-V2-OmniXiaomi | 85.0 | 35 | +22 | +16 |
| 10 | GPT-4o Search PreviewOpenAI | 63.4 | 35 | +7 | +7 |
| 11 | gpt-oss-20b (free)OpenAI | 73.7 | 35 | +25 | +20 |
| 12 | Llama 3.2 11B Vision InstructMeta | 54.3 | 36 | -7 | +7 |
| 13 | o4 Mini Deep ResearchOpenAI | 85.0 | 36 | -7 | +11 |
| 14 | GPT-5 NanoOpenAI | 75.5 | 36 | -8 | +18 |
| 15 | Seed 1.6ByteDance | 85.0 | 36 | -6 | +12 |
| 16 | GPT-4.1 NanoOpenAI | 80.5 | 36 | +12 | +12 |
| 17 | GPT-5.3 ChatOpenAI | 85.0 | 36 | +31 | +16 |
| 18 | gpt-oss-120b (free)OpenAI | 73.7 | 36 | -6 | +28 |
| 19 | Qwen Plus 0728 (thinking)Alibaba | 82.7 | 36 | -9 | +7 |
| 20 | Claude Opus 4.1Anthropic | 81.9 | 36 | -25 | +8 |
Aggregated stability metrics per provider. Providers are ranked by their average stability score across all models.
| Provider | Models | Avg Stability |
|---|---|---|
| eleutherai | 1 | 98.0 |
| Windsurf | 1 | 93.7 |
| Inflection | 2 | 91.4 |
| Microsoft | 2 | 90.7 |
| Vercel | 1 | 87.5 |
| JetBrains | 1 | 82.9 |
| Cohere | 4 | 75.0 |
| Meituan | 1 | 74.6 |
| Meta | 14 | 71.5 |
| Anthropic | 13 | 68.7 |
| OpenAI | 60 | 63.9 |
| StepFun | 2 | 63.9 |
| Liquid AI | 5 | 63.8 |
| Allen AI | 4 | 62.1 |
| xAI | 10 | 61.4 |
| DeepSeek | 11 | 59.6 |
| Mistral AI | 25 | 59.6 |
| Upstage | 1 | 58.2 |
| Perplexity | 5 | 57.8 |
| Amazon | 5 | 57.5 |
| Alibaba | 50 | 57.0 |
| 23 | 56.5 | |
| Cursor | 2 | 56.5 |
| aion-labs | 3 | 56.3 |
| reka | 1 | 55.9 |
| arcee-ai | 7 | 55.0 |
| NVIDIA | 11 | 52.8 |
| MiniMax | 8 | 52.1 |
| Xiaomi | 3 | 50.8 |
| Inception | 3 | 49.9 |
| ByteDance | 5 | 48.9 |
| Baidu | 5 | 47.8 |
| AI21 Labs | 1 | 45.7 |
| IBM | 1 | 44.1 |
| essentialai | 1 | 40.8 |
| Moonshot AI | 4 | 40.1 |
| Writer | 1 | 37.7 |
| Tencent | 1 | 37.5 |
| Kuaishou | 1 | 37.4 |
| deepcogito | 1 | 36.0 |
How stability scores are distributed across all 300 tracked models.
Our stability scoring system uses three key signals to measure how consistently a model performs over time.
The most direct measure of stability. Models lose up to 25 points for large 24-hour rank changes (5 points per rank position moved) and up to 21 points for 7-day changes (3 points per position). Models that hold their rank tightly score higher.
Each model has a state reflecting its overall reliability. Models in a "stable" state receive a 10-point bonus, while "fragile" models are penalized 15 points. This captures systemic reliability beyond simple rank movement.
The 14-day sparkline data reveals hidden volatility. We compute the standard deviation of the sparkline and subtract up to 20 points. Even models that end where they started can be penalized if they oscillated wildly along the way.
The stability score starts at 100 and is reduced based on three factors: 24-hour rank changes (up to -25 points, at 5 per position moved), 7-day rank changes (up to -21 points, at 3 per position), and sparkline volatility measured by standard deviation (up to -20 points). Models in a "stable" state get a +10 bonus, while "fragile" models lose 15 points.
Models are classified into four tiers based on their stability score: "Rock Solid" (85-100) means extremely consistent performance with minimal fluctuation. "Consistent" (70-84) means generally reliable with minor variations. "Variable" (50-69) shows noticeable ranking fluctuations. "Volatile" (below 50) indicates significant instability and unpredictable performance.
Stability indicates how predictably a model will perform over time. A highly rated but volatile model may deliver inconsistent results, which is problematic for production applications requiring reliable output quality. Stable models provide more predictable performance, making them safer choices for mission-critical workloads even if they do not always hold the top rank.