AI models ranked by coding ability across benchmarks, real-world usage, and developer sentiment. 排名每小时更新,使用实时数据,包括基准测试、Elo评级、社区情绪和采用指标。
| Rank | Model | Provider | Score | Status | Actions |
|---|---|---|---|---|---|
1 | OpenAI | 94 | |||
2 | GPT-5.42nd | OpenAI | 94 | ||
3 | GPT-5.4 Mini3rd | OpenAI | 93 | ||
4 | OpenAI | 93 | |||
5 | OpenAI | 93 | |||
6 | Anthropic | 92 | |||
7 | OpenAI | 92 | |||
8 | OpenAI | 92 | |||
9 | Anthropic | 90 | |||
10 | OpenAI | 90 | |||
11 | 89 | May change | |||
12 | Anthropic | 89 | |||
13 | Anthropic | 89 | |||
14 | OpenAI | 88 | |||
15 | xAI | 87 | |||
16 | xAI | 86 | |||
17 | xAI | 86 | May change | ||
18 | 86 | May change | |||
19 | OpenAI | 86 | May change | ||
20 | OpenAI | 85 | |||
21 | Xiaomi | 85 | May change | ||
22 | Xiaomi | 85 | May change | ||
23 | OpenAI | 85 | May change | ||
24 | ByteDance | 85 | May change | ||
25 | Alibaba | 85 | |||
26 | OpenAI | 85 | May change | ||
27 | ByteDance | 85 | May change | ||
28 | 85 | ||||
29 | OpenAI | 85 | |||
30 | Alibaba | 85 | |||
31 | Moonshot AI | 85 | May change | ||
32 | OpenAI | 85 | May change | ||
33 | ByteDance | 85 | |||
34 | ByteDance | 85 | May change | ||
35 | OpenAI | 85 | May change | ||
36 | OpenAI | 85 | May change | ||
37 | OpenAI | 85 | May change | ||
38 | OpenAI | 85 | May change | ||
39 | Perplexity | 85 | May change | ||
40 | Alibaba | 85 | |||
41 | OpenAI | 85 | May change | ||
42 | Alibaba | 85 | May change | ||
43 | OpenAI | 85 | May change | ||
44 | OpenAI | 85 | May change | ||
45 | 85 | May change | |||
46 | xAI | 85 | May change | ||
47 | NVIDIA | 84 | May change | ||
48 | 84 | May change | |||
49 | 84 | May change | |||
50 | OpenAI | 84 | May change |
我们的llm模型排名使用综合评分系统,结合多个信号为您提供每个模型优缺点的最完整图景。
Performance on standardized coding, reasoning, and category-specific benchmarks.
Head-to-head comparison ratings from AI chatbot arenas and blind testing.
Analysis of discussions on Reddit, Twitter/X, and developer forums.
Real-world usage data, API traffic patterns, and growth trajectories.
Search volume and interest trends for model-related queries.
Stars, forks, and contributor activity for open-source models and integrations.
Performance-per-dollar analysis based on API pricing and output quality.
Real-time API latency measurements and throughput testing.
分数归一化为0-100分制。排名每小时更新。了解更多关于our methodology.
As of our latest rankings, GPT-5.4 Pro leads the llm category with a composite score of 94. Rankings are updated hourly based on benchmarks, adoption metrics, community sentiment, and more.
We use a composite scoring system that combines multiple signals: benchmark performance, Elo ratings, repository popularity, community sentiment, API latency, cost efficiency, adoption rates, and expert reviews. Each signal is normalized and weighted to produce a final score.
We currently track 50 AI models in the llm category. Our coverage is expanding as new models are released.
Rankings refresh every hour. We continuously monitor all signals and recalculate scores to ensure you always see the most up-to-date information.
Yes! Click on any two models to see a detailed head-to-head comparison, including signal-by-signal breakdowns, pricing calculators, and personalized recommendations.