Zed is a high-performance code editor with built-in AI assistant support. It connects to multiple model providers for inline completions and chat.
Best Models for Zed
Top 15 by tool-optimized score
Scored by: benchmark performance (90%) from MMLU, GPQA, HumanEval, SWE-bench, and 15+ standardized evaluations, with capabilities and context as tiebreakers (10%).
| # | 模型 | 评分 | 输出$/百万 |
|---|---|---|---|
| 1 | Claude Opus 4.7 Arena Elo: 1491 | 87 | $25.00 |
| 2 | Gemini 3.1 Pro Preview Arena Elo: 1494 | 87 | $12.00 |
| 3 | GPT-5.5 Arena Elo: 1475 | 85 | $30.00 |
| 4 | MiMo-V2.5-Pro Arena Elo: 1464 | 85 | $3.00 |
| 5 | GLM 5.1 Arena Elo: 1471 | 85 | $3.50 |
| 6 | Grok 4.1 Fast Arena Elo: 1467 | 85 | $0.500 |
| 7 | Grok 4.3 Arena Elo: 1455 | 84 | $2.50 |
| 8 | Qwen3.6 Max Preview Arena Elo: 1457 | 84 | $6.24 |
| 9 | DeepSeek V4 Pro Arena Elo: 1463 | 84 | $0.870 |
| 10 | Kimi K2.6 Arena Elo: 1462 | 84 | $3.50 |
| 11 | GLM 5 Arena Elo: 1457 | 84 | $1.92 |
| 12 | Claude Opus 4.6 (Fast) | 83 | $150.00 |
| 13 | Gemma 4 31B Arena Elo: 1451 | 83 | $0.380 |
| 14 | Qwen3.6 Plus Arena Elo: 1448 | 83 | $1.95 |
| 15 | MiMo-V2-Pro Arena Elo: 1447 | 83 | $3.00 |
| 16 | GPT-5.4 Pro | 83 | $180.00 |
| 17 | Qwen3.5 397B A17B Arena Elo: 1446 | 83 | $2.34 |
| 18 | GLM 4.7 Arena Elo: 1443 | 83 | $1.75 |
| 19 | GPT-5.2 Pro | 83 | $168.00 |
| 20 | Claude Opus 4.1 Arena Elo: 1449 | 83 | $75.00 |
| 21 | DeepSeek V4 Flash Arena Elo: 1433 | 82 | $0.280 |
| 22 | Gemma 4 26B A4B Arena Elo: 1438 | 82 | $0.330 |
| 23 | Grok 4.20 | 82 | $2.50 |
| 24 | Gemini 3.1 Flash Lite Preview Arena Elo: 1438 | 82 | $1.50 |
| 25 | GPT-5.3-Codex | 82 | $14.00 |
| 26 | GPT-5.2-Codex | 82 | $14.00 |
| 27 | GPT-5.2 Chat Arena Elo: 1477 | 82 | $14.00 |
| 28 | GPT-5 Pro | 82 | $120.00 |
| 29 | Claude 3.5 Haiku HumanEval: 88.1% | 82 | $4.00 |
| 30 | Hy3 preview Arena Elo: 1418 | 81 | $0.260 |
| 31 | MiMo-V2.5 Arena Elo: 1423 | 81 | $2.00 |
| 32 | Qwen3.5-122B-A10B Arena Elo: 1418 | 81 | $2.08 |
| 33 | GPT-5.1-Codex-Max | 81 | $10.00 |
| 34 | GPT-5.1-Codex | 81 | $10.00 |
| 35 | GPT-5.1-Codex-Mini | 81 | $2.00 |
| 36 | o3 Deep Research | 81 | $40.00 |
| 37 | GLM 4.6 Arena Elo: 1426 | 81 | $1.90 |
| 38 | GPT-5 Codex | 81 | $10.00 |
| 39 | Grok 4 Fast Arena Elo: 1421 | 81 | $0.500 |
| 40 | o3 Pro | 81 | $80.00 |
| 41 | MiniMax M2.7 Arena Elo: 1407 | 80 | $1.20 |
| 42 | Qwen3.5-27B Arena Elo: 1406 | 80 | $1.56 |
| 43 | Qwen3 VL 235B A22B Instruct Arena Elo: 1415 | 80 | $0.880 |
| 44 | Mistral Large HumanEval: 92% | 80 | $6.00 |
| 45 | Qwen3.5-35B-A3B Arena Elo: 1397 | 79 | $1.00 |
| 46 | Qwen3.5-Flash Arena Elo: 1398 | 79 | $0.260 |
| 47 | Claude Opus 4.6 SWE-bench: 83.7% | 79 | $25.00 |
| 48 | Step 3.5 Flash Arena Elo: 1393 | 79 | $0.300 |
| 49 | DeepSeek V3.2 Exp Arena Elo: 1423 | 79 | $0.410 |
| 50 | DeepSeek V3.1 Terminus Arena Elo: 1416 | 79 | $0.950 |
Based on our analysis of coding benchmarks, capability matching, and pricing, Claude Opus 4.7 currently ranks #1 for Zed. Rankings are rebuilt as benchmark, pricing, and provider data refresh.
We score models using benchmark performance (90%) from LMArena, HumanEval, SWE-bench, MMLU, and 15+ standardized evaluations. Capabilities and context serve as tiebreakers (10%). Only models with the capabilities Zed needs are included in the tool-specific rankings.
We currently track 341 AI models compatible with Zed. This includes models from OpenAI, Anthropic, Google, DeepSeek, and other providers accessible via API.
Many open-source models are compatible with Zed through API providers like OpenRouter, Together AI, and Groq. Check our rankings to see which open-source models perform best.
Rankings refresh whenever the underlying benchmark, pricing, and catalog sources refresh. That means some signals update faster than others, and the page reflects the latest verified source data available.