NVIDIA Nemotron 3 Super is a 120B-parameter open hybrid MoE model, activating just 12B parameters for maximum compute efficiency and accuracy in complex multi-agent applications. Built on a hybrid Mamba-Transformer Mixture-of-Experts architecture with multi-token prediction (MTP), it delivers over 50% higher token generation compared to leading open models. The model features a 1M token context window for long-term agent coherence, cross-document reasoning, and multi-step task planning. Latent MoE enables calling 4 experts for the inference cost of only one, improving intelligence and generalization. Multi-environment RL training across 10+ environments delivers leading accuracy on benchmarks including AIME 2025, TerminalBench, and SWE-Bench Verified. Fully open with weights, datasets, and recipes under the NVIDIA Open License, Nemotron 3 Super allows easy customization and secure deployment anywhere — from workstation to cloud.
| 信号 | 强度 | 权重 | 影响 |
|---|---|---|---|
| Capabilitiesjust now | 67 | 30% | +20.0 |
| Recencyjust now | 100 | 15% | +15.0 |
| Context Windowjust now | 86 | 15% | +12.9 |
| Output Capacityjust now | 20 | 15% | +3.0 |
| Pricingjust now | 1 | 25% | +0.1 |
社区和从业者反馈在基准测试和价格之上增加了真实世界的信号。
Share your experience with Nemotron 3 Super and help the community make better decisions.
成本估算器
每月比类别平均节省$40.34