Technique · architecture
Mamba / Selective State Space Models
A state-space sequence model with input-dependent selection that matches Transformer quality with linear inference cost and unlimited context.
1
Products deploying
2y
Avg research → prod
2y
First commercial deploy
Deployment timeline
- Nemotron 3 Superhigh
Deployed 2026-03-11 · Velocity 2y
“Nemotron 3 Super uses a hybrid Mamba-Transformer MoE architecture.”