Mixture of Experts (Sparse MoE for LLMs)
An architecture where a router activates only a subset of expert sub-networks per token, scaling parameter count without proportional compute cost.
Signal Radar
Five-axis snapshot of this entity's footprint
Mentions × Lab Attention
Weekly mentions (solid) and average article relevance (dotted)
Timeline
1- Research MilestoneMar 11, 2026
New research reveals structural inference disadvantage via 'qs inequality', showing MoE models can be 4.5x slower than dense models
View source
Relationships
21Invented By
Uses
Introduces
- ←paper1 mention100% conf.
Prior Art
Deploys
Recent Articles
3NVIDIA Nemotron 3 Super: 120B Hybrid Mamba-Transformer MoE with 1M Context
~NVIDIA has released Nemotron 3 Super, a 120B parameter open hybrid Mamba-Transformer Mixture of Experts model with 12B active parameters and 1M token
95 relevanceCursor AI Claims 1.84x Faster MoE Inference on NVIDIA Blackwell GPUs
~Cursor AI announced a rebuilt inference engine for Mixture-of-Experts models on NVIDIA's new Blackwell GPUs, resulting in a claimed 1.84x speedup and
85 relevanceGoogle Releases Gemma 4 Family Under Apache 2.0, Featuring 2B to 31B Models with MoE and Multimodal Capabilities
~Google has released the Gemma 4 family of open-weight models, derived from Gemini 3 technology. The four models, ranging from 2B to 31B parameters and
100 relevance
Predictions
No predictions linked to this entity.
AI Discoveries
1- observationactiveMar 27, 2026
Lifecycle: Mixture-of-Experts
Mixture-of-Experts is in 'active' phase (0 mentions/3d, 5/14d, 9 total)
90% confidence
Sentiment History
| Week | Avg Sentiment | Mentions |
|---|---|---|
| 2026-W11 | 0.15 | 2 |
| 2026-W12 | 0.15 | 4 |
| 2026-W13 | 0.60 | 1 |
| 2026-W14 | 0.10 | 1 |
| 2026-W15 | 0.20 | 1 |
| 2026-W16 | 0.20 | 1 |