Cerebras REAP Collection Sparse MoE models compressed using REAP (Router-weighted Expert Activation Pruning) method • 19 items • Updated 8 days ago • 59
REAP the Experts: Why Pruning Prevails for One-Shot MoE compression Paper • 2510.13999 • Published Oct 15 • 6
The Dragon Hatchling: The Missing Link between the Transformer and Models of the Brain Paper • 2509.26507 • Published Sep 30 • 535
Bark Collection Bark is a transformer-based text-to-audio model created by Suno. Currently, two checkpoints are supported: a small and a large version. • 3 items • Updated Sep 14, 2023 • 20
MultiSlav Collection Multilingual Machine Translation Open-Source Slavic models • 19 items • Updated Mar 7 • 9
Kimi-VL-A3B Collection Moonshot's efficient MoE VLMs, exceptional on agent, long-context, and thinking • 7 items • Updated Oct 30 • 77