SlimQwen: Exploring the Pruning and Distillation in Large MoE Model Pre-training Paper • 2605.08738 • Published 4 days ago • 7
SlimQwen: Exploring the Pruning and Distillation in Large MoE Model Pre-training Paper • 2605.08738 • Published 4 days ago • 7
Canzona: A Unified, Asynchronous, and Load-Balanced Framework for Distributed Matrix-based Optimizers Paper • 2602.06079 • Published Feb 4 • 21
Learnable Multipliers: Freeing the Scale of Language Model Matrix Layers Paper • 2601.04890 • Published Jan 8 • 44
PRIMA.CPP: Speeding Up 70B-Scale LLM Inference on Low-Resource Everyday Home Clusters Paper • 2504.08791 • Published Apr 7, 2025 • 140