DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter Paper • 1910.01108 • Published Oct 2, 2019 • 23
MemPrivacy: Privacy-Preserving Personalized Memory Management for Edge-Cloud Agents Paper • 2605.09530 • Published 6 days ago • 139
MoESD: Unveil Speculative Decoding's Potential for Accelerating Sparse MoE Paper • 2505.19645 • Published Feb 16 • 1
view article Article What’s MXFP4? The 4-Bit Secret Powering OpenAI’s GPT‑OSS Models on Modest Hardware RakshitAralimatti • Aug 8, 2025 • 35
SD-MoE: Spectral Decomposition for Effective Expert Specialization Paper • 2602.12556 • Published Feb 13 • 1
Mean Mode Screaming: Mean--Variance Split Residuals for 1000-Layer Diffusion Transformers Paper • 2605.06169 • Published 9 days ago • 182