MoE-CAP: Benchmarking Cost, Accuracy and Performance of Sparse Mixture-of-Experts Systems Paper • 2412.07067 • Published Dec 10, 2024
Towards Stable and Effective Reinforcement Learning for Mixture-of-Experts Paper • 2510.23027 • Published Oct 27, 2025 • 1
Multiplex Thinking: Reasoning via Token-wise Branch-and-Merge Paper • 2601.08808 • Published 13 days ago • 38
MT6: Multilingual Pretrained Text-to-Text Transformer with Translation Pairs Paper • 2104.08692 • Published Apr 18, 2021
Multiplex Thinking: Reasoning via Token-wise Branch-and-Merge Paper • 2601.08808 • Published 13 days ago • 38
Black-Box On-Policy Distillation of Large Language Models Paper • 2511.10643 • Published Nov 13, 2025 • 51
Benefits and Pitfalls of Reinforcement Learning for Language Model Planning: A Theoretical Perspective Paper • 2509.22613 • Published Sep 26, 2025 • 10
DocReward: A Document Reward Model for Structuring and Stylizing Paper • 2510.11391 • Published Oct 13, 2025 • 27
Information-Preserving Reformulation of Reasoning Traces for Antidistillation Paper • 2510.11545 • Published Oct 13, 2025 • 2
Latent Sketchpad: Sketching Visual Thoughts to Elicit Multimodal Reasoning in MLLMs Paper • 2510.24514 • Published Oct 28, 2025 • 22
The Era of Agentic Organization: Learning to Organize with Language Models Paper • 2510.26658 • Published Oct 30, 2025 • 28
SeerAttention-R: Sparse Attention Adaptation for Long Reasoning Paper • 2506.08889 • Published Jun 10, 2025 • 23
Model as a Game: On Numerical and Spatial Consistency for Generative Games Paper • 2503.21172 • Published Mar 27, 2025
Think Only When You Need with Large Hybrid-Reasoning Models Paper • 2505.14631 • Published May 20, 2025 • 20