| | --- |
| | pipeline_tag: robotics |
| | license: apache-2.0 |
| | tags: |
| | - reinforcement-learning |
| | - robotic-manipulation |
| | - action-chunking |
| | --- |
| | |
| | # This repository provided the torch-version openpi checkpoints (pi0base, pi0.5base) coverted from JAX-version! |
| |
|
| | # [Mixture of Horizons in Action Chunking](https://huggingface.co/papers/2511.19433) |
| |
|
| | Vision-language-action (VLA) models for robotic manipulation are highly sensitive to the chosen **action chunk length**, termed **horizon** in this work. A fixed horizon presents an inherent trade-off: longer horizons offer superior global foresight but compromise fine-grained accuracy, while shorter ones provide precise local control but struggle with long-term tasks. |
| |
|
| | To address this challenge, we propose **Mixture of Horizons (MoH)**, a novel, plug-and-play strategy that fuses multiple horizons within a single policy. MoH processes action chunks in parallel segments with different horizons and integrates their outputs. This approach simultaneously leverages long-term foresight and short-term precision with minimal overhead, and enables **Dynamic Inference** through cross-horizon consensus for enhanced efficiency and robustness in complex robotic tasks. |
| |
|
| | - π [Paper](https://huggingface.co/papers/2511.19433) |
| | - π [Project Page](https://timsty1.github.io/moh/) |
| | - π» [Code](https://github.com/Timsty1/MixtureOfHorizons/tree/main) |
| |
|
| | ## Introduction |
| |
|
| | <div align="center"> |
| | <table border="0" cellspacing="0" cellpadding="0"> |
| | <tr> |
| | <td align="center" width="50%"> |
| | <img src="https://github.com/Timsty1/MixtureOfHorizons/raw/main/figure/study_of_horizons_pi0.png" alt="Trade-off Effect" width="100%"> |
| | </td> |
| | <td align="center" width="50%"> |
| | <img src="https://github.com/Timsty1/MixtureOfHorizons/raw/main/figure/intro_motivation_v2.png" alt="Mixture of Horizons" width="100%"> |
| | </td> |
| | </tr> |
| | <tr> |
| | <td align="center" valign="top"> |
| | Figure 1: Trade-off between long-term foresight and short-term precision induced by single horizon |
| | </td> |
| | <td align="center" valign="top"> |
| | Figure 2: Overview of the proposed mixture-of-horizons strategy |
| | </td> |
| | </tr> |
| | </table> |
| | </div> |
| | |
| | <br> |
| |
|
| | * **Mitigates Trade-off**: Addresses the inherent trade-off between long-term foresight and short-term precision induced by single action chunk horizons. |
| | * **Plug-and-Play**: Easily integrates into existing full-attention action modules with minimal training or inference overhead. |
| | * **Dynamic Inference**: Achieves higher efficiency and robustness by selecting stable actions through cross-horizon consensus. |
| |
|
| | #### More results on LIBERO |
| | <div align="center"> |
| | <img src="https://github.com/Timsty1/MixtureOfHorizons/raw/main/figure/libero_main.jpg" width="90%" /> |
| | </div> |
| |
|
| | ## Usage |
| |
|
| | For detailed instructions on environment setup, training, and evaluation, please refer to the [GitHub repository](https://github.com/Timsty1/MixtureOfHorizons/tree/main). |
| |
|
| | ## β€οΈ Acknowledgment |
| |
|
| | We express our gratitude to [OpenPi](https://github.com/Physical-Intelligence/openpi/tree/main), [LIBERO](https://github.com/Lifelong-Robot-Learning/LIBERO), and [RoboTwin](https://robotwin-platform.github.io/) for their open-source contributions. |
| |
|
| | ## π Citation |
| |
|
| | If you feel that this paper, models, or codes are helpful, please cite our paper, thanks for your support! |
| |
|
| | ```bibtex |
| | @article{jing2025mixture_of_horizons, |
| | title={Mixture of Horizons in Action Chunking}, |
| | author={Jing, Dong and Wang, Gang and Liu, Jiaqi and Tang, Weiliang and Sun, Zelong and Yao, Yunchao and Wei, Zhenyu and Liu, Yunhui and Lu, Zhiwu and Ding, Mingyu}, |
| | journal={arXiv preprint arXiv:2511.19433}, |
| | year={2025} |
| | } |
| | ``` |