HypeNet
Collection
The models for the paper: Hybrid Linear Attention Done Right: Efficient Distillation and Effective Architectures for Extremely Long Contexts • 2 items • Updated
Links:
This is the final HypeNet-5B checkpoint from the paper Hybrid Linear Attention Done Right: Efficient Distillation and Effective Architectures for Extremely Long Contexts, distilled from Qwen3-4B using the HALO pipeline proposed in our paper. For more information, please refer to our GitHub repo.