The models for the paper: Hybrid Linear Attention Done Right: Efficient Distillation and Effective Architectures for Extremely Long Contexts
Yingfa Chen
chen-yingfa
AI & ML interests
Long-context modeling, continual learning, architectures
Recent Activity
liked a model 2 days ago
chen-yingfa/HypeNet-5B updated a model 2 days ago
chen-yingfa/HypeNet-5B updated a collection 2 days ago
HypeNetOrganizations
None yet