The models for the paper: Hybrid Linear Attention Done Right: Efficient Distillation and Effective Architectures for Extremely Long Contexts
Yingfa Chen
chen-yingfa
AI & ML interests
Long-context modeling, continual learning, architectures
Recent Activity
updated a collection about 7 hours ago
HypeNet liked a model about 7 hours ago
chen-yingfa/HypeNet-2B updated a model about 7 hours ago
chen-yingfa/HypeNet-2BOrganizations
None yet