Submitted by
Yingfa Chen
AI & ML interests
Large Language Models
Recent Activity
View all activity
Papers
Hybrid Linear Attention Done Right: Efficient Distillation and Effective Architectures for Extremely Long Contexts
InfLLM-V2: Dense-Sparse Switchable Attention for Seamless Short-to-Long Adaptation