Skip to yearly menu bar Skip to main content


Poster

Long-Context Modeling with Dynamic Hierarchical Sparse Attention for Memory-Constrained LLM Inference

Siheng Xiong ⋅ Joe Zou ⋅ Faramarz Fekri ⋅ Yae Jee Cho

Abstract

Log in and register to view live content