Skip to yearly menu bar Skip to main content


Poster

Reasoning-preserved Efficient Distillation of Large Language Models via Activation-aware Initialization

Junlin He ⋅ Yihong Tang ⋅ Tong Nie ⋅ Guilong Li ⋅ Binyu Yang ⋅ Jinxiao Du ⋅ Lijun Sun ⋅ Wei Ma

Abstract

Log in and register to view live content