Skip to yearly menu bar Skip to main content


Poster

Don't Drop Dropout: Optimizing Layer Sparsity for Efficient LLM Training and Inference

Mostafa Elhoushi ⋅ Nolan Dey ⋅ Alexander Pretko ⋅ Bin Zhang ⋅ Gavia Gray ⋅ Gurpreet Gosal ⋅ Abdulrahman Mahmoud ⋅ Shane Bergsma ⋅ Joel Hestness

Abstract

Log in and register to view live content