Skip to yearly menu bar Skip to main content


Poster

MoEQuant: Enhancing Quantization for Mixture-of-Experts Large Language Models via Expert-Balanced Sampling and Affinity Guidance

Zhixuan Chen ⋅ Xing Hu ⋅ Dawei Yang ⋅ Zukang Xu ⋅ XUCHEN ⋅ Zhihang Yuan ⋅ Sifan Zhou ⋅ JiangyongYu
2025 Poster

Abstract

Lay Summary

Video

Chat is not available.