Unveiling And Addressing Dimensional Collapse In Vector Quantization Models Via Codebook Regularization
Abstract
While recent advancements in Vector Quantization (VQ) models have successfully achieved complete codebook utilization, a critical bottleneck remains largely unexplored: the effective dimensionality of the codebook embedding space. We observe that discrete codebook representations tend to degenerate into low-dimensional subspaces, characterized by significantly lower effective rank than continuous representations during quantization. Through comprehensive spectral analysis, we identify that this dimensional collapse stems from the suppression of low-variance components inherent to the vector quantization process, thereby severely limiting the expressive capacity of VQ models. To mitigate this fundamental issue, we propose a simple yet effective codebook regularization strategy designed to restore low-variance components, effectively bridging the spectral gap between discrete codebook spaces and continuous representations. Extensive experiments demonstrate that this regularization objective is compatible with diverse VQ training paradigms, yielding significant improvements in reconstruction fidelity and downstream performance in autoregressive image generative models.