An Information Theoretic Analysis of Maximum Likelihood Mixture Estimation for Exponential Families
Arindam Banerjee - University of Texas at Austin
Inderjit Dhillon - University of Texas at Austin
Joydeep Ghosh - University of Texas at Austin
Srujana Merugu - University of Texas at Austin
An important task in unsupervised learning is maximum likelihood mixture estimation (MLME) for exponential families. In this paper, we prove a mathematical equivalence between this MLME problem and the rate distortion problem for Bregman divergences. We also present new theoretical results in rate distortion theory for Bregman divergences. Further, an analysis of the problems as a trade-off between compression and preservation of information is presented that yields the information bottleneck method as an interesting special case.