Skip to yearly menu bar Skip to main content


Poster

Quantum Expectation-Maximization for Gaussian mixture models

Alessandro Luongo · Iordanis Kerenidis · Anupam Prakash

Keywords: [ Speech Processing ] [ Unsupervised Learning ] [ Unsupervised and Semi-supervised Learning ]


Abstract:

We define a quantum version of Expectation-Maximization (QEM), a fundamental tool in unsupervised machine learning, often used to solve Maximum Likelihood (ML) and Maximum A Posteriori (MAP) estimation problems. We use QEM to fit a Gaussian Mixture Model, and show how to generalize it to fit mixture models with base distributions in the exponential family. Given quantum access to a dataset, our algorithm has convergence and precision guarantees similar to the classical algorithm, while the runtime is polylogarithmic in the number of elements in the training set and polynomial in other parameters, such as the dimension of the feature space and the number of components in the mixture. We discuss the performance of the algorithm on a dataset that is expected to be classified successfully by classical EM and provide guarantees for its runtime.

Chat is not available.