Skip to yearly menu bar Skip to main content


On Scalable and Efficient Computation of Large Scale Optimal Transport

Yujia Xie · Minshuo Chen · Haoming Jiang · Tuo Zhao · Hongyuan Zha

Pacific Ballroom #10

Keywords: [ Unsupervised Learning ] [ Large Scale Learning and Big Data ] [ Generative Models ] [ Deep Generative Models ] [ Computer Vision ]


Optimal Transport (OT) naturally arises in many machine learning applications, yet the heavy computational burden limits its wide-spread uses. To address the scalability issue, we propose an implicit generative learning-based framework called SPOT (Scalable Push-forward of Optimal Transport). Specifically, we approximate the optimal transport plan by a pushforward of a reference distribution, and cast the optimal transport problem into a minimax problem. We then can solve OT problems efficiently using primal dual stochastic gradient-type algorithms. We also show that we can recover the density of the optimal transport plan using neural ordinary differential equations. Numerical experiments on both synthetic and real datasets illustrate that SPOT is robust and has favorable convergence behavior. SPOT also allows us to efficiently sample from the optimal transport plan, which benefits downstream applications such as domain adaptation.

Live content is unavailable. Log in and register to view live content