Timezone: »
Self-training is a standard approach to semi-supervised learning where the learner's own predictions on unlabeled data are used as supervision during training. In this paper, we reinterpret this label assignment process as an optimal transportation problem between examples and classes, wherein the cost of assigning an example to a class is mediated by the current predictions of the classifier. This formulation facilitates a practical annealing strategy for label assignment and allows for the inclusion of prior knowledge on class proportions via flexible upper bound constraints. The solutions to these assignment problems can be efficiently approximated using Sinkhorn iteration, thus enabling their use in the inner loop of standard stochastic optimization algorithms. We demonstrate the effectiveness of our algorithm on the CIFAR-10, CIFAR-100, and SVHN datasets in comparison with FixMatch, a state-of-the-art self-training algorithm.
Author Information
Kai Sheng Tai (Stanford University)
Peter Bailis (Stanford University)
Gregory Valiant (Stanford University)
Related Events (a corresponding poster, oral, or spotlight)
-
2021 Poster: Sinkhorn Label Allocation: Semi-Supervised Classification via Annealed Self-Training »
Thu. Jul 22nd 04:00 -- 06:00 PM Room
More from the Same Authors
-
2023 Poster: One-sided Matrix Completion from Two Observations Per Row »
Steven Cao · Percy Liang · Greg Valiant -
2020 Poster: Sample Amplification: Increasing Dataset Size even when Learning is Impossible »
Brian Axelrod · Shivam Garg · Vatsal Sharan · Gregory Valiant -
2019 Poster: LIT: Learned Intermediate Representation Training for Model Compression »
Animesh Koratana · Daniel Kang · Peter Bailis · Matei Zaharia -
2019 Oral: LIT: Learned Intermediate Representation Training for Model Compression »
Animesh Koratana · Daniel Kang · Peter Bailis · Matei Zaharia -
2019 Poster: Compressed Factorization: Fast and Accurate Low-Rank Factorization of Compressively-Sensed Data »
Vatsal Sharan · Kai Sheng Tai · Peter Bailis · Gregory Valiant -
2019 Poster: Equivariant Transformer Networks »
Kai Sheng Tai · Peter Bailis · Gregory Valiant -
2019 Oral: Equivariant Transformer Networks »
Kai Sheng Tai · Peter Bailis · Gregory Valiant -
2019 Oral: Compressed Factorization: Fast and Accurate Low-Rank Factorization of Compressively-Sensed Data »
Vatsal Sharan · Kai Sheng Tai · Peter Bailis · Gregory Valiant -
2019 Poster: Rehashing Kernel Evaluation in High Dimensions »
Paris Siminelakis · Kexin Rong · Peter Bailis · Moses Charikar · Philip Levis -
2019 Poster: Maximum Likelihood Estimation for Learning Populations of Parameters »
Ramya Korlakai Vinayak · Weihao Kong · Gregory Valiant · Sham Kakade -
2019 Oral: Rehashing Kernel Evaluation in High Dimensions »
Paris Siminelakis · Kexin Rong · Peter Bailis · Moses Charikar · Philip Levis -
2019 Oral: Maximum Likelihood Estimation for Learning Populations of Parameters »
Ramya Korlakai Vinayak · Weihao Kong · Gregory Valiant · Sham Kakade -
2017 Poster: Estimating the unseen from multiple populations »
Aditi Raghunathan · Greg Valiant · James Zou -
2017 Poster: Orthogonalized ALS: A Theoretically Principled Tensor Decomposition Algorithm for Practical Use »
Vatsal Sharan · Gregory Valiant -
2017 Talk: Estimating the unseen from multiple populations »
Aditi Raghunathan · Greg Valiant · James Zou -
2017 Talk: Orthogonalized ALS: A Theoretically Principled Tensor Decomposition Algorithm for Practical Use »
Vatsal Sharan · Gregory Valiant