Test-Time Training with Self-Supervision for Generalization under Distribution Shifts

Yu Sun · Xiaolong Wang · Zhuang Liu · John Miller · Alexei Efros · Moritz Hardt


Keywords: [ Algorithms ] [ Robust Statistics and Machine Learning ] [ Trustworthy Machine Learning ]

[ Abstract ]
[ Slides
Wed 15 Jul 10 a.m. PDT — 10:45 a.m. PDT
Wed 15 Jul 11 p.m. PDT — 11:45 p.m. PDT


In this paper, we propose Test-Time Training, a general approach for improving the performance of predictive models when training and test data come from different distributions. We turn a single unlabeled test sample into a self-supervised learning problem, on which we update the model parameters before making a prediction. This also extends naturally to data in an online stream. Our simple approach leads to improvements on diverse image classification benchmarks aimed at evaluating robustness to distribution shifts.

Chat is not available.