Test-Time Training for Out-of-Distribution Generalization

09/29/2019
by   Yu Sun, et al.
8

We introduce a general approach, called test-time training, for improving the performance of predictive models when test and training data come from different distributions. Test-time training turns a single unlabeled test instance into a self-supervised learning problem, on which we update the model parameters before making a prediction on the test sample. We show that this simple idea leads to surprising improvements on diverse image classification benchmarks aimed at evaluating robustness to distribution shifts. Theoretical investigations on a convex model reveal helpful intuitions for when we can expect our approach to help.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset