54: Minimum-Norm Interpolation Under Covariate Shift

Neil Mallinar Co-Author
University of California, San Diego
 
Spencer Frei Co-Author
 
Bin Yu Co-Author
University of California at Berkeley
 
Austin Zane First Author
University of California, Berkeley
 
Austin Zane Presenting Author
University of California, Berkeley
 
Tuesday, Aug 5: 10:30 AM - 12:20 PM
2480 
Contributed Posters 
Music City Center 
Transfer learning is a critical part of real-world machine learning deployments and has been extensively studied in experimental works with overparameterized neural networks. However, even in the simplest setting of linear regression a notable gap still exists in the theoretical understanding of transfer learning. In-distribution research on high-dimensional linear regression has led to the identification of a phenomenon known as benign overfitting, in which linear interpolators overfit to noisy training labels and yet still generalize well. This behavior occurs under specific conditions on the source covariance matrix and input data dimension. Therefore, it is natural to wonder how such high-dimensional linear models behave under transfer learning. We prove the first non-asymptotic excess risk bounds for benignly-overfit linear interpolators in the transfer learning setting. From our analysis, we propose a taxonomy of beneficial and malignant covariate shifts based on the degree of overparameterization. We follow our analysis with empirical studies showing these beneficial and malignant covariate shifts for linear interpolators and simple neural networks in certain settings.

Keywords

high-dimensional statistics

statistical learning

distribution shift

generalization

deep learning theory

transfer learning 

Main Sponsor

Section on Statistical Learning and Data Science