54: Minimum-Norm Interpolation Under Covariate Shift
Bin Yu
Co-Author
University of California at Berkeley
Austin Zane
First Author
University of California, Berkeley
Austin Zane
Presenting Author
University of California, Berkeley
Tuesday, Aug 5: 10:30 AM - 12:20 PM
2480
Contributed Posters
Music City Center
Transfer learning is a critical part of real-world machine learning deployments and has been extensively studied in experimental works with overparameterized neural networks. However, even in the simplest setting of linear regression a notable gap still exists in the theoretical understanding of transfer learning. In-distribution research on high-dimensional linear regression has led to the identification of a phenomenon known as benign overfitting, in which linear interpolators overfit to noisy training labels and yet still generalize well. This behavior occurs under specific conditions on the source covariance matrix and input data dimension. Therefore, it is natural to wonder how such high-dimensional linear models behave under transfer learning. We prove the first non-asymptotic excess risk bounds for benignly-overfit linear interpolators in the transfer learning setting. From our analysis, we propose a taxonomy of beneficial and malignant covariate shifts based on the degree of overparameterization. We follow our analysis with empirical studies showing these beneficial and malignant covariate shifts for linear interpolators and simple neural networks in certain settings.
high-dimensional statistics
statistical learning
distribution shift
generalization
deep learning theory
transfer learning
Main Sponsor
Section on Statistical Learning and Data Science
You have unsaved changes.