Case Sensitivity in Regression and Beyond

Yoonkyung Lee Co-Author
The Ohio State University
 
Haozhen Yu First Author
 
Haozhen Yu Presenting Author
 
Wednesday, Aug 7: 10:50 AM - 11:05 AM
2411 
Contributed Papers 
Oregon Convention Center 
The sensitivity of a model to data perturbations is key to model diagnostics and understanding model stability and complexity. Case deletion has been primarily considered for sensitivity analysis in linear regression, where the notions of leverage and residual are central to the influence of a case on the model. Instead of case deletion, we examine the change in the model due to an infinitesimal data perturbation, known as local influence, for various machine learning methods. This local influence analysis reveals a notable commonality in the form of case influence across different methods, allowing us to generalize the concepts of leverage and residual far beyond linear regression. At the same time, the results show differences in the mode of case influence, depending on the method. Through the lens of local influence, we provide a generalized and convergent perspective on case sensitivity in modeling that includes regularized regression, large margin classification, generalized linear models, and quantile regression.

Keywords

Case Influence

Leverage

Model Diagnostics

Residual

Sensitivity Analysis 

Main Sponsor

Section on Statistical Learning and Data Science