Case Sensitivity in Regression and Beyond
Wednesday, Aug 7: 10:50 AM - 11:05 AM
2411
Contributed Papers
Oregon Convention Center
The sensitivity of a model to data perturbations is key to model diagnostics and understanding model stability and complexity. Case deletion has been primarily considered for sensitivity analysis in linear regression, where the notions of leverage and residual are central to the influence of a case on the model. Instead of case deletion, we examine the change in the model due to an infinitesimal data perturbation, known as local influence, for various machine learning methods. This local influence analysis reveals a notable commonality in the form of case influence across different methods, allowing us to generalize the concepts of leverage and residual far beyond linear regression. At the same time, the results show differences in the mode of case influence, depending on the method. Through the lens of local influence, we provide a generalized and convergent perspective on case sensitivity in modeling that includes regularized regression, large margin classification, generalized linear models, and quantile regression.
Case Influence
Leverage
Model Diagnostics
Residual
Sensitivity Analysis
Main Sponsor
Section on Statistical Learning and Data Science
You have unsaved changes.