Stability of a Generalized Debiased Lasso with Applications to Resampling-Based Variable Selection

Jingbo Liu First Author
UIUC
 
Jingbo Liu Presenting Author
UIUC
 
Wednesday, Aug 6: 2:50 PM - 3:05 PM
2058 
Contributed Papers 
Music City Center 
Suppose that we first apply the Lasso to a design matrix, and then update one of its columns. In general, the signs of the Lasso coefficients may change, and there is no closed-form expression for updating the Lasso solution exactly. In this work, we propose an approximate formula for updating a debiased Lasso coefficient. We provide general nonasymptotic error bounds in terms of the norms and correlations of a given design matrix's columns, and then prove asymptotic convergence results for the case of a random design matrix with i.i.d.\ sub-Gaussian row vectors and i.i.d.\ Gaussian noise. Notably, the approximate formula is asymptotically correct for most coordinates in the proportional growth regime, under the mild assumption that each row of the design matrix is sub-Gaussian with a covariance matrix having a bounded condition number. Our proof only requires certain concentration and anti-concentration properties to control various error terms and the number of sign changes. In contrast, rigorously establishing distributional limit properties (e.g.\ Gaussian limits for the debiased Lasso) under similarly general assumptions has been considered open problem in the universality

Keywords

debiased Lasso

inference in high-dimensional regression models

knockoffs

false discovery rate

university

small ball probability 

Main Sponsor

Section on Statistical Learning and Data Science