Out-of-sample risk estimation in no time flat
Abstract Number:
2344
Submission Type:
Contributed Abstract
Contributed Abstract Type:
Paper
Participants:
Daniel LeJeune (1), Parth Nobel (1), Emmanuel Candes (1)
Institutions:
(1) Stanford University, N/A
Co-Author(s):
First Author:
Presenting Author:
Abstract Text:
Hyperparameter tuning is an essential part of statistical machine learning pipelines, and becomes more computationally challenging as datasets become large. Furthermore, the standard method of k-fold cross-validation is known to be inconsistent for high-dimensional problems. We propose instead an efficient implementation of approximate leave-one-out (ALO) risk estimation, providing consistent risk estimation in high-dimensions at a fraction of the cost of k-fold cross-validation. We leverage randomized numerical linear algebra and reduce the computational task to a handful of quasisemidefinite linear systems, equivalent to equality-constrained quadratic programs, for any convex non-smooth loss and linear-separable regularizer.
Keywords:
Risk estimation|Cross-validation|High dimensions|Convex optimization|Randomized methods|
Sponsors:
Section on Statistical Learning and Data Science
Tracks:
Machine Learning
Can this be considered for alternate subtype?
Yes
Are you interested in volunteering to serve as a session chair?
No
I have read and understand that JSM participants must abide by the Participant Guidelines.
Yes
I understand that JSM participants must register and pay the appropriate registration fee by June 1, 2024. The registration fee is non-refundable.
I understand
You have unsaved changes.