Bayes beats Cross Validation: Fast and Accurate Ridge Regression via Expectation Maximization

Shu Yu Tew, Mario Boley, Daniel F. Schmidt

Research output: Contribution to journalConference articlepeer-review

Abstract

We present a novel method for tuning the regularization hyper-parameter, λ, of a ridge regression that is faster to compute than leave-one-out cross-validation (LOOCV) while yielding estimates of the regression parameters of equal, or particularly in the setting of sparse covariates, superior quality to those obtained by minimising the LOOCV risk. The LOOCV risk can suffer from multiple and bad local minima for finite n and thus requires the specification of a set of candidate λ, which can fail to provide good solutions. In contrast, we show that the proposed method is guaranteed to find a unique optimal solution for large enough n, under relatively mild conditions, without requiring the specification of any difficult to determine hyper-parameters. This is based on a Bayesian formulation of ridge regression that we prove to have a unimodal posterior for large enough n, allowing for both the optimal λ and the regression coefficients to be jointly learned within an iterative expectation maximization (EM) procedure. Importantly, we show that by utilizing an appropriate preprocessing step, a single iteration of the main EM loop can be implemented in O(min(n, p)) operations, for input data with n rows and p columns. In contrast, evaluating a single value of λ using fast LOOCV costs O(n min(n, p)) operations when using the same preprocessing. This advantage amounts to an asymptotic improvement of a factor of l for l candidate values for λ (in the regime q, p ∈ O(√n) where q is the number of regression targets).

Original languageEnglish
JournalAdvances in Neural Information Processing Systems
Volume36
StatePublished - 2023
Externally publishedYes
Event37th Conference on Neural Information Processing Systems, NeurIPS 2023 - New Orleans, United States
Duration: 10 Dec 202316 Dec 2023

Bibliographical note

Publisher Copyright:
© 2023 Neural information processing systems foundation. All rights reserved.

ASJC Scopus subject areas

  • Computer Networks and Communications
  • Information Systems
  • Signal Processing

Fingerprint

Dive into the research topics of 'Bayes beats Cross Validation: Fast and Accurate Ridge Regression via Expectation Maximization'. Together they form a unique fingerprint.

Cite this