ID: 2206.12832

Prediction Errors for Penalized Regressions based on Generalized Approximate Message Passing

June 26, 2022

View on ArXiv
Ayaka Sakata
Statistics
Condensed Matter
Computer Science
Machine Learning
Disordered Systems and Neura...
Machine Learning

We discuss the prediction accuracy of assumed statistical models in terms of prediction errors for the generalized linear model and penalized maximum likelihood methods. We derive the forms of estimators for the prediction errors, such as $C_p$ criterion, information criteria, and leave-one-out cross validation (LOOCV) error, using the generalized approximate message passing (GAMP) algorithm and replica method. These estimators coincide with each other when the number of model parameters is sufficiently small; however, there is a discrepancy between them in particular in the parameter region where the number of model parameters is larger than the data dimension. In this paper, we review the prediction errors and corresponding estimators, and discuss their differences. In the framework of GAMP, we show that the information criteria can be expressed by using the variance of the estimates. Further, we demonstrate how to approach LOOCV error from the information criteria by utilizing the expression provided by GAMP.

Similar papers 1

Estimator of Prediction Error Based on Approximate Message Passing for Penalized Linear Regression

February 20, 2018

93% Match
Ayaka Sakata
Machine Learning
Machine Learning

We propose an estimator of prediction error using an approximate message passing (AMP) algorithm that can be applied to a broad range of sparse penalties. Following Stein's lemma, the estimator of the generalized degrees of freedom, which is a key quantity for the construction of the estimator of the prediction error, is calculated at the AMP fixed point. The resulting form of the AMP-based estimator does not depend on the penalty function, and its value can be further improv...

Find SimilarView on arXiv

A computationally fast alternative to cross-validation in penalized Gaussian graphical models

September 24, 2013

89% Match
Ivan Vujacic, Antonino Abbruzzo, Ernst Wit
Methodology

We study the problem of selection of regularization parameter in penalized Gaussian graphical models. When the goal is to obtain the model with good predicting power, cross validation is the gold standard. We present a new estimator of Kullback-Leibler loss in Gaussian Graphical model which provides a computationally fast alternative to cross-validation. The estimator is obtained by approximating leave-one-out-cross validation. Our approach is demonstrated on simulated data s...

Find SimilarView on arXiv

Shrinkage Tuning Parameter Selection in Precision Matrices Estimation

September 7, 2009

88% Match
Heng Lian
Methodology

Recent literature provides many computational and modeling approaches for covariance matrices estimation in a penalized Gaussian graphical models but relatively little study has been carried out on the choice of the tuning parameter. This paper tries to fill this gap by focusing on the problem of shrinkage parameter selection when estimating sparse precision matrices using the penalized likelihood approach. Previous approaches typically used K-fold cross-validation in this re...

Find SimilarView on arXiv

Consistent information criteria for regularized regression and loss-based learning problems

April 26, 2024

88% Match
Qingyuan Zhang, Hien Duy Nguyen
Methodology

Many problems in statistics and machine learning can be formulated as model selection problems, where the goal is to choose an optimal parsimonious model among a set of candidate models. It is typical to conduct model selection by penalizing the objective function via information criteria (IC), as with the pioneering work by Akaike and Schwarz. Via recent work, we propose a generalized IC framework to consistently estimate general loss-based learning problems. In this work, w...

Find SimilarView on arXiv

Finite-sample and asymptotic analysis of generalization ability with an application to penalized regression

September 12, 2016

88% Match
Ning Xu, Jian Hong, Timothy C. G. Fisher
stat.ML
cs.LG
math.ST
q-fin.EC
stat.CO
stat.TH

In this paper, we study the performance of extremum estimators from the perspective of generalization ability (GA): the ability of a model to predict outcomes in new samples from the same population. By adapting the classical concentration inequalities, we derive upper bounds on the empirical out-of-sample prediction errors as a function of the in-sample errors, in-sample data size, heaviness in the tails of the error distribution, and model complexity. We show that the error...

Find SimilarView on arXiv

Error bounds in estimating the out-of-sample prediction error using leave-one-out cross validation in high-dimensions

March 3, 2020

88% Match
Kamiar Rahnama Rad, Wenda Zhou, Arian Maleki
Machine Learning
Machine Learning

We study the problem of out-of-sample risk estimation in the high dimensional regime where both the sample size $n$ and number of features $p$ are large, and $n/p$ can be less than one. Extensive empirical evidence confirms the accuracy of leave-one-out cross validation (LO) for out-of-sample risk estimation. Yet, a unifying theoretical evaluation of the accuracy of LO in high-dimensional problems has remained an open problem. This paper aims to fill this gap for penalized re...

Find SimilarView on arXiv

Generalization error minimization: a new approach to model evaluation and selection with an application to penalized regression

October 18, 2016

88% Match
Ning Xu, Jian Hong, Timothy C. G. Fisher
Machine Learning
Statistics Theory
Economics
Statistics Theory

We study model evaluation and model selection from the perspective of generalization ability (GA): the ability of a model to predict outcomes in new samples from the same population. We believe that GA is one way formally to address concerns about the external validity of a model. The GA of a model estimated on a sample can be measured by its empirical out-of-sample errors, called the generalization errors (GE). We derive upper bounds for the GE, which depend on sample sizes,...

Find SimilarView on arXiv

Cross validation in LASSO and its acceleration

December 29, 2015

88% Match
Tomoyuki Obuchi, Yoshiyuki Kabashima
Information Theory
Disordered Systems and Neura...
Statistical Mechanics
Information Theory

We investigate leave-one-out cross validation (CV) as a determinator of the weight of the penalty term in the least absolute shrinkage and selection operator (LASSO). First, on the basis of the message passing algorithm and a perturbative discussion assuming that the number of observations is sufficiently large, we provide simple formulas for approximately assessing two types of CV errors, which enable us to significantly reduce the necessary cost of computation. These formul...

Find SimilarView on arXiv

Asymptotic Errors for Teacher-Student Convex Generalized Linear Models (or : How to Prove Kabashima's Replica Formula)

June 11, 2020

87% Match
Cedric Gerbelot, Alia Abbara, Florent Krzakala
stat.ML
cond-mat.dis-nn
cs.IT
cs.LG
math.IT
math.PR

There has been a recent surge of interest in the study of asymptotic reconstruction performance in various cases of generalized linear estimation problems in the teacher-student setting, especially for the case of i.i.d standard normal matrices. Here, we go beyond these matrices, and prove an analytical formula for the reconstruction performance of convex generalized linear models with rotationally-invariant data matrices with arbitrary bounded spectrum, rigorously confirming...

Find SimilarView on arXiv

On Bootstrapping Lasso in Generalized Linear Models and the Cross Validation

March 28, 2024

87% Match
Mayukh Choudhury, Debraj Das
Methodology

Generalized linear models or GLM constitutes an important set of models which generalizes the ordinary linear regression by connecting the response variable with the covariates through arbitrary link functions. On the other hand, Lasso is a popular and easy to implement penalization method in regression when all the covariates are not relevant. However, Lasso generally has non-tractable asymptotic distribution and hence development of an alternative method of distributional a...

Find SimilarView on arXiv