ID: 2310.09931

Sub-optimality of the Naive Mean Field approximation for proportional high-dimensional Linear Regression

October 15, 2023

View on ArXiv
Jiaze Qiu
Mathematics
Statistics
Statistics Theory
Statistics Theory

The Na\"ive Mean Field (NMF) approximation is widely employed in modern Machine Learning due to the huge computational gains it bestows on the statistician. Despite its popularity in practice, theoretical guarantees for high-dimensional problems are only available under strong structural assumptions (e.g., sparsity). Moreover, existing theory often does not explain empirical observations noted in the existing literature. In this paper, we take a step towards addressing these problems by deriving sharp asymptotic characterizations for the NMF approximation in high-dimensional linear regression. Our results apply to a wide class of natural priors and allow for model mismatch (i.e., the underlying statistical model can be different from the fitted model). We work under an \textit{iid} Gaussian design and the proportional asymptotic regime, where the number of features and the number of observations grow at a proportional rate. As a consequence of our asymptotic characterization, we establish two concrete corollaries: (a) we establish the inaccuracy of the NMF approximation for the log-normalizing constant in this regime, and (b) we provide theoretical results backing the empirical observation that the NMF approximation can be overconfident in terms of uncertainty quantification. Our results utilize recent advances in the theory of Gaussian comparison inequalities. To the best of our knowledge, this is the first application of these ideas to the analysis of Bayesian variational inference problems. Our theoretical results are corroborated by numerical experiments. Lastly, we believe our results can be generalized to non-Gaussian designs and provide empirical evidence to support it.

Similar papers 1

On Naive Mean-Field Approximation for high-dimensional canonical GLMs

June 21, 2024

91% Match
Sumit Mukherjee, Jiaze Qiu, Subhabrata Sen
Statistics Theory
Information Theory
Information Theory
Probability
Statistics Theory

We study the validity of the Naive Mean Field (NMF) approximation for canonical GLMs with product priors. This setting is challenging due to the non-conjugacy of the likelihood and the prior. Using the theory of non-linear large deviations (Austin 2019, Chatterjee, Dembo 2016, Eldan 2018), we derive sufficient conditions for the tightness of the NMF approximation to the log-normalizing constant of the posterior distribution. As a second contribution, we establish that under m...

Find SimilarView on arXiv

A Mean Field Approach to Empirical Bayes Estimation in High-dimensional Linear Regression

September 28, 2023

90% Match
Sumit Mukherjee, Bodhisattva Sen, Subhabrata Sen
Statistics Theory
Methodology
Machine Learning
Statistics Theory

We study empirical Bayes estimation in high-dimensional linear regression. To facilitate computationally efficient estimation of the underlying prior, we adopt a variational empirical Bayes approach, introduced originally in Carbonetto and Stephens (2012) and Kim et al. (2022). We establish asymptotic consistency of the nonparametric maximum likelihood estimator (NPMLE) and its (computable) naive mean field variational surrogate under mild assumptions on the design and the pr...

Find SimilarView on arXiv

Fundamental Limits of Ridge-Regularized Empirical Risk Minimization in High Dimensions

June 16, 2020

89% Match
Hossein Taheri, Ramtin Pedarsani, Christos Thrampoulidis
Machine Learning
Information Theory
Machine Learning
Signal Processing
Information Theory

Empirical Risk Minimization (ERM) algorithms are widely used in a variety of estimation and prediction tasks in signal-processing and machine learning applications. Despite their popularity, a theory that explains their statistical properties in modern regimes where both the number of measurements and the number of unknown parameters is large is only recently emerging. In this paper, we characterize for the first time the fundamental limits on the statistical accuracy of conv...

Find SimilarView on arXiv

Performance of Bayesian linear regression in a model with mismatch

July 14, 2021

88% Match
Jean Barbier, Wei-Kuo Chen, ... , Sáenz Manuel
math.PR
cond-mat.dis-nn
cs.IT
cs.LG
math.IT
math.MP
math.ST
stat.TH

In this paper we analyze, for a model of linear regression with gaussian covariates, the performance of a Bayesian estimator given by the mean of a log-concave posterior distribution with gaussian prior, in the high-dimensional limit where the number of samples and the covariates' dimension are large and proportional. Although the high-dimensional analysis of Bayesian estimators has been previously studied for Bayesian-optimal linear regression where the correct posterior is ...

Find SimilarView on arXiv
Alexander Mozeika, Mansoor Sheikh, Fabian Aguirre-Lopez, ... , Coolen Anthony CC
Statistics Theory
Disordered Systems and Neura...
Statistics Theory

It is clear that conventional statistical inference protocols need to be revised to deal correctly with the high-dimensional data that are now common. Most recent studies aimed at achieving this revision rely on powerful approximation techniques, that call for rigorous results against which they can be tested. In this context, the simplest case of high-dimensional linear regression has acquired significant new relevance and attention. In this paper we use the statistical phys...

Scalable and Accurate Variational Bayes for High-Dimensional Binary Regression Models

November 15, 2019

88% Match
Augusto Fasano, Daniele Durante, Giacomo Zanella
Methodology
Computation

Modern methods for Bayesian regression beyond the Gaussian response setting are often computationally impractical or inaccurate in high dimensions. In fact, as discussed in recent literature, bypassing such a trade-off is still an open problem even in routine binary regression models, and there is limited theory on the quality of variational approximations in high-dimensional settings. To address this gap, we study the approximation accuracy of routinely-used mean-field varia...

Find SimilarView on arXiv

Fundamental Barriers to High-Dimensional Regression with Convex Penalties

March 25, 2019

87% Match
Michael Celentano, Andrea Montanari
Statistics Theory
Statistics Theory

In high-dimensional regression, we attempt to estimate a parameter vector $\beta_0\in\mathbb{R}^p$ from $n\lesssim p$ observations $\{(y_i,x_i)\}_{i\leq n}$ where $x_i\in\mathbb{R}^p$ is a vector of predictors and $y_i$ is a response variable. A well-established approach uses convex regularizers to promote specific structures (e.g. sparsity) of the estimate $\widehat{\beta}$, while allowing for practical algorithms. Theoretical analysis implies that convex penalization scheme...

Find SimilarView on arXiv

Variational Inference in high-dimensional linear regression

April 25, 2021

87% Match
Sumit Mukherjee, Subhabrata Sen
Statistics Theory
Probability
Machine Learning
Statistics Theory

We study high-dimensional Bayesian linear regression with product priors. Using the nascent theory of non-linear large deviations (Chatterjee and Dembo,2016), we derive sufficient conditions for the leading-order correctness of the naive mean-field approximation to the log-normalizing constant of the posterior distribution. Subsequently, assuming a true linear model for the observed data, we derive a limiting infinite dimensional variational formula for the log normalizing co...

Find SimilarView on arXiv

A variational Bayes approach to debiased inference for low-dimensional parameters in high-dimensional linear regression

June 18, 2024

87% Match
Ismaël Castillo, Alice L'Huillier, ... , Travis Luke
Machine Learning
Machine Learning
Statistics Theory
Statistics Theory

We propose a scalable variational Bayes method for statistical inference for a single or low-dimensional subset of the coordinates of a high-dimensional parameter in sparse linear regression. Our approach relies on assigning a mean-field approximation to the nuisance coordinates and carefully modelling the conditional distribution of the target given the nuisance. This requires only a preprocessing step and preserves the computational advantages of mean-field variational Baye...

Find SimilarView on arXiv

Bayesian Model Selection via Mean-Field Variational Approximation

December 17, 2023

86% Match
Yangfan Zhang, Yun Yang
Methodology
Statistics Theory
Computation
Machine Learning
Statistics Theory

This article considers Bayesian model selection via mean-field (MF) variational approximation. Towards this goal, we study the non-asymptotic properties of MF inference under the Bayesian framework that allows latent variables and model mis-specification. Concretely, we show a Bernstein von-Mises (BvM) theorem for the variational distribution from MF under possible model mis-specification, which implies the distributional convergence of MF variational approximation to a norma...

Find SimilarView on arXiv