ID: math/0611589

High Dimensional Statistical Inference and Random Matrices

November 19, 2006

View on ArXiv

Similar papers 4

A shrinkage estimation for large dimensional precision matrices using random matrix theory

November 11, 2012

88% Match
Cheng Wang, Guangming Pan, Longbing Cao
Methodology

In this paper, a new ridge-type shrinkage estimator for the precision matrix has been proposed. The asymptotic optimal shrinkage coefficients and the theoretical loss were derived. Data-driven estimators for the shrinkage coefficients were also conducted based on the asymptotic results deriving from random matrix theories. The new estimator which has a simple explicit formula is distribution-free and applicable to situation where the dimension of observation is greater than t...

Find SimilarView on arXiv

The High-Dimensional Asymptotics of Principal Component Regression

May 19, 2024

88% Match
Alden Green, Elad Romanov
Statistics Theory
Statistics Theory

We study principal components regression (PCR) in an asymptotic high-dimensional regression setting, where the number of data points is proportional to the dimension. We derive exact limiting formulas for the estimation and prediction risks, which depend in a complicated manner on the eigenvalues of the population covariance, the alignment between the population PCs and the true signal, and the number of selected PCs. A key challenge in the high-dimensional setting stems from...

Find SimilarView on arXiv

CLT for linear spectral statistics of large dimensional sample covariance matrices with dependent data

August 12, 2017

88% Match
Shurong Zheng, Zhidong Bai, ... , Zhu Hongtu
Probability

This paper investigates the central limit theorem for linear spectral statistics of high dimensional sample covariance matrices of the form $\mathbf{B}_n=n^{-1}\sum_{j=1}^{n}\mathbf{Q}\mathbf{x}_j\mathbf{x}_j^{*}\mathbf{Q}^{*}$ where $\mathbf{Q}$ is a nonrandom matrix of dimension $p\times k$, and $\{\mathbf{x}_j\}$ is a sequence of independent $k$-dimensional random vector with independent entries, under the assumption that $p/n\to y>0$. A key novelty here is that the dimens...

Find SimilarView on arXiv

On convergence of the sample correlation matrices in high-dimensional data

June 20, 2017

88% Match
Sévérien Nkurunziza, Yueleng Wang
Statistics Theory
Statistics Theory

In this paper, we consider an estimation problem concerning the matrix of correlation coefficients in context of high dimensional data settings. In particular, we revisit some results in Li and Rolsalsky [Li, D. and Rolsalsky, A. (2006). Some strong limit theorems for the largest entries of sample correlation matrices, The Annals of Applied Probability, 16, 1, 423-447]. Four of the main theorems of Li and Rolsalsky (2006) are established in their full generalities and we simp...

Find SimilarView on arXiv

Inference in High-dimensional Multivariate Response Regression with Hidden Variables

January 20, 2022

88% Match
Xin Bing, Wei Cheng, ... , Ning Yang
Methodology

This paper studies the inference of the regression coefficient matrix under multivariate response linear regressions in the presence of hidden variables. A novel procedure for constructing confidence intervals of entries of the coefficient matrix is proposed. Our method first utilizes the multivariate nature of the responses by estimating and adjusting the hidden effect to construct an initial estimator of the coefficient matrix. By further deploying a low-dimensional project...

Find SimilarView on arXiv

Developments in Random Matrix Theory

March 11, 2003

88% Match
N. C. Snaith, P. J. Forrester, J. J. M. Verbaarschot
Disordered Systems and Neura...
Mathematical Physics

In this preface to the Journal of Physics A, Special Edition on Random Matrix Theory, we give a review of the main historical developments of random matrix theory. A short summary of the papers that appear in this special edition is also given.

Find SimilarView on arXiv

On Identity Tests for High Dimensional Data Using RMT

March 15, 2012

88% Match
Cheng Wang, Jing Yang, ... , Cao Longbing
Methodology
Applications

In this work, we redefined two important statistics, the CLRT test (Bai et.al., Ann. Stat. 37 (2009) 3822-3840) and the LW test (Ledoit and Wolf, Ann. Stat. 30 (2002) 1081-1102) on identity tests for high dimensional data using random matrix theories. Compared with existing CLRT and LW tests, the new tests can accommodate data which has unknown means and non-Gaussian distributions. Simulations demonstrate that the new tests have good properties in terms of size and power. Wha...

Find SimilarView on arXiv

Robust high dimensional factor models with applications to statistical machine learning

August 12, 2018

88% Match
Jianqing Fan, Kaizheng Wang, ... , Zhu Ziwei
Methodology
Statistics Theory
Machine Learning
Statistics Theory

Factor models are a class of powerful statistical models that have been widely used to deal with dependent measurements that arise frequently from various applications from genomics and neuroscience to economics and finance. As data are collected at an ever-growing scale, statistical machine learning faces some new challenges: high dimensionality, strong dependence among observed variables, heavy-tailed variables and heterogeneity. High-dimensional robust factor analysis serv...

Find SimilarView on arXiv

Optimal Shrinkage Estimator for High-Dimensional Mean Vector

October 28, 2016

88% Match
Taras Bodnar, Ostap Okhrin, Nestor Parolya
Statistics Theory
Statistical Finance
Statistics Theory

In this paper we derive the optimal linear shrinkage estimator for the high-dimensional mean vector using random matrix theory. The results are obtained under the assumption that both the dimension $p$ and the sample size $n$ tend to infinity in such a way that $p/n \to c\in(0,\infty)$. Under weak conditions imposed on the underlying data generating mechanism, we find the asymptotic equivalents to the optimal shrinkage intensities and estimate them consistently. The proposed ...

Find SimilarView on arXiv

Improved Gaussian Mean Matrix Estimators In High-Dimensional Data

November 24, 2023

87% Match
Arash A. Foroushani, Severien Nkurunziza
Statistics Theory
Statistics Theory

In this paper, we introduce a class of improved estimators for the mean parameter matrix of a multivariate normal distribution with an unknown variance-covariance matrix. In particular, the main results of [D.Ch\'etelat and M. T. Wells(2012). Improved Multivariate Normal Mean Estimation with Unknown Covariance when $p$ is Greater than $n$. The Annals of Statistics, Vol. 40, No.6, 3137--3160] are established in their full generalities and we provide the corrected version of th...

Find SimilarView on arXiv