ID: math/0611589

High Dimensional Statistical Inference and Random Matrices

November 19, 2006

View on ArXiv

Similar papers 2

Introduction to the non-asymptotic analysis of random matrices

November 12, 2010

89% Match
Roman Vershynin
Probability
Numerical Analysis
Functional Analysis

This is a tutorial on some basic non-asymptotic methods and concepts in random matrix theory. The reader will learn several tools for the analysis of the extreme singular values of random matrices with independent rows or columns. Many of these methods sprung off from the development of geometric functional analysis since the 1970's. They have applications in several fields, most notably in theoretical computer science, statistics and signal processing. A few basic applicatio...

Find SimilarView on arXiv

Testing for Independence of Large Dimensional Vectors

August 13, 2017

89% Match
Taras Bodnar, Holger Dette, Nestor Parolya
Statistics Theory
Probability
Statistics Theory

In this paper new tests for the independence of two high-dimensional vectors are investigated. We consider the case where the dimension of the vectors increases with the sample size and propose multivariate analysis of variance-type statistics for the hypothesis of a block diagonal covariance matrix. The asymptotic properties of the new test statistics are investigated under the null hypothesis and the alternative hypothesis using random matrix theory. For this purpose we stu...

Find SimilarView on arXiv

Randomized algorithms for matrices and data

April 29, 2011

89% Match
Michael W. Mahoney
Data Structures and Algorith...

Randomized algorithms for very large matrix problems have received a great deal of attention in recent years. Much of this work was motivated by problems in large-scale data analysis, and this work was performed by individuals from many different research communities. This monograph will provide a detailed overview of recent work on the theory of randomized matrix algorithms as well as the application of those ideas to the solution of practical problems in large-scale data an...

Find SimilarView on arXiv

Adaptive Randomized Dimension Reduction on Massive Data

April 13, 2015

89% Match
Gregory Darnell, Stoyan Georgiev, ... , Engelhardt Barbara E
Machine Learning
Quantitative Methods

The scalability of statistical estimators is of increasing importance in modern applications. One approach to implementing scalable algorithms is to compress data into a low dimensional latent space using dimension reduction methods. In this paper we develop an approach for dimension reduction that exploits the assumption of low rank structure in high dimensional data to gain both computational and statistical advantages. We adapt recent randomized low-rank approximation algo...

Find SimilarView on arXiv

High Dimensional Correlation Matrices: CLT and Its Applications

November 1, 2014

89% Match
Jiti Gao, Xiao Han, ... , Yang Yanrong
Statistics Theory
Statistics Theory

Statistical inferences for sample correlation matrices are important in high dimensional data analysis. Motivated by this, this paper establishes a new central limit theorem (CLT) for a linear spectral statistic (LSS) of high dimensional sample correlation matrices for the case where the dimension p and the sample size $n$ are comparable. This result is of independent interest in large dimensional random matrix theory. Meanwhile, we apply the linear spectral statistic to an i...

Find SimilarView on arXiv

Testing linear hypotheses in high-dimensional regressions

June 5, 2012

89% Match
Z. Bai, D. Jiang, ... , Zheng S.
Methodology
Statistics Theory
Statistics Theory

For a multivariate linear model, Wilk's likelihood ratio test (LRT) constitutes one of the cornerstone tools. However, the computation of its quantiles under the null or the alternative requires complex analytic approximations and more importantly, these distributional approximations are feasible only for moderate dimension of the dependent variable, say $p\le 20$. On the other hand, assuming that the data dimension $p$ as well as the number $q$ of regression variables are fi...

Find SimilarView on arXiv

Statistical applications of Random matrix theory: comparison of two populations II

February 28, 2020

88% Match
Rémy Mariétan, Stephan Morgenthaler
Statistics Theory
Methodology
Statistics Theory

This paper investigates a statistical procedure for testing the equality of two independent estimated covariance matrices when the number of potentially dependent data vectors is large and proportional to the size of the vectors, that is, the number of variables. Inspired by the spike models used in random matrix theory, we concentrate on the largest eigenvalues of the matrices in order to determine significance. To avoid false rejections we must guard against residual spikes...

Find SimilarView on arXiv

Random matrix theory and multivariate statistics

July 6, 2009

88% Match
Jose A. Diaz-Garcia, Ramon Gutiérrez Jáimez
Statistics Theory
Statistics Theory

Some tools and ideas are interchanged between random matrix theory and multivariate statistics. In the context of the random matrix theory, classes of spherical and generalised Wishart random matrix ensemble, containing as particular cases the classical random matrix ensembles, are proposed. Some properties of these classes of ensemble are analysed. In addition, the random matrix ensemble approach is extended and a unified theory proposed for the study of distributions for re...

Find SimilarView on arXiv

Asymptotic inference for high-dimensional data

February 24, 2010

88% Match
Jim Kuelbs, Anand N. Vidyashankar
Statistics Theory
Statistics Theory

In this paper, we study inference for high-dimensional data characterized by small sample sizes relative to the dimension of the data. In particular, we provide an infinite-dimensional framework to study statistical models that involve situations in which (i) the number of parameters increase with the sample size (that is, allowed to be random) and (ii) there is a possibility of missing data. Under a variety of tail conditions on the components of the data, we provide precise...

Find SimilarView on arXiv

High-dimensional analysis of double descent for linear regression with random projections

March 2, 2023

88% Match
Francis SIERRA Bach
Machine Learning
Machine Learning

We consider linear regression problems with a varying number of random projections, where we provably exhibit a double descent curve for a fixed prediction problem, with a high-dimensional analysis based on random matrix theory. We first consider the ridge regression estimator and review earlier results using classical notions from non-parametric statistics, namely degrees of freedom, also known as effective dimensionality. We then compute asymptotic equivalents of the genera...

Find SimilarView on arXiv