ID: 2405.13690

The effect of regularization in high dimensional Cox regression

May 22, 2024

View on ArXiv

Similar papers 4

On the Properties of Simulation-based Estimators in High Dimensions

October 10, 2018

86% Match
Stéphane Guerrier, Mucyo Karemera, ... , Victoria-Feser Maria-Pia
Statistics Theory
Computation
Methodology
Statistics Theory

Considering the increasing size of available data, the need for statistical methods that control the finite sample bias is growing. This is mainly due to the frequent settings where the number of variables is large and allowed to increase with the sample size bringing standard inferential procedures to incur significant loss in terms of performance. Moreover, the complexity of statistical models is also increasing thereby entailing important computational challenges in constr...

Find SimilarView on arXiv

Statistical Inference for Cox Proportional Hazards Models with a Diverging Number of Covariates

June 6, 2021

86% Match
Lu Xia, Bin Nan, Yi Li
Methodology

For statistical inference on regression models with a diverging number of covariates, the existing literature typically makes sparsity assumptions on the inverse of the Fisher information matrix. Such assumptions, however, are often violated under Cox proportion hazards models, leading to biased estimates with under-coverage confidence intervals. We propose a modified debiased lasso approach, which solves a series of quadratic programming problems to approximate the inverse i...

Find SimilarView on arXiv

Adaptive kernel estimation of the baseline function in the Cox model, with high-dimensional covariates

July 6, 2015

86% Match
Agathe LSTA Guilloux, Sarah LaMME Lemler, Marie-Luce LaMME, Unité MIAJ Taupin
Applications
Methodology

The aim of this article is to propose a novel kernel estimator of the baseline function in a general high-dimensional Cox model, for which we derive non-asymptotic rates of convergence. To construct our estimator, we first estimate the regression parameter in the Cox model via a Lasso procedure. We then plug this estimator into the classical kernel estimator of the baseline function, obtained by smoothing the so-called Breslow estimator of the cumulative baseline function. We...

Find SimilarView on arXiv

Non-asymptotic Oracle Inequalities for the High-Dimensional Cox Regression via Lasso

April 9, 2012

86% Match
Shengchun Kong, Bin Nan
Statistics Theory
Machine Learning
Statistics Theory

We consider the finite sample properties of the regularized high-dimensional Cox regression via lasso. Existing literature focuses on linear models or generalized linear models with Lipschitz loss functions, where the empirical risk functions are the summations of independent and identically distributed (iid) losses. The summands in the negative log partial likelihood function for censored survival data, however, are neither iid nor Lipschitz. We first approximate the negativ...

Find SimilarView on arXiv

Penalized variable selection procedure for Cox models with semiparametric relative risk

October 19, 2010

86% Match
Pang Du, Shuangge Ma, Hua Liang
Statistics Theory
Statistics Theory

We study the Cox models with semiparametric relative risk, which can be partially linear with one nonparametric component, or multiple additive or nonadditive nonparametric components. A penalized partial likelihood procedure is proposed to simultaneously estimate the parameters and select variables for both the parametric and the nonparametric parts. Two penalties are applied sequentially. The first penalty, governing the smoothness of the multivariate nonlinear covariate ef...

Find SimilarView on arXiv

High-dimensional variable selection for Cox's proportional hazards model

February 17, 2010

86% Match
Jianqing Fan, Yang Feng, Yichao Wu
Machine Learning
Methodology

Variable selection in high dimensional space has challenged many contemporary statistical problems from many frontiers of scientific disciplines. Recent technology advance has made it possible to collect a huge amount of covariate information such as microarray, proteomic and SNP data via bioimaging technology while observing survival information on patients in clinical studies. Thus, the same challenge applies to the survival analysis in order to understand the association b...

Find SimilarView on arXiv

Optimal Estimation for the Functional Cox Model

January 27, 2016

86% Match
Simeng Qu, Jane-Ling Wang, Xiao Wang
Methodology
Statistics Theory
Statistics Theory

Functional covariates are common in many medical, biodemographic, and neuroimaging studies. The aim of this paper is to study functional Cox models with right-censored data in the presence of both functional and scalar covariates. We study the asymptotic properties of the maximum partial likelihood estimator and establish the asymptotic normality and efficiency of the estimator of the finite-dimensional estimator. Under the framework of reproducing kernel Hilbert space, the e...

Find SimilarView on arXiv

Robust and sparse estimation methods for high dimensional linear and logistic regression

March 15, 2017

86% Match
Fatma Sevinc Kurnaz, Irene Hoffmann, Peter Filzmoser
Methodology

Fully robust versions of the elastic net estimator are introduced for linear and logistic regression. The algorithms to compute the estimators are based on the idea of repeatedly applying the non-robust classical estimators to data subsets only. It is shown how outlier-free subsets can be identified efficiently, and how appropriate tuning parameters for the elastic net penalties can be selected. A final reweighting step improves the efficiency of the estimators. Simulation st...

Find SimilarView on arXiv

The CDF penalty:sparse and quasi unbiased estimation in regression models

December 16, 2022

86% Match
Daniele Cuntrera, Luigi Augugliaro, Vito M. R. Muggeo
Methodology

In high-dimensional regression modelling, the number of candidate covariates to be included in the predictor is quite large, and variable selection is crucial. In this work, we propose a new penalty able to guarantee both sparse variable selection, i.e. exactly zero regression coefficient estimates, and quasi-unbiasedness for the coefficients of 'selected' variables in high dimensional regression models. Simulation results suggest that our proposal performs no worse than its ...

Find SimilarView on arXiv

Regularized estimation for highly multivariate log Gaussian Cox processes

May 4, 2019

86% Match
Achmad Choiruddin, Francisco Cuevas-Pacheco, ... , Waagepetersen Rasmus
Methodology
Computation

Statistical inference for highly multivariate point pattern data is challenging due to complex models with large numbers of parameters. In this paper, we develop numerically stable and efficient parameter estimation and model selection algorithms for a class of multivariate log Gaussian Cox processes. The methodology is applied to a highly multivariate point pattern data set from tropical rain forest ecology.

Find SimilarView on arXiv