High dimensional thresholded regression and shrinkage effect

05/11/2016
by   Zemin Zheng, et al.
0

High-dimensional sparse modeling via regularization provides a powerful tool for analyzing large-scale data sets and obtaining meaningful, interpretable models. The use of nonconvex penalty functions shows advantage in selecting important features in high dimensions, but the global optimality of such methods still demands more understanding. In this paper, we consider sparse regression with hard-thresholding penalty, which we show to give rise to thresholded regression. This approach is motivated by its close connection with the L_0-regularization, which can be unrealistic to implement in practice but of appealing sampling properties, and its computational advantage. Under some mild regularity conditions allowing possibly exponentially growing dimensionality, we establish the oracle inequalities of the resulting regularized estimator, as the global minimizer, under various prediction and variable selection losses, as well as the oracle risk inequalities of the hard-thresholded estimator followed by a further L_2-regularization. The risk properties exhibit interesting shrinkage effects under both estimation and prediction losses. We identify the optimal choice of the ridge parameter, which is shown to have simultaneous advantages to both the L_2-loss and prediction loss. These new results and phenomena are evidenced by simulation and real data examples.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/11/2016

Asymptotic equivalence of regularization methods in thresholded parameter space

High-dimensional data analysis has motivated a spectrum of regularizatio...
research
05/11/2016

Asymptotic properties for combined L_1 and concave regularization

Two important goals of high-dimensional modeling are prediction and vari...
research
07/28/2014

Efficient Regularized Regression for Variable Selection with L0 Penalty

Variable (feature, gene, model, which we use interchangeably) selections...
research
11/09/2019

Influence of single observations on the choice of the penalty parameter in ridge regression

Penalized regression methods, such as ridge regression, heavily rely on ...
research
07/28/2017

Empirical Bayes Estimators for High-Dimensional Sparse Vectors

The problem of estimating a high-dimensional sparse vector θ∈R^n from an...
research
01/16/2019

Smooth Adjustment for Correlated Effects

This paper considers a high dimensional linear regression model with cor...
research
06/23/2016

Non-convex regularization in remote sensing

In this paper, we study the effect of different regularizers and their i...

Please sign up or login with your details

Forgot password? Click here to reset