A Selective Overview of Variable Selection in High Dimensional Feature Space (Invited Review Article)
Abstract
High dimensional statistical problems arise from diverse fields of scientific research and technological development. Variable selection plays a pivotal role in contemporary statistical learning and scientific discoveries. The traditional idea of best subset selection methods, which can be regarded as a specific form of penalized likelihood, is computationally too expensive for many modern statistical applications. Other forms of penalized likelihood methods have been successfully developed over the last decade to cope with high dimensionality. They have been widely applied for simultaneously selecting important variables and estimating their effects in high dimensional statistical inference. In this article, we present a brief account of the recent developments of theory, methods, and implementations for high dimensional variable selection. What limits of the dimensionality such methods can handle, what the role of penalty functions is, and what the statistical properties are rapidly drive the advances of the field. The properties of non-concave penalized likelihood and its roles in high dimensional statistical modeling are emphasized. We also review some recent advances in ultra-high dimensional variable selection, with emphasis on independence screening and two-scale methods.
- Publication:
-
arXiv e-prints
- Pub Date:
- October 2009
- DOI:
- 10.48550/arXiv.0910.1122
- arXiv:
- arXiv:0910.1122
- Bibcode:
- 2009arXiv0910.1122F
- Keywords:
-
- Mathematics - Statistics;
- Primary 62J99;
- secondary 62F12;
- 68Q32
- E-Print:
- To appear in Statistica Sinica, 44 pages, 5 figures