Open Access
2009 Forest Garrote
Nicolai Meinshausen
Electron. J. Statist. 3: 1288-1304 (2009). DOI: 10.1214/09-EJS434

Abstract

Variable selection for high-dimensional linear models has received a lot of attention lately, mostly in the context of 1-regularization. Part of the attraction is the variable selection effect: parsimonious models are obtained, which are very suitable for interpretation. In terms of predictive power, however, these regularized linear models are often slightly inferior to machine learning procedures like tree ensembles. Tree ensembles, on the other hand, lack usually a formal way of variable selection and are difficult to visualize. A Garrote-style convex penalty for trees ensembles, in particular Random Forests, is proposed. The penalty selects functional groups of nodes in the trees. These could be as simple as monotone functions of individual predictor variables. This yields a parsimonious function fit, which lends itself easily to visualization and interpretation. The predictive power is maintained at least at the same level as the original tree ensemble. A key feature of the method is that, once a tree ensemble is fitted, no further tuning parameter needs to be selected. The empirical performance is demonstrated on a wide array of datasets.

Citation

Download Citation

Nicolai Meinshausen. "Forest Garrote." Electron. J. Statist. 3 1288 - 1304, 2009. https://doi.org/10.1214/09-EJS434

Information

Published: 2009
First available in Project Euclid: 4 December 2009

zbMATH: 1326.62093
MathSciNet: MR2566188
Digital Object Identifier: 10.1214/09-EJS434

Subjects:
Primary: 62G08

Keywords: nonnegative garrote , random forests , Sparsity , tree ensembles

Rights: Copyright © 2009 The Institute of Mathematical Statistics and the Bernoulli Society

Back to Top