Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2019 Feb 20;38(4):558-582.
doi: 10.1002/sim.7803. Epub 2018 Jun 4.

Standard errors and confidence intervals for variable importance in random forest regression, classification, and survival

Affiliations

Standard errors and confidence intervals for variable importance in random forest regression, classification, and survival

Hemant Ishwaran et al. Stat Med. .

Abstract

Random forests are a popular nonparametric tree ensemble procedure with broad applications to data analysis. While its widespread popularity stems from its prediction performance, an equally important feature is that it provides a fully nonparametric measure of variable importance (VIMP). A current limitation of VIMP, however, is that no systematic method exists for estimating its variance. As a solution, we propose a subsampling approach that can be used to estimate the variance of VIMP and for constructing confidence intervals. The method is general enough that it can be applied to many useful settings, including regression, classification, and survival problems. Using extensive simulations, we demonstrate the effectiveness of the subsampling estimator and in particular find that the delete-d jackknife variance estimator, a close cousin, is especially effective under low subsampling rates due to its bias correction properties. These 2 estimators are highly competitive when compared with the .164 bootstrap estimator, a modified bootstrap procedure designed to deal with ties in out-of-sample data. Most importantly, subsampling is computationally fast, thus making it especially attractive for big data settings.

Keywords: VIMP; bootstrap; delete-d jackknife; permutation importance; prediction error; subsampling.

PubMed Disclaimer

Figures

FIGURE 10
FIGURE 10
Assessing asymptotic normality of VIMP from RF-C, RSF, and RSF competing risk simulations. Figure displays bias of standardized VIMP quantiles compared to standard normal quantiles. Values are displayed for 5,10,25,50,75,90,95 percentile values.
FIGURE 11
FIGURE 11
Results from competing risk simulations showing performance of subsampling, delete-d jackknife, and .164 bootstrap. Left and right hand side figures are based on subsampling rates b = n1/2 and b = n3/4. Top and middle figures display bias and standardized MSE for estimating VIMP standard error. Bottom figure displays coverage for VIMP 90% asymptotic normal confidence intervals. Results have been stratified into 6 groups based on 10, 25, 50, 75, and 90th percentiles of true finite VIMP.
FIGURE 1
FIGURE 1
Bias and standardized mean-squared-error (SMSE) performance for estimating VIMP standard error from RF-R (RF-regression) simulations. In total there are 240 variables (12 simulations, p = 20 variables in each simulation). These 240 variables have been stratified into 6 groups based on 10, 25, 50, 75, and 90th percentiles of true finite VIMP. Extreme right boxplots labeled “ALL” display performance for all 240 variables simultaneously.
FIGURE 2
FIGURE 2
Results from RF-R simulations but with increased subsampling rate b = n3/4. Notice the improvement in bias and SMSE for the subsampling estimator.
FIGURE 3
FIGURE 3
Assessing asymptotic normality of VIMP from RF-R simulations. Left-hand figure displays normal quantile plots for standardized VIMP for each of the 12 simulations. Right-hand figure displays bias of VIMP quantiles compared to standard normal quantiles for all 240 variables from all 12 simulations.
FIGURE 4
FIGURE 4
Coverage of VIMP 90% asymptotic normal confidence intervals from RF-R simulations. Left and right hand side figures based on subsampling rates b = n1/2 and b = n3/4 respectively. Confidence regions for the 240 variables from the 12 simulation experiments have been stratified into 6 groups based on 10, 25, 50, 75, and 90th percentiles of true finite VIMP values.
FIGURE 5
FIGURE 5
OOB misclassification error rate versus OOB normalized Brier score for Wisconsin breast cancer data (obtained from the mlbench R-package). Note the fluctuations in misclassification error even after 20,000 trees in contrast to the stable behavior of the Brier score.
FIGURE 6
FIGURE 6
Results from RF-C (RF-classification) simulations showing performance of subsampling, delete-d jackknife, and .164 bootstrap. Left and right hand side figures are based on subsampling rates b = n1/2 and b = n3/4. Top and middle figures display bias and standardized MSE for estimating VIMP standard error. Bottom figure displays coverage for VIMP 90% asymptotic normal confidence intervals. Results have been stratified into 6 groups based on 10, 25, 50, 75, and 90th percentiles of true finite VIMP.
FIGURE 7
FIGURE 7
Delete-d jackknife 95% asymptotic normal confidence intervals from RSF analysis of systolic heart failure data.
FIGURE 8
FIGURE 8
Results from RSF simulations showing performance of subsampling, delete-d jackknife, and .164 bootstrap. Left and right hand side figures are based on subsampling rates b = n1/2 and b = n3/4. Top and middle figures display bias and standardized MSE for estimating VIMP standard error. Bottom figure displays coverage for VIMP 90% asymptotic normal confidence intervals. Results have been stratified into 6 groups based on 10, 25, 50, 75, and 90th percentiles of true finite VIMP.
FIGURE 9
FIGURE 9
Results from variable selection experiment using RF-R simulations of Section 4. Displayed are the true positive rate (TPR) and true negative rate (TNR) for variables selected using 100(1 − α)% delete-d jackknife confidence regions where α = .25, .1, .01, .001. Top and bottom figures are based on subsampling rates b = n1/2 and b = n3/4.

Similar articles

Cited by

References

    1. Breiman L. Random forests. Machine Learning. 2001;45:5–32.
    1. Ishwaran H, Kogalur UB, Blackstone EH, Lauer MS. Random survival forests. The Annals of Applied Statistics. 2008;2(3):841–860.
    1. Ishwaran H, Gerds TA, Kogalur UB, Moore RD, Gange SJ, Lau BM. Random survival forests for competing risks. Biostatistics. 2014;15(4):757–773. - PMC - PubMed
    1. Hothorn T, Hornik K, Zeileis A. Unbiased recursive partitioning: A conditional inference framework. Journal of Computational and Graphical Statistics. 2006;15(3):651–674.
    1. Zhu R, Kosorok MR. Recursively imputed survival trees. Journal of the American Statistical Association. 2012;107(497):331–340. - PMC - PubMed

Publication types