Title :
A Very Simple Safe-Bayesian Random Forest
Author :
Quadrianto, Novi ; Ghahramani, Zoubin
Author_Institution :
Dept. of Inf., Univ. of Sussex, Brighton, UK
Abstract :
Random forests works by averaging several predictions of de-correlated trees. We show a conceptually radical approach to generate a random forest: random sampling of many trees from a prior distribution, and subsequently performing a weighted ensemble of predictive probabilities. Our approach uses priors that allow sampling of decision trees even before looking at the data, and a power likelihood that explores the space spanned by combination of decision trees. While each tree performs Bayesian inference to compute its predictions, our aggregation procedure uses the power likelihood rather than the likelihood and is therefore strictly speaking not Bayesian. Nonetheless, we refer to it as a Bayesian random forest but with a built-in safety. The safeness comes as it has good predictive performance even if the underlying probabilistic model is wrong. We demonstrate empirically that our Safe-Bayesian random forest outperforms MCMC or SMC based Bayesian decision trees in term of speed and accuracy, and achieves competitive performance to entropy or Gini optimised random forest, yet is very simple to construct.
Keywords :
belief networks; decision trees; entropy; inference mechanisms; Bayesian inference; Gini optimised random forest; MCMC based Bayesian decision trees; SMC based Bayesian decision trees; built-in safety; conceptually radical approach; decorrelated trees; entropy; power likelihood; predictive probabilities; random sampling; safe-Bayesian random forest; Bayes methods; Decision trees; Equations; Mathematical model; Monte Carlo methods; Training; Vegetation; Bayesian methods; decision trees; random forest;
Journal_Title :
Pattern Analysis and Machine Intelligence, IEEE Transactions on
DOI :
10.1109/TPAMI.2014.2362751