Jump to content

Random naive Bayes

From Wikipedia, the free encyclopedia

This is an old revision of this page, as edited by 192.35.44.24 (talk) at 17:17, 5 August 2011. The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

Random naive Bayes extends the Naive Bayes classifier by adopting the random forest principles: random input selection, bagging (i.e. bootstrap aggregating), and random feature selection ([1]).

Naive Bayes classifier

The naive Bayes classifier is a probabilistic classifier simplifying Bayes' theorem by naively assuming class conditional independence. Although this assumption leads to biased posterior probabilities, the ordered probabilities of Naive Bayes result in a classification performance comparable to that of classification trees and neural networks. [2] Notwithstanding Naive Bayes' popularity due to its simplicity combined with high accuracy and speed, its conditional independence assumption rarely holds. There are mainly two approaches to alleviate this naivety:

  1. Selecting attribute subsets in which attributes are conditionally independent (cf. Selective Bayesian Classifier [3]).
  2. Extending the structure of Naive Bayes to represent attribute dependencies (cf. Averaged One-Dependence Estimators (AODE) [4]).

Random naive Bayes' alleviation of the class conditional independence assumption

Random Naive Bayes adopts the first approach by randomly selecting a subset of attributes in which attributes are assumed to be conditionally independent. Naive Bayes' performance might benefit from this random feature selection. Analogous to AODE, Random Naive Bayes builds an ensemble, but unlike AODE, the ensemble combines zero-dependence classifiers.

Random naive Bayes and random forest

Generalizing Random Forest to Naive Bayes, Random Naive Bayes (Random NB), is a bagged classifier combining a forest of B Naive Bayes. Each bth Naive Bayes is estimated on a bootstrap sample Sb with m randomly selected features. To classify an observation put the input vector down the B Naive Bayes in the forest. Each Naive Bayes generates posterior class probabilities. Unlike Random Forest, the predicted class of the ensemble is assessed by adjusted majority voting rather than majority voting, as each bth Naive Bayes delivers continuous posterior probabilities. Similar to Random Forests, the importance of each feature is estimated on the out-of-bag (oob) data.

Notes

  1. ^ Breiman, 2001
  2. ^ Langley, Iba and Thomas, 1992
  3. ^ Langley and Sage, 1994
  4. ^ Webb et al. 2005

References

See also