|Machine learning and
In machine learning, a probabilistic classifier is a classifier that is able to predict, given a sample input, a probability distribution over a set of classes, rather than only predicting a class for the sample. Probabilistic classifiers provide classification with a degree of certainty, which can be useful when combining them into larger systems, for example in ensemble classifiers.
Formally, a probabilistic classifier is a conditional distribution over a finite set of classes Y, given inputs X. Deciding on the best class label for X can then be done using the optimal decision rule:39–40
Some classification models, such as naive Bayes, logistic regression and multilayer perceptrons (when trained under an appropriate loss function) are naturally probabilistic. Other models such as support vector machines are not, but methods exist to turn them into probabilistic classifiers.
Generative and conditional training
Some models, such as logistic regression, are conditionally trained: they optimize the conditional probability directly on a training set (see empirical risk minimization). Other classifiers, such as naive Bayes, are trained generatively: at training time, the class-conditional distribution and the class prior are found, and the conditional distribution is derived using Bayes' rule.:43
Not all classification models are naturally probabilistic, and some that are, notably naive Bayes classifiers and boosting methods, produce distorted class probability distributions. However, for classification models that produce some kind of "score" on their outputs (such as a distorted probability distribution or the "signed distance to the hyperplane" in a support vector machine), there are several methods that turn these scores into properly calibrated class membership probabilities.
For the binary case, a common approach is to apply Platt scaling, which learns a logistic regression model on the scores. An alternative method using isotonic regression is generally superior to Platt's method when sufficient training data is available.
In the multiclass case, one can use a reduction to binary tasks, followed by univariate calibration with an algorithm as described above and further application of the pairwise coupling algorithm by Hastie and Tibshirani. An alternative one-step method, the Dirichlet calibration, is introduced by Gebel and Weihs.
Evaluating probabilistic classification
Commonly used loss functions for probabilistic classification include log loss and the mean squared error between the predicted and the true probability distributions. The former of these is commonly used to train logistic models.
- Bishop, Christopher M. (2006). Pattern Recognition and Machine Learning. Springer.
- Niculescu-Mizil, Alexandru; Caruana, Rich (2005). "Predicting good probabilities with supervised learning". ICML.
- Platt, John (1999). "Probabilistic outputs for support vector machines and comparisons to regularized likelihood methods". Advances in large margin classifiers 10 (3): 61–74.
- Zadrozny, Bianca; Elkan, Charles (2002). "Transforming classifier scores into accurate multiclass probability estimates". Proceedings of the eighth ACM SIGKDD international conference on Knowledge discovery and data mining - KDD '02. pp. 694–699. doi:10.1145/775047.775151. ISBN 1-58113-567-X. CiteSeerX: 10.1.1.13.7457.
- Hastie, Trevor; Tibshirani, Robert (1998). "Classification by pairwise coupling". The Annals of Statistics 26 (2): 451–471. doi:10.1214/aos/1028144844. Zbl 0932.62071. CiteSeerX: 10.1.1.46.6032.
- Gebel, Martin; Weihs, Claus (2008). "Calibrating Margin-Based Classifier Scores into Polychotomous Probabilities". In Preisach, Christine; Burkhardt, Hans; Schmidt-Thieme, Lars; Decker, Reinhold. Data Analysis, Machine Learning and Applications. Studies in Classification, Data Analysis, and Knowledge Organization. pp. 29–36. doi:10.1007/978-3-540-78246-9_4. ISBN 978-3-540-78239-1.
|This artificial intelligence-related article is a stub. You can help Wikipedia by expanding it.|