vignettes/Class_Probability_Calcs.Rmd
Class_Probability_Calcs.Rmd
This document describes exactly how the model computes class probabilities using the data in the terminal nodes. Here is an example model using the iris data:
> library(C50)
> mod <- C5.0(Species ~ ., data = iris)
> summary(mod)
Call:
C5.0.formula(formula = Species ~ ., data = iris)
C5.0 [Release 2.07 GPL Edition] Wed Feb 8 19:59:28 2023
-------------------------------
Class specified by attribute `outcome'
Read 150 cases (5 attributes) from undefined.data
Decision tree:
Petal.Length <= 1.9: setosa (50)
Petal.Length > 1.9:
:...Petal.Width > 1.7: virginica (46/1)
Petal.Width <= 1.7:
:...Petal.Length <= 4.9: versicolor (48/1)
Petal.Length > 4.9: virginica (6/2)
Evaluation on training data (150 cases):
Decision Tree
----------------
Size Errors
4 4( 2.7%) <<
(a) (b) (c) <-classified as
---- ---- ----
50 (a): class setosa
47 3 (b): class versicolor
1 49 (c): class virginica
Attribute usage:
100.00% Petal.Length
66.67% Petal.Width
Time: 0.0 secs
Suppose that we are predicting the sample in row 130 with a petal
length of 5.8 and a petal width of 1.6. From this tree, the terminal
node shows virginica (6/2)
which means a predicted class of
the virginica species with a probability of 4/6 = 0.66667. However, we
get a different predicted probability:
> predict(mod, iris[130,], type = "prob")
setosa versicolor virginica
130 0.04761905 0.3333333 0.6190476
When we wanted to describe the technical aspects of the C5.0 and cubist models, the main source of information on these models was the raw C source code from the RuleQuest website. For many years, both of these models were proprietary commercial products and we only recently open-sourced. Our intuition is that Quinlan quietly evolved these models from the versions described in the most recent publications to what they are today. For example, it would not be unreasonable to assume that C5.0 uses AdaBoost. From the sources, a similar reweighting scheme is used but it does not appear to be the same.
For classifying new samples, the C sources have
(DataRec Case, Tree DecisionTree){
ClassNo PredictTreeClassify, C;
ClassNo cdouble Prior;
/* Save total leaf count in ClassSum[0] */
(c, 0, MaxClass) {
ForEach[c] = 0;
ClassSum}
(Case, DecisionTree, Nil, 1.0);
PredictFindLeaf
= SelectClassGen(DecisionTree->Leaf, (Boolean)(MCost != Nil), ClassSum);
C
/* Set all confidence values in ClassSum */
(c, 1, MaxClass){
ForEach= DecisionTree->ClassDist[c] / DecisionTree->Cases;
Prior [c] = (ClassSum[0] * ClassSum[c] + Prior) / (ClassSum[0] + 1);
ClassSum}
= ClassSum[C];
Confidence
return C;
}
Here:
ClassSum
is the probabilities of each class
in the terminal node although ClassSum[0]
is the number of
samples in the terminal node (which, if there are missing values, can be
fractional).For sample 130, the virginica values are:
(ClassSum[0] * ClassSum[c] + Prior) / (ClassSum[0] + 1)
= ( 6 * (4/6) + (1/3)) / ( 6 + 1)
= 0.6190476
Why is it doing this? This will tend to avoid class predictions that are absolute zero or one.
Basically, it can be viewed to be similar to how Bayesian
methods operate where the simple probability estimates are “shrunken”
towards the prior probabilities. Note that, as the number of samples in
the terminal nodes (ClassSum[0]
) becomes large, this
operation has less effect on the final results. Suppose
ClassSum[0] = 10000
, then the predicted virginica
probability would be 0.6663337, which is closer to the simple
estimate.
This is very much related to the Laplace
Correction. Traditionally, we would add a value of one to the
denominator of the simple estimate and add the number of classes to the
bottom, resulting in (4+1)/(6+3) = 0.5555556
. C5.0 is
substituting the prior probabilities and their sum (always one) into
this equation instead.
To be fair, there are well known Bayesian estimates of the sample proportions under different prior distributions for the two class case. For example, if there were two classes, the estimate of the class probability under a uniform prior would be the same as the basic Laplace correction (using the integers and not the fractions). A more flexible Bayesian approach is the Beta-Binomial model, which uses a Beta prior instead of the uniform. The downside here is that two extra parameters need to be estimated (and it only is defined for two classes)