Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

Q2 (Coefficient of determination)

MAE

Classification models

The most common case of classification models is binary classification, where the instances belong to either positive (active) or negative (inactive) class. Some statistical measures are applicable to binary classification models only.

For binary classification models the accepted notion is to discriminate:

TP = true positives  - number of instances of active class, that were correctly predicted by the model as actives

FP = false positives - number of instances of inactive class, that were incorrectly predicted by the model as actives

TN = true negatives - number of instances of inactive class, that were correctly predicted by the model as inactives

FN = false negatives - number of instances of active class, that were incorrectly predicted by the model as inactives

Accuracy

"Accuracy" is merely the percentage of correctly classified samples. For binary classification accuracy can be calculated as follows:

ACC = (TP + TN) / (TP + FP + TN + FN)

Class hit rate

Hit rate is a measure that is applicable to a single class in a classification model and denotes a ratio of instances of a specific class that were correctly identified as belonging to this class.

For binary classification tasks class hit rate for positive class is called sensitivity, and for negative class - specificity.

 

Precision

Precision in the context of classification models is a measure applicable to a single class of a model and denotes a ratio between the instances correctly identified as belonging to a particular class and a total number of instances identified as belonging to this class.

For binary classification models precision for positive class is called positive predictive value, and for negative class - negative predictive value.

 

Sensitivity

Sensitivity is a measure applicable to binary classifications and denotes a ratio of positive instances that were correctly identified as such.

SENS = TP / (TP + FN)

Specificity

Specificity is a measure applicable to binary classifications and denotes a ratio of negative instances that were correctly identified as such.

SPEC = TN / (TN + FP)

Positive predictive value

Positive predictive value is a binary classification measure and shows a ratio of true positive to all instances that were classified as positive.

PPV = TP / (TP + FP)

Negative predictive value

Negative predictive value is a binary classification measure and shows a ratio of true positive to all instances that were classified as positive.

NPV = TN / (TN + FN)

Balanced accuracy

Balanced accuracy is the averaged accuracy for each class, e.g. (positive_class_accuracy + negative_class_accuracy) / 2.
This parameter is important for imbalanced datasets, which have significantly different number of samples in different classes.
If a classifier has a similar performance for both negative and positive classes, accuracy and balanced accuracy are also similar.

BACC = 0.5 * (TP / (TP + FN) + TN / (TN + FP)) = 0.5 * (SENS + SPEC)

 

Confusion matrix

Confusion matrix shows the number of samples from a particular class classified as another particular class.

For binary classification models with "yes" and "no" classes, the confusion matrix shows the number of:

  • true positives
  • true negatives
  • false positives
  • false negatives


Confidence intervals

The confidence intervals (95%) for all statistical parameters are evaluated using bootstrap procedure with n = 1000 samples.
Following model calculation we get predicted values zi for training (or test) samples with experimental values yi, i.e. {zi, yi} are pairs of values with i = 1,..,N.
We randomly sample pairs {zi, yi} to form n = 1000 bootstrap sets of the same size as the analyzed set. For each bootstrap set we calculate
statistical parameters and use their respective distributions to determine respective confidence intervals.
Since the intervals are in general non-symmetric with respect to the values calculated for the analyzed sets, the average values are reported.

...