Confusion Matrix
The confusion matrix is a table with the number of correct and incorrect predictions broken down by class
False Positive (Type 1 Error): You predicted Positive and it's False
False Negative (Type 2 Error): You predicted Negative and it's False
We describe predicted values as Positive and Negative and actual values as True and False.
Accuracy
Accuracy is the ratio between the number of correct
predictions to total samples
Precision
From all samples we have predicted as positive, how many are actually positive.
Precision should be high as possible.
Recall/Sensitivity (True Positive Rate)
From all the Positive samples, how many we predicted correctly.
Recall should be high as possible.
F1 Score
It is a combination of Precision and Accuracy
Best value reaches to 1 when Precision = Recall = 100%
Worst value = 0
ROC Curve
A receiver operating characteristic (ROC) curve is a
graphical plot that illustrates the diagnostic ability of a binary classifier
system as its discrimination threshold is varied
Provides an aggregate measure of performance across all
positive classification thresholds
No comments:
Post a Comment