Evaluation Notes
Evaluation Notes
The stage of testing the models before deployment in the real world.
Importance of Evaluation
Prediction
Reality
True Positive
True Negative
Models that use the training dataset during testing, will always result in correct
output.
This is known as overfitting.
To evaluate the AI model, it is not necessary to use the data that is used to build the
model because AI Model remembers the whole training data set, therefore it always
predicts the correct label for any point in the training dataset.
ACCURACY
Percentage of correct predictions out of all the predictions.
PRECISION
Percentage of true positives out of all the true predictions
RECALL
Measure of balance between precision and recall
Includes a fraction of positive cases that are correctly identified.
F1 SCORE
A measure of the test accuracy
False Negative cases (VIRAL OUTBREAK) are more crucial and dangerous when compared to
FALSE POSITIVE cases.
3. Calculate Accuracy, Precision, Recall and F1 Score
for the following Confusion Matrix on Heart Attack
Risk. Also suggest which metric would not be a good
evaluation parameter here and why?
Calculation:
Accuracy:
Calculation:
ACCURACY
A = (75+15) / (75+15+5+5) = (90 / 100) =0.9
Precision:
Accuracy= 0.9%
Precision=0.9375%
Recall=0.9375%
F1 Score=0.
Here precision, recall, accuracy, f1 score all are same.
Calculate Accuracy, Precision, Recall and F1 Score for the following Confusion
Matrix on SPAM FILTERING: Also suggest which metric would not be a good
evaluation parameter here and why?
Calculation
ACCURACY
Accuracy= 0.35
Precision= 0.15
Recall= 0.5
F1 Score= 0.23