Data Mining - Lecture Performance Measures Flashcards
What do errors based on the training set tell us?
About the model’s fit.
What do errors based on the validation set tell us?
The ability to predict new data.
These errors are called prediction errors.
Which three types of outcome do we deal with in this course?
and can we also evaluate
- A predicted numerical value
- A predicted class membership
- The probability of class membership
How do we measure prediction accuracy for numerical prediction?
We use the error for each record and compute one of the following measures:
- Mean absolute error (MAE)
- Mean error
- Mean percentage error (MPE)
- Mean absolute percentage error (MAPE)
- Root mean squared error
- Lift Chart
How do you compute the mean absolute error?
For each record you get the error. You disregard the sign before the value and you sum up all errors.
You multiply the sum with (1/n)
How do you compute the mean error?
You sum up all errors (including the signs before the value). You multiply the sum of errors with (1/n).
How do you compute the mean percentage error?
You divide each record’s error by the records actual value (Yi). You do this for all records and sum them up.
You multiply the sum with (1/N) and you do that * 100
How do you compute the mean absolute percentage error?
Same as the mean percentage error, except in the first part you do not take the sign into considerations before the error value/Yi.
How do you compute the Root mean squared error?
You square each individual error and sum them up. You multiply it by (1/n)/ The answer you put in a square root.
What is a lift chart?
It is a chart that to compare a model with a baseline model with no predictions, to see which subset of records gives the highest cumulative predicted values.
On the x-axis you put percentage of samples. On the y-axis you put the percentage where the model predicts the positive class well.
Then you can see how your model performs. It can be that for a small percentage of the sample you will get a more then average prediction/response rate.
How can we evaluate classifier performance?
- We make a confusion matrix.
- Based on that confusion matrix we can compute accuracy, prediction, recall, F1
- We can also make a ROC Curve based on the confusion matrix.
What is a misclassification?
If your model puts records in the wrong class:
False Negative and False Positive
What is a confusion matrix?
A matrix with the predicted classes on the x-axis and the actual classes on the y classes. It lists the TP, FP, TN, FN.
It reads like:
True Negative | False Positive
False Negative | True Positive
What is a type I error?
False Positive
What is a type II error?
False Negative