Evaluation Flashcards

Notes on Evaluation Strategies that may help with the exam

1
Q

What three methods are commonly used to measure Generalisation Error?

A

Training/Validation/Test

Cross-Validation

Independent Test Set

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

How does K-fold Cross Validation work? Go through step-by-step.

A
  1. Split training data into K amount of folds
  2. For each fold, train on all other folds and make predictions on the held-out test fold
  3. Iterate all folds as the test fold
  4. Combine all the results of K folds, so that every data point has been used as a test data point
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the difference between Nested K-Fold Cross Validation and K-Fold Cross Validation?

A

The Nested K-Fold updates its hyperparameters as it goes along, whereas traditional K-Fold does not.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

How does Nested K-Fold Cross Validation work?

A

Splits the data into K sections.

For each section, tests each part, using the rest of the data in the section as the training dataset. The hyperparameters stay consistent throughout

Updates the hyperparameters each time the inner loop completes, using the results found from the subsequent analysis.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the equation for Classification Accuracy?

A

Classification Accuracy = (TP + TN) / (TP + FP + FN + TN)
Where:
TP = True Positive
TN = True Negative
FP = False Positive
FN = False Negative

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is the equation for Balanced Classification Accuracy?

A

(Sensitivity + Specificity) / 2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is the equation for Precision in regards to Evaluation?

A

Precision = TP / (TP + FP)
Where:
TP = True Positive
FP = False Positive

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the equation for Recall in regards to Evaluation?

A

Recall = TP / (TP + FN)
Where:
TP - True Positive
FN = False Negative

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What does Precision mean in regards to Evaluation?

A

Percentage of correct positive predictions among all predicted positive predictions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What does Recall mean in regards to Evaluation?

A

Percentage of correct positive predictions among all real positive cases e.g. don’t want to miss any diseased cases

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

How is Sensitivity calculated?

A

Sensitivity = Recall

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

How is Specificity calculated?

A

Specificity = TN / (TN + FP)
Where:
TN = True Negative
FP = False Positive

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What does Specificity mean in regards to Evaluation?

A

Percentage of correct negative predictions among all real negative cases

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is F Measure in regards to Evaluation?

A

F Measure combines Recall and Precision into a single measure, and they range from 0 to 1, where 1 indicates perfect precision and recall.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is the equation for F Measure?

A

f_b = (1 + b^2) * (Precision * Recall) / ((b^2 * Precision) + Recall)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How does the Receiver Operator Characteristic evaluate the performance of a model?

A

The larger the area under a curve, the better the classifier is at that problem.

17
Q

What is a Confusion Matrix?

A

A visualisation tool used to present the results attained by a learner, as it’s easy to see if the system is commonly mislabelling one class as another.

18
Q

How is Confidence Interval calculated?

A

The mean of the estimate +- the variation in that estimate