Evaluation Metrics Flashcards

1
Q

In Hypothesis Testing, what is the null hypothesis?

A

Null Hypothesis H_0 states the assumption to be tested

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What are the two types of error in hypothesis testing

A

Type I Error “false positive”: Rejecting H_0 when it is in fact true

Type II Error “false negative”: Failing to reject H_0 when it is in fact false

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Give the formulas for misclassification rate and accuracy in Classifier Accuracy

A

MR = # incorrect predictions /
total predictions

ACC = # correct predictions /
total predictions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What do confusion matrices summarise

A

the performance of an algorithm when compared with the real classes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Give the formulas for ACC, TPR, FPR, TNR

A

ACC = TP+TN / TP+FP+TN+FN
TPR = TP / TP + FN
FPR = FP / FP + TN
TNR = TN / FP + TN

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Give the formulas for precision and recall

A

A = Retrieved and relevant
B = Retrieved results
C = Relevant Results
Precision = A / A+B
Recall = A / A+C

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is a Precision-Recall (PR) curve used for?

A
  • To study the output of a binary classifier
  • Measure precision at fixed recall intervals
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Give the formulas for Balance Accuracy Rate (BAR) and Balance Error Rate (BER)

A

BAR: Mean of TPR and TNR
BER: Mean of FPR and FNR

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Give the formula of the F1 Measure

A

2 * Precision * Recall /
Precision + Recall

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is a decision threshold and what is its most common value?

A

the value (theta) used to discriminate when selecting between a positive and negative outcome.
Most common value = 0.5

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What does a ROC plot visualise?

A

How the TPR and FPR change over many different thresholds.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is overfitting

A

Model is fitted too closely to the training data (including its noise). The model cannot generalise to situations not presented during training, so it is not useful when applied to unseen data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Possible causes of overfitting

A
  • Small training set
  • Complex Model
  • noise
  • high dimensionality
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is peeking and what can you use to avoid it

A

When the performance of a model is evaluated using the same data used to train it.

Avoid peeking by using a hold-out set

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What are some drawbacks to using a random split as the hold-out strategy

A
  • Sometimes we don’t have the “luxury” of setting aside data for testing.
  • Since it is a single experiment, the hold-out estimate of error rate can be misleading if we get an “unfortunate” split of the data.
  • Even if we use multiple splits, some examples will never be included for training or testing, while others might be selected many times.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Briefly explain the steps of k-Fold Cross Validation

A
  • Divide the data into k disjoint subsets - “folds”
  • For each of k experiments, use k-1 folds for training and the selected one fold for testing.
  • Repeat for all k folds, average the accuracy/error rates.
17
Q

What is the validation set in the Three-Way-Hold-Out Strategy?

A

The subset of examples used to tune the classifier

18
Q

What is the main advantage of the Three-Way-Hold-Out Strategy?

A

It avoids a bias in evaluation of the model