Measuring Performance Flashcards

1
Q

What is Classification Accuracy?

A

number of correctly classified samples

/

total samples

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is Classification Error Rate?

A

number of wrongly classified samples

/

total samples

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is Recall?

A

TP

/

TP + FN

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is Precision?

A

TP

/

TP + FP

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What are some measures we can use to measure performance in regression?

A
  1. Root Mean Squared Error
  2. Mean Absolute Error
  3. Mean Absolute Percentage Error
  4. Coefficient of Determination
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is Root Mean Squared Error?

A

Take the average of the square differences, and then square root

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is Mean Absolute Error?

A

Take the average of the absolute differences

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the Mean Absolute Percentage Error?

A

Take the average of ( the absolute differences divided by the true value)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

When is Mean Absolute Percentage Error useful?

A

When different classes in our output might give drastically different range of values.

E.g one output might be temperature, and the other might be kilo calories. Therefore we should normalise the error for each before summing them up

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the Bias Issue?

A

The idea that the accuracy of the training samples can be a poor estimator of the accuracy on unseen samples

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the Variance Issue?

A

The idea that the accuracy on a new set of test samples can still vary from the true accuracy, depending on the makeup of the test samples

Smaller set of test samples can result in a higher variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is the Holdout Method?

A

Splitting the dataset into training and testing dataset

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is Random Sub-Sampling?

A
  1. Split the data in to K parts
  2. Randomly select a fixed number for training and testing
  3. Train the classifier from scratch using the training data, and then test to compute an error
  4. Repeat the experiment from steps 1 and 2, K times
  5. The final error is the average error of all the experiments
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is K-Fold Cross Validation?

A
  1. Divide the dataset into K partitions
  2. For each of the K experiments, use K-1 partitions for training, and estimate the error using the remaining partition
  3. The final error estimate is the average error of all K experiments
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is an advantage of using K-Fold Cross Validation?

A

All the examples in the dataset are eventually used for both training and testing

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is Leave One Out Cross Validation?

A

This is an extreme case of K-Fold Cross Validation.

Where if we have a dataset of N samples, we do N-Fold Cross Validation.

17
Q

What is Bootstrap?

A

For K experiments, select the training samples using sampling with replacement
And use the data that was never selected as testing data

The final error is the average error of the K experiments

18
Q

Comparing the errors of two models using Z-Tests

A

???

19
Q

What is F-1 Score?

A

2 * P * R

/

P + R

20
Q

What is the Variance Error?

A

This tells you how much the predictions changes under different realisations of the same model

21
Q

What is the Bias Error?

A

This is a measure of how the close the average prediction is to the true value, under different realisations of the same model.

22
Q

If I have high variance error, and low bias error, will I suffer from over-fitting or under-fitting?

A

Over-fitting

23
Q

If I have low-variance error, and high bias error, will I suffer from over-fitting or under-fitting?

A

Under-fitting