Model Evaluation Flashcards

1
Q

Formula for Accuracy

A

True Positive + True Negative / Number of Items

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Formula for Precision (p)

A

p = TP / TP + FP

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Formula for Recall (r)

A

r = TP / TP + FN

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Formula for F-Measure

A

Precision = p
Recall = r

fm = 2rp / r + p

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

compare the accuracy of the classifier with a random classifier.

A

Kappa Statistics

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

was developed in 1950 for signal detection theory.’

Works only for binary classification.

A

Receiver Operating Characteristic (ROC)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

to estimate performance of classifier on previously unseen data.

A

Purpose of Model Evaluation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

reserve k% for training and (100 - k) % for testing.

A

Holdout

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

partition data into k disjoint subset.

A

Cross Validation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

train on k-1 partition, test on the remaining one.

A

K- Fold

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

shows how accuracy on unseen examples changes with varying training sample size.

A

Learning Curve

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

many algorithms allow choices for learning.

A

Hyperparameters

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

3 STEPS IN TRAINING THE MODEL

A
  1. Train
  2. Model Selection
  3. Test
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

learn models on the training data using different hyperparameters.

A

Train

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

evaluate the models using the validation data and choose the hyperparameters with the best accuracy.

A

Model Selection

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

test the final model using the test data.

A

Test

17
Q

3 TYPES OF CLASSIFICATIONS ERRORS

A
  • Training Errors
  • Test Errors
  • Generalization Errors
18
Q

errors committed on the training set.

A

Training Errors

19
Q

errors committed on the test set.

A

Test Errors

20
Q

expected errors of a model over random selection of records from same distribution.

A

Generalization Errors

21
Q

is when a model is too simple, both training and test errors are large.

A

Underfitting

22
Q

when model is too complex, training error is small, but test error is large.

A

Overfitting

23
Q

2 REASONS FOR OVERFITTING

A
  1. Not enough training data
  2. High model complexity
24
Q

2 MODEL SELECTION FOR DECISION TREE

A
  1. Pre-Pruning (Early Stopping Rule)
  2. Post-Pruning
25
Q

stops the algorithm before it becomes a fully grown tree.

A

Pre-Pruning (Early Stopping Rule)

26
Q

grow decision tree to its entirety.

A

Post-Pruning