Lecture 10 - Model Evaluation Flashcards

1
Q

What does it mean when a model has a high bias?

A

The model does not match the training data closely enough to be useful.

Bias - Limited flexibility

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What does it mean if a model has a high variance?

A

It means that it matches the training data too closely.

Variance - sensitivity to specific sets

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What does it mean if we’re too “fit”?

A

If we are too “fit” then the model conforms too much to this one data set so we can’t generalize.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the bias-variance trade off?

A

It is the effort to minimize two sources of error that prevents supervised learning algorithms from generalizing beyond their training set.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the irreducible error?

A

The bias-variance decomposition is composed of tree terms: Bias, variance, and irreducible error.

error = bais + Variance + irreducible error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

How do you detect overfitting?

A

Use a separate set of holdout data. We spit the labelled data into two collections: training and evaluation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What are the two important properties for detecting overfitting using holdout data?

A

The data were not used int he training so they cant have been memorized.

They have labels, so we can review models accuracy without labeling costs.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is cross-validation and why is it important?

A

It allows us to see how our model does, on average across a number of randomized trials.

This will tend toward the population average.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Why do we need to be careful using train-test splits?

A

We don’t want to end up with all one class in training and not in evaluation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

How do we make sure that we don’t end up with all one class in our training set?

A

Make several random splits.

We do this with k-fold cross validation where k can be 3, 5, 10 different splits.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Describe underfitting?

A

It means our model has not captured the complexity present in our training data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

You will see excellent performance on the training data and much worse performance on the test data if

A

The model is overfit

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

If you see bad performance on both the train and the test sets

A

The model is underfit

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Training performance is almost always better than

A

test performance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

If you have overfit data?

A

Collect more data
Try a similer model
Apply regularization

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

If you underfit the data?

A

Try more complex model

Try engineering additional features that may help explain

17
Q

Explain regularization in a sentence

A

Regularization is a way to penalize models as they become more complex to discourage over fitting during the training process.

18
Q

What does regularization encourage?

A

A “simpler” hypotheses

19
Q

What effect does regularization have?

A

Shrinking the coefficients of the model towards zero.

20
Q

Name two common approaches to regularization.

A

L1 and L2

21
Q

Describe L1 and L2 regularization.

A

In L1 regularization we add the absolute values

In L2 regularization we use the squared term

22
Q

Which regularization results in better predictions?

A

L2

23
Q

What is hyperparameter optimization?

A

Some models have parameters of their own.

Use use grid search over the possible values of parameters in conjunction with cross validation to discover the optimal combination of those parameters.

We learn how best to learn.