lecture 7: Over-fitting and bias/variance trade-off Flashcards

1
Q

review: what is the main aim of regression?

A
given feature(s) x, we want to predict target y
note: x can be 1D or multi-D, y is 1-D
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

what is the number one rule for train and test sets

A

they should never overlap, test set should always be unseen data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

what is overfitting

A

fit is very good for the training set but very bad for the test set, usually means the model used is too complex

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

what is underfitting

A

the fit is very bad for both the training and test set, usually means the model used is too simple

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

what are the reasons for overfitting

A

model is too complex or too many features and not enough training samples

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

what are the solutions for overfitting

A

use simpler models(eg. lower order polynomial) or use regularisation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

what are the reasons for underfitting

A

model is too simple or features are not informative enough

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

what is regularisation

A

it is an umbrella term that includes methods that force learning algorithms to build less complex models
recall in previous lecture 𝜆

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

what does adding the regularisation term 𝜆reg(w) do?

A

encourages w to be small - called weight decay(L2 regularisation), which penalises more complex models
visually, it makes complex models flatter

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

what does lambda signify?

A

the trade-off between data-loss and regularisation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

what is the difference between bias and variance

A

low bias represents the predictions being close to the target on average while low variance represents the spread of the predictions being small

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

in general, very simple models exhibit what bias and variance

A

high bias and low variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

in general, very complex models exhibit what bias and variance

A

low bias an high variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

according to the bias variance trade-off theorem, the mse of a new test sample x is given by

A

test error = bias squared + variance + irreducible noise
Bias(f)² + Var(f) + σ²
Bias = favg(x) - f(x)
Var = E[(f(x) - favg(x))²]

How well did you know this?
1
Not at all
2
3
4
5
Perfectly