Capacity, Overfitting And Underfitting Flashcards

1
Q

The ability to perform well on previously unobserved inputs is called

A

Generalisation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Generalisation error

A

Aka test error

For linear regression (image)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Training error (linear regression)

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Data generating process

A

Making IID assumptions of training and test data collectively

We assume we are drawing test and train data from sample distribution (data generating distribution demoted pdata )

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

caveat to data generating process in ML

A

We do not fix parameters ahead of time

We sample training set then use it to choose parameters to reduce training set error THEN sample test set

Therefore, Etest error >= Etraining error

(Where without this process they would be equal as they’re from same dist)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Factors determining how well an ML algorithm performs

A

1) make training error small (large error is underfitting)

2) make gap between training error and test error small (large gap is overfitting)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

How do we control overfitting/underfitting

A

Altering capacity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Capacity def

A

(Informally) a model’s capacity is its ability to fit a wide variety of functions

Low capacity models may struggle to fit training set

High capacity may over fit by ‘memorising’ training set

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Hypothesis space

A

The set of functions that the learning algorithm is allowed to select as a solution

Altering this is one way of altering capacity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Increasing capacity of linear regression example

A

Adding polynomial terms of the inputs does not require polynomial terms in the output

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Draw and explain the underfitting over fitting diagram

A

Self explanatory diagram

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Def regularization

A

Any modification that we make to a learning algorithm to intent reduce its generalisation error but not its training error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

K fold cross validation algorithm

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Point estimator

A

Any function of the parameters

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Frequentlist perspective on statistics

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Weight decay

A

Form of regularisation for linear regression

(Image is quantity to minimise)

17
Q

Asymptotically unbiased

A
18
Q

Variance of estimator and standard error

A
19
Q

Navigating bias variance trade off

A

(Can use cross validation)

Or

20
Q

Show how bias and variance relate to over and underfitting in a diagram

A