Capacity, Overfitting And Underfitting Flashcards
The ability to perform well on previously unobserved inputs is called
Generalisation
Generalisation error
Aka test error
For linear regression (image)
Training error (linear regression)
Data generating process
Making IID assumptions of training and test data collectively
We assume we are drawing test and train data from sample distribution (data generating distribution demoted pdata )
caveat to data generating process in ML
We do not fix parameters ahead of time
We sample training set then use it to choose parameters to reduce training set error THEN sample test set
Therefore, Etest error >= Etraining error
(Where without this process they would be equal as they’re from same dist)
Factors determining how well an ML algorithm performs
1) make training error small (large error is underfitting)
2) make gap between training error and test error small (large gap is overfitting)
How do we control overfitting/underfitting
Altering capacity
Capacity def
(Informally) a model’s capacity is its ability to fit a wide variety of functions
Low capacity models may struggle to fit training set
High capacity may over fit by ‘memorising’ training set
Hypothesis space
The set of functions that the learning algorithm is allowed to select as a solution
Altering this is one way of altering capacity
Increasing capacity of linear regression example
Adding polynomial terms of the inputs does not require polynomial terms in the output
Draw and explain the underfitting over fitting diagram
Self explanatory diagram
Def regularization
Any modification that we make to a learning algorithm to intent reduce its generalisation error but not its training error
K fold cross validation algorithm
Point estimator
Any function of the parameters
Frequentlist perspective on statistics