L5 - Regression - Bias-Variance Trade Off Flashcards
Define bias…
Bias refers to the extent that a model produces errors due to under-fitting. High bias means that high model error is encountered due to underfitting.
Define variance…
Variance refers to the extent that a model has been overfit. High variance means a model will not perform accurately on unseen data due to overfitting the training data.
What is the ideal bias-variance trade off for a regression model?
Low bias and low variance.
What are the 3 causes of model error? Define each…
Bias - The extent to which a model predicts wrong.
Variance - The extent to which a model has learned data, and thus is overfit.
Irreducible Error - Refers to the unavoidable randomness of real world data that can cause errors.
How can we tune bias and variance?
To increase bias we must reduce variance and vice versa.
What is another way of wording the bias-variance trade off of a model?
Bias-variance trade off can also be called a Complexity trade off.
If we achieve the optimal bias-variance ratio, what model complexity does this give us?
Optimal model complexity.
What technique can we use to prevent our model from overfitting, and to encourage generalisation?
Linear Model Regularisation (Shrinkage)
This reduces the magnitude of the coefficients in the polynomial regression model.
What is the objective of Linear Model Regularisation?
Establish a trade off between bias and variance, resulting in optimal model complexity.
How does Linear Model Regularisation work?
- Introduces a penalty to the models loss function
- Penalty can be increased or decreased to increase or decrease complexity.
- We want to push all coefficients towards 0.
What happens if the Tuning parameter of the Regression Models cost function is increased?
- Results in less regularisation, which leans to overfitting ( increases variance )
What happens if the Tuning parameter of the Regression Models cost function is decreased?
- Results in more regularisation
- Decreased model complexity which leans to underfitting ( increases bias )
What is L1 regularisation?
- L1 ( LASSO ) applies a penalty value that is proportional to the sum of the absolute coefficient values.
- Prevents overfitting and performs feature selection.
What is L2 regularisation?
- Ridge Regression
- Prevents overfitting and improves model stability.
- A penalty is applied that is proportional to the squared coefficient values.
- Penalty imposes a bias, and thus can be used to control the bias-variance trade off.
How do we establish the penalty value?
- Through cross-validation, to find the lambda with the lowest variance.