Bias Variance Trade off and Inductive Bias Flashcards
Reducing bias when having a large variance:
When you have a large complex model and training data S is small, small changes to S lead to large changes in the model. This is a risk of overfitting.
Reduce variance having a large bias:
When you have a restricted, simple model, and the training data is small, small changes to it don’t lead to large changes in the model. This leads to a risk of underfitting (strong generalisation without sufficiently fitting the model)
What do machine learning algorithms need to be able to generalise to new data?
It needs bias (inductive bias)
What are the two sources of bias?
Representation bias
Search bias
What is inductive bias?
It is a criterion (preference or assumption) different from consistency with the data, used to favour a hypothesis over another.
What must inductive algorithms have?
Inductive bias
What is the approximation error?
How much error we have because we restrict ourselves to a specific class of models (i.e., how much inductive bias we have)
What is estimation error?
The training error is only an estimation of the true error.
What does every classification algorithm have?
Inductive bias
Is bias suitable for all data?
Yes and no. Every bias is suitable for some data sets and unsuitable for others.
What is the best way to find the best result?
Trying several different classification algorithms and using the best result (called the Toolbox Approach)