Introduction Flashcards
How is defined a learning task?
By a performance P, a task T and an experience E : <P, T, E>.
What is a linear model?
It’s a model which is linear in terms of the parameters.
How to solve overfitting?
-train on more data
-regularization
What is regularization?
It consists in adding a penalty term to the error function to discourage coefficients from reaching large values:
E(w) = […] + λ/2 • ||w||².
What is model selection?
It’s a procedure to choose the value of an hyper-parameter λ:
1. For different values of λ:
- train the model
- compute the performance on the validation set
2. Choose the value of λ that has the best validation performance.
3. Compute the test performance for the model with chosen value of λ.
What is cross-validation?
It’s splitting the training set into a training set and a validation set for model selection.
What is k-fold cross-validation?
k-fold cross-validation consists in dividing the train set in k folds and doing the training and validation steps using iteratively each of the folds for validation (while using the k-1 others for training). All performance are then average. This ensures a statistically representative training/validation set split but is very costly.
What is the Machine Learning Pipeline?
- Define the input and output.
- Collect examples for the task.
- Divide the examples into training, validation and testing sets.
- Do data preprocessing.
- Define your model (including parameters and hyper-parameters).
- Define the error and loss functions you want to minimize.
- For different values of hyper-parameters:
- learn model parameters by minimizing the loss function
- compute validation performance
- Pick the best model based on validation performance.
- Test the model on the testing set.
What is the assumption of a Least squares classifier?
The decision boundary is linear.
What is the assumption of a k-Nearest-Neighbors (k-NN) classifier?
The class distribution is locally smooth.
Compare linear regression and k-NN regression in terms of:
-dependance to the model
-learning tasks
-prediction speed
-smoothness of the solution
-stability of the solution
-inductive bias
-bias, variance
linear regression VS k-NN regression:
model-based VS model-free
learning of parameters VS no learning
fast VS slow
smooth VS not smooth
more stable VS less stable
f(x) is well approximated by a globally linear function VS f(x) is well approximated by a locally constant function
high bias, low variance VS low bias, high variance