Introduction Flashcards

1
Q

How is defined a learning task?

A

By a performance P, a task T and an experience E : <P, T, E>.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is a linear model?

A

It’s a model which is linear in terms of the parameters.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

How to solve overfitting?

A

-train on more data
-regularization

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is regularization?

A

It consists in adding a penalty term to the error function to discourage coefficients from reaching large values:
E(w) = […] + λ/2 • ||w||².

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is model selection?

A

It’s a procedure to choose the value of an hyper-parameter λ:
1. For different values of λ:
- train the model
- compute the performance on the validation set
2. Choose the value of λ that has the best validation performance.
3. Compute the test performance for the model with chosen value of λ.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is cross-validation?

A

It’s splitting the training set into a training set and a validation set for model selection.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is k-fold cross-validation?

A

k-fold cross-validation consists in dividing the train set in k folds and doing the training and validation steps using iteratively each of the folds for validation (while using the k-1 others for training). All performance are then average. This ensures a statistically representative training/validation set split but is very costly.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the Machine Learning Pipeline?

A
  1. Define the input and output.
  2. Collect examples for the task.
  3. Divide the examples into training, validation and testing sets.
  4. Do data preprocessing.
  5. Define your model (including parameters and hyper-parameters).
  6. Define the error and loss functions you want to minimize.
  7. For different values of hyper-parameters:
    • learn model parameters by minimizing the loss function
    • compute validation performance
  8. Pick the best model based on validation performance.
  9. Test the model on the testing set.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the assumption of a Least squares classifier?

A

The decision boundary is linear.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the assumption of a k-Nearest-Neighbors (k-NN) classifier?

A

The class distribution is locally smooth.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Compare linear regression and k-NN regression in terms of:
-dependance to the model
-learning tasks
-prediction speed
-smoothness of the solution
-stability of the solution
-inductive bias
-bias, variance

A

linear regression VS k-NN regression:
model-based VS model-free
learning of parameters VS no learning
fast VS slow
smooth VS not smooth
more stable VS less stable
f(x) is well approximated by a globally linear function VS f(x) is well approximated by a locally constant function
high bias, low variance VS low bias, high variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly