ML-06 - Model and feature selection Flashcards

1
Q

ML-06 - Model and feature selection

How do you use validation data to check if your problem is due to high bias vs. high variance?

A

(See image)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

ML-06 - Model and feature selection

Which area is high bias and which is high variance?

A

(See image)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

ML-06 - Model and feature selection

If you have high bias, is your model underfit/overfit?

A

Underfit

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

ML-06 - Model and feature selection

If you have high variance, is your model underfit/overfit?

A

Overfit

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

ML-06 - Model and feature selection

If your model is underfit, do you have high bias or variance?

A

High bias

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

ML-06 - Model and feature selection

If your model is overfit, do you have high bias or variance?

A

High variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

ML-06 - Model and feature selection

What is the first tool to try for overfitting problems?

A

Regularization

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

ML-06 - Model and feature selection

What does regularization prevent?

A

Overfitting.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

ML-06 - Model and feature selection

Rescribe the bias/variance as a function of the regularization lamba parameter.

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

ML-06 - Model and feature selection

Describe how the error vs. training set size looks for a situation with a good bias/variance trade-off.

A

(See image)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

ML-06 - Model and feature selection

Describe how the error vs. training set size looks for a situation high bias.

A

(See image)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

ML-06 - Model and feature selection

Describe how the error vs. training set size looks for a situation with high variance.

A

(See image)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

ML-06 - Model and feature selection

What should you try if you have high variance? (3)

A
  • Get more data
  • Smaller sets of features (or smaller NN)
  • Try increasing regularization lambda
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

ML-06 - Model and feature selection

What should you try if you have high bias? (3)

A
  • Get more features
  • Feature engineering, add polynomial features
  • Try decreasing regularization lambda
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

ML-06 - Model and feature selection

What are the 3 steps of the ML design guideline?

A

1) Start with a small model (baseline) that’s quick to implement.
2) Decide if more data or features will help (guided by learning curves)
3) Error analysis, manually examine samples where model made errors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

ML-06 - Model and feature selection

In the ML design guideline, how do you perform error analysis? (3)

A
  • Look at data your model predicted wrongly
  • Look for systematic trends in the type of errors made
  • Hypothesize what cues (features) could have helped
17
Q

ML-06 - Model and feature selection

What is feature selection in ML?

A

Selecting which features are necessary, e.g. because they are redundant or not correlated with the labels (e.g. ID column).

18
Q

ML-06 - Model and feature selection

What is the goal of feature selection?

A

Find an optimal set of features that results in a “best model” for a problem.

19
Q

ML-06 - Model and feature selection

For feature selection, what are the 4 classes of feature selection mentioned?

A
  • Filter methods
  • Wrapper methods
  • Embedded methods
  • Dimensionality reduction methods
20
Q

ML-06 - Model and feature selection

Describe the gist of feature selection filter methods.

A

From the set of all features, find a subset based on some selection criteria, e.g. correlation coefficient.

21
Q

ML-06 - Model and feature selection

If you use a selection criteria to reduce a set of features to a subset, what type of feature selection method is that?

A

Filter methods

22
Q

ML-06 - Model and feature selection

What does the correlation coefficient measure?

A

Linear relationships between two or more features relative to each other or the output label.

23
Q

ML-06 - Model and feature selection

How would you use the correlation coefficient to filter out unnecessary features?

A

If two features are correlated, you don’t need both.

24
Q

ML-06 - Model and feature selection

Describe how wrapper methods work.

A

Search
- Loop until stopping
- Generate a feature set
- Test performance
- Select best performer

25
Q

ML-06 - Model and feature selection

What are the most commonly used wrapper methods? (4)

A
  • Exhaustive feature selection
  • Forward/sequential feature selection
  • backward feature elimination
  • recursive feature elimination (RFE)
26
Q

ML-06 - Model and feature selection

Describe exhaustive feature selection.

A

Try every combination of feature subsets and use the best one.

27
Q

ML-06 - Model and feature selection

Describe forward/sequential feature selection.

A

An iterative method starting with a single feature, then add more.

28
Q

ML-06 - Model and feature selection

Describe backward feature elimination

A

Start with all features, remove the worst performers.

29
Q

ML-06 - Model and feature selection

Describe recursive feature elimination (RFE)

A

is a type of back feature elimination method, which removes the weakest feature (or features) recursively one at a time until the specified number of features is reached.

30
Q

ML-06 - Model and feature selection

What is RFE short for?

A

recursive feature elimination

31
Q

ML-06 - Model and feature selection

What is RFECV short for?

A

recursive feature elimination with cross validation

32
Q

ML-06 - Model and feature selection

What are embedded methods?

A

An iterative method where in each iteration, a method extracts the features that contributed the most.

33
Q

ML-06 - Model and feature selection

Describe how regularization is used for feature selection.

A

L1 - Penalizes features to change their importance in the model. Promotes sparsity, i.e. feature is not used.

L2 - Maintains all variables, but assigns importance to improve performance.

34
Q

ML-06 - Model and feature selection

Describe how dimensionality reduction methods are used for feature selection.

A

Analyses input and projects it into a lower dimension, meaning it transforms the inputs into a new form.

35
Q

ML-06 - Model and feature selection

What are 3 common dimensionality reduction methods?

A
  • PCA (Principal Component Analysis)
  • SVD (Singular Value Decomposition)
  • LDA (Linear Discriminant Analysis)
36
Q

ML-06 - Model and feature selection

What is PCA short for?

A

Principal Component Analysis

37
Q

What is SVD short for?

A

Singular Value Decomposition

38
Q

What is LDA short for?

A

Linear Discriminant Analysis