Lecture 14 - Feature Selection Flashcards

1
Q

What is the curse of dimensionality?

A

The paradox of the more features you use in your model, the more data you need to help sort though the various permutations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What does feature selection do?

A

Helps us quickly identify features that don’t offer predictive power which allows us to iterate faster and with less data so we can experiment more.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Which method of feature selection analyzes the variance of each feature and removes any features that have a low variance.

A

Variance Threshold

Bike seat example

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

SelectKBest does what?

A

Removes all but the k highest scoring features.

It is configurable using any function of (X, y) that returns a set of scores per feature.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Chi2 is

A

A predefined scoring function that tests the likelihood that the relationship between each variable and the outcome is due to that feature

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is RFE?

A

Recursive feature elimination

Uses an external estimator that assigns weights to features (coefficients of a linear model).

It selects features by recursively considering smaller and smaller sets of features.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

How does RFE work?

A

It trains an external estimator and then drops the least important features.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

How does SelectFromModel differ from RFE?

A

Select from model using a threshold on the coef or feature_importances_ values instead.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly