Dimension Reduction Flashcards

1
Q

Reasons for dimension reduction

A
  1. Computational Cost
  2. Financial Cost
  3. Interpretability
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Explain the Filter strategy for feature selection

A

Its a pre-processing step that ranks and filters features independently of the choice of classifier. It assigns a score to the different feature subsets using an evaluation function of choice

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is a good strategy for selecting the top features from a Filter

A

evaluate classifier performance using feature subsets of increasing size

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Disadvantages of Filters

A
  1. No model bias: doesn’t account for feature suitability across models
  2. No feature dependencies
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Explain the Wrapper strategy for feature selection

A

The classifier is “wrapped” in the feature selection mechanism. Feature subsets are evaluated directly based on their performance when used with that specific classifier

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Advantages of Wrappers

A
  1. Accounts for bias
  2. considers features in context (feature dependencies)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

List the types of search used in feature subset search

A
  • exhaustive
  • exponential (heuristics)
  • sequential (add/remove one f at a time)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Describe the steps of Forward Sequential search

A
  1. Start with an empty subset
  2. find the most informative feature and add it to the subset
  3. Repeat until there us no improvement by adding features
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Describe the steps of Backward Elimination

A
  1. Start with the complete set of features
  2. remove the least informative feature
  3. repeat until there is no improvement by dropping features
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Compare Forward Sequential Search (FSS) to Backward Elimination (BE)

A
  • FSS requires less running time if completed early
  • BE tends to find better models, can find subsets with interacting features, but tends to be slower
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Disadvantages of Wrappers

A
  • Computational cost
  • risk of overfitting
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is the general idea of projection methods

A

They are used in feature transformation to map the original d-dimensional space to a new (k-d)-dimensional space, with the minimum loss of information

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is Principal Component Analysis (PCA)

A

an unsupervised projection method which aims to keep as much of the variance in the data as possible

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What are principal components in PCA

A

new dimensions constructed (from eigenvectors) as linear combinations of the original features, which are uncorrelated with one another.
The first PC accounts for the most variability in the data, and so on…

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Give context as to what eigenvectors and eigenvalues are

A

Given a matrix X, an eigenvector of the matrix is a non-zero vector v that satisfies the equation: Xv = λv, where λ is the eigenvalue.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is eigendecomposition

A

the factorisation of a matrix into its eigenvalues and eigenvectors

17
Q

Eigenvectors in the context of PCA

A
  • orthogonal matrices which have direction
  • each eigenvalue represents the variance in its direction
18
Q

What is a Covariance Matrix in PCA

A

Symmetric matrix which measures the tendency for features to vary in the same direction

19
Q

Give the steps of the PCA Algorithm (given dataset matrix X with n examples)

A
  1. Calculate the mean of the columns of X
  2. Subtract the column means from each row of X, to create the centred matrix Y
  3. Calculate the Covariance matrix C
  4. Calculate the eigenvectors of C
  5. The PCs are given by the eigenvalues
  6. Select PCs as a new reduced representation of X