HW 5 Flashcards

1
Q

It is good practice to perform variable selection based on the statistical significance of the regression coefficients

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

The training risk is an unbiased estimator of the prediction risk

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

When the number of predicting variables is large, both backward and forward step wise regressions will always select the same set of variables

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

It is not required to standardize or rescale the predicting variables when performing regularized regression

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Complex models with many predictors are often extremely biased, but have low variance

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Variable selection is a simple and solved statistical problem since we can implement it using the R statistical software

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Backward stepwise regression is preferable over forward stepwise regression because it starts with larger models

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Stepwise regression is a greedy algorithm searching through all possible combinations of the predicting variables to find the model with the best score

A

False, not all possible combinations

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Akaike Information Criterion (AIC) is an estimate for the prediction risk

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Mallow’s CP statistic penalizes complexity for the model more than leave-one-out CV and BIC

A

False, BIC penalizes more than other approaches

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Ridge regression is a regularized approach that can be used for variable selection

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

The lasso regression requires a numerical algorithm to minimize the penalized sum of least squares

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

The L1 penalty measures the sparsity of a vector and forces regression coefficients to be zero

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Elastic net regression use both penalties of ridge regression and hence combines the benefits of both

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

In regularized regression, the penalization is generally applied to all regression coeffs where p = number of predictors

A

False, the shrinkage penalty is not applied to the intercept

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

If there are specific variables that are required to control the bias selection in the model, they should be forced into the model and not be part of the variable selection process

A

True

17
Q

The penalty constant lambda in penalized regression controls the trade-off between lack of fit and model complexity

A

true

18
Q

In ridge regression, when the penalty lambda is zero, the corresponding ridge regression estimates are the same as the ordinary least squares estimates

A

True

19
Q

Ridge regression can be used to deal with problems caused by high correlation among the predictors

A

True

20
Q

When selecting variables for explanatory purpose, one might consider including predicting variables which are correlated if it would help answer your research hypothesis

A

True