HW 5 Flashcards
It is good practice to perform variable selection based on the statistical significance of the regression coefficients
False
The training risk is an unbiased estimator of the prediction risk
False
When the number of predicting variables is large, both backward and forward step wise regressions will always select the same set of variables
False
It is not required to standardize or rescale the predicting variables when performing regularized regression
False
Complex models with many predictors are often extremely biased, but have low variance
False
Variable selection is a simple and solved statistical problem since we can implement it using the R statistical software
False
Backward stepwise regression is preferable over forward stepwise regression because it starts with larger models
False
Stepwise regression is a greedy algorithm searching through all possible combinations of the predicting variables to find the model with the best score
False, not all possible combinations
Akaike Information Criterion (AIC) is an estimate for the prediction risk
True
Mallow’s CP statistic penalizes complexity for the model more than leave-one-out CV and BIC
False, BIC penalizes more than other approaches
Ridge regression is a regularized approach that can be used for variable selection
False
The lasso regression requires a numerical algorithm to minimize the penalized sum of least squares
True
The L1 penalty measures the sparsity of a vector and forces regression coefficients to be zero
True
Elastic net regression use both penalties of ridge regression and hence combines the benefits of both
True
In regularized regression, the penalization is generally applied to all regression coeffs where p = number of predictors
False, the shrinkage penalty is not applied to the intercept