Wronged Questions: Statistical Learning Flashcards

1
Q

Adding more predictions will never (increase/decrease) R^2

A

Decrease

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

T/F: The expected test MSE formula applies to both quantitative and qualitative responses

A

False. It only applies to quantitative responses

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

T/F: In the classification setting, the bias-variance trade-off does not apply since y_i is quantitative.

A

False. The bias-variance trade-off does indeed apply in the classification setting, albeit with some modifications due to
being categorical rather than quantitative.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

T/F: The training error rate is defined as the proportion of correct classifications made when applying our estimate
to the training observations.

A

False. The training error rate is defined as the proportion of incorrect classifications made when applying our estimate
to the training observations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

T/F: A classifier’s effectiveness is determined by the magnitude of its training error rate rather than its test error rate.

A

False. A classifier’s effectiveness is determined by the magnitude of its test error rate rather than its training error rate.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

T/F: The Bayes classifier is known to produce the highest possible test error rate, known as the Bayes error rate.

A

False. The Bayes classifier is known to produce the lowest possible test error rate, known as the Bayes error rate.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

T/F: The Bayes error rate serves a role similar to that of the irreducible error in the classification setting.

A

True. The Bayes error rate is analogous to the irreducible error in classification, representing the lowest error rate that can be achieved by any classifier and is due to the noise in the data itself.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Less flexible, more interpretable

A

Lasso, subset selection

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Moderately flexible and interpretable

A

Least squares, regression trees, classification trees

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

More flexible, less interpretable

A

Bagging, boosting

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

T/F: The K-Nearest Neighbor algorithm determines the function for f
that includes free parameters.

A

False. KNN is non-parametric

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

T/F: Random forest makes no assumption about f’s function form.

A

True, random forest is non-parametric

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

T/F: Compared to non-parametric methods, parametric methods are more versatile in fitting various forms of f.

A

False. Non-parametric methods are more versatile because they make no assumptions about f’s form.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

T/F: Compared to non-parametric methods, parametric methods typically need a larger number of observations to accurately estimate f.

A

False. Parametric methods need smaller sample sizes because they already assume a pattern

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

T/F: Compared to non-parametric methods, parametric methods are generally more difficult to interpret.

A

False, parametric methods are easier to interpret due to them already having built-in structure

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

T/F: Decision boundaries become overly flexible and capture noise, especially at lower K values.

A

True

17
Q

T/F: Larger K values lead to smoother, less flexible boundaries that better generalize to new data.

A

True

18
Q

T/F: Smaller K values result in boundaries that closely follow the training data, potentially capturing noise.

A

True

19
Q

T/F: Larger K values result in highly complex boundaries that can easily overfit the data.

A

False. Smaller K values result in highly complex boundaries that can easily overfit the data.

20
Q

T/F: Smaller K values create smoother and more generalized boundaries.

A

False. Larger K values create smoother and more generalized boundaries.

21
Q

Inference

A

Exploring + confirming the significant associations between predictors and the response

22
Q

Non-parametric learning methods

A

1) KNN
2) Decision trees
3) Bagging, boosting, random forest

23
Q

Bias

A

Error arising from the assumptions made in the statistical learning tool

24
Q

Variance

A

Error arising from the model’s sensitivity towards the training data

25
Q

T/F: If we are mainly interested in inference, then restrictive models are much more interpretable.

A

True

26
Q

T/F: Subset selection is more interpretable than linear regression.

A

True

27
Q

T/F: Bagging and boosting are more restrictive than linear regression.

A

False

28
Q

T/F: Highly flexible methods make it difficult to discern how changes in predictors affect the response.

A

True

29
Q

T/F: The reducible error provides a lower bound for the test error.

A

False

30
Q

T/F: In the classification setting, the bias-variance trade-off does not apply since yi is quantitative.

A

False. The bias-variance trade-off does indeed apply in the classification setting, albeit with some modifications due to yi being categorical rather than quantitative.

31
Q

T/F: KNN can often produce classifiers that are close to the optimal Bayes classifier.

A

True