Complexity/Selection/Regularization Flashcards

1
Q

What is the bias-variance decomposition?

A

E[(Y - f̂_D(x))^2] = (Bias^2) + (Variance)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What does high bias in a model indicate?

A

A bias towards a particular kind of solution (e.g., linear model), also known as inductive bias.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What does high variance in a model indicate?

A

: The estimated model changes significantly when trained on different data sets, indicating overfitting.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the VC dimension?

A

The maximum number of points that can be correctly classified by at least one member of a set of classifiers.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the VC dimension of a linear classifier on R^p?

A

VC = p + 1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

How are Degrees of Freedom (DF) defined for an estimate ŷ = f̂(X)?

A

Back: df(ŷ) = (1/σ^2) Σ cov(ŷ_i, y_i) = (1/σ^2) tr(cov(ŷ, y))

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is an intuition/interpretation of degrees of freedom in model complexity defined for an estimate ŷ = f̂(X)??

A

Degrees of freedom in model complexity represent the number of independent parameters that can be adjusted to fit the data, influencing the model’s flexibility.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Relationship between complexity, bias, variance and total error

A

The higher the model complexity, the lower the bias, the higher the variance and total error does a convex in between.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What does a good bias require?

A

Domain knowledge.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the relationship between degrees of freedom, number of samples, number of features, and λλ in ridge regression?

A

In ridge regression, the degrees of freedom are influenced by the number of features and the penalty parameter λλ. As λλ increases, the degrees of freedom decrease, reducing model complexity and helping to prevent overfitting, especially when the number of samples is limited relative to the number of features.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the PRESS statistic?

A

Predicted Residual Error Sum of Squares: PRESS = Σ(y_i - ŷ_-i)^2, where ŷ_-i is the prediction for the i-th sample when the model is estimated on all but the i-th sample.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Higher model complexity effect on performance? What to do?

A

It will always lead to better fit on training data but not on test data therefore we need to select a model by estimating its performance for train and validation/test data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is a method for validation?

A

Cross validation: estimate generalization error on different train/test splits.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is Leave-one-out Cross-Validation (LOO-CV)?

A

A method where the model is trained on all but one sample and tested on the left-out sample, repeated for all samples. The average prediction error is reported.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is k-fold Cross-Validation?

A

A method where data is split into k subsets, the model is trained on k-1 subsets and tested on the remaining one, repeated k times. Often k=5 or k=10 is used in practice.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is the relationship between expected prediction error and expected training error?

A

expected pred error = expected train error + constant (>0) * DF(model)

17
Q

What is a better selection criteria?

A

training error + complexity penalty. The larger the complexity (which tends to increase train error and lower generalization), the larger the penalty to keep it down!

18
Q

If two models fit the data equally well? which to select?

A

One with less complexity!

19
Q

What is the problem with cross validation?

A

If we test too many models, we will reach overfitting..

20
Q

What is the Bayes factor in model selection?

A

The ratio of marginal likelihoods: pr(x|m_i) / pr(x|m_j)

21
Q

What is the Bayes Information Criterion (BIC)?

A

BIC(x;m) = -2 log pr(x|θ̂,m) + p log(n), where θ̂ is the maximum likelihood estimate, p is the number of parameters, and n is the number of samples.

22
Q

What is the Fisher Information Approximation (FIA)?

A

FIA(x;m) = -log pr(x|θ̂,m) + (p/2)log(n/2π) + log C_m, where C_m is the geometric complexity

23
Q

What is the objective of l_k-penalized regression?

A

Minimize ω(θ) + λ||θ||_k^k, where ω(θ) is the loss function and ||θ||_k is the l_k norm of the parameters.

24
Q

What is the objective of l_k-penalized regression?

A

As the number of parameters increases past the interpolation threshold, test error first increases then decreases again, forming a double descent curve.

25
Q

What is implicit regularization in the context of gradient descent?

A

The gradient descent optimization process itself can act as a regularizer, especially in deep neural networks.

25
Q

How can increasing the number of features reduce model complexity?

A

When using a minimum l_2-norm estimator, increasing features can lead to implicit regularization.

26
Q

Why might overparameterized models be beneficial for complex data?

A

They can capture complex patterns while still benefiting from regularization, which is part of the success behind deep learning.

27
Q

How does the l_1 norm penalty (Lasso) differ from the l_2 norm penalty (Ridge) in regularization?

A

L_1 tends to produce sparse solutions (some coefficients exactly zero), while l_2 shrinks all coefficients but rarely to exactly zero.

28
Q

What is the “No Free Lunch” theorem in machine learning?

A

There is no single model that works best for every problem; the best model depends on the specific data and domain.

29
Q

What is Occam’s razor in the context of model selection?

A

If two models fit the data equally well, we should select the simpler one.

30
Q

How does the AIC (Akaike Information Criterion) differ from BIC?

A

AIC uses a fixed penalty of 2p, while BIC uses p log(n), making BIC more conservative for large n.

31
Q

What is the marginal likelihood in Bayesian model selection?

A

pr(x|m) = ∫ pr(x|θ,m)pr(θ|m)dθ, integrating over all possible parameter values.

32
Q

How can gradient descent act as an implicit regularizer?

A

It can bias the solution towards certain regions of the parameter space, often simpler or smoother solutions.

33
Q

What is the geometric interpretation of l_1 and l_2 regularization?

A

L_2 forms circular (2D) or spherical (higher dimensions) constraint regions, while l_1 forms diamond-shaped regions with corners on the axes.

34
Q

How does the effective degrees of freedom change as the ridge penalty λ increases?

A

The effective degrees of freedom decrease as λ increases, indicating a simpler model.