DS | Statistics | Priority Flashcards

1
Q

Sampling distribution?

A

tak “The sampling distribution of an estimator is the distribution of results we would see if we applied the estimator multiple times to different datasets sampled from some distribution;”

Probabilistic Machine Learning: An Introduction 4.7.1 Sampling distributions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Sampling distribution?

A

tak “The sampling distribution of an estimator is the distribution of results we would see if we applied the estimator multiple times to different datasets sampled from some distribution;”

Probabilistic Machine Learning: An Introduction 4.7.1 Sampling distributions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Treatment of data and parameter in frequentist vs. Bayesian approaches?

A

tak “In the frequentist approach, theta is treated as an unknown fixed constant, and the data is treated as random. In the Bayesian approach, we treat the data as fixed (since it is known) and the parameter as random (since it is unknown).“

Probabilistic Machine Learning: An Introduction 4.7.5 Caution: Confidence intervals are not credible

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Basic idea of a confidence interval.

A

tak “This means that, if we repeatedly sampled data, and compute I(~D) for each such dataset, then about 95% of such intervals will contain the true parameter” theta

Probabilistic Machine Learning: An Introduction 4.7.4 Confidence intervals

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Treatment of data and parameter in frequentist vs. Bayesian approaches?

A

tak “In the frequentist approach, theta is treated as an unknown fixed constant, and the data is treated as random. In the Bayesian approach, we treat the data as fixed (since it is known) and the parameter as random (since it is unknown).“

Probabilistic Machine Learning: An Introduction 4.7.5 Caution: Confidence intervals are not credible

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is te equation for bias/variance for MSE?

A

MSE = variance + bias^2

Probabilistic Machine Learning: An Introduction 4.7.6.3 The bias-variance tradeoff

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

How does regularization affect bias/variance?

A

tak “as we increase the strength of the regularizer, the variance decreases, but the bias increases.“

Probabilistic Machine Learning: An Introduction 4.7.6.5 Example: MAP estimator for linear regression

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Illustration of the bias variance tradeoff.

A

Figure 4.26: Cartoon illustration of the bias variance tradeoff.

Probabilistic Machine Learning: An Introduction 4.7.6.5 Example: MAP estimator for linear regression

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Bias-variance tradeoff for classification?

A

“If we use 0-1 loss … the bias and variance combine multiplicatively.“; “This little known fact illustrates that the bias-variance tradeoff is not very useful for classification. It is better to focus on expected loss … We can approximate the expected loss using cross validation“

Probabilistic Machine Learning: An Introduction 4.7.6.6 Bias-variance tradeoff for classification

How well did you know this?
1
Not at all
2
3
4
5
Perfectly