DS | Statistics | Priority Flashcards
Sampling distribution?
tak “The sampling distribution of an estimator is the distribution of results we would see if we applied the estimator multiple times to different datasets sampled from some distribution;”
Probabilistic Machine Learning: An Introduction 4.7.1 Sampling distributions
Sampling distribution?
tak “The sampling distribution of an estimator is the distribution of results we would see if we applied the estimator multiple times to different datasets sampled from some distribution;”
Probabilistic Machine Learning: An Introduction 4.7.1 Sampling distributions
Treatment of data and parameter in frequentist vs. Bayesian approaches?
tak “In the frequentist approach, theta is treated as an unknown fixed constant, and the data is treated as random. In the Bayesian approach, we treat the data as fixed (since it is known) and the parameter as random (since it is unknown).“
Probabilistic Machine Learning: An Introduction 4.7.5 Caution: Confidence intervals are not credible
Basic idea of a confidence interval.
tak “This means that, if we repeatedly sampled data, and compute I(~D) for each such dataset, then about 95% of such intervals will contain the true parameter” theta
Probabilistic Machine Learning: An Introduction 4.7.4 Confidence intervals
Treatment of data and parameter in frequentist vs. Bayesian approaches?
tak “In the frequentist approach, theta is treated as an unknown fixed constant, and the data is treated as random. In the Bayesian approach, we treat the data as fixed (since it is known) and the parameter as random (since it is unknown).“
Probabilistic Machine Learning: An Introduction 4.7.5 Caution: Confidence intervals are not credible
What is te equation for bias/variance for MSE?
MSE = variance + bias^2
Probabilistic Machine Learning: An Introduction 4.7.6.3 The bias-variance tradeoff
How does regularization affect bias/variance?
tak “as we increase the strength of the regularizer, the variance decreases, but the bias increases.“
Probabilistic Machine Learning: An Introduction 4.7.6.5 Example: MAP estimator for linear regression
Illustration of the bias variance tradeoff.
Figure 4.26: Cartoon illustration of the bias variance tradeoff.
Probabilistic Machine Learning: An Introduction 4.7.6.5 Example: MAP estimator for linear regression
Bias-variance tradeoff for classification?
“If we use 0-1 loss … the bias and variance combine multiplicatively.“; “This little known fact illustrates that the bias-variance tradeoff is not very useful for classification. It is better to focus on expected loss … We can approximate the expected loss using cross validation“
Probabilistic Machine Learning: An Introduction 4.7.6.6 Bias-variance tradeoff for classification