Lecture 9 - Bayesian inference: approximation & sampling Flashcards
What is Bayesian inference?
It involves updating the probability of a hypothesis as more evidence becomes available, using Bayesβ theorem.
What is the Laplace approximation?
It approximates a complicated posterior distribution with a simpler multivariate Gaussian π(π,Ξ£).
What are the key parameters in the Laplace approximation?
- ΞΌ is the mode of the posterior (maximum of π(π€)).
- Ξ£=βπ»β1 , where π» is the Hessian matrix at w^\hat{}.
How is the mean π of the Gaussian chosen in the Laplace approximation?
The mean π is set to the maximum of the posterior distribution (π€^ ).
How is Ξ£ determined in the Laplace approximation?
Ξ£ =βπ»β1, where π» is the Hessian matrix (second derivative of the log-posterior) at π€^
How does the Laplace approximation perform with the Gamma distribution?
It approximates well near the mode π¦^ but diverges significantly further away (Page 4, visual example).
What is Monte Carlo sampling?
A technique to estimate integrals or expectations by averaging values from random samples
How do you estimate the Bayesian predictive distribution?
By sampling π€ from π(π,Ξ£) and averaging π(π‘newβ£π€)
What is the Metropolis-Hastings algorithm?
It generates samples from the posterior by proposing and accepting/rejecting steps based on an acceptance ratio.
Mnemonic: βPropose, compare, accept (or stay).β
How is the acceptance ratio calculated in Metropolis-Hastings?
It combines the posterior ratio and the proposal ratio:
What does the Metropolis-Hastings algorithm achieve?
It generates samples from a posterior distribution even when it cannot be computed analytically
What are common challenges in sampling?
High-dimensional spaces require many burn-in samples.
Risk of exploring only local maxima.
High rejection rates without careful proposal densities.
What is the βburn-inβ period in sampling?
The initial samples discarded because they may not represent the posterior accurately.
What are the main steps of the Metropolis-Hastings algorithm?
- Propose a new sample π€~π based on the previous sample π€π β1
- Compute acceptance ratio π.
- Accept π€~π with probability
min(π,1); otherwise, keep π€π β1
What are some challenges with the Metropolis-Hastings algorithm?
Requires discarding βburn-inβ samples.
Risk of exploring only local maxima.
May reject most proposals if the proposal distribution is poorly chosen.