week 5 Flashcards

1
Q

How does the interpretation of a Bayesian credible interval differ from a frequentist confidence interval?

A

A 100(1-α)% credible interval (L, U) contains the parameter θ with probability (1-α) given the observed data x, i.e., P(L ≤ θ ≤ U | x) = 1-α. A frequentist confidence interval (L(X), U(X)) is a random interval such that in the long run, 100(1-α)% of such intervals would contain the fixed, unknown true parameter value.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the mathematical definition of a 100(1-α)% Bayesian Credible Interval I = (L, U)?

A

It is an interval such that the integral of the posterior density π(θ|x) from L to U equals 1-α: ∫[L, U] π(θ|x) dθ = 1-α.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Are Bayesian credible intervals unique for a given posterior distribution and confidence level α?

A

No, there are generally infinitely many intervals (L, U) that satisfy the definition for a given α and posterior.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is a Quantile Credible Interval?

A

An interval (L, U) where L is the α/2 quantile and U is the 1-α/2 quantile of the posterior distribution π(θ|x).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What defines a Highest Posterior Density (HPD) credible interval (L, U)?

A

For any θ’ inside the interval [L, U] and any θ’’ outside the interval, the posterior density at θ’ is greater than or equal to the posterior density at θ’’: π(θ’|x) ≥ π(θ’‘|x).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is a key characteristic of HPD intervals compared to other credible intervals of the same size?

A

They are the shortest possible interval containing 100(1-α)% posterior probability.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

If the posterior distribution is symmetric and unimodal (like a Normal distribution), how does the HPD interval relate to the quantile interval?

A

They coincide, and π(L|x) = π(U|x).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

How does a quantile credible interval transform under a bijective (monotonic) transformation g(θ)?

A

If [L, U] is a quantile interval for θ, then [g(L), g(U)] is the corresponding quantile interval for g(θ).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Is the same transformation property true for HPD intervals?

A

No, only if the transformation g is linear. Otherwise, the HPD interval for g(θ) must be recalculated.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

In a multivariate setting with parameter θ = (δ, ξ), where δ is the parameter of interest and ξ is a nuisance parameter, how is the marginal posterior distribution for δ, π(δ|x), obtained?

A

By integrating the joint posterior distribution π(δ, ξ|x) over the nuisance parameter ξ: π(δ|x) = ∫ π(δ, ξ|x) dξ.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

If we have N samples θ¹, …, θᴺ drawn from the posterior distribution π(θ|x), how can we approximate the posterior mean E[θ|x]?

A

Using the sample mean: (1/N) * Σ_{j=1}^N θʲ.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

How can we approximate the posterior probability P(L ≤ θ ≤ U | x) using N posterior samples θ¹, …, θᴺ?

A

By calculating the proportion of samples that fall within the interval [L, U]: (1/N) * Σ_{j=1}^N 1_{[L,U]}(θʲ), where 1 is the indicator function.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

How can Monte Carlo samples be used to estimate a quantile (e.g., the α-th quantile) of the posterior distribution?

A

Draw N samples, sort them, and find the sample corresponding to the desired quantile index (e.g., k = ceil(α * N)). (Algorithm 1)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What general class of methods is mentioned for sampling from posterior distributions, especially when they are not standard or conjugate, often without needing the normalizing constant?

A

Markov Chain Monte Carlo (MCMC) methods (e.g., Metropolis-Hastings, Gibbs sampler).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

If MCMC provides samples from the joint posterior π(δ, ξ|x), how can samples from the marginal posterior π(δ|x) be obtained?

A

By simply discarding the ξ components of the joint samples.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is the Prior Predictive Distribution f(x*)? What does it represent?

A

f(x) = ∫ f(x|θ) π₀(θ) dθ. It represents the expected distribution of a new observation x* before collecting any data (x), averaging over the prior uncertainty about θ.

17
Q

What is the Posterior Predictive Distribution f(x*|x)? What does it represent?

A

f(x|x) = ∫ f(x|θ) π(θ|x) dθ. It represents the expected distribution of a new observation x* after observing data x, averaging over the remaining (posterior) uncertainty about θ.

18
Q

How is the posterior predictive distribution f(x*|x) derived or justified intuitively?

A

It’s the likelihood f(x*|θ) for a new observation, weighted by the posterior beliefs about θ, π(θ|x), integrated over all possible θ.