Bayesian Analysis Flashcards

1
Q

[Derive the marginal distribution of β from a regression model with an (adjusted) conjugate prior.]

Consider the conjugate linear regression Bayesian framework. Here, the prior of sigma^2 follows an IG-2(delta,v) distribution. We need to derive the marginal distribution of beta. The question is partly how the delta and v will be processed into the posterior.

A

We use the standard Bayesian approach to solve a linear regression framework for a conjugate prior. v is introduced in the power and delta is introduced as intercept besides the (constructed) w and V. delta is introduced as a constant in sigma^2 tilde and v is introduced as extra degrees of freedom.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

[Derive the marginal distribution of sigma^2 from a regression model with an (adjusted) conjugate prior.]

Consider the conjugate linear regression Bayesian framework. Here, the prior of sigma^2 follows an IG-2(delta,v) distribution. We need to derive the marginal distribution of beta. The question is partly how the delta and v will be processed into the posterior.

A

We use the standard Bayesian approach to solve a linear regression framework for a conjugate prior. The delta enters the scale parameter of the prior marginal distribution IG-2 as a constant and v as an addition to the degrees of freedom. In the derivation, we make use of a few tricks: integral equal to a constant, determinant of a vector times matrix and the decomposition rule.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

To avoid the inclusion of prior influence into a Bayesian analysis, researchers often
use a non-informative (flat) prior specification, that is,
p(θ) ∝ 1,
where p(θ) is the prior density for the model parameters θ. A researcher claims that
this flat prior is not completely uninformative as it is informative about nonlinear
functions of the parameters θ. Is he right? Motivate your answer!

A

He is right. When you are uninformative about θ it does not have to be the case
that you are uninformative on a nonlinear transformation of θ as p(θ) ∝ 1 does not
automatically imply that p(h(θ)) ∝ 1 where h is a nonlinear function of θ. This is
due to the Jacobian of the transformation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Let the prior be: β∣σ2∼N(b,σ2/γN)​) and p(σ2)∝σ−2
Linear regression: y = eβ + ε
The posterior mean of β is a weighted average of b and the mean of y. Derive
the weight!

A

This prior is a conjugate prior where B = 1/γN. Use the results from the slides that the marginal posterior density β|y follows a t-distribution with a defined location and scale parameter. Take the location parameter. Use that X=e, e’e = N and e’y = Nmean(y). Then, arithmetically solve the formula of the location parameter.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Let the prior be: β∣σ2∼N(b,σ2/γN)​) and p(σ2)∝σ−2
It is often stated that the prior does not influence the posterior mean for large
number of observations? Explain why this is not the case for this prior.

A

The prior variance decreases with N and hence the prior information gets larger
when the sample size increases.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly