Week 5 Flashcards

1
Q

MAP (abbreviation)

A

Maximum A Posteriori method

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What does the MAP method output?

A

It finds only a single value of a parameter that maximizes the posterior, so it forgets the uncertainty in the posterior.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the Laplace approximation?

A

When you don’t have a formula for the posterior, the Laplace approximation finds a different distribution that tries to approximate the posterior.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

In what ways should the approximation agree with the posterior in Laplace approximation?

A

The maximum likelihood, w^ needs to be the same, so this fixes the choice of the mean for the approximation. Plus the area around w^ needs to behave the same in both functions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What property does the multivariate Gaussian of a Laplace approximation need to have in order to satisfy agreement of the area around w^?

A

It needs to have the same Hessian of w^ as the actual posterior.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What two things of the posterior do you need in order to construct the laplace approximation for a particular posterior?

A

The point of maximization and the 2nd derivatives.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

T/F:
The gamma distribution is a continuous distribution

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Over what is the density of the gamma distribution defined?

A

It is continuous, so on positive real numbers.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Over what is a Gaussian distribution defined?

A

Over positive and negative reals.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Over what is the beta distribution defined?

A

between 0 and 1.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

In Bayesian statistics, what is the distribution called that you need to have when you want to compute the probability that a specific new datapoint has a certain value?

A

The Bayesian predictive distribution.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What happens in Monte Carlo sampling?

A

Draw many random independent samples of w, compute P(t.new|w,…) for each w and take the average of these.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

How can you write the expected value of a discrete random variable?

A

as a summation of all possible values with their probabilities.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is the output of the Metropolis-Hastings algorithm?

A

a sequence of random samples, w1, …, wNs, with Ns= the number that we determine beforehand.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What requirement of samples do you drop when using the MH algorithm?

A

The requirement that samples are independent.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What does the MH algorithm do?

A

It explores the space of w’s and if it finds a region with large density of w’s it picks more before it continues to another region.

17
Q

Describe the MH algorithm in pseudo-code:

A

choose w1
for s=2,3,…,Ns
{ propose new sample ~ws
compute acceptance ratio r
if probability is min(1,r)
{ accept proposal, so ws= ~ws}
else
{ reject proposal, so ws= ws-1}
}

18
Q

How do you pick w1 in the MH algorithm?

A

By sampling from the prior.

19
Q

is MH an optimization algorithm?

A

no

20
Q

What is the acceptance ratio a product of?

A

The posterior ratio and the proposal ratio.

21
Q

How is the posterior ratio calculated in the MH algorithm?

A

The posterior of the proposed point is divided by the posterior of the previous point.

22
Q

what is the goal of the proposal ratio?

A

it compensates for tendencies of the proposal distribution. If the proposal density has a tendency to go to the upper right, the ratio should not always accept upper right points.

23
Q

Why is the proposal ratio for a multivariate gaussian always equal to 1?

A

Stepping in either direction is equally likely

24
Q

What is ||x||?

A

The Euclidian norm of x

25
Q

What is a Euclidian norm?

A

The length of the vector when measured from (0,0).

26
Q

For MH with a symmetric proposal distribution, the acceptance ratio depends on…

A

the posterior density at the proposed point and the posterior density at the previous point.

27
Q

Why can the MAP estimate be considered a step up from the maximum likelihood solution?

A

it incorporates the prior.

28
Q

Burn-in

A

The first x samples in the MH algorithm, called burn-in because it may take a while to find regions where the posterior is large.