Max likelihood & Bayes Flashcards

1
Q

How does a nugget change our covariance matrix

A

goes from sigma^2 Sigma to sigma^2 Sigma + sigma^2-n I

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

why can we not just miltiply the pdfs of the data points to get the Likelihood

A

Our data are not I.I.d and we have to consider the joint distribution instead

See equation sheet to see likelihood of MVN

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are the difference between classical stats and bayesian

A

In classical/frequentist statistics:
▶ Probability of an event is the number of times the event happens divided by the number of trials (relative frequency)
▶ Parameters are fixed constants we try to estimate
In Bayesian statisics
▶ Probability is a measure of our degree of belief
▶ Parameters have probability distributions which we update when we collect data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

States Bayes theorem

A

p(theta|x) = p(x|theta)p(theta)/p(x)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Why can we get rid of the denominator in bayes theorem in favor of proportionality

A

Our densities have to integreate to 1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Say Bayes theorem in words

A

Posterior is proportional to the prior times the likelihood

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Loss function

A

The posterior is a distribution, to get a point estimate we specify a loss function and minimise the loss

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Types of loss functions

A
  • squared loss = mean of the posterior;
  • absolute loss = median;
  • (0, 1) loss = mode (also known as maximum a posteriori (MAP) estimates
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Subjective bayes

A

By using a Prior we specify Bayes is subjective

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Elciting beliefs

A

Turning prior opinons into usable distributions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Objective bayes

A

If we don’t like subjectivity/lazy we use objective priors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Conjugate prior

A

Having the form of the posterior and prior to be the same form

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Conjugate prior for Normal

A

Also Normal for the mean
Inverse Gamma for the variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Non-informative Prior

A

One option is the non-informative prior
This is either flat from −∞ to ∞ or 1/x from 0 to ∞
This prior does not alter the likelihood
The prior can be improper
Non-informative priors + (0,1) loss = MLE

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

How can we use informative priors that aren’t conjugate

A

MCMC

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

MCMC

A

The idea is simple
We create a Markov chain that has the same asymptotic distribution as the posterior
Sampling from the chain is the same as sampling from the posterior
Gives a sample from the posterior
Can we guarantee convergence?

17
Q

Discretising the Prior

A

If we discretise the prior on δ over a limited range we can sample from the posterior directly using Monte Carlo methods (not MCMC)

18
Q

Leave One Out

A

Leave out a point
Predict the point
validate through methods given in lectures