L.9 - Bayesian ANOVA chapters Flashcards
1
Q
Predictive quality
A
- how well did model, or parameter value, predict the observed data?
- we use this predictive quality to update our knowledge about the world
> we use updated knowledge to make predictions about tomorrow’s world - see Bayesian learning circle in image 1
2
Q
what are the values of parameters across different models?
A
- θ → binomial test
- δ → t-test
- ρ → correlation
3
Q
Cauchy distribution
- when do we use it?
- why?
A
- difference in means is on a continuous scale without hard bounds
> as opposed to correlation or proportion - we use this distribution to characterize each model’s prediction
4
Q
how can we write a t-test model as a linear regression?
A
y1 = b0 + b1*xi
- do the beer tastings differ meaningfully if the beer is alcoholic or not?
- is b1=0?
5
Q
how can we write down the hypothesis of the Cauchy distribution?
A
H0: b1 = 0
H1: b1 ~ Cauchy(0.707)
- 0.707 is the set scale of Cauchy’s distribution
6
Q
what is the 0.707 in the Cauchy distribution?
A
- the alternative model bets around 50% on values between -0.707 and 0.707
- this number is conventional, we don’t really change it
7
Q
how can we interpret the hypothesis of the Cauchy’s distribution?
A
- H0 goes all-in on 0 being the true value
- H1 spreads its bets across a range of values of b1
> how likely are the data under each hypothesis?
8
Q
priors in ANOVA
A
- in expanding ANOVA (more than one effect or more than 2 groups), we add “b” for each parameter of the model
- each parameter will need a prior distribution, to make concrete what the model is predicting
> average quality of a model’s prediction is its marginal likelihood - see images 2 & 3
9
Q
how can we use JASP for the example above?
A
- repeated measures ANOVA
- between (correctedness) and within (alcohol & non-alcohol)
- see image 4
10
Q
P(M)
A
- P(M): prior model probability
> how likely is each model, before seeing the data?
> usually we divide the odds between models (here 0.25 each)
→ BeerType model: before looking at the data, there is a 25% probability that this model is the true model, out of these four models - see image 5
11
Q
P(M|data)
A
- P(M|data): posterior model probability
> how likely is each model, after seeing the data?
> all values sum to 1
→ BeerType model: after looking at the data, there is a 44.6% probability that this model is the true model, out of all these models - see image 5
12
Q
BFM
A
- posterior model odds
> updating factor from prior to posterior model probability
→ for BeerType model, we calculate it through posterior and prior odds
→ the data are 2.42 times more likely under this model, than under all the other models combined
13
Q
BF10
A
- pairwise Bayes factor
> how likely are the data under this model, compared to another model?
> to compute BF10, we take ratio of posterior model probability of specific model divided by posterior model probability of best model
→ BeerType model: the data are 0.81 times more likely under this model, than under the best model
14
Q
BFM vs BF10
A
- BFM: comparison between single model and all other models combines
- BF10: pairwise comparisons between single models
15
Q
Bayes factor transitivity
A
- we can use BF10 column to conduct additional model comparisons
- e.g. we can compare BeerType model directly to Null model
→ we take those two models’ BF10 values and divide them by each other (beertype BF10 / null BF10)
→ the data are 97296 times more likely under beertype model than under null model