Module 4 - Power & Effect Sizes Flashcards
Statistical power
What is the probability that a study will detect an
effect when there is an effect there to be detected
Sampling to Populations
So when we estimate samples from populations – we accept a probability of making an error
Typically we set this as .05 (or 5%) probability that we will say there is an effect where there isn’t one to be found
We want to be 95% sure that two (or more) samples have been drawn from different populations
Alpha
The probability that we will reject the null hypothesis when we shouldn’t
That we say there is an effect when there isn’t one
This is our Type I Error
<5% (<.05) means a low chance we have made this error
Statistical power
The likelihood that a study will detect an effect when there is an effect there to be detected
What is the probability of a correct decision of rejecting the null
hypothesis when it is false
Power = 1 – Probability of a false negative (Type II error)
Power = 1 – β
Factors Affecting Statistical Power
*Alpha level
*Error variance
*Sample size
*Effect size
How Sample Size affects Power
Sample size works in the same way that error variance works
As we test more people we are able to better describe a distribution
Our hypothetical distributions (based on our samples) gets smaller/narrower
Effect Size
Effect size is the relative distance between our null and true distributions
This distance is measured in standard deviation units
An effect size of 0 (zero) would mean no difference between groups (a “perfect” null result)
Effect size increases as two or more groups become “more” different from each other
This can help tell us if differences are practically meaningful
Effect Size Measurements
Main Effect (ANOVA)
- Eta Squared
- Omega Squared
Multiple Comparisons (Planned contrast or Post-hoc)
- r
- Cohen’s d
ETA squared
- Used for main effect
- Small (.01); Medium (.09); Large (.25)
n2 = SSbetween/SStotal
Omega squared
- Used for main effect
- Most accurate measure of effect size for main effect
- Small (.01); Medium (.06); Large (.14)
w^2 = SSb -(df b*MSw)/ SSt+MSw
Effect Sizes for Planned-Contrasts
- r
- Used for follow-up tests
- Particularly useful for planned contrasts
- Small (.10); Medium (.30); Large (.50)
r = Square root (t^2 / (t^2+df)
Effect Size for Post-Hoc Tests
Cohen’s d
* Used for follow-up tests
* Can be used for Tukey’s post-hoc tests
* Small (.20); Medium (.50); Large (.80)
Post-Hoc Tests - Cohen’s d
Step 1. Spooled = Square root of (n1-1)s1^2+(n2-1)s2^2 / n1+n2
Step 2. D = x1-x2/Spooled
Cohen’s d - Reporting example
Post-hoc tests using Tukey HSD revealed that taking no drugs (M = 13.5) led to significantly higher rated Ikea build quality than taking Marijuana (M = 3.20, p < .001, d = 8.25) or LSD (M = 6.75, p < .001, d = 4.75). Furniture ratings after taking LSD were significantly higher than Marijuana (p < .001, d = 2.88). All effects were large.
Calculating required sample size
Before running an experiment we want to ensure that if there is an effect present, that we will observe it (Power)
Part of that will rely on testing sufficient numbers of participants (see effect of Sample Size on Power)
We can use effect size and desired Power to estimate how many people we will need to test