Module 4 - Power & Effect Sizes Flashcards

1
Q

Statistical power

A

What is the probability that a study will detect an
effect when there is an effect there to be detected

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Sampling to Populations

A

So when we estimate samples from populations – we accept a probability of making an error
Typically we set this as .05 (or 5%) probability that we will say there is an effect where there isn’t one to be found
We want to be 95% sure that two (or more) samples have been drawn from different populations

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Alpha

A

The probability that we will reject the null hypothesis when we shouldn’t
That we say there is an effect when there isn’t one
This is our Type I Error
<5% (<.05) means a low chance we have made this error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Statistical power

A

The likelihood that a study will detect an effect when there is an effect there to be detected
What is the probability of a correct decision of rejecting the null
hypothesis when it is false
Power = 1 – Probability of a false negative (Type II error)
Power = 1 – β

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Factors Affecting Statistical Power

A

*Alpha level
*Error variance
*Sample size
*Effect size

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

How Sample Size affects Power

A

Sample size works in the same way that error variance works
As we test more people we are able to better describe a distribution
Our hypothetical distributions (based on our samples) gets smaller/narrower

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Effect Size

A

Effect size is the relative distance between our null and true distributions
This distance is measured in standard deviation units
An effect size of 0 (zero) would mean no difference between groups (a “perfect” null result)
Effect size increases as two or more groups become “more” different from each other
This can help tell us if differences are practically meaningful

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Effect Size Measurements

A

Main Effect (ANOVA)
- Eta Squared
- Omega Squared
Multiple Comparisons (Planned contrast or Post-hoc)
- r
- Cohen’s d

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

ETA squared

A
  • Used for main effect
  • Small (.01); Medium (.09); Large (.25)
    n2 = SSbetween/SStotal
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Omega squared

A
  • Used for main effect
  • Most accurate measure of effect size for main effect
  • Small (.01); Medium (.06); Large (.14)
    w^2 = SSb -(df b*MSw)/ SSt+MSw
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Effect Sizes for Planned-Contrasts

A
  • r
  • Used for follow-up tests
  • Particularly useful for planned contrasts
  • Small (.10); Medium (.30); Large (.50)
    r = Square root (t^2 / (t^2+df)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Effect Size for Post-Hoc Tests

A

Cohen’s d
* Used for follow-up tests
* Can be used for Tukey’s post-hoc tests
* Small (.20); Medium (.50); Large (.80)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Post-Hoc Tests - Cohen’s d

A

Step 1. Spooled = Square root of (n1-1)s1^2+(n2-1)s2^2 / n1+n2
Step 2. D = x1-x2/Spooled

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Cohen’s d - Reporting example

A

Post-hoc tests using Tukey HSD revealed that taking no drugs (M = 13.5) led to significantly higher rated Ikea build quality than taking Marijuana (M = 3.20, p < .001, d = 8.25) or LSD (M = 6.75, p < .001, d = 4.75). Furniture ratings after taking LSD were significantly higher than Marijuana (p < .001, d = 2.88). All effects were large.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Calculating required sample size

A

Before running an experiment we want to ensure that if there is an effect present, that we will observe it (Power)
Part of that will rely on testing sufficient numbers of participants (see effect of Sample Size on Power)
We can use effect size and desired Power to estimate how many people we will need to test

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Calculating required sample size in other words…

A

In other words, by making predictions of what our null, and alternative hypothesis distributions will look like (i.e., Mean, SD), and using that to estimate effect size (difference between null and alternative), we can estimate how many people we will need to describe those distributions sufficiently, to meet a certain power specification
You might not need to do this in this course, but you may need this for your honours project
Free software known as G*Power

17
Q

Power and Effect Sizes Wrap-up

A

The probability that we will find an effect if there is one to be found is known as statistical power
The probabilities of making inferential errors are related to observed power
Certain factors affect power (e.g., sample size)
Effect sizes are standardised metrics of difference
For ANOVA we can use Eta Squared, or Omega Squared
For multiple comparisons, we have different effect sizes
We can calculate the sample size we would need to observe a certain effect (size) at a certain statistical power

18
Q

G* Power - Calculating Sample Size

A

We can determine our mean difference using different types of tests (including ANOVA)
If we are looking for a smaller mean difference between groups, we need a larger sample size
If we are looking for a larger mean difference between groups, we need a smaller sample size