Week 4 - Lecture 4 Flashcards

1
Q

what is statistical power?

A

The probability that a test will detect an effect if one exists (ie. correctly rejecting a false null hypothesis)
- formula: power = 1 - β (Type II error rate)
- power increases the chances of detecting a true effect

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

what does the null hypothesis (H0) state in a study?

A

That there is no effect or no difference

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

what does the alternative hypothesis (H1) state?

A

That there is an effect or a difference exists

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is a Type I error?

A

Saying there is an effect when there isn’t one (false positive)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the symbol used to represent a Type I error?

A

Alpha (α)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is the common value set for alpha (α) in research?

A

0.05 (or 5% chance of making a Type I error)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Give an example of a Type I error

A

A pregnancy test says someone **is pregnant **when they are **not **

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is a Type II error?

A

Saying there is no effect when there actually is one (false negative)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the symbol for a Type II error?

A

Beta (β)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Give an example of a Type II error

A

A pregancy test says someone is not pregnant when they actually **are **

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What does **power **refer to in significance testing?

A

The ability to correctly detect an effect when it exists (1 - β).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What happens to power when beta increases?

A

power decreases (you’re more likely to miss real effects)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

In terms of errors, why do we not always reduce alpha to 0.01?

A

Because it would increase beta and make us more likely to miss real effects (Type II errors)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What does it mean if power = 0.80?

A

There is an 80% chance of correctly finding a real effect if it exists

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

complete the table:

A

refer to photo

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Does statistical significance always mean the result is practically important?

A

No. A result can be statistically significant but still be** too small to matter in real life**

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

How does a large sample size affect statistical significance?

A

It can make tiny effects appear **statistically significant **

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What can happen with a small sample size?

A

It might miss real effects because there’s not enough data

19
Q

What is an example of a study with a statistically significant but practically unimportant result?

A

Facebook’s 2014 study - effect size was d = 0.02, which is very small.

20
Q

What happens to power if you lower alpha (eg. from 0.05 to 0.01)?

A

Power decreases (but you reduce the chance of a Type I error).

21
Q

What happens to power if you increase alpha (eg. from 0.05 to 0.10)?

A

power increases (but so does the chance of a Type I error)

22
Q

What does “effect size” tell us?

A

How big the difference or relationship is, measured in standard deviation units

23
Q

What are Cohen’s effect size benchmarks for small, medium and large effects?

A
  • small = 0.2
  • medium = 0.5
  • large = 0.8
24
Q

How does sample size affect power?

A

larger sample size gives more power by reducing standard error

25
Q

What is the effect of increasing sample size from N = 25 to N = 100 on significance?

A

It can change a non-significant result (Z=1.5) to a significant one (Z=3)

26
Q

What happens to power when variance in the population is low?

A

power increases because it becomes easier to detect real differences?

27
Q

How can you reduce variance in your sample?

A

By using a homogeneous (similar) group of participants

28
Q

What does effect size (d) measure?

A

The difference between group means, expressed in standard deviation (SD) units

29
Q

What does statistical effect (δ or delta) represent?

A

How many standard errors apart the populations are

30
Q

What’s the formula for δ in a one-sample t-test?

A

refer to photo

31
Q

What’s the formula for δ in a two-sample t-test (equal group sizes)?

A

refer to photo

32
Q

What is the formula to calculate effect size (d)?

A

refer to photo

33
Q

What do we use δ for once it’s calculated?

A

We use it to look up power in a power table, given a chosen α level.

34
Q

If d = 0.333 and N = 25, what is δ and power approximately?

A
  • δ = 1.665
  • Power ≈ 0.36 (not sufficient)
35
Q

If d increases to 0.667, what happens to power (with same N)?

A

δ increases to 3.335 and power ≈ 0.91 → power improves

36
Q

What is the formula to find sample size N for one-sample tests?

A

refer to photo

37
Q

What value of δ is commonly used to achieve 80% power?

A

δ = 2.8 (when α = 0.05)

38
Q

If d = 0.333 and δ = 2.8, what is the required N?

A

refer to photo

39
Q

Why must you always round up when calculating N?

A

You can’t test a fraction of a person, and rounding down gives less power than needed

40
Q

In two-sample tests, what does lowercase n represent?

A

The number of people in each group

41
Q

What is the formula for n in a two-sample t-test?

A

refer to photo

42
Q

If d = 0.5 and δ = 2.8, how many people are needed in each group?

How many total participants are needed in that example?

A

63 per group

Total N = 63 x 2 = 126 participants