Statistical Inference Flashcards

1
Q

What is hypothesis testing?

A

procedures to decide if hypotheses about a population statistic can be accepted

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is a hypothesis?

A

Proposition of fact that will be tested

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the alternative hypothesis?

A

H1 = experimental hypothesis, the thing we would like to be true

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the null hypothesis?

A

H0 = the opposite of the experiemental hypothesis
e.g. there is no benefit of a new treatment compared to the current standard (this does not mean that the new treatment is worse but that they are the same)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Why use null hypotheses?

A

This is ‘safe’ - we won’t change practice unless the data suggests the current practice (null hypothesis) is incorrect and we reject it

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What are the two types of hypotheses?

A

Can be 1 or 2 sided

1 sided - one direction e.g. better or worse
2 sided - different in either direction

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is a parallel group study?

A

A study where 1 group has 1 intervention the other has the other

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is a crossover trial?

A

In this study one group starts with one intervention and then switches to the other intervention (and the other gorup does the same in a different order)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is a p value?

A

The probability, assuming the null hypothesis is true, that the data (test statistic) you see is at least as big as, or larger than, observed

aka

The probability of coming to the wrong conclusion that we are happy to accept
- there is a small random chance that the sample we took doesn’t represent the population

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

When is a p value statistically significant?

A

With a 95% ‘signifience level’, any p
< 0.05 is statistically significant

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the difference between clinical and statistical significance?

A

e.g. progression free survival of 0.42 weeks might be stastisitically significant but isn’t clinically signficiant

whereas a pain score of 85 vs 65 could not be statistically significant but is clincially signficiant

sometimes a clincally significant result with a statistically insignificant result can be restudied with a larger trial

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is a type 1 error?

A

aka
* ‘false positive’ rate
* significance level
* alpha

the study finds a differene due to the random sample of the data and wrongly rejects the null hypothesis when the null hypothesis is true and there actually isn’t a difference
therefore clinical practice might get chnaged nincorrectly

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is a type II error?

A

aka
* false negative
* beta

the study finds no difference and wrongly accepts the null hyothesis when in reality there is a difference
therefore clinical practice doesn’t change when it should

This is due to statistical power - the study needs to be big enough

We never actually know if we have made this error we just try to power the study enough that we don’t make it

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is statistical power?

A

Power = 1-B
set at study design stage

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is the issue caused by type I errors?

A

We are wrong 5% of the time but we don’t know when!!

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How do multiple tests affect Type I errors?

A

Multiple tests (comparisons) increases the risk of a type I error (false positive) - each time there’s a 5% chance of error

17
Q

How can Type I errors be avoided or dealt with?

A
  1. Ignore them
  2. Control the familywise error rate - aim for a 5% error rate across all the tests together
  3. Control the false discovery rate - make sure each individual test has a 5% of error
18
Q

Ignoring type I errors

A

leave other people to apply their own correction criteria on your data
but leaves the risk of Type I error unmodified

19
Q

What is familywise error rate?

A

family means tests in the same sample

this controls the overall type I error rate of all the tests combined

This is a very strong control and means results will be conservaive

20
Q

What are the two corrections that can control the familywise error rate?

A

Bonferri correction - divide alpha by number of tests and use that as new threshold. E.g. if 6 tests 0.05/6
Sidak correction - less conservative

21
Q

What is ‘false discovery’ rate?

A

Make each individual test error rate 5% to control the overall rate of type I errors

so instead of 5% chance of type I error overall a 5% chance of all tests being a tyoe I error is maintained, so overall could be wrong 5% of the time

Benjamin-Hochberg procedure