3 - Interpreting the Results Flashcards
check out the error chart
- truth means the truth of the universe
- we don’t know this for sure
what is power (p-value)?
- p = 1-Beta (ie 80%) - lower right in chart
- the probability we have of making the correct decision that there is a difference when in fact a difference occured
what is the p-value?
- p-value means there is a statistical difference, reject null (generally if p <0.05) - NOT whether this difference is significant
- OR the probability of the observed result arising by chance
- people think this means there is an important finding, wrong
- a p value less than 0.05 is good, basically this is the probability that you will incorrectly conclude that the null hypothesis is false (make T2 error - ie B)
- tells you nothing about the probability of replication (reproducability) or magnature of an effect
- do not put too much weight on p-value (highly dependant on small n sizes - bc random sampling error can occur!)
what is the worst type of error?
Type 1 - false positive
what is a type 1 error vs type 2 error
1 - alpha - false positive (reject null even though null is true)
2 - beta - false negatve (fail to reject null even though reject null is true)
what is null-hypothesis significance testing?
- the idea of presenting us with a p-value in the article
what is confirmation bias?
- we see a result just bc we are looking for it
name/describe the 4 types of data, most often seen = *
- nominal: discrete categories w no order, dichotomous and categorical (y/n, dead/alive)
- ordinal: ordered categories w difference btw categories not assumed to be equal, categorical (mild, moderate, severe)
- interval: equal distances btw values and 0 is arbitrary, continuous (IQ)
- *ratio: equal intervals and meaningful q, continuous (height, rom, weight)
- ratio treated the same as interval
- don’t use ordinal much
what does it mean if out of 10 patients, every patient showed improvement but it is still not statistically significant?
- p value is not low enough to indicate a stat. sig. diff.
- just means that MAGNITUDE of change is not S.S. (if based on improvement/no improvement it would be)
* note that the state of reality cannot be changed, only the results of the null hypothesis sig. test can be changed
what should you be wary of when looking at null hypothesis significance testing (NHST)?
- look at effect sizes not p values ofr results (how diff are the groups and what is our confidence that they are different)
- pay attention to alpha and power
- use MCID when possible (ie are the results/differences meaningful to patients/clinicians)
- power can be ower bc of small n-size
what are the 2 important questions you should ask when interpreting the results?
1) what is the single value most likely to represent the truth? (effect size/summary measures)
2) what is the plausible range of values within which our true value may lie? (C.I. - how conifdent we are about summary measure)
- note we almost always find summary measures (best guess at validity)
what are some common summary measures? (3)
1) measures of central tendancy (mean, median, mode)
2) measures of dispersion (SD, SE, variance, range)
3) statistical tests (t-test, anova, ancova, regression, etc)
when do we use mean vs median vs statistical tests?
mean = with normally distributed data
median = not normally distributed, small n size, interquartel ranges
statistical tests - normally distributed data, larger n size
what is the difference btw anova and ancova?
- the only differnce is that ancova can adjust for certain things and anova cant, other than that just comparing btw 2 groups (ie if we want to adjust the score based on how someone is doing in the baseline)
what does a t-test compare?
- it compares btw 2 groups, where they probably used the mean or standard deviation and continuous data
what is a common standardized effect size?
- cohen’s d
- difference btw 2 means/SD
- see yellow (control) vs purple (treatment)
- small = 0.2 SD, 0.5 med, 0.8 large
what are non-standardized effect sizes?
- mean (SD,SE) - for normal dist
- and median (25 and 75th quartile) - for non-normal dist
- t-test, anova, ancova, regression (normal dist)
- mann-whitney (non-normal dist)
What is incidence vs prevalence? - statistical tests? *DI*
indidence: proportin of NEW events (AKA absolute risk) - (# of new events/number exposed) - for prospective studies!
prevalance: the proportion of events (# of events/number exposed) - for retrospective studies!
- tests = chi-square, regression
what is a case control study?
- follow people with event, don’t know incidence, see who was exposed to treatment or control
what are summary measures/effect sizes? (6) *DI*
- absolute risk reduction (ARR)
- number needed to treat (NNT)
- relative risk (RR)
- relative risk reduction (RRR)
- odds ratio (OR)
- survival
what is absolute risk? *DI*
- AKA risk - the event rate in the control group (baseline risk - risk in original group, incidence)
- incidence in group does not tell us anything about comparing btw groups!
what is absolute risk reduction? *DI*
- aka risk differnce
- absolute risk in control group - absolute risk in treatment group
- no effect, ARR = 0
what is number needed to treat NNT? *DI*
number of patients one would need to treat in order to prevent one event
= 1/ARR
- higher means not as effective!
what is relative risk RR? *DI*
- proportion of original risk still remaining after therapy - basically a fraction
- ARtx/ARct
- no indicatoin of what baseline risk was
- less than 1 means treatment more effective, 0.5 means risk of death cut in half (still no understanding for importance bc could be half of 20 or half of 1)
what is relative risk reduction RRR? *DI*
- proportion of original risk removed by therapy
= ARR/ARct = 1-RR
- ratio instead of difference
- recreate relative risk and odds ratio chart *DI*
- which is used for prospective vs retrospective?
- RR for prospective
- OR for retrospective