Lecture 31-33: More Biostats Flashcards
What are the 4ish measures of central tendency and dispersion
- MODE/MEDIAN/MEAN
- OUTLIERS
- MINIMUM / MAXIMUM / RANGE
- INTERQUARTILE RANGE (IQR)
- See Slide 20
Define Variance and Standard Deviation
- VARIANCE (from Mean)
- The average of the squared differences in each individual measurement value and the groups’ mean
(There is No way I can type this but I’ll try:)
SUM OF: ((x-x)squared) /n - STANDARD DEVIATION (SD)
- square root of variance value (restores units of mean)
Describe a normally distributed graphical representation of data
- Graphical representation shows SHAPE of data
- NORMALLY DISTRIBUTED = Symmetrical
- When a dataset is normally-distributed the following values (PARAMETERS) are EQUAL/NEAR EQUAL: Mean / Median / Mode (Stats tests useful for normally-distributed data are called “PARAMETRIC” tests)
- Equal dispersion of curve “tails” to both sides of mean, median, & mode
Describe a positively skewed graphical representation
POSITIVELY SKEWED
- Asymmetrical distribution with one “tail” longer than another
- A distribution is skewed anytime the Median Differs From The Mean
- When mean is higher than median, “positive skew”.
- Tail pointing to the right
- Positive skew (skew to right): mean > median
Describe a negatively skewed graphical representation
NEGATIVELY SKEWED
- Asymmetrical distribution with one “tail” longer than another
- A distribution is skewed anytime the Median Differs From the Mean
- When mean is LOWER than MEDIAN, “negative skew”.
- Tail pointing to the left
- negative skew (skew to left): mean < median
Define skewness
- A measure of the asymmetry of a distribution
- The perfectly-normal distribution is symmetric and has a SKEWNESS VALUE OF 0
Define kurtosis
- A measure of the extent to which observations cluster around the mean. For a normal distribution, the value of the kurtosis statistic is 0
- Positive kurtosis – more cluster
- Negative kurtosis - less cluster
What are the required assumptions of interval data for proper selection of a parametric studies
REQUIRED ASSUMPTIONS OF INTERVAL DATA (for proper selection of a PARAMETRIC Test):
- NORMALLY-DISTRIBUTED
- Equal variances
* Multiple tests available to assess for equal variances between groups
- LEVENE’S TEST**
- Kolmogorov-Smirnoff
- Bartlett’s or F-Test - RANDOMLY-DERIVED & INDEPENDENT
What is the protocol for handling data that is Not normally distributed?
- HANDLING INTERVAL DATA NOT NORMALLY-DISTRIBUTED
- Use a statistical test that Does Not Require the data to be normally-distributed (NON-PARAMETRIC TESTS), or
- Transform data to a standardized value (Z-SCORE OR LOG)
- hoping transformation allows data to be normally-distributed
Describe the 2 impacts of statistical significance
- POWER (1-β)
- The ability of a study design to detect a true difference if one truly exists between group-comparisons, and therefore…
- The level of accuracy in correctly accepting/rejecting the Null Hypothesis (analogous to Sensitivity in screenings)
- Sample Size
- The larger the sample size, the greater the likelihood (ability) of detecting a difference if one truly exists
- Increase in Power
What are the 3 statistical elements to consider when determining a sample size?
- MINIMUM DIFFERENCE BETWEEN THE GROUPS DEEMED SIGNIFICANT
- The smaller the difference between groups necessary to be considered “significant” (important), the greater sample size (number; or ‘N’) needed - EXPECTED VARIATION OF MEASUREMENT (KNOWN OR ESTIMATED)
- ALPHA (TYPE 1) & BETA (TYPE 2) ERROR RATES & CONFIDENCE INTERVAL
- Add in anticipated drop-outs or loss to follow-ups ***
Describe Null Hypothesis (H0)
- A research perspective which states there will be No (true) difference between the groups being compared
- Most conservative and commonly utilized
- Various statistical-perspectives can be taken by the researcher:
- Superiority
- Noninferiority
- Equivalency
- Researchers either ACCEPT or REJECT this perspective, based on STATISTICAL ANALYSIS
Describe alternate hypothesis (h1)
A research perspective which states there Will Be a (true) difference between the groups being compared
Describe a Type I Error
Type 1 Error (a.k.a. ; alpha)
- REJECTING the Null Hypothesis when it is actually TRUE, and you Should Have Accepted It!
- There really is no true differences between the groups being compared but you (in error) reject the Null Hypothesis thereby ultimately stating that you believe there is a difference between groups (when there really is NOT!)
- Somewhat analogous to the concept of a FALSE POSITIVE in medical screenings
Describe a Type II Error
Type II Error = Beta Error
- NOT REJECTING the Null Hypothesis when it is actually FALSE, and you Should Have Rejected It!
- There really IS a true difference between the groups being compared but you (in error) do NOT reject the Null Hypothesis thereby ultimately stating that you believe there is no difference between groups (when there really IS!)
- Somewhat analogous to the concept of a FALSE NEGATIVE in medical screenings
- See table on slide 50