Quiz #3 Flashcards
ANCOVA what it does
Variables used
shows the effect of the independent variable after covariates have been removed or accounted for
Uses 2+ categorical independent variable and 1 continuous dependent
ANCOVA assumptions (4)
Absence of multicollinearity = correlation of independent variables to each other
Normal distribution of residuals (errors or how far away line is from your error)
- output in SPSS can be plotted to assess normality
Homogeneity of residual variance
- Levene’s test for homogeneity of variance
Linear relationship between covariates and outcome variable at each level of independent variable
ANCOVA outputs
Least squared mean (LSM) or estimated marginal means = mean output which has been adjusted for covariates
Partial eta = effect size, above .15 —> considered good
Beta coefficient = is slope significantly different from 0 (strength of relationship between variables + covariates)
What are dummy variables and how are they used?
Dummy variables - method of dichotomizing variables, independent continuous or multi-categorical variables which must be split into 2 categories for ANCOVA
Base case is ignored (000), then other groups coded using 0 or 1
Number of dummy variables is groups - 1
Used because regressions can’t have categorical variables with more than 2 groups (ex. 4 categories of physical activity –> med + high with base case low)
Correlation
Types
What to report
measures strength of relationship between 2 continuous variables with a line of best fit through data
Pearson = parametric
Spearman = nonparametric
df = n - 2 (2 parameters), r vaue, p value
Qualities of correlations to measure effect
Direction: positive or inverse
Strength: coefficient of correlation (r) or distance of observed data from line
r =between 0 to ±1, ±1 = perfect correlation, 0 = no correlation
Slope ≠ r value
Strength brackets for correlation
Strength brackets (can depend on model used):
Small strength = 0.1-0.3
Medium strength = >0.3 - 0.5
Strong strength = >0.5 - 1.0
Assumptions for correlations
Variables are continuous (no categorical data)
Variables are approximately normally distributed → otherwise use Spearman’s
Linear relationship between variables
No extreme outliers - they drastically change r value
Association vs causation
correlation/regressions assess relatedness/association and NOT causation due to error and confounding variables
Linear regressions
assesses relationships between 2+ continuous variables