HRSS SEM 2 Flashcards
Cross-sectional study design
Observational or descriptive
Collects data from a population at 1 specific point in time
Groups determined by existing differences, not random allocation
Advantages of cross-sectional study design (5)
- Snapshot of a population at one time
- Can draw inferences from existing relationships or differences
- Large numbers of subjects
- Relatively inexpensive
- Can generate odds ratio, absolute/relative risk and prevalence
Disadvantages of cross-sectional study designs
- Results are static: no sequence of events
- Doesn’t randomly sample
- Can’t establish cause and effect relationship
Pearson’s correlation co-efficient
Measures linear relationship between 2 variables
P=0 suggests no linear relationship
Coefficients offer crude linear association and are unable to adjust for other variables
Regression modelling
Investigates if an association exists between variables of interest
Measures strength and direction of association between variables
Studies the form of relationships
How are continuous linear relationships examined
By linear or non-linear regression models
How are categorial outcomes examined
By logistic regression
Describe Linear Regression
Ho = no relationship between DV and IV
IV and DV must be continuous; IV can be continuous or categorical
Assumptions for linear regression
Linear relationship between DV and iV
Observation independently and randomly selected
Effects are additive
Homogeneity of variances
Residuals are independent + normally distributed
Absence of outliers and multi-collinearity
Describe skewness and kurtosis, mean/median/mode for a normally distributed variable
both = 0
- the further the value is from 0 the more likely it is that the variable isn’t normally distributed
mean median and mode should be equal for a normally distributed variable
Tests of normality
Tests of normality (Shapiro-Wilk) compare the shape of the sample distribution to the shape of a normally distributed curve
- non significant tests suggest distribution of sample isn’t sig different from normal distribution
- significant tests suggest distribution in question is sig different from a normal distribution
Multicollinearity
Refers to IV’s that are correlated with other IV’s
In presence of multi-collinearity, regression models may not give valid estimates of individual predictors
Variance inflation factor (VIF)
Measure of how much the variance of the estimated regression coefficient is inflated by the existence of correlation among IV’s in the model
VIF = 1 : no correlation among predictors
VIF >4 : warrants further investigation
VIF > 10: signs of serious multicollinearity
In a simple linear regression model if B > 0 …
Positive association between IV and DV
For each unit increase in IV, the DV would increase by (B) value units.
Building a regression model
If IV associated with outcome (DV) and no affected by multi-collinearity, then can build multivariable multiple linear regression model for DV
Fitted regression model presents regression coefficients representing adjusted associations between DV and IV; adjusted for each other
Interpreting a regression model
For each unit increase in IV, the estimated DV unit would increase by (B) value, after adjusting for other IV’s
May change association
R^2
% of variability explained by fitted model
Observational studies
Subjects observed in natural state; can be measure and tested by no intervention or treatment