Chapter 8: Advanced Correlational Strategies Flashcards
Regression Equation
An equation that provides a mathematical description of how variables are related; allows us to predict scores on one variable based on one or more other variables.
Linear Regression
Involves 1 predictor variable.
Multiple Regression
Involves 2+ predictor variables.
Linear Regression Equation
An equation that defines the straight line that best represents the linear relationship between two variables.
y = β₀ + β₁x
y = dependent/criterion/outcome variable x = predictor variable β₀ = regression constant (y-intercept) β₁ = regression coefficient (slope)
b
The unstandardized regression coefficient; you can interpret b as the predicted change in y given a one unit change in x (use the original scales of each variable).
β
The standardized regression coefficient; the predicted change in y (in standard deviations) for a one standard deviation change in x (independent of the original scales of the variables); allows you to compare the size of different slopes.
Residual
The distance of each data point from the regression line
(y - ŷ); residuals represent unexplained error.
“Least Squares Criterion”
Minimize distance between all data points and the line (i.e. minimize “residual error”).
Three Types of Multiple Regression
- Standard (or Simultaneous)
- Stepwise
- Hierarchical
Standard (Simultaneous) Multiple Regression
All of the predictor variables are entered into the regression analysis at the same time. The resulting equation provides a regression constant (β₀ or intercept) and separate regression coefficients for each predictor (β₁, β₂, β₃, …).
Stepwise Multiple Regression
Builds the regression equation by entering predictor variables one at a time based on their ability to predict the outcome variable. Each step looks at unique associations (unique variance “above and beyond” the other predictors).
Hierarchical Multiple Regression
The predictor variables are entered into the equation in an order that is predetermined by the researcher. As each new variable is entered into the equation, the researcher tests whether the new variable significantly predicts unique variance in the criterion variable. Can be used to control for confounding variables, to test interactions with continuous variables (moderation), and to test for mediation.
R
The multiple correlation coefficient; it describes the degree of the relationship between the criterion variable and the set of ALL predictor variables. The larger the value of R, the better job the regression equation does of predicting the criterion variable from the predictor variables.
R²
The proportion of variance in the criterion variable that can be accounted for by the set of all predictor variables.
How do you control for confounding variables using hierarchical multiple regression (steps)?
Step 1: Enter the “confound” or control variable.
Step 2: Enter the predictor you are interested in to test its unique effects, over and above the control variable.