Multiple Regression Flashcards
What does our model y hat equation look like when we add an additional predictor?
y hat = b0 + b1X1 + b2X2
We have our intercept (b0)
Then our first regression coefficient going along with the first predictor
Then our second regression coefficient going along with the second predictor
Residual (+e) is added for observed ys
Interpret the univariate information of each of the variables, from a correlation matrix: n, mean, sd, median, trimmed, mad, min, max, range, skew, kurtosis, and se.
n = group sizes
mean = relatively similar; if it’s not, it’s measured on some other range (i.e., phyhealth mean is 3.6 and stress mean is 172).
median = identify that data are symmetrically distributed - match it up with mean.
trimmed mean = takes off outer quartiles, and extreme scores to take out skewness and outliers.
MEAN, MEDIAN, TRIMMED MEAN = should be similar
mad (median absolute deviation) = compares absolute value from from the median. It’s another measure of dispersion. SD and MAD should be similar to one another.
min and max = just to check the coding error
skew = right around ZERO (between -3 and 3)
kurtosis = around ZERO (between -3 and 3)
se = standard error of the mean distribution (to be used for standard error for a t-test).
In the assumption of Multicollinearity, predictors cannot be correlated… so what does it mean if an rcorr function produces a correlation table with an .80 correlation between 2 predictors (mental health and physical health)?
Values above .80 is a concern for multicollinearity.
It makes it difficult to pinpoint the true predictor due to the shared variability.
What is the first thing we do when looking at a regression model?
We look at the correlations with our predictors and the outcome variable - if the predictors and outcome variables are NOT correlated, they’re not good predictors to add to the regression model.
What are the steps for a regression write-up?
We first report our correlations in a table, then indicate that the correlations are significant, direction, hypothesized direction…
How does our interpretation for a multiple regression change from SLR?
b0 = Our intercept is the point at which the regression plane intersects the Y-axis; or expected value of Y when both X1 and X2 are = 0.
ex) The expected # of doctor visits when ind. reports NO mental AND physical health problems.
b1 = The change in the EXPECTED value of Y associated with a 1 unit increase in X1, OVER AND ABOVE THE EFFECT OF X2 (or when X2 = 0).
ex) Holding mental health problems constant…
b2 = The change in the EXPECTED value of Y associated with a 1 unit increase in X2, OVER AND ABOVE THE EFFECT OF X1 (or when X1 = 0).
ex) Holding physical health problems constant…
What does the introduction to statistical control mean for our interpretation?
What is the technical term of this statistical control?
“Over and above the effect of…”
We are holding 1 variable constant while watching how the other one changes.
It is officially called ‘Partial Regression Coefficient’
Conceptually, with the b1 coefficient equation, what are we trying to end up with?
And what are the steps towards getting rid of the variable?
What are we left with?
How is this process similar to Factorial ANOVA?
We just want the stuff associated between X1 and Y, over and above the effect of X2 - so we incorporate statistical control and take the variability of all of X2 and multiply it by the shared portion, then we subtract what’s shared between X1 and X2, multiplied what’s shared between X and Y.
What’s left is the what’s shared between X1 and Y without considering X2.
This is similar to main effect of FANOVA. The effect of X1 on Y, ignoring the levels of X2.
Conceptually, with the b2 coefficient equation, what are we trying to end up with?
It’s exactly the same as b1 coefficient, but instead of X2, we plug in X1.
We just want to effect of X2 and Y, over and above the effect of X1.
What would a negative b2 coefficient indicate if X1 is physical health problems and X2 is mental health problems
It indicates the shared variability between x1 and x2.
What is the b0 equation for MR?
b0 = Y-bar - b1(X1) - b1(X2)
*THE SIGN MATTERS. IF THE B1 OR B2 COEFFICIENT IS NEGATIVE, THIS EQUATION WILL CHANGE BETWEEN PLUS AND MINUS.
What is the equation for Full Model for MR?
ŷ = b0 +b1(X1) +b2(X2)
*THE SIGN MATTERS. IF THE B1 OR B2 COEFFICIENT IS NEGATIVE, THIS EQUATION WILL CHANGE TO MINUS.
We ran a regression on R… last week in SLR, physical health problems was a GREAT predictor. Now that we have added mental health problems to the model, both predictors are non-significant. What does this mean?
What could you do if this happens?
It isn’t good to keep both the predictors in the model. The addition of the mental health predictor did not improve our model fit.
We could either choose one of the variables OR because they’re redundant, make them into 1 predictor as “general health problems”.
What is the point of adding multiple predictors in a MR?
We are trying to see if adding predictors improves our model fit in regards to predicting our Y (or doctor visits).
Conceptually, what are we comparing in a t-test?
The estimate divided by the standard error of the b.
What is the residual standard error in MR?
What does it tell us?
Residual standard error = Standard error or the estimate
The standard error of the estimate tells us the variability of the standard deviation of the points around the regression plane.
Conceptually, it is the measureof MISFIT. Or variability of what’s left unpredicted from our model.
What is the df for multiple regression for 2 predictors and 10 individuals?
df = n - k - 1
= 10 - 2 - 1
What does the Multiple R² tell us?
How is it computed?
Explain conceptually, each component of what’s used to compute the Multiple R².
It’s an effect size that tells us that taken together, X1 and X2 accounts for % of variability in Y.
ex) Taken together, physical health problems and mental health problems accounts for 43% of variability in doctor visits.
Multiple R-squared is computed by SSreg/SStot.
*SSregression = ŷ - y-bar
>This tells me the IMPROVEMENT we make in predicting Y by adding the predictors
*SStotal = y - y-bar
>This is the amount of Y we can predict with the mean of y
So essentially, we are seeing the accountability of JUST the improvement of the predictors over JUST the mean…
What’s the degrees of freedom for the regression model?
*think about the model fit anova table
k, or the number of predictors
What type of problem do we face with the Multiple r-squared when we add predictors?
k goes up - by adding more predictors, we are essentially inflating the model r-squared, so the Adjusted r-squared incorporates df to penalize us for crappy predictors, thereby, lowering the effect size.
When would the adjusted r-square go up?
The adjusted r-squared will only go up if the predictors we include are worth their weight in df.
How should the multiple and adjusted r-square compare to each other?
Having similar multiple r-squared and adjusted will signify that our predictors are a good measure of variability. If a predictor doesn’t influence the model greatly, we should expect the adjusted predicted value to be similar to the predictor value.
What happens if there is a large gap between the multiple r-squared and adjusted r-squared?
We would ethically report the adjusted r-squared.
What is the statistical sentence and interpretation for MR?
F (k, n-k-1) = F-value, P ≤ .05
Taken together, physical and mental health problems are not good predictors for doctor visits.
What is the idea regression model?
Orthogonality between the predictors - that there is NO correlation or overlap between predictors.