W5: RQ for Predictions 2 Flashcards
What is a full regression equation involving 2 IVs
Yi = a + b1X1i + b2X2i + ei
- b1 and b2
- partial regerssion coefficients
- a
- intercept
What is a model regression equation involving 2 IVs
Y^i = a + b1X1i + b2X2i
What is an intercept in a regression equation. What is it signified by
- a
- Expected score on DV when all IVs = 0
What is a partial regression coefficient in a regression equation. What is it signified by
- b1, b2,…,
- Expected change in DV for each unit change in IV, holding constant scores on all other IV
What is the Sum of Squares in a regression equation. What is SS Total Indicated by
SStotal = SSreg + SSres
- OLS gurantees SSreg will be as large as possible because it ensures SSres is as small as possible!
- Unbiased in this regard
What is the aim of OLS. Use sum of squares to explain
Maximize SSreg; Minimize SSres
What is R^2. Alterative name. Is it an effect size?
-
Coefficient of Determination
- Effect size
- Strength of prediction in a regression analysis
- R^2 = SSreg/SStotal
- Proportion of SStotal accounted for by SSreg
- Proportion of variance in DV that is predictable from IVs
What is the range of R^2
0 to 1.
Closer to 1 = Stronger = Greater proportion of DV explained by regression model (IVs) = SSreg great
To calculate a confidence interval on R, what are the 4 things we require
- Estimated R2 value
- Numerator df on F statistic (no. of IVs)
- Denominator df on F statistic (n-IVs-1)
- Desired confidence level
- Smaller width = Greater precision
Under which conditions will the observed R-squared value be most biased (all other aspects being equal)?
When sample size is small and the number of IVs is large
Typically what is the biasness/consistency of R^2
OLS estimate of R2 is biased (often overestimated), but consistent (if sample size increase, it will get increasingly closer to 95%)
- Note
- OLS estimate of slope is unbiased, but it is bias for R2
How is an adjusted R2 better than R2
It is usually less biased, but we should always report both values
What are the two ways we can make meaningful comparison between IVs in a multiple regression
- Transform regression coefficient to standardised partial regression coefficient
- Z-scores
- Semi-Partial and Squared Semi-Partial Correlations
Making meaningful comparison between IVs in a multiple regression: Method 1. When interpreting coefficients, what is the difference. When is this method useful?
- Coefficients are interpreted in SD units
- Example
- One SD increase in variable X will increase in 0.5 SD decrease in variable Y, holding constant scores on all other IV
- Example
- No (intercept)
- Only useful when IV has an arbitrary scaling
Making meaningful comparison between IVs in a multiple regression: Method 1. What is the intercept
Always 0
Making meaningful comparison between IVs in a multiple regression: Semi-Partial and Squared semi-partial Correlations. Overview.
- Semi-partial correlatons
- Correlation between DV and each focal IV, when effects of other IVs have been removed from that focal IV
- Correlation between observed (not predicted) DV and scores on focal IV that is not accounted for by all other IV in the regression analysis
- Effect size estimate
- Squared semi-partial correlation
- Proportion of variation in observed (not predicted) DV uniquely explained by each IV
- Directly analgous to R2 of the overall model