Linear Regression Flashcards
Linear Regression
Examine linear relationship between independent/predictor variables and a continuous dependent/outcome variable
Simple Linear Regression
1 independent/predictor variable
Multiple Linear Regression
> 1 independent/predictor variable
Fitted Line (best fit) Equation
Y= b0 + b1X
b0 = intercept
b1 = slope
Slope
The average amount of change in the outcome variable for each one-unit increase in the predictor variable
-Characterizes the relationship or marginal effect
Residual
Relationship between each individual observation and the trend line can be measured by the vertical distance between each data point and the trend line
Residual is also known as what?
Estimated Error (u)
Can the residual be negative and positive values?
YES.
Positive = above the trend line
Below = under the trend line
What is the Regression Equation mean?
Sum of the fitted line and the error term
Sum of Squares Regression
How far the predicted values on the fitted line differ from the overall mean
SSB between on ANOVA
Sum of Squares Residual
Difference between the original data and the predicted values on the fitted line
SSW within on ANOVA
Residual and Regression Sum of Squares demonstrates what?
Variance
For a Regression calculation what must be done first?
ANOVA and F-Stat
-confirm or deny significance prior to proceeding with the rest of the regression calculations
ANOVA and F-Stat state whether or not the regression model is significant, but what does it NOT tell us?
- Positive or Negative Relationship
- Extent of change in Dependent Variable based on Independent Variable
b1 the SLOPE is what?
UNSTANDARDIZED estimate of slop coefficient
The t-stat in a regression where the p-value is less than alpha you would what?
REJECT null hypothesis that b=0
The 95% CI for a regression cannot include what if you want to reject the null hypothesis?
ZERO
b1 the Slope tells you the direction of the relationship (+/-), and extent of change in dependent variable based on independent variable, but it does NOT tell us what?
- Proportion of variance in dependent variable explained by independent variable
Coefficient of Determination
Proportion of the variance in the dependent variable explained by the independent variable
R^2 is the Coefficient of Determination (square of Pearson coefficient), and that R^2 value is interpreted how?
R^2 = variance in outcome explained by the exposure
1-R^2 = what?
Proportion of variance in outcome NOT explained by exposure
R^2 is bounded between 0 and 1 what do the values imply?
1 = excellent model fit
0 = no model fit
Adjusted R^2
- Always lower than R^2
- Adjusts for inherent increase in R^2 that occurs every time we add an independent variable to regression equation
- Preferable when comparing models and multiple linear regression
Multiple Linear Regression
Examine linear relationships between two or more independent variables and a continuous dependent variable