Linear Regression Model & OLS Flashcards
what does the error term (u_i) capture
-the effect of other variables
-unpredictable elements in human behaviour
-measurement error
this gives us the linear regression model
population regression line formula
π_π= π+πβπ_π+ π’_π
what is u_i
the error term
OLS assumption 1
the regression function is linear in parameters
Examples β OK, linear
π_π=π½_0+π½_1 π_π^2+π’_π
π_π=π½_0+π½_1 1/π_π +π’_π
Examples β Not OK, non-linear
π_π=π½_0+π_π^(π½_1 )+π’_π
OLS assumption 2
random sampling
OLS assumption 3
zero conditional mean
E(u|X) = 0
Since πΆππ£(π’,π)β 0 implies E(π’|π)β 0, we can think of this assumption as lack of correlation between the error term and the regressor(s)
OLS assumption 4
error term has constant variance (homescedasticity)
V(u|X) = ο³2
i.e. variance of the error term does not depend on X.
the smaller the residualβ¦
the closer the estimated value of Y is to the actual value of Y
π_π=(π_π )Μ+(π’_π )Μ
what does OLS stand for
Ordinary Least Squares
linear regression
TSS = ESS + RSS
TSS - Total sum of squares
ESS - Explained sum of squares
RSS - Residual sum of squares
OLS is BLUE - what does BLUE stand for
Best Linear Unbiased Estimator
Gauss-Markov theorem
when you take all 4 OLS assumptions or properties together you have what is referred to as the Gauss-Markov theorem
theorem states that OLS estimators are BLUE when the OLS assumptions are satisfied
the bigger the difference between the estimate and the actual valueβ¦
the worse the fit of your model
we generate this measure by decomposing the variance of the OLS estimates