L6 - Interpreting the OLS regression Flashcards
What happens to the t-distribution for the OLS estimator when the sample become larger?
As the sample size becomes larger the t-distribution becomes more like the normal distribution.
What happens to the variance for the OLS estimator when the sample become larger?
As the sample size become very large the variance of the OLS estimator will tend to zero.
V(β(hat)) = (σu2)/(ΣNi=1(Xi-X(bar))2) –> 0 as N –> ∞
as N gets larger the positive dominator (due to its squared nature) gets larger thus tending towards 0
What happens when the sample of the OLS estimator gets too large?
The asymptotic distribution of the OLS estimator is degenerate i.e. the PDF collapses onto a single value as the sample size becomes arbitrarily large
How does increasing the sample size affect the OLS estimator PDF on a graph?
- gets thinner and tall around the mean get closer and closer to a vertical line.
How can we compare estimators on the basis of their asymptotic variance?
Sometimes we wish to compare estimators on the basis of their-asymptotic variance (why?) - to see if the estimator is consistent –> this being that is the number of observations tended to infinity would we get the population estimator, or the difference between the estimate and the actual value is so small that is is basically zero
but we can’t do this if their distributions are degenerate.
- Instead we look at a transformation of the distribution which produces a non-zero finite variance.
sqrt(N)(β(hat)-β) ~N(0, (σu2)/(1/N *ΣNi=1(Xi-X(bar))2)
As N –> ∞
sqrt(N)(β(hat)-β) –> ~N(0, (σu2)/(σX2))
What is the OLS Regression Residuals?
The OLS regression residuals can be defined as the difference between the actual values of Y and the fitted values from the regression equation.
ui(hat)= Yi-α(hat)-β(hat)Xi
- The regression residuals are not the same thing as the equation errors which are unobservable –> as in a general regression formual we don’t know the value of α or β
- The residual sum of squares from the regression modelcan be written as:
RSS= ΣNi=1(ui(hat))2
- Where as the equation errors are the vertical distances from the scatter plot to the population regression function.
- The regression residuals are the vertical distances from the scatter plot to the sample regression function.
What is the 1st important property of the OLS residuals?
the OLS residuals sum to zero:
Σui(hat)= Σ(Yi-α(hat)-β(hat)Xi)
= ΣYi-nα(hat)-β(hat)ΣXi
= N(Y(bar)-α(hat)-β(hat)X(bar)) = 0
by virtue of the property that the OLS regression passes through the sample means of the data. –> Y(bar)-α(hat)-β(hat)X(bar) is what we used to calculate the RSS which is equal to 0
- This is not true of the equation errors. Suppose u~N(0,σu2), it follows that:
W= ΣNi=1(ui)~ N(0,Nσu2)
- as Nσu2 there will be some point that are different from 0
What is the 2nd important property of the OLS residuals?
The OLS residuals are uncorrelated with the X variable. Note:
ui(hat)= (Yi-α(hat)-β(hat)Xi)
sub in for α(hat)
= Yi-Y(bar)+β(hat)X(bar)- β(hat)Xi
= Yi-Y(bar)-β(hat)(Xi-X(bar))
Therefore:
Σ(Xi-X(bar))ui(hat) = Σ(Xi-X(bar))(Yi-Y(bar) -β(hat)Σ(Xi-X(bar))2
Recall that β(hat) =Σ(Xi-X(bar))(Yi-Y(bar)/Σ(Xi-X(bar))2
Therefore:
cov(X,u(hat))= (1/N-1)Σ(Xi-X(bar))ui(hat)= 0
Note cov(X,u) = 0 may or may not be true. This is a matter of assumption rather than a mathematical property of the model.
What is Covariance?
- Covariance is a measure of how changes in one variable are associated with changes in a second variable.
- Specifically, covariance measures the degree to which two variables are linearly associated. However, it is also often used informally as a general measure of how monotonically related two variables are
How can we interpret the Slope Coefficient of the OLS regression model?
- The slope coefficient gives the marginal effect on the endogenous variable of an increase in the exogenous variable
- The interpretation of the slope coefficient depends on the units of measurement.
- in most cases we use constant or real prices as we must take inflation into account to produce the best estimator
Why do we use Log-linear equations for an OLS regression?
The slope coefficient of a log-linear regression gives us the elasticity of y with respect to x.
- so if there was a 1% percent increase in x it will give us the incease/decrease in y as a percentage
Note:
dLn(Yt)/dYt=1/Yt
dLn(Xt)/dXt=1/Xt
dLn(Yt)/dLn(Xt)= (dYt/dXt) * (Yt/Xt) = ηY,X
How do we interpret the intercept of the OLS regression?
- The intercept is chosen so that the regression line passes through the sample means of the data
- It does not make much sense to think of the intercept as the value of Y when X is zero.
- Although this is mathematically true, the zero value of X often lies well outside the range of the data –> cant be explained by econometrics
How do you Predict Error Variance?
- Suppose we wish to predict Y for a given value of X.
Yp = βXp+ up
Yp(hat) = β(hat)Xp
Yp - Yp(hat) = (β -β(hat)) Xp+ up )
If we take the expect value of both sides of the equation, as E(β(hat)) = β and E(Yp - Yp(hat))= 0 –> since the OLS estimates of the intercept and the slope are unbiased and the expected value of the error is zero.
Therefore, the prediction error for Yp is partly due to the random disturbance u and partly due to the inaccurate estimate of β.
V(Yp - Yp(hat)) = (β -β(hat)) Xp+ up)2 = ((σu2Xp2)/(ΣXi2)) + σu2
The prediction error variance is smallest when X equals its mean value (in this case zero). It gets larger the further X is from its mean.
- What this means is that we are more likely to get accurate predictions when the value of the x variable used is a ‘typical’ value in the sense that it is close to the sample mean.
- The more extreme the value of x we use, i.e. the further from the sample mean, then the less reliable will be the prediction. The importance of this effect will vary depending on the nature of the data used
Why is differencing useful?
- Economic variables often contain a trend. Differencing will remove the trend from the data.
- Regressions in levels may give an unrealistic impression of the explanatory power of equations when the data is trended.
- you can still see if there is a relationship between related variables
- Differencing or taking the difference of the logarithms helps because we are effectively regressing the growth rate on one variable on another
- An equation estimated in the first differences of the logarithms of the variables constitutes a relationship in growth rates:
- dLn(yt)/dt = (1/yt) *(dyt/dt)
- gI,t= α + βgY,t + ut(hat)
What is a summary of all the regressions so far?
1: Yi= α(hat)+β(hat)Xi + ui(hat)
Marginal effect of Y on X
2: Ln(Yi)= α(hat)+β(hat)Ln(Xi) + ui(hat)
elasticity effect of Y on X –> direct relationship
- d(Yi)= α(hat)+β(hat)d(Xi) + ui(hat)
marginal effect of Y on X
- dLn(Yi)= α(hat)+β(hat)dLn(Xi) + ui(hat)
growth rate effect of Y on X