The Simple Linear Regression Model Flashcards
What are the SLRM and the error term?
A regression model that estimates the relationship between one independent variable and one dependent variable using a straight line.
y = β0 + β1x + u
The variable u is called the error term, random disturbance
(”disturbs an otherwise stable relationship”).
Represents unobserved factors other than x that affect y.
What is the Conditional mean assumption?
E [u | x] = 0
It states that unobserved factors do not contain information on explanatory variables.
What is the population regression function (PFR)?
The conditional mean independence assumption implies that:
E [y | x] = β0 + β1x
This means that the average value of the dependent variable can be expressed as a linear function of the explanatory variable.
What are the orthogonality conditions?
E [u | x] = 0
⇒
E [xu] = 0
E [u] = 0
⇔
E [x (y − β0 − β1x)] = 0
E [y − β0 − β1x] = 0
When independent variables are orthogonal, they are uncorrelated, which is beneficial.
These conditions enable us to estimate the coefficients.
What is the method of the moments approach?
The idea behind the Method of Moments (MoM) estimation is that: to find a good estimator, we should have the true and sample moments match as best we can. That is, I should choose the parameter β such that the first true moment E[X] is equal to the first sample moment ¯x.
The equations given by the orthogonality conditions might be used to obtain good estimators for the parameters β0 and β1:
(1/n).[Σxi(ˆy −ˆ β0 −ˆβ1xi)] = 0
and
(1/n).[Σ(ˆy − ˆβ0 − ˆβ1xi)] = 0
(numerically the same as OLS)
What are the OLS estimators?
Using the basic properties of the summation operator we have the ordinary least squares (OLS) estimates of β0 and β1:
1. ˆβ0= ¯y − ˆβ1.¯x
and
2. ˆ β1 = [(1/n).(Σ(xi - ¯x) (yi − ¯y)] / [(1/n).(Σ(xi - ¯x)^2] or, the sample covariance between x and y divided by the sample variance of x.
These parameters are chosen to minimize the sum of squared residuals.
How do we minimize the SSR?
we choose ˆβ0 and ˆβ1 to make the sum of squared
residuals minimal:
SSR(ˆβ0, ˆβ1) = Σ(u^2) = Σ(y − ˆβ0 − ˆβ1xi)^2
The FOC (First-order condition) for the minimum is found by setting the partial derivatives of the previous equation equal to zero:
(δ/δˆβ0) . SSR(ˆβ0, ˆβ1) = −2.Σ(yi − ˆβ0 − ˆβ1xi)
(δ/δˆβ1) . SSR(ˆβ0, ˆβ1) = −2.Σ(yi − ˆβ0 − ˆβ1xi).xi
What are the so-called normal equations?
The FOC equations when solved equal to zero, provide the so-called normal equations:
- n.ˆβ0 + ˆβ1.Σxi = Σyi
- ˆβ0.Σxi + ˆβ1.Σ(xi^2) = Σyi.xi
which ultimately gives us:
- ˆβ0= ¯y − ˆβ1.¯x
- ˆ β1 = [(1/n).(Σ(xi - ¯x) (yi − ¯y)] / [(1/n).(Σ(xi - ¯x)^2]
What are the algebraic properties of the OLS?
- Σ(ˆui^2) = 0
- Σ(xi.ˆui^2) = 0
- ¯y = ˆβ0 + ˆβ1.¯x
What are the SST, SSE, and SSR?
The sum of squares total, the sum of squares regression, and the sum of squares error.
STT = Σ(yi - ¯y)^2
SSE = Σ(ˆyi - ¯y)^2
SSR = Σ(ˆu^2)
Therefore, it is straightforward to show that:
SST = SSE + SSR
What is the R-square?
R^2 = SSE/SST = 1 − SSR/SST
It is the ratio of the explained variation compared to the total variation;
It is interpreted as the fraction of the sample variation in y that is explained by x.
A value of R^2 that is nearly equal to zero indicates a poor fit of the OLS line: very little of the variation in the yi is captured by the variation in the ˆyi .
It can be shown that R2 is equal to the square of the sample
the correlation coefficient between yi and ˆyi.