L4 - Distribution of Regression Estimates Flashcards
What can we think of the OLS estimator?
- We can think of the OLS estimator as a mapping from the sample moments of the data to the parameters of interest.
- Yi = α+βXi+ui
- Parameters of interest –> α, β,(σu)2
- Sample Moments –> Y(bar), X(bar), (σY)2(hat), (σX)2(hat),σXY(hat)
The mapping is as follows:
- β(hat)=(σXY(hat))/((σX)2(hat))
- α(hat)= Y(bar)- β(hat)X(bar)
- (σu)2= ΣNi=1(Yi - α(hat)-β(hat)Xi)2= (N-1/N-2)*((σY)(hat) -((β(hat))2(σX(hat))2)
What are properties of estimators?
1 - An estimator is said to be unbiased if its expected value is equal to the (unknown) true value. E(β(hat)) = β
- An estimator is said to be efficient if it has the lowest possible variance in the class of estimators under consideration –> Lowest value of d
- In some circumstances there may be a trade-off between bias and efficiency
What is the distribution of the OLS estimator?
We have shown that the OLS estimator for the slope coefficient can be written:
β(hat)= (Σ(Xi-X(bar))(Yi-Y(bar))/(Σ(Xi-X(bar)2)
= (Σ(Xi-X(bar))Yi/(Σ(Xi-X(bar)2) since Y(bar)(Σ(Xi-X(bar)) = 0
substiuting for Y and rearranging yields:
β(hat)=β + (Σ(Xi-X(bar))ui/(Σ(Xi-X(bar)2)
The OLS estimator is therefore a random variable whose distribution depends on the properties of the random variable u.
What are the Gauss-Markov assumptions?
Under a specific set of assumptions the OLS estimates can be shown to be the best linear unbiased estimates (BLUE).
1 . The error has expected value zero - E(u{i}) = 0, ∀i
- The errors are serially uncorrelated - E(u{i}u{j})= 0, ∀i≠j
- The errors have constant variance - E(u{i}^2) = σ{u}^2 0, ∀i
- The X variable is non-stochastic (fixed in repeated samples) - E(X{i}u{i})=X{i}E(u{i})
errors follow a normal distribution
What does it mean by OLS estimates can be shown to be the best linear unbiased estimates (BLUE)?
E - Estimator - α(hat) and β(hat) are estimators of the true values of α and β
L - Linear - α(hat) and β(hat) are linear estimates - i.e.e the formulae gfor α(hat) and β(hat) are linear combinations of the random variables (Y and possibly X)
U - Unbiased - on avaerage, the actual values of α(hat) and β(hat) will be equal to their true values
B - Best - the OLS estimator β(hat) has minimum variance among the class of linear unbiased estimaors: Gauss-Markov theorem proves that the OLS estimator is best by examining an arbitary alternative linear unbiased estimator and showing in all cases that it must have a variance no smaller than the OLS estimator
What can we derive from the Gauss-Markov assumptions?
- Using the Gauss-Markov assumptions we can derive the mean and the variance of the OLS estimator.
- E(β(hat)) = β
- V(β(hat)) = (σu)2/(Σ(Xi -X(bar))2)
- Under the GM assumptions OLS is the Best Linear Unbiased Estimator (BLUE).
- This means that the OLS estimator has the lowest possible variance in the class of linear unbiased estimators.
- Alternatively OLS is the most efficient estimator we can use when these assumptions are satisfied.
What is a Statistical Inference from the OLS estimator?
- If the errors in the regression model follow a normal distribution then the OLS estimator is a linear combination of normally distributed variables.
- Therefore the OLS estimator also follows a normal distribution (β^hat)~N ( β , (σu)2/(Σ(Xi-X(bar))2)
- Statistical inference is the process of using data to make inferences about unknown population parameters
- Examples of statistical inference are hypothesis tests about the parameters or the derivation of intervals within which parameters are likely to lie
What do you need to conduct a hypothesis test?
To conduct a hypothesis test we need the following:
- A hypothesis to be tested (usually described as the null hypothesis) and an alternative against which it can be tested.
- A test statistic whose distribution is known under the null hypothesis.
- A decision rule which tells us when to reject the null hypothesis and when not to reject it.
What is the test statistic for β when performing a hypothesis test?
under the null hypothesis the following random variable is N(0,1):
(β(hat) - β(bar))/ (σu/sqrt((Σ(Xi-X(bar))2)) ~N(0,1)
Z=(x-μ/σ)
If the error variance is known then we can calculate this test statistic and compare it with a critcal value from the normal table to decide if we should reject the null
What is a Type I Error?
- An error-type I is the error of rejecting a true null hypothesis. - Large significance level (large tails) => larger probability of error-type I
What is a Type II Error?
- cant make the tails as small as possible as increase the likelihood of anerror type II: the probability of not rejecting a false null hypothesis
What is the desicion rule in hypothesis testing?
- Involves fixing the size of the test to avoid type I and II errors
- Fix the probability of when we get type I and II
What is a critical value?
- A critical value is a value corresponding to a predetermined p-value. For example a 5% critical value is a value of the test statistic which would yield a p-value of 0.05.
Critical values are often set at 10%, 5% or 1% levels. For example, for the standard normal distribution critical values for a 1-tailed test are:
10% - 1.282
5% - 1.645
1%- 2.326
- If the test statistic exceeds the critical value then the test is said to reject the null hypothesis at that particular critical value.
What is a P-value?
The P value, or calculated probability, is the probability of finding the observed, or more extreme, results when the null hypothesis (H0) of a study question is true – the definition of ‘extreme’ depends on how the hypothesis is being tested. P is also described in terms of rejecting H0 when it is actually true, however, it is not a direct probability of this state.
- It shows the level of statistical significance –> how probably true the event will be