L5 - Confidence Intervals Flashcards
How do you conduct a Hypothesis Test?
1-Hypothesis: Decide on your hypothesis. You need a null hypothesis and an alternative hypothesis.
2-Significance: Decide on the level of significance unless there is a good reason not to, people generally use 0.05 as the significance level also known as the alpha value. The significance level is the probability that you will say that the null hypothesis is wrong when really it is correct (TYPE I error).
3-Sample: Take a sample from the population to provide the statistics you need
4-P-value : Calculate the p-value. This is almost always done by a computer package. Based on the determination of suitable test statistics i.e. Z, t, Chi-squared, or F-statistics
5-Decision: Use the p-value to decide whether to reject the null hypothesis. If the p-value is less than significance level you chose earlier, you will reject the null hypothesis. The sample has given you evidence that the null hypothesis is wrong.
What is the P-value?
- Let X be a random variable which can take on any real value between limits a and b.
- Let xc be a real number which lies somewhere in the range a to b.
- The p-value of xc is the probability that a random draw of X is
‘at least as extreme as xc’ under a given null hypothesis.
Depending on the context, this may be:
- P(X>xc), P(Xc) or P(X>xc or X < -xc)
Essentially, we are asking how far out in the tail of the distribution is the number xc –> looking at the shaded area of the tails
- For the case P(X> xc) we can evaluate the p-value as the area under the tail of the distribution to the right of xc, The p-value gives the probability of that the random variable X will be as least as large as the value given under the given null.
- For the case P(Xc) we evaluate the p-value as the area under the left tail of the distribution.
What are Critical Values?
- A critical value is a value corresponding to a predetermined p-value. For example a 5% critical value is a value of the test statistic which would yield a p-value of 0.05 ( also called signficance level –> α
- Critical values are often set at 10%, 5% or 1% levels:
10% 1.282
5% 1.645
1% 2.326
- If the test statistic exceeds the critical value then the test is said to reject the null hypothesis at that particular critical value.
- Critical values are chosen so as to fix the probability of making a Type I error i.e. we choose a critical value which determines aparticular probability of falsely rejecting the null.
- This is known as the size of the test. Note that the smaller we make the probability of a Type I error then the larger is the probability of a Type II error.
Why do we fix the size of a hypothesis test?
- The size of the test fixes the acceptable probability of making a type I error i.e.
P(Reject H0|H0 is true)
- Therefore if we fix the size of the test at 5% and repeat the test 100 time we would expect to make a Type I error in five cases. A test is said to be stricter the harder we make to reject the null hypothesis.
- For example, adopting a 1% significance level results in a stricter test. Similarly, adopting a two-sided alternative implies a stricter test for the same significance level.
What is the Power of the test?
The power of a test is defined as:
Power = 1 - P(Type II error)
- Therefore the stricter is the test (i.e. the smaller is its size) then the lower is its power
- It is hard to attach a specific number to the power of a test but we can often rank different tests in terms of their relative power.
- If a number of different tests are available then we would normally choose the most powerful test.
- If we have a test against a specific alternative then we can put a numerical value on the power of the test.
What is a confidence interval?
- The confidence interval is an interval estimate of a parameter of interest. It indicates an interval within which the unknown parameter is likely to lie.
- Confidence intervals are particularly associated with classical or frequentist statistical theory.
- A confidence interval is normally expressed in percentage terms e.g. ‘we are 95% confident that the population mean lies between the limits L and U’.
The way to interpret a 95% confidence interval is that if we were to repeat a random experiment 100 times then 95 of the interval estimates we calculate would contain the true population parameter.
- this calculation assumes that the estimate is normally distributed. This is a reasonable assumption in a large sample but may not be appropriate in small samples
How can you derive a confidence interval for slope coefficent?
β(hat)~N(β, (σ2u)/(Σ(Xi-X(bar))2)
(β(hat)-β)/ (σu)/sqrt(Σ(Xi-X(bar))2)~N(0,1)
This allows us to derive a condifence interval for a slope at 95% confidence:
p(-1.96 < (β(hat)-β)/ (σu)/sqrt(Σ(Xi-X(bar))2) <1.96) = 0.95
p(β(hat)-1.96 *(σu)/sqrt(Σ(Xi-X(bar))2) < β <β(hat)+1.96 *(σu)/sqrt(Σ(Xi-X(bar))2) = 0.95
How do we perform a hypothesis test without the value of variance?
- using a t-test
- one-sample t-test is used to compare the mean of a population to a specified theoretical mean (μ).
(β(hat)-β)/ (σu)/sqrt(Σ(Xi-X(bar))2)
- we replace σu with an unbiased estimate then we can show that: (β(hat)-β)/ (σu(hat))/sqrt(Σ(Xi-X(bar))2)~ tn-2
- Where (σu(hat))2 = (1/N-2) ΣNi=1(Yi -α(hat) -β(hat)Xi)2
- in a regression model degrees of freedom = n - p - 1 where n is number of events and p is the number of explanatory parameter in the model ( in this case it will probably be N - 2)
- Thus we have determined the small sample distribution of the OLS estimator for the slope coefficient. We can use this to test hypotheses about the unknown population parameter.
What is the standard error of the slope coefficient?
SE(β(hat)= σu(hat)sqrt(1/ΣNi=1(Xi-X(bar))2)
What is the standard error of the intercept?
SE(α(hat)= σu(hat)sqrt((1/N) + (X(bar))2/(ΣNi=1(Xi-X(bar))2
How can we use the standard error test statistic?
H0 : β = β(bar) –> β(hat)- β(bar)/SE(β(hat) ~ tN-2
H0 : α = α(bar) –> α(hat)- α(bar)/SE(α(hat) ~ tN-2