Week 2 Flashcards
Unbiasedness
This concept says that we will estimate the true population coefficient on average, since our distribution will be normally distributed and centered around the true value, even though we sometimes make over/underestimations.
E(B-hatt)=B.
We say that the estimator is an unbiased.
(Holds under OLS1-4)
Consistency
As the sample size increases, we observe decreased variability of the estimates. Becomes more concentrated around the true value –> so for very large samples we can say that B-hatt=B. This is called consistency. B-hat is consistent for B.
(Holds under OLS1-4)
The power of the test
It is the ability to detect false hypotheses by triggering rejections. Detect deviations from the null hypothesis. If the difference between B-hat and B (which is the numerator in the T-statistic) is positive this means a shift of the N(0,1) to the right - IF this movement increases the probability (the gained critical region in the positive side>the loss in the neg region) it means that the overall prob of rejection increases. –> Assuring that more rejections are done if H0 is wrong –> POWER.
How to get around the linearity assumption when using non-linear functions?
We create a new variable for the squared x1, so that x2=x1^2. This will let our regression model to look linear and we can therefore estimate our betas. But we must be carfeul when interpreting the betas, because we cannot use the ceteris paribus - cannot hold x2 still while changing x1 because x2 contains x1 as well.
Interpret:
- level-level
- level-log
- log-level
- log-log
- Level-level: a 1 unit change in x leads to B change in Y.
- Level-log: a 1% change in x leads to B/100 change in Y.
- Log-level: a 1 unit change in x leads to B*100 change in Y.
- Log-log: a 1% change in x leads to B% change in Y (elasticity!!)
Dummy variable
Random variable that takes the value 1 for a subpopulation satisfying a certain property and the value 0 for everyone else. Used to model group membership. The zero group is called the base or benchmark group.
Also common to use with interaction term when the marginal effects are different in the different groups.
Interpret the coefficient for the dummy variable?
It is the difference between the intercepts corresponding to the two groups. The curve will shift with this parameter.
Interpret the coefficient for the dummy variable when it is multiplied by another regressor?
Then we have a dummy variable that also changes the effect/slope of that regressor. When taking the partial derivative w/r to that regressor, say x1, we will see that the the parameter will give us the difference in the marginal effect of x1.
Modeling complementarities between regressors:
It is the same as for dummy variables, we multiply two regressors. The difference here is that both variables are measured on a real scale (dummy only 0 or 1). This will give us a coefficient measuring how they work together: one direct effect of x1 and one indirect effect of x1 that goes through this third combined regressor.
Asymptotic approximation
The statements are true only in very large samples.
3 things that will reduce the var(B-hat1):
- decreasing the variance of U
- increasing the variance of x1
- decreasing the correlation between x1 and other regressors.
All of these will help making better predictions of U-hat and therefore reducing the variance in our estimates.
Decision rule in hypothesis testing with a test statistic:
If the test statistic in absolute value is larger than the critical value calculated from the level of significance (within the critical region), we reject the null hypothesis and stop believing it’s true. Instead we make a statement about the population that is in line with the alternative hypothesis. This rule is equivalent to finding out that the p-value is smaller than alpha.
Important difference between F-test and two-sided t-test:
That the t-test is two sided so you have to take into account that the alpha (probability of type-1 error) must be divided on the two sides, while in the F-test we only consider the right hand side as it is skewed to the right.
Confidence interval for a coefficient:
A random set of possible values for B such that the true value is contained in this set with a pre-specified probability. That probability is called the “confidence level”.
What is the numerator/denominator degrees of freedom?
The numerator is q=the number of tested coefficients. The denominator is n-(k+1).