Week 2 Flashcards

1
Q

Unbiasedness

A

This concept says that we will estimate the true population coefficient on average, since our distribution will be normally distributed and centered around the true value, even though we sometimes make over/underestimations.
E(B-hatt)=B.
We say that the estimator is an unbiased.
(Holds under OLS1-4)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Consistency

A

As the sample size increases, we observe decreased variability of the estimates. Becomes more concentrated around the true value –> so for very large samples we can say that B-hatt=B. This is called consistency. B-hat is consistent for B.
(Holds under OLS1-4)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

The power of the test

A

It is the ability to detect false hypotheses by triggering rejections. Detect deviations from the null hypothesis. If the difference between B-hat and B (which is the numerator in the T-statistic) is positive this means a shift of the N(0,1) to the right - IF this movement increases the probability (the gained critical region in the positive side>the loss in the neg region) it means that the overall prob of rejection increases. –> Assuring that more rejections are done if H0 is wrong –> POWER.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

How to get around the linearity assumption when using non-linear functions?

A

We create a new variable for the squared x1, so that x2=x1^2. This will let our regression model to look linear and we can therefore estimate our betas. But we must be carfeul when interpreting the betas, because we cannot use the ceteris paribus - cannot hold x2 still while changing x1 because x2 contains x1 as well.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Interpret:

  • level-level
  • level-log
  • log-level
  • log-log
A
  • Level-level: a 1 unit change in x leads to B change in Y.
  • Level-log: a 1% change in x leads to B/100 change in Y.
  • Log-level: a 1 unit change in x leads to B*100 change in Y.
  • Log-log: a 1% change in x leads to B% change in Y (elasticity!!)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Dummy variable

A

Random variable that takes the value 1 for a subpopulation satisfying a certain property and the value 0 for everyone else. Used to model group membership. The zero group is called the base or benchmark group.

Also common to use with interaction term when the marginal effects are different in the different groups.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Interpret the coefficient for the dummy variable?

A

It is the difference between the intercepts corresponding to the two groups. The curve will shift with this parameter.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Interpret the coefficient for the dummy variable when it is multiplied by another regressor?

A

Then we have a dummy variable that also changes the effect/slope of that regressor. When taking the partial derivative w/r to that regressor, say x1, we will see that the the parameter will give us the difference in the marginal effect of x1.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Modeling complementarities between regressors:

A

It is the same as for dummy variables, we multiply two regressors. The difference here is that both variables are measured on a real scale (dummy only 0 or 1). This will give us a coefficient measuring how they work together: one direct effect of x1 and one indirect effect of x1 that goes through this third combined regressor.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Asymptotic approximation

A

The statements are true only in very large samples.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

3 things that will reduce the var(B-hat1):

A
  • decreasing the variance of U
  • increasing the variance of x1
  • decreasing the correlation between x1 and other regressors.

All of these will help making better predictions of U-hat and therefore reducing the variance in our estimates.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Decision rule in hypothesis testing with a test statistic:

A

If the test statistic in absolute value is larger than the critical value calculated from the level of significance (within the critical region), we reject the null hypothesis and stop believing it’s true. Instead we make a statement about the population that is in line with the alternative hypothesis. This rule is equivalent to finding out that the p-value is smaller than alpha.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Important difference between F-test and two-sided t-test:

A

That the t-test is two sided so you have to take into account that the alpha (probability of type-1 error) must be divided on the two sides, while in the F-test we only consider the right hand side as it is skewed to the right.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Confidence interval for a coefficient:

A

A random set of possible values for B such that the true value is contained in this set with a pre-specified probability. That probability is called the “confidence level”.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is the numerator/denominator degrees of freedom?

A

The numerator is q=the number of tested coefficients. The denominator is n-(k+1).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Problem with p-values:

A

That it is so easy to use that people without statistical knowledge think that they can use them, but they can’t. They are prone to manipulate the test by changing the level of significance afterwards just to reach a statement…But the sample is only random before drawing it, after it is realized and then we can no longer make probabilistic statements. Cannot set up alpha=0.05 but then find p-value =0.09 and say that the test rejects at 10%…

17
Q

Short vs long regression:

A

The long regression regress Y on X1 and X2 while the short regression is one that recover the estimates B for a regression that only regesses on X1. The coefficient in the short regression will combine the direct effect of X1 (as measured by B1 in the long one) plus the indirect effect (as measured by B2 in the long) which will be the scaled ability of X1 to capture var in X2. If no co-movements, no contribution.

18
Q

Specification

A

The variables that we include in the regression, also called the dgp.

19
Q

Why do we transform regressors?

A

Because we think that the true effect of the variable is different than linear, so we can for ex transform into a log or a quadratic term.

20
Q

Interaction term

A

A regressor that is defined as the product of two variables.

21
Q

Does our estimator B-hat following a normal dist?

A

Yes it does, if we transform it like this:
(B-hat - B)/sd(B-hat).
This gives us N(0,1). Mean 0, sd=1.

22
Q

Standard error

A

SE is the estimator of the standard deviation of an OLS estimator.

23
Q

Robust standard error

A

A standard error that will “work” if the unobserved component is heteroscedastic. It includes outliers at a larger extent and therefore the SE is expected to increase in this model.

24
Q

Null hypothesis

A

A claim about our model. We are trying to disprove this claim.

25
Q

Type-I error:

A

Rejecting the null hypothesis even though it is true. The probability of doing this error is equal to a. Ex 5%.

26
Q

Type-II error:

A

Not rejecting the null hypothesis even though it is false.

27
Q

What is the prob of doing type-II error?

A

It depends on a lot of things and there is no clear answer, depends on the way in which the H0 is violated.