Cross Sectional Flashcards

1
Q

What are two requirements for linear regression

A

1) Additive errors 2) One multiplicative parameter per term

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What are some important requirements for dummy variables?

A

They cannot depend on Y You cannot include intercept and all dummies. Would give perfect collinearity (Xs are not linearly independent)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Derive the OLS beta estimator

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the formula for R^2 centered?

A

RSS/ TSS = 1 - SSE/TSS

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is TSS, RSS and SSE

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

When to use centered and uncentered R^2?

A

When model does not contains constant, use uncentered

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What are the large-sample asymptotic assumptions of regression?

A

1) Stationary ergodicity 2) Full rank (E[Xi’Xi]) 3) Martingale difference sequence (error times regressor) 4) Moment existence. Regressors have finite 4th moments and errors have finite variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is estimator bias?

A

Difference between expected value of estimator and true value of parameter

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is estimator consistency?

A

And estimator is consistent if its probability limit is the true parameter

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the problem with ommited variables?

A

If the estimated excludes relevant variables that are correlated with the included variables, the parameter estimates will be biased. Also, parameter variance will not be estimated consistently. It is safe to exclude dummy variables since they are orthogonal to regressors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What happens if you include extraneous (non-relevant variables)

A

You parameter estimates will still be unbiased and variance will still be consistent. But you will have more variance (potentially too low t-stat)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What are the problems of working with heteroscedastic data?

A

Parameters will be unbiased, but variance estimator will be inconsistent. One solution is to use White’s robust variance estimator. Using White’s estimator on homoscedastic data will however give worse finite sample properties and increases likelihood of size distortions. Another solution to heteroscedastcity is to use GLS

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

How can use test for heteroscedasticity?

A

Use White’s test. Estimate your model. Then regress squared errors on squares and cross products of all regressors (including constants). Null hypothesis: all parameters (except for intercept) are zero. Test statistic n*R^2. It is chi^2 squared distributed with df = number of regressors in auxiliary regression (excluding intercept)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What happens when errors are correlated with regressors?

A

If measuring regressors with noise. Regressors and errors may be correlated. This gives downwards bias. Endogeneity will also give bias.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is a type 1 error

A

Rejecting a true null

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is a type 2 error?

A

Failure to reject the null when the alternative is true

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What is the size of a test?

A

Pr(Type 1 error) = alpha

18
Q

What is the power of a test?

A

Pr(1- type 2 error). Probability of rejecting a null when the opposite is true

19
Q

What is a linear equiality hypothesis?

A

R*beta - r = 0 Where R is an mxk matrix r is mx1 vector m i number of restrictions k is number of regressors

20
Q

What are the 3 types of hypothesis tests?

A

Wald test Lagrange Multiplier test Likelihood Ratio test

21
Q

What are the differences in power between the 3 types of hypothesis tests?

A

W is about the same as LR which is larger than LM in terms of test statistics. Since they follow the same distribution, larger test statistics gives more power

22
Q

How to implement a Wald test

A

Run the unrestricted regression and estimate parameters and covariances. Compute test statistics using null hypothesis. Compare against chi-squared distribution

23
Q

How to implement a Lagrange Multiplier test

A

That is, the Lagrange Multiplier Test examines the size of the “shadow price” of the constraint regarding we are trying to test in the OLS optimization framework. Estimate the restricted model and compute its errors. Calculate the score using errors from the restricted model and the regressors from the unrestricted model Calculate the average score and the varianc of the score Compute the test statistic

24
Q

How to implement a Likelihood ratio test

A

The Likelihood Ratio test is based on testing whether the difference of the scores, evaluated at the restricted and unrestricted parameters, is large Estimate the unrestricted model and its parameter covariance Compute the test-statistic using the restrictions Compare against CVs from Chi^2_m distribution

25
Q

What are the main model selection techniques?

A

General to specific (GTS) Specific to general (STG) Information Criteria Cross-valiadation

26
Q

Explain general to specific

A

Start with the largest possible number of possible models If at least one regressor is insignificant, remove the regressors with the lowest t-stat. Restimate the model. Continue until all regressors are significant

27
Q

Explain specific to general

A

Start with the variable with the lowest p-value. At new variables sequentially as long as most recently added variable is significant

28
Q

Discuss pros and cons of GTS and STG

A

GTS has a positive probability of including irrelevant variables, it will however never exclude relevant variables. In STG, the variance is NOT consistently estimated in the beginning, which can lead to wrong inference. For both, we have the problem that t-stats do not follow standard distributions when used sequentially.

29
Q

Explain Information Criteria

A

If possible, search through all models and pick the one where the information criteria is lowest. Typical information criteria are AIC and BIC. If not feasible, run GTS or STG type search through models. Information criteria are of the form: - log-likelihood + penalty term for including regressors. In OLS use ln(error variance) instead of ll.

30
Q

Discuss pros and cons of different information criteria

A

The AIC asymptotically may select a model with irrelevant regressors

31
Q

Explain k-fold cross validation

A

Divide the data into k group. For each group, estimate beta excluding the data from the group. Calculate SSE error using the beta estimate on the data in the group. Sum SSE for all group. Select the model with lowest total SSE.

32
Q

What are some important specification tests?

A

Chow test. For stability / structural breaks

RESET test: for misspecified model

Rollig/recursive parameters

Residual plots

33
Q

What is a chow test?

A

it is a test for parameter stability. Regress y on xs and x*indicator(s) Null is that indicator-parameter is zero. Test using Wald, LR or LM

34
Q

What is a RESET test?

A

Test for a misspecified model. Estimate the model. Calculate y_hat Run regression of y on original regressors, plut squared y_hat, cubed y_hat etc. Null: parameters for y_hat are zero. If rejected, some non-linear model may be better

35
Q

How should outliers in regression by defined?

A

In terms of errors. Could upper an lower quantiles (e.g. 0.025). Defining errors in terms y would lead to downwards bias

36
Q

Explain two methods of deadling with outliers?

A

Trimming. Removing outliers observations Winzorisations. Setting observations above threshold to threshold (in terms of errors) After this, rerun the regression

37
Q

Explain residual bootstrap and when to use it

A

Estimate beta_hat and errors in full sample Resample errors and xs independetly. Compute y_tilde from resample x and errors and beta_hat Estiamte beta_tilde Repeat k times (e.g. 500) Take variance of beta-tildes Can only be used if data is homoscedastic

38
Q

Explain non-parametric bootstrap

A

Resample x and y pairwise. Estimate beta_tilde. Repeat k times (e.g. 500). Take variance of beta_tildes

39
Q

What is the test statistic in a Wald test?

A

Distributed Chi^2 m, where m is number of restrictions

40
Q

What are some key factor that affect the power of a test?

A

Sample size

Distance between null and true value

Estimator precision