Econometrics II Flashcards

1
Q

Explain the Classic Linear Model Assumptions and what they mean for estimation

A
  1. Linear in parameters
  2. Zero conditional mean E(Ut|X)=0
  3. No perfect collinearity
    (These three mean the OLS estimators are unbiased)
  4. Homoskedasticity
  5. No serial correlation
    (These five are the Gauss-Markov assumptions. Under this, the estimation is BLUE)
  6. Normality
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Explain the conditions for a time series being covariance stationary (useful for variance etc.)

A

1) Constant mean
2) Constant variance
3) Constant covariance for all lags k

So the key properties of distribution don’t change over time

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Explain the conditions for {et} to be a white noise process

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Derive the mean, variance, covariance, and ACF of an MA(1) model

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Derive the mean, variance, covariance, and ACF of an AR(1) model

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Describe the ACF and PACF of MA(1) and AR(1) processes. How do you decide if different models provide a similarly good fit?

A

We need to use a more parsimonious model. Implement diagnostic tests: either through overfitting = ARMA(p,q+1) or ARMA(p+1,q) and test significance, or by Ljung-Box tests = A Q-test where if rehected, the model building process has to be repeated.

Also test goodness of fit with AIC and BIC which measures the trade-off between fit and parsimony. The lower the values of this, the better.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Explain the forecasting principles, and how forecasts work with ARMA models. How do you determine forecast error?

How do you evaluate which forecast to choose?

A

An optimal forecast should minimize the mean squared error representing the loss function.

Forecasting with ARMA models are provided based on linear projection: AR represented as linear functions of p recent values of Y, MA represented as linear functions of infinite values of Y.

Forecast error is the difference between the realised values and its forecast at time t

To evaluate which forecast to choose, split the sample in two, get the one-step ahead until the end. Then calculate the MSPE, and pick the lowest

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Explain the nonstationary processes

A

1) Random walk. Change in the variable between today and tomorrow is completely random
2) Random walk with drift. We insert a constant (i.e. a drift). Yt consists of two nonstationary components: a deterministic trend component, and Σεi a stochastic trend
3) Trend stationary process. Yt = α0 + α1t + εt

Test for unit roots using the Dickey-Fuller test. Under null, unit root (theta) = 1

  • I hate this section
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Outline SEM and the relevant assumptions

A

Allow some RHS variables to be endogenous. After getting the structural equations, we see the system is interdependent. The completeness of a system requires the number of endogenous variables is equal to the number of equations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Why can you not use OLS with SEM? What do you do instead?

A

After putting into reduced form, the endogenous variables are correlated with the error terms. This means the assumptions of the classical regression model are violated, so the OLS estimator is inconsistent.

Instead, we use an instrumental variable (IV). The only variable uncorrelated with ε is x, so we use this as an instrument for β.

  • Triangular system rubbish can bugger off
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Explain the problem of identification with SEM

A

A structure is unidentified is you have theories that are observationally equivalent. This is when you derive the same reduced form from different structures. This is NOT affected by the size of the sample.

Resultingly, we get “false” structural matrices and disturbances. We get a reduced-form coefficient matrix corresponding to the “false” structure

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Explain how to solve the identification problem with SEM

A

We need non-sample information.

Additional information comes in several forms:
1) Normalisations: In each equation, one variable has a coefficient of 1. With this, there are M(M-1), not M^2 undetermined values in Γ
2) Identities: Variable definitions or equilibrium conditions imply all coefficients are known
3) Exclusions: Omission of variables from equations - places zeroes in B and Γ
4) Linear restrictions: restrictions on structural parameters may also rule out false structures
5) Restrictions on the disturbance covariance matrix: similar to restrictions on the slope parameters

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Explain the rank and order conditions for identifying SEM

A

For there to be a solution, there must be at least as many equations as unknowns.

Order condition for identification of equation j:
- Kj(star)>=Mj
- This condition is necessary but not sufficient. It ensures there is at least one solution, but doesn’t mean there is only one

Rank condition for identification:
- (see equation below)
- Imposes a restriction on a submatrix of the reduced-form coefficient matrix. It ensures there is exzctly one solution for the structural paramters, given the reduced-form

A rule-of-thumb for checking the rank and order conditions is to check whether every equation has its own predetermined variable. There are three cases:
1) Underidentified: Kj(star) < Mj or rank condition fails
2) Exactly identified: Kj(star)=Mj rank condition is met
3) Overidentified: Kj(star)>Mj rank condition is met

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Explain two fundamental methods of estimating SEM

A

1) Single equation methods: the limited information estimators are constructed for each equation individually:
- 2SLS. Endogenous variable regressed on exogenous in first stage. In second stage, system equations esimated by OLS. Consistently estimates parameters, so this is almost universally used
- LIML. A least-variance estimator. Disadvantages: based on normal distribution, computationally more expensive. Advantages: Invariant to normalization
- Less sensitive to specification error

2) System methods: the full information estimators are used for all equations simultaneously:
- 3SLS. Residuals used to estimate the cross-equation error-covariance matrix. Constructs generalized least squares estimator (GLS)
- FIML. Assumes errors are normall distributed
- Asymptotically more efficient with no specification error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

How would you analyse stability properties of a VAR(1) model?

Xt= A0 + A1Xt-1 + et

A

VAR in reduced form can be written using the lag polynomial:
(I-A1L)Xt = A0 + et
So we analyse stability based on the lag polynomial (I-A1L). We can get the characteristic equation in our bivariate system: det(I-A1z)=0
Stable when |z1|>1 and |z2|>1

It has a VMA representation:
xt = μ +ΣA1et

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How would you estimate a VAR model?

A

Cannot consistently estimate using OLS, unless in the reduced form, as it contains only predetermined values. With lots of variables, risk overparameterisation.

17
Q

How would you identify a VAR reduced form

A

We cannot recover the structural form from reduced without restrictions. One approach is Choleski decomposition. This is a restriction on the structural form (i.e. b21=0). Put into matrix, and reduce as normal with this value. Notably has an implication for reduced form errors.

18
Q

Explain how we can determine impulse response functions from VAR models

A

et is reduced-form shocks. To trace out the effects of pure structural shocks, we need to decompose et (see graph) from the VMA representation.

The coefficients of Φ are the impulse-response functions. They are impact multipliers.

The total cumulated effect is the cumulated effect when n goes to infinity. Since (usually) the series are stationary, impulse-response coefficients will converge to zero.

Note how this affects e

19
Q

Explain the types of restrictions on VAR models

A

1) Coefficient restriction. These are short-run restrictions. The most common is zero restriction, or setting a coefficient to unity
2) Variance restriction. Not often used and difficult to justify. The most common case is var(εi)=1
3) Symmetry restriction. Popular in open-economy models: shocks have equal effects across countries. For example b21 = b12
4) Long-run restrictions. Importantly, the Blanchard-Quah decomposition. wtf is this :(

20
Q

Consider the VAR in structural form.

Put it in matrix form, then reduced form, and write it as a function of the true structural shocks

A
21
Q

Describe the Box-Jenkins method

A

1) Identification. Derive variance etc. and finding ACF/PACF)
2) Estimation. Estimate parameters usually by maximum likelihood. Select best model using AIC/BIC for goodness of fit. Check ACF/PACF residuals to see if they are white noise using portmanteau test (Ljung-Box)
3) Application. Use for forecasting

22
Q

Write the following SEM model in reduced form

A
23
Q

Show the VAR reduced form errors are white noise

A
24
Q

Show how this AR(1) model can be represented as MA(∞)

A
25
Q

Explain the concept of a strong stationary process

A

A process is strong stationary if its joint probability distribution does not change over time. The statistical properties between values at different times are the same regardless

26
Q

Show how this MA(1) model can be represented as AR(∞)

A