Lecture 16 Flashcards

1
Q

What is time series

A

You observe one unit repeatedly over time
- e.g. quarterly GDP

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Whats the point of time series?

A

Time series analysis is incredibly useful in real-world applications where things evolve over time
- forecasting: predict future GDP, inflation
- dynamic causal effects: understand not just if something has an affect, but when
- modelling risks

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Technical challenges in time series

A
  1. Not a random sample, same unit so can we generalist?
  2. Observations are correlated over time, if GDP is high today, probably high tomorrow too, so OLS assumption of independent errors violated
  3. Lagged variables matter, past values often help explain current outcomes
  4. Serial correlation in errors, SEs will be wrong unless corrected
  5. Small samples
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Key notation for time series:

A

Yt is the value of some variable
- dataset is a sequence {y1,y2,…,yt} is a single variable observed over time
- first loaf of yt is just the value from the previous time, yt-1
- when we apply the log before differencing, we approximate the percentage change in y

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Autovariance

A

Measures how two values of the same series at different times move together
- e.g. cov(yt,yt-1)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Autocorrelation

A

Normalises the autocovariance, i.e., scales it so its between -1 and 1
- p1 = corr(yt,yt-1) = ((cov(yt,yt-1))/ROOT(var(yt).var(yt-1)))
- tells us how strongly yt is linearly related to its own past
- extends to any lag k, so corr(yt,yt-j) and cov(yt,yt-j)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

How to estimate autocorrelation from data

A

Sample autocorrelation at lag j:
Pj^ = (cov^(yt,yt-j))/(var(yt))
- Whats the calculation for the numerator? The average of product of deviations from the mean, at times t and t-j, across the sample
- note, you’re summing from t = j+1 to T, asa you need both yt and yt-j to exist. Normalizing by the variance of yt in the denominator

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Why does the autocorrelation matter?

A
  • high autocorrelation means past values are informative, might want to include them in models
  • low or no autocorrelation means the series is more random or less predictable
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Definition of autocovariance

A

Measures how much a time series value at time t and a lagged value at time t-j co-vary. Captures the strength and direction of the linear relationship between those two points in time

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Definition of autocorrelation

A

The normalised version of autocovariance, scales it to lie between -1 and 1
- 1 is perfect positive autocorrelation and vice versa, 0 is no autocorrelation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

The least squares assumptions for time series data
- denote Xt = (x1,x2,…,xt)’
- yt = B0 + X’.B1 + X’t-1.B2 + … X’t-p.Bp+1 + ut

A
  • MLR.1 - linear in parameters, yes
  • MLR.3 - no perfect collinearity, yes
  • MLR.4 - zero conditional mean, we can go beyond and say error at time t is uncorrelated with entire history and future of the regressors
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

MLR.2 for time series regression

A

I.i.d no longer make sense
- the replacement is ‘stationarity’, a time series yt is stationary if its probability distribution doesn’t change over time
- mean and variance stays constant, and covariance between lags and yt depends on lag length not time itself.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Now consider joint stationarity

A

Means the joint distribution of a collection of time series variables does not depend on time
- so if you slide your time window forward or back, statistical relationships don’t change
E.g. if you’re tracking the temp at noon every day, if the distribution of temp on days 1-5 is the same as 10–105, then series is jointly stationary

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What does ‘covariance only depends on how far apart’ means

A

Let’s say:
- cov(y100,y99) = 50, so covariance between adjacent days
Then, cov(y78,y77) = 50 also as only the lag matters.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Whats second order stationarity

A
  • only care about mean and variance being constant not the next moments
  • covariance between two still depends only on the time gap
  • much more realistic, we dont care about skewness, Kurtosis or whether the ENTIRE distribution stays the same.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What do we do with the independence assumption

A

We relax and assume weak dependence,
- still correlation across time
- that correlation decays, as the time gap increases.
Assume Ergodicity:
- averages over time behave like averages over a large number of individuals, where the idea is that as you observe the process for long enough, dependence becomes negligible.

17
Q

So finally, what are the properties of OLS in time series (5)

A
  1. TS1 - linear in parameters
  2. TS2 - strict exogeneity
  3. TS3 - no perfect collinearity
  4. TS4 - stationarity
  5. TS5 - weak dependence/ ergodicity
    Under 1-3, OLS is unbiased and under all 6, OLS is consistent and asymptotically normal, due to CLT
18
Q

Issue with strict exogeneity

A

Hardest assumption to meet
- variables evolve over time
- current or future variables may depend on past chocks
- as OLS becomes biased, RHS variables no longer predetermined
- reverse causality or autoregression can violate strict exogeneity

19
Q

Proof assumption of normal exogeneity is not enough in time series

A

We have bias, and assuming Xt^ depend on past shocks, then numerator and denominator are correlated, so expectation does not simplify to 0 and OLS is biased
- via ergodic theorem, which says time averages converge to expectations, so even with bias in small samples, OLS is consistent and asymptotically normal
- without strict exogeneity, OLS is biased due to time dependence, but with stationarity and ergodicity, we can rely on large samples to recover right answer

20
Q

CLT and ergodicity

A

If time series data is ergodic, and stationary, then by a time series version of the CLT, the OLS estimator’s distribution looks more and more like a normal distribution as T tends to infinity.

21
Q

Is OLS BLUE in time series?

A

Even if you have homoskedastic errors, OLS is not BLUE unless, errors are serially uncorrelated, so ut and ut-1 are independent.