Financial time series - compelte Flashcards

1
Q

why use returns vs prices regarding time series

A

Returns are complete and have a scale-free characteristic. Returns also have favorable statistical characteristicas

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Define simple gross return

A

1+R_t = P_t / P_{t-1}

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Define simple net return

A

R_t = P_t / P_{t-1}

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Define a time series in the context of asset returns

A

A time series is a collection of random variables over time, where the random variables are the asset returns.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

how do we denote a time sereis

A

{r_t}

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

define autocorrelation

A

correlation between past values of the same random variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q
A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

what is the foundation of time series analysis and why

A

Stationarity.

Important to be able to make predictions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

elaborate on stationarity

A

Strict: The joint distribution of a subset of the time series is invariant under time shifts.

Weak: The mean of r_t is time invariant and the covariance between r_{t} and r_{t-l} is also time invariant and depends only on l. meaning the covariance of lag-l is constant.

Weak stationarity should show constant fluctuation arounda fixed level.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

when is weak stationarity equal to strict?

A

If the random variables are normally distributed, and weakly distriubted, then it is also strict.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

what do we call “cov(r_t, r_{t-l})”?

A

lag-l autocovariance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

two important properteis from lag-l autocovariance

A

1) cov(r_t, r_t) = var(r_t) = gamma_0

2) gamma_k = gamma_(-k)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

how do we check for stationarity, broadly speaking

A

divide tiem series into subsections. Check for consistency among each period

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

when is autocorrelation between two variables 0?

A

if they are normally distirubted, they have correlation 0 if they are indpendent.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

important regarding samples of time series

A

we need to use the sample counterpart formulas

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

how do we denote lag-l autocorrelation

A

p_l

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

important result regarding variance in time sries

A

variance is constant, so variance of various r_k is the same:

var(r_k) = var(r_l) etc

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

why cant we use this on a sample time series?

A

Becasue it requires that we know the true values of covariances. These are not observable, so we need to estimate them. And estimation is subject to error, so with small samples there is a differnece between those formulas. However, in the limit they approach the true value.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

what can we say about the properties of the sample autocorrelation function for some lag?

A

if {r_t} is iid sequence, then the estimate is asymptotically normal with mean zero and variance 1/T.

This is the base we use to test for autocorrelation differnet to 0, by using a regular t-ratio

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

what do we get from Bartletts formula?

A

gives the variance of sample lag-l autocorrelation, given that r_t satisfy the linear time series requirement (r_t = my + ∑constant_i a_{t-i}), where a_t is a time series of its own with zero mean.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

how do we test for individual lag-l autocorrelations

A

Some assumptions first: We need to use an estimator of the lag-l correlation because this is not a directly observable size. The estimator we use is the sample correlation function. A great property of this estimator is that it is asymptotically normal with mean 0 and variance 1/T if the returns are IID.

This allows us to use a t-ratio test. Subtract the mean (0) and divide on standardized variance or whatever and we can use Bartletts formula for this.

We use t-ratio, with Bartlets fromula for variance to get the standard deviation. Since t-ratio, the null hyp is that the lag-l autocorreletion is 0.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

what do we need to know regarding the sampel lag-l autocorrelation function+

A

it is biased wiht order 1/T. Substantial in smaller samples, but is negligible with larger samples.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

more effective way of testing autocorrelation functions

A

Portmanteau test.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

elaborate on Portmanteau test

A

Asymptotically chi squared with m degrees of freedom. Null hyp is that all m autocorrelations are 0.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

better version of Portmanteau test

A

Ljung box

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

what do we actualyl get from ljung box and portmanteau testing

A

Indication to whether there is “some” lag with non-zero autocorrelation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

what is ACF

A

autocorrelation function, we take AC of lags from 0 to M, and plot.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
28
Q

elaborate on white noise

A

white noise is a time series, iid, finite mean and finite variance.

We typically work with gaussian white noise, whcih simply means that the random variables are normally distriubted.

Important regarding white noise: ACF is zero entirely. This means that it is common to test whether something is white noise to see if there are meaningful patterns in the data or not.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
29
Q

define a linear time seires

A

IMPORTANT: a_t is a white noise series.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
30
Q

a_t

A

shock, innovation, new information

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
31
Q

does the variance of a time series depend on the weights?

A

yes obvioisuly. if the time series react violently to shocks, then it is more volatile and the variance should be larger.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
32
Q

what can we say about the varaince of a linear time series

A

The variance is given as:

var(r_t) = sigma_a^2 ∑w_i^2

The sum goes from 0 to infinity. As a result, if the time series is weakly stationariy, this is a convergent property that makes sure that hte variance if finite and bounded. This is a result of how the squared weights approach 0 as the iteration number grow.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
33
Q

what can we say about the autocorrelation function (ACF) of a weakly stationariy linear time series

A

Since the squared weights approach 0 as the iteration number grow larger, the covariance is smaller and smaller for larger lags. As a result, the corrleation will be smaller and smaller as well. As a result, the ACF is expected to diminish. Converge to 0.

Reasonable, but very important result.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
34
Q

important regarding the white noise series variance

35
Q

give AR model

A

autoregressive:

r_t = ø_0 + ø_1 r_{t-1} + a_t

IMP: the shock is that of the same time as the varialbe we are predicting.

36
Q

what is the conditional mean of the simpel AR(1) model

A

E[r_t | r_t-1, r_t-2, ….] = ø_0 + ø_1r_{t-1}

37
Q

what is the conditional variance of simple AR(1) model (conditional on r_t-1, …)

A

var(r_t | r_t-1, r_t-1,….) =

38
Q

what is the property we get from the conditiona lmean and variance of the AR model?

A

Markov property. Conditional on the return r_t-1, the return r_t is not correlated with r_{t-i}, for i>1.

we only care about the order of the model.

39
Q

how do we find the unconditional mean of the AR(1) model

A

E(r_t) = ø_0 + ø_1 E(r_{t-1})

now, if the model is stationary, the mean (true mean) is constant. AS a result, the expectaiton we take on both sides are equal, which allow us to simplify:

mu = ø_0 + ø_1 mu

mu (1 - ø_1) = ø_0

mu = ø_0 / (1 - ø_1)

Thus, we have arrived at an expression for the unconditional mean of a simple AR(1) model. The crucial part is stationarity. if we do not have it, we are not able to make this nice closed form result.

40
Q

how can we prove that AR(1) is a true linear (valid) time series model+

A

We use the result from the unconditional mean, and by repeated substitutions we can show that the model can be written as an infinite sequence of weights and shock variables from the past.

Specifically for this case, the weights are powered to the sum iteration number, so they converge rapidly to 0.

41
Q

what is the stationarity condition for AR(p) models

A

We find the characteristic equation. Its roots are not characteristic roots, but their inverses (the inverses of the roots) are the characteristic roots.

In order to have stationarity, all characteristic roots must be less than 1.
alternatively, all the regular roots must be greater than 1.

42
Q

2 ways to determine orders of AR models

A

1) PACF
2) Information criteria

43
Q

what is PACF

A

Added contribution of only the direct relationship between a specific lag and the variable of interest. This means, we do not want all of the intermediary effects.

We want to isolate for instance how the returns on Friday and correlated with the returns on Monday without accounting for ripple effects.

The way we can do this for AR models, is to create consecutive models of icnrementally increasing orders. For a true AR(p) process, lags greater than p should have zero influence, and thus zero PACF. We therefore want to identify the cutoff point.

44
Q

what can we say about PACF is regards to a correlation coefficeint

A

PACF is more than just a correlation. It is not limited to [-1, 1]. Therefore, it represent more of a direct effect relationship. If it is 2.5, we interpret this as saying that each unit increase in the specific lagged variable has a direct influence of 2.5 on the variable we are estimating. At the same time, the overall effect on the main variable as a result of increasing the lagged variable by a unit, might not be exactly 2.5, because of the regular correlation as well.

45
Q

elaobrate on using informaiton criteria

A

Select one or several criteria. AIC, etc.

then compute the value for the criteria for each lagged variable we test for. So, we get one value per added order.

We select the order that gave the lowest information criteria value.

46
Q

how do we typically estimate hte paramters of a AR(p) model

A

conditional least squares.

47
Q

what is important regarding the fitted AR(p) model

A

it doesnt include the white noise error term. We need to compute this based on the difference between fitted and observed r_t values.

WE call the residuals the “residual series”

48
Q

how do we do goodness of fit of some model on AR(p) models?

A

R^2 = 1 - RSS/TSS

Requires stationarity

49
Q

elaborate on forecasting

A

We use a model to forecast, which means that we are not actually inserting the noise terms.

The noise is the error. We can measure the expected error, but generally we care about hte forecast itself.

Forecasting is done using the conditional expectation. We can take unconditional as well, but this is much more involved.

Using the conditional expectation, we can generalize it to multi-step forecasting, where we use the recursive structure to build the sequence of forecasts. Meaning, to forecast 3 periods into the future, we must first forecast 1 period, then use this result to forecast 2 periods, then use this new result to forecast 3 periods.

From analyzing the variance of forecasting, we know the variance of the forecasts will increase with the number of steps into the future (variance increase with the forecasting horizon). This is also common sense.

50
Q

general form MA(q) model

51
Q

what is required for stationarity?

A

MA models are always weakly stationary because of how they are finite linear combinations of white noise series.

52
Q

characteristics of ACF of MA model

A

cuts of sharply at order

53
Q

how to determine order of MA model

A

Because it cuts off hard at some ACF point, we use the ACF to determine order.

Likewise, for AR models, since PACF cuts off sharp at some order, we use PACF to determine order of the AR model.

54
Q

Moving onto volatility

55
Q

what do we mean by volatility and why?

A

Conditional standard deviation of the underlying asset returns.

This is important, because it highlights a crucial point we make: conditional. In this context, conditional does not mean exactly rhe same as in regular statistics. It refers to the fact that we are assuming that there is a possibility that volatility depends on its earlier, previous values.

56
Q

what is special about volatility

A

not observable

57
Q

what is the result of volatility not being observable in the context of volatility models

A

It is difficult to assess their performance

58
Q

what is IV

A

implied volatility refers specifically to the volatility we get from reversing the BS equation

59
Q

what can we say about IV vs volatility frm GARCH models?

A

IV tends to be higher

60
Q

There are some characteristics that are curcial to understadning volatility. elaborateo n them

A

1) Volatility tends to cluster

2) Volatility moves in a continuous way, meaning that there are no jumps.

3) Volatility does not diverge to infinity. It stays within a range. Mean reverting. typically this means that volatility is close to stationariy.

4) Leverage effect: volatility react more to negative returns than to positive returns.

These 4 characteristics are empirical results mainly. The models we consider will be strong on some of them, and weak on others.

61
Q

what is the “idea” of volatility study in regards to correlaton?

A

There is very little autocorrelation, but volatility as a time series is still a dependent series.

What we can see, is that when plotting ACF of log returns, ACF of absolute log returns, ACF of squared log returns, we see that volatility is not serially correlated. however, the squared and absolute ACF show relation. But this is not correlation in the typical sense. It is correlation of the sign of the volatility. This means that volatility is not autocorrelated in the typical way, but is clearly dependent on earlier values.

These dependencies is the key issue of volatility models.

62
Q

what is the mean equation and what is the volatility equation?

A

mean equation is a model for the mean mu.

Volatility equation is mostl ikely built on top of the mean equation, and gives the way volatility evolve.

Both are conditional equations.

63
Q

discuss model building of volatility models

A

We must remove linear dependence from the model. we do this by building a model for the conditional mean.

Use residuals from the mean equation to test for ARCH effects.

Specify a volatility model if ARCH effects are present (statistically significant).

Check model and refine etc.

64
Q

elaborate on removing linear dependenceis from a model

A

“Removing dependencies” means that we are identifying the dependency, and accounting for it. The aim is to produce a model that create residuals that behave like white noise. If there is a predictable pattern in the residual series, then we know that there is a relationship we have not accounted for.

For most asset return series, the serial correlations are weak, if any.

Therefore, building a mean equation amounts to removing the mean (sample mean) from the data, if the sample mean is statistically significant differnet from 0.

The mean equaiton would look like this:

r_t = mu + a_t

If there are no other dependencies, the idea is that by assuming that the best way to predict r_t is by the simple mean constant, the a_t residual series (resulting from fitting a model and obtaining hte residuals), can be proved to be white noise.

In some cases, we might need more than this to remove linear dependencies.

In terms of the volatility equation, since this builds on the mean equation (volatility of the returns etc) we will only be able to accurately predict volatility to the highest level of precision if the mean equation accounts for all the dependencies.

Note on why we say “remove dependencies”: I’d assume that if we take r_t, and subtract the mean and whatever relationships we have accounted for, we “remove dependencies from the return series”. The resulting variable is the a_t residual. So therefore, we can look at the process as trying to reduce a_t to “nothing”.

65
Q

elaborate on the thought process of volatility models

A

The conditional volatility of a model is given as Var(model | all prior info). Given all prior info, this quickly becomes Var(a_t | all prior info), as the other terms become constants. Therefore, we are taking the variance of the residuals. The question is, is there a pattern here? Can we predict sigma_t^2 more precisely than the unconditional variance?

And to answer this, we must first try to reduce the residuals down to as close to noise as possible. This is because if we dont, the residuals reflect certain relationships. While possibly important, if we study for volatility, we dont give a fuck about them. Therefore, we want to remove them from consideration. This actually gives us an incentive to create a very good model for the asset return series in itself. If we are able to capture more actual relationships between various variables, our volatility model will also become better. Therefore, if we build a great asset return series model, we also have a great foundation for a volatility model.
Typically, when working with volatility, we say that we want to remove linear dependencies. In the simple case where there are no dependencies, we simply remove the mean, and we have the white noise series as our residuals.

The residuals are white noise, which means that their variance is finite over time, and they have ZERO autocorrelation for all lags. It is important to understand that white noise does NOT have constant variance. We want to see if the variance follow a specific pattern over time that depends on previous values. Volatility is typically close to stationary when looking at long time frames, but the dependent patterns are also present. These patterns are called ARCH effects, and represent what we want to capture with the volatility models

66
Q

very briefly, how do we test for ARCH effects?

A

Acquire a mean equation, and produce their residuals

Then square the residual series. The squared residual series is the basis we use to test for ARCH effects.

67
Q

another word for ARCH effects

A

conditional heteroskedasticity effects

68
Q

what tests are available for us to test for ARCH effects?

A

1) Ljung box

2) Lagrange multiplier test

69
Q

elaborate on the lagrange multiplier test for ARCH effects

A

equivalent to the usual F test used in linear regression. But we use a_t squared series.

70
Q

elaborate on the idea of ARCH models

A

ARCH is built on the idea that there is no serial autocorrelation between the shocks of an asset return, but dependent, and that this dependence can be described by a simple quadratic relationship.

71
Q

what are ARCH effects?

A

ARCH effects refer to the fact that we can use past squared residuals to predict the next squared residual BETTER than the simple mean. IF we can do this, we say that there are ARCH effects present.

We test for ARCH effects by lagrange method that use an F-distribution to test for a number of parameters (correspondng to squared “lags” of the residulas”) and checks whether some of hte coefficients are non-zero.

IF some are non-zero, there are arch effects (statistically significant non-zero, not actually purely non-zero).

72
Q

what is the general form of ARCH(q) model=

73
Q

negatives with ARCH models

A

not accounting for leverage effects.

limited in general.

likely to overpredict.

74
Q

does the GARCH model have ARCH parameters?

A

Yes. GARCH is basically the same as ARCH, but with another summation term that includes lagged variables of the volatility as well.

We refer to these components as ARCH and GARCH parameters.

75
Q

what can we say about the fourth moment of ARCH and GARCH models?

A

Fatter tails than normal distribution. This is nice because this is more in line with outliers, which we know appear empirically.

76
Q

what do we mean by this? What defines the residuals?

A

We obtain residuals from the mean equation.

However, if the process is a true garch(1,1)-m process, the residuals will follow the structure provided by the equation with the epsilon.

77
Q

what is special about garch m models?

A

allow returns to be dependent on the volatility.

78
Q

what is a multi variate time series model?

A

We are no longer in the world of volatility.

we consider for instance logged asset returns for a number of differnet assets. we put them together into a vector, at treat is as a multivariate model.

79
Q

what is VAR

A

Vector AutoRegressive models.

in vector-matrix notation it looks exactly like the regular AR model. but now, we have vectors and matrices.

80
Q

This is the VAR(1) model. Elaborate on what it actually entails and represent

A

most significant is the matrix.

The matrix holds the contribution that the other return series has on each other, conditionally on their own values.

So, if we have VAR(1), and we have 2 asset return series, then this matrix is a 2x2 matrix.
For r_{1t}, which represent the return series 1 at time t, the matrix has the row “a_11 r_{1, t-1} + a_12 r_{2, t-1}” and for r_{2,t}:
“a_21 r_{1, t-1} + a_22 r_{2, t-1}”.

the order of the terms is determiend by the matrix, so there is no “sense” here.

The important part is the coefficients that actually establish whether there is a relationship between some asset return at time t-1 and the other asset return at time t.

81
Q

difference between concurrent and dynamic relationships in multivariate models?

A

Concurrent refers to relationships at the same tiem period.

Dynamic refers to relationships at different time periods.

82
Q

how do we find concurrent relationships?

A

We use the covariance metrix, ∑, of the residuals a_t.

If there is a non zero eleemnt at some place other than the main diagonal, we know that there is a concurrent relationship.