Quantitative Methods Flashcards

You may prefer our related Brainscape-certified flashcards:
1
Q

What is the sample correlation coefficient (r) for 2 variables?

A

r = Cov(X,Y) / SD X * SD Y

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the t-test formula?

A

t = (r * √(n-2)) / (√1 – r^2)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the b formula?

A

b = Cov(X,Y) / Var X

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

How to interpret t-test value?

A

If calculated test statistic has higher absolute value than critical value, the value is significant.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What are the six assumptions of the classic normal linear regression model?

A

1) Linear relation exists between dependent and independent variables
2) Independent variable is not random
3) Expected value of error term in 0
4) Variance of error term is the same for all observations (homoskedasticity)
5) Error term is uncorrelated across observations
6) Error term is normally distributed

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What does the standard error of estimate (SEE) measure?

A

How well the regression model fits the data. If SEE is small, the model fits well.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is the formula for SEE?

A

SEE = ( (Unexplained variation / (n -2 ) )^0.5

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the coefficient of determination and what is the formula

A

r^2 = (Total variation - Unexplained variation) / Total variation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the formula for F-statistic?

A

F-statistic = Regression MSS / Residual MSS

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the formula for sample variance of dependent variable?

A

Total variation / (n-1)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

How to calculate the interval confidence?

A

Coefficient +/- α * Standard Error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is the formula for the f-test?

A

f-test = (RSS/k) / (SEE / (n-(k+1))) or MSR/MSE

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

In a multiple linear regression model, what is the t-test and how can we interpret its result?

A

t = b - 0 / Standard Error

The lower the p-value, the more significant the result

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What are the six assumptions of the classical normal multiple linear regression model?

A

1) Linear relation exists between dependent and independent variables
2) Independent variables are not random. No exact linear relation exists between 2 or more independent variables.
3) Expected value of error term in 0
4) Variance of error term is the same for all observations (homoskedasticity)
5) Error term is uncorrelated across observations
6) Error term is normally distributed

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

When predicting the dependent variable using a linear regression model, what are the two types of uncertainty we encounter?

A

1) Uncertainty in the regression model itself (SEE)

2) Uncertainty about the estimates of the regression coefficients

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is the formula for adjusted R^2

A

Adjusted R^2 = 1 - ((n-1)/(n-k-1)) (1 - R^2)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What is conditional heteroskedasticity?

A

1) Variance of the errors differs across observations: error term is correlated with the values of the independent variables
2) F-Test is unreliable
3) SEE are underestimated and t-stats are inflated
4) If ignored, we tend to find significant relationships when none actually exists

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

How to correct for heteroskedasticity?

A

1) Computing robust standard errors

2) Generalized least squares (modifies original equation)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

What is a simple test for conditional heteroskedasticity?

A

Breusch-Pagan test

20
Q

What is serial correlation?

A

Positive serial correlation typically inflates the t-statistics of estimated regression coefficients as well as the F-statistic for the overall significance of the regression.

21
Q

What is the test for serial correlation?

A

Durbin-Watson test
If DW > 2 : Negatively correlated
If DW = 2 : Not correlated
If DW < 2 : Positively correlated

IF DW > du, we conclude no evidence of positive serial correlation for the error term

22
Q

How to correct for serial correlation?

A

1) Adjust the coefficient standard errors

2) Modify the equation

23
Q

What is multicollinearity?

A

2 or more independent variables are highly (but not perfectly) correlated with each other.
Classic symptoms are: High R^2, t-test not significant

24
Q

What is a probit model?

A

Estimates the probability of a discrete outcome given the values of the independent variables used to explain that outcome. It is based on the normal distribution.

25
Q

What is a logit model?

A

Estimates the probability of a discrete outcome given the values of the independent variables used to explain that outcome. It is based on the logistic distribution.

26
Q

How to correct for multicollinearity?

A

Remove one or more independent variables.

27
Q

Time series that tend to grow by a constant amount should be modeled by…

A

Linear trend

28
Q

Time series that tend to grow at a constant rate should be modeled by

A

Log linear trend

29
Q

A time series is covariance stationary if:

A

1) Expected value of the time series must be constant and finite in all periods
2) Variance of the time series must be constant and finite in all periods
3) Covariance of the time series with itself for a fixed number of periods in the past or future must be constant and finite in all periods.

30
Q

How to calculate the mean-reverting level?

A

Intercept / (1 - Coefficient)

31
Q

What is the purpose of root mean squared error (RMSE)?

A

Compare the forecast accuracy of different time-series models ; smaller RMSE implies greater forecast accuracy.

32
Q

What is a random walk?

A

Time-series in which the value of the series in one period is the value of the series in the previous period + unpredictable random error. A random walk is NOT covariance stationary. All random walks have unit roots.
Intercept near 0. Slope near 1.

33
Q

What is ARCH and ARCH(1) errors?

A

Autoregressive conditional heteroskedasticity.
ARCH(1) errors: coefficient on the squared residual is statistically significant
C1 is significantly different form 0
Unit root: both independent and dependent variables need to be tested.

34
Q

What is the conclusion if the Engle-Granger test rejects the null hypothesis that the error term has a unit root?

A

Covariance stationary ; time series are cointegrated; estimates of intercept and slope are valid.

35
Q

What is a first-differenced random walk?

A

Special case of AR(1) model with b0 = 0 and b1 = 0

It is covariance stationary.

36
Q

What are the assumptions for using a decision tree approach?

A

Discrete, Independent, Sequential

37
Q

What are the assumptions for using a scenario analysis approach?

A

Discrete, Correlated, Concurrent

38
Q

What are the assumptions for using a simulation approach?

A

Continuous, Independent or Correlated, Sequential or Concurrent

39
Q

What is the most efficient choice to focus on when defining the probability distribution?

A

Directly estimating the statistical parameters of the variables.

40
Q

If ARCH exists, how should we model the variables?

A

IF ARCH exists, it means that there is a heteroskedasticity. We must correct using:

1) Generalized least squares model
2) Computing robust standard errors

41
Q

How can we interpret the Durbin-Watson stats?

A

If DW > upper critical value: fail to reject null hypothesis of no positive serial correlation
If DW < lower critical value: reject null hypothesis
If DW is in between: test is inconclusive

42
Q

What is the purpose of adjusted R2?

A

Adjusted R2 adjusts for the loss of degrees of freedom when additional independent variables are added to a regression. It does not adjust for the effects of serial correlation in the data, nor does it adjust for heteroskedasticity.

43
Q

How is the presence of heteroskedasticity indicated?

A

Systematic relationship between the residuals and the independent variable

44
Q

How can we identify negative serial correlation?

A

Significantly large values of DW stats.

If DW > 4 - Dl

45
Q

How can we interpret the unit root test?

A

If unit root test is smaller than critical value: model does not exhibit a unit root.

46
Q

How can we interpret heteroskedasticity test?

A

If heteroskedasticity test stat is smaller than critical value: model does not have heteroskedasticity.

47
Q

How can we calculate the prediction interval for a regression?

A

1) Prediction interval = Intercept + Slope

2) Prediction interval = Predicted value ± (critical t-statistic × standard error of forecast)