Quantitative Methods Flashcards

You may prefer our related Brainscape-certified flashcards:
1
Q

Sum of Squared Errors (SSE)

A

Measures the unexplained variation in the dependent variable. Is the sum of the squared vertical differences between the actual values and the predicted values on the regression line.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Breusch-Pagan (BP) Chi-Square Test

A

=n*residualR^2 with k degrees of freedom

If larger than the one-tailed critical value for a chi-square distribution, you should reject the null and conclude heteroskedasticity is present.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Effect of Serial Correlation

A
  • Positive serial correlation typically results in coefficient standard errors that are too small. This will cause the t-statistic to be too high and the null to be rejected too often.
  • F-test will be unreliable because MSE will be underestimated
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

F-Statistic

A

=MSR/MSE

If F-statistic > F-value, then you reject the null hypothesis, indicating at least one coefficient is statistically significant.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Unconditional Heteroskedasticity

A

Occurs when the heteroskedasticity is not related to the level of independent variables. It usually causes no major problems with regression.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Conditional Heteroskedasticity

A

Exists if the variance of the residual term increases as the value of the independent variable increases. Creates significant problems for regression and statistical inference.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Adjusted R^2

A

=1 - { [ (n-1) / (n-k-1) ] x ( 1-R^2) }

Will always be equal to or less than R^2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Root Mean Squared Error (RMSE)

A

Used to compare the accuracy of autoregressive models in forecasting out-of-sample values. The model with lower RMSE for the out-of-sample data will have lower forecast error and will be expected to have better predictive power in the future.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Correcting Multicollinearity

A

Omit one or more of the correlated independent variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Random Walk with a Drift

A

The time series is expected to increase or decrease by a constant amount each period (b0).

Xt = b0 + b1*Xt-1 +εt

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Correcting Heteroskedasticity

A
  • Calculate robust standard errors used to recalculate the t-statistics using the original regression coefficient if evidence of heteroskedasticity.
  • Using generalized least squares, which attempts to eliminate heteroskedasticity, by modifying the original equation.

(If both heteroskedasticity and serial correlation, use the Hansen method)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Detecting Multicollinearty

A

Fail to reject the null in a t-test, but fail to reject the null in an F-test while R^2 is high.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Multicollinearty

A

The condition when. two or more independent variables are highly correlated with each other. This distorts the standard error of the estimate and the coefficient standard errors, leading to unreliable t-tests. Slope coefficients will be unreliable as well.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Serial Correlation

A

The situation in which the residual terms are correlated with one another. Positive serial correlation exists when a positive regression error in one period increases the probability of observing a positive regression error in the next period. Negative serial correlation exists when a negative regression error in one period increases the probability of observing a negative regression error in the next period.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Durbin-Watson Statistic

A

Detects the presence of serial correlation.

DW > 2 – Positive serial correlation
DW < 2 – Negative serial correlation
DW = 2 – No serial correlation
DW < dL – Reject null; positive serial correlation
DW > dU – Fail to reject null; no evidence of SC
dL < DW < dU – Test is inconclusive

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Confidence interval for Regression Coefficient

A

=b1 ± ( t x Standard Error)

t is from the t-table
Standard Error is = (coefficient / t-statistic) using n-k-1 degrees of freedom

17
Q

t-statistic

A

= ( Estimated Regression Coefficient - Hypothesized Value) / Standard Error of Coefficient
with n-k-1 degrees of freedom

Two-tailed when hypothesized value is = or ≠
One-tailed when hypothesized value is > or <

If t-statistic > t-value, then the coefficient is statistically significant and you reject the null hypothesis.

18
Q

Mean Square Error (MSE)

A

= SSE / (n-k-1)

19
Q

Regression Sum of Squares (RSS)

A

Measures the variation in the dependent variable that is explained by the independent variable. Is the sum of squared differences between predicted values and the mean.

20
Q

First Differencing

A

Used to transform a time-series with a random walk to a covariance stationary time series. Involves subtracting the value of the time series in the immediately preceding period from the current value of the time series to define a new dependent variable.

21
Q

p-value

A

The smallest level of significance for which the null hypothesis can be rejected. If the p-value is less than the significance level, the null hypothesis can be rejected. If the p-value is greater than the significance level the null hypothesis cannot be rejected (fail to reject).

22
Q

Coefficient of Determination (R^2)

A

The percentage of variation in the dependent variable that is collectively explained by all of the independent variables. R^2 will increase with more variables. √R^2 is the correlation.

= (SST - SSE) / SST

= RSS / SST

=Correlation Coefficient^2

23
Q

Standard Error of the Estimate (SEE)

A

The standard deviation of the residual terms in the regression. SEE will be low relative to total variability if the relationship between dependent and independent variables is strong, and will be high if the relationship its weak (the smaller the standard error, the better the fit)

=√MSE

=√[ SSE / (n-k-1) ]

24
Q

Cointegration

A

Two time-series are economically linked (related to the same macro variables) or follow the same trend and that relationship is not expected to change. If cointegrated, the error term from regressing one on the other is covariance stationary and the t-tests are reliable.

25
Q

Autoregressive Conditional Heteroskedasticity (ARCH)

A

Exists if the variance of the residuals in one period is dependent on the variance of the residuals in the previous period. When this condition exists, the standard errors of the regression coefficients in AR models and the hypothesis tests of these coefficients are invalid.

26
Q

Correcting Seasonality

A

An additional lag of the dependent variable (corresponding to the same period in the previous year) is added to the original model as another independent variable.

27
Q

Mean Reverting Level

A

Xt = b0 / (1 - b1)

If Xt > b0 / (1 - b1), the model predicts that Xt+1 < Xt
If Xt < b0 / (1 - b1), the model predicts that Xt+1 > Xt

28
Q

Type I Error

A

Incorrectly rejecting the null when it should not be rejected.

29
Q

Type II Error

A

Incorrectly failing to reject the null when out should be rejected.

30
Q

Unit Root

A

If the lag coefficient is equal to one. Will follow a random walk process. Since a time series follows a random walk is not covariance stationary, modeling such a time series in an AR model can lead to incorrect inferences.

31
Q

Linear Regression Assumptions

A
  1. A linear relationship exists between the dependent and independent variable.
  2. The variable is uncorrelated with the residuals.
  3. Expected value of the residual terms is 0.
  4. The variance of the residual term is constant for all observations.
  5. The residual term is independently distributed; the residual term for one observation is not correlated with that of another observation.
  6. The residual term is normally distributed.
  7. Independent variables are not random, and there is no exact linear relationship between two or more independent variables.