Quantitative Methods Flashcards
Sum of Squared Errors (SSE)
Measures the unexplained variation in the dependent variable. Is the sum of the squared vertical differences between the actual values and the predicted values on the regression line.
Breusch-Pagan (BP) Chi-Square Test
=n*residualR^2 with k degrees of freedom
If larger than the one-tailed critical value for a chi-square distribution, you should reject the null and conclude heteroskedasticity is present.
Effect of Serial Correlation
- Positive serial correlation typically results in coefficient standard errors that are too small. This will cause the t-statistic to be too high and the null to be rejected too often.
- F-test will be unreliable because MSE will be underestimated
F-Statistic
=MSR/MSE
If F-statistic > F-value, then you reject the null hypothesis, indicating at least one coefficient is statistically significant.
Unconditional Heteroskedasticity
Occurs when the heteroskedasticity is not related to the level of independent variables. It usually causes no major problems with regression.
Conditional Heteroskedasticity
Exists if the variance of the residual term increases as the value of the independent variable increases. Creates significant problems for regression and statistical inference.
Adjusted R^2
=1 - { [ (n-1) / (n-k-1) ] x ( 1-R^2) }
Will always be equal to or less than R^2
Root Mean Squared Error (RMSE)
Used to compare the accuracy of autoregressive models in forecasting out-of-sample values. The model with lower RMSE for the out-of-sample data will have lower forecast error and will be expected to have better predictive power in the future.
Correcting Multicollinearity
Omit one or more of the correlated independent variables
Random Walk with a Drift
The time series is expected to increase or decrease by a constant amount each period (b0).
Xt = b0 + b1*Xt-1 +εt
Correcting Heteroskedasticity
- Calculate robust standard errors used to recalculate the t-statistics using the original regression coefficient if evidence of heteroskedasticity.
- Using generalized least squares, which attempts to eliminate heteroskedasticity, by modifying the original equation.
(If both heteroskedasticity and serial correlation, use the Hansen method)
Detecting Multicollinearty
Fail to reject the null in a t-test, but fail to reject the null in an F-test while R^2 is high.
Multicollinearty
The condition when. two or more independent variables are highly correlated with each other. This distorts the standard error of the estimate and the coefficient standard errors, leading to unreliable t-tests. Slope coefficients will be unreliable as well.
Serial Correlation
The situation in which the residual terms are correlated with one another. Positive serial correlation exists when a positive regression error in one period increases the probability of observing a positive regression error in the next period. Negative serial correlation exists when a negative regression error in one period increases the probability of observing a negative regression error in the next period.
Durbin-Watson Statistic
Detects the presence of serial correlation.
DW > 2 – Positive serial correlation
DW < 2 – Negative serial correlation
DW = 2 – No serial correlation
DW < dL – Reject null; positive serial correlation
DW > dU – Fail to reject null; no evidence of SC
dL < DW < dU – Test is inconclusive