Quantitative Methods Flashcards
MSE Formula
SSE / (n-k-1)
AIC
n x ln(SSE/n) + 2(k+1)
Better when goal is better forecast
BIC
n x ln(SSE/n) + ln(n) x (k+1)
Better when goal is a better goodness is fit
F-statistic
MSR/MSE
Assumes all slope coefficients simultaneously 0
Rejection if F>F(critical)
F-statistic joint hypothesis
((SSE_R - SSE_U)/q) / (SSE_U/(n-k-1))
q is number of excluded variables in restricted model
Dickey-Fuller test for unit roots could be used to test whether the data is covariance non-stationarity. The Durbin-Watson test is used for detecting serial correlation in the residuals of trend models but cannot be used in AR models. A t-test is used to test for residual autocorrelation in AR models.
Test for unit root to test whether data is covariance non-stationary
Durbin-Watson
Test for serial correlation in residuals of trend models but cannot be used in AR models
T-test for AR models
Used to test residuals in AR models
Breusch-Pagan
Used for conditional heteoskedasticity
Breusch-Godfrey
Used for positive serial correlation
White-corrected standard errors
Used to correct for conditional heteroskedasticity
Impact of conditional heteroskedasticity (overestimate)
No effect on coefficient estimate
Std Err of coefficient overestimated
More type II
Impact of conditional heteroskedasticity (underestimate)
No effect on coefficient estimate
Std Err of coefficient overestimated
More type II
Newey-West Standard Errors
Used to correct positive serial correlation
Impact of serial correlation
No impact on coefficient estimate
Std Err underestimated
More type I error
Dummy variable misspecification
If we use too many dummy variables (e.g. >n-1), there will be multicollinearity
What happens to the coefficients of correlated independent variables when a new correlated variable is added to the model?
Adding the new variable will change the coefficient for the other correlated variables.
What does the intercept term (b0) represent in a multiple linear regression model?
It shows the value of the dependent variable when all independent variables are 0.
What do the slope coefficients (bi) represent in a multiple linear regression model?
They are the estimated changes in the dependent variable for a one-unit change in the corresponding independent variable, holding all other independent variables constant. Also called partial slope coefficients.
What are the assumptions underlying a multiple linear regression model?
- Linearity between dependent and independent variables. 2. No significant multicollinearity. 3. Expected error is 0. 4. Homoscedasticity (constant error variance). 5. No serial correlation (errors are independent). 6. Errors are normally distributed.
How is the Total Sum of Squares (SST) calculated?
- Subtract the mean from each individual observation. 2. Square each result. 3. Sum the squared results. Degrees of freedom = n-1.
How is the Regression Sum of Squares (RSS) calculated?
- Subtract the mean from each predicted observation. 2. Square each result. 3. Sum the squared results. Degrees of freedom = k.
How is the Sum of Squared Errors (SSE) calculated?
- Subtract the predicted value from each observed value. 2. Square each result. 3. Sum the squared results. Degrees of freedom = n-k-1.
What is the formula for R2 (Coefficient of Determination)?
R2 = SST/RSS (Explained Variation / Total Variation).