Reading 12: Multiple Regression Flashcards

You may prefer our related Brainscape-certified flashcards:
1
Q

Multiple Regression Equation

A

This shows several Betas for several independent variables that “predict” the dependent variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What type of test needs to be performed in order to test the significance of one of the Betas? And what is the formula?

A

A t-statistic is done to check the significance

H0: Beta = 0

If t > t-critical : Reject H0

t-statistic =

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

How can the significance be tested of individual components in the multiple regression?

A

Perform a t-statistic.

if t > t-critical reject H0

In the formula Beta tested agains the idea that the real beta = 0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

How can the significance be tested of a regression analysis as a whole?

A

An F-Test is to be performed. An F-test is performed as following:

F= MSR / MSE

MSR= RSS / k

MSE = SSE / [n-(k+1)]

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

How is the Sum of Squared Errors (Residuals) calculated?

A

(SSE) with df = k

= Actual observation - Expected Observantion

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

How is the Regression Sum of Squares calculated?

A

RSS is the explained variation by the independent variables.

= Expected Y - average observation Y

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

F - Test is done how?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is R2?

A

R2 is a measure of goodness of fit of the estimated regression of the data.

R2 = Explained Varation / Total Variation

R2 = RSS / SSE

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the R2 - Adjusted?

A

This adjusted version of the R2 is done as this statistic is not automatically increasing when another variablle is added and is adjusted for degrees of freedom.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

A. What is Hetroskedasticity?

B. What are the consecquences?

C. How can we test for it?

D. How can it be corrected?

A

A. Hetroskedasticity is a non-constant variance of the errors accross the data set.

B. The Regression coefficients are not affected. ONLY the standard error of regression. -> in general lower erros lead to higher t-values -> leads to lower p-values.

C. Tested by the Breusch-Pagan test.

D. Corrected by:

  • White corrected standard erros -> adjusts the standard errors of the LRM.
  • Generalized least squares
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

How and why is the Breusch - Pagan Test performed?

A

Done to test for conditional Hetroskedasticity in the regression.

dit is an Chi-statistic

X2 = n R2

with k degrees of freedom (independant variables)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is Serial Correlation (Autocorrelation)?

A

This means that there is correlation among the error terms. This mostly happens in a time-series regression.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Serial Correlation (Autoregression) A:

  • Consequences?
A

There is Positive Serial Correlation:

  • A positive error for one observation increases the chance of a positive error for another observation.
  • We assume that the sign of the error term tends to persist from one period to the other.

Consequences of serial correlation: is an incorrect standard error of regression.

F-test may be inflated by underestimated MSE

T-test will be inflated and p-values show significance where not appropriate.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Serial Correlation (Autoregression) B:

How to test for this?

A

We test with the Durbin and Watson test.

DW can be estimated by

DW = 2 (1-r)

H0 = NO serial correlation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Serial Correlation (Autocorrelation) C:

When do we reject the H0: No serial correlation?

A

DW > Du -> Fail to reject the null hypothesis

DW < Dl -> Reject the null hyphotesis

Dl u -> inconclusive test.

IF WE REJECT: UNDERESTIMATED ST. ERRORS

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How to correct for Serial Correlation (Autocorrelation)?

A

Two Ways:

  1. Adjust the coefficient standard error (Hansen Method)
  2. Modify the regression equation itself.

Hansen’ Method corrects for both Serial Correlation & conditional Hetroskedasticity

17
Q

Multicollinearity (A)

What is it?

A

Multicollinearity occurs when two or more independent variables are highly correlated

If the independent variables are perfectly correlated regression is impossible and is called perfect collinearity.

18
Q

Multicollinearity (B)

What is the problem generated by multicollinearity in a regression analysis?

A
  • Presence does not affect the consistency of the OLS estimates / regression Coefficients
  • Due to multicollinearity its impossible to distinguish the individual impacts of the independent variables on the dependent variable.
  • Inflated OLS standard errors for the regression coefficients. Hence T-statistic very small -> no ability to reject H0
19
Q

Multicollinearity (C)

How to detect Multicollinearity?

How to correct?

A

No statistic to measure it however:

If F-test is significant and High R2 and t-tests are insignificant

Correct by ommitting one variable.

20
Q

Summary of Heteroskedasticity, Serial Correlation, Multicollinearity:

A
21
Q

Model Misspecification?

A
  1. One or more important variables could be omitted from the regression
  2. One or more of the regression variables may need to be transformed before estimating the regression
  3. The regression model pools data from different samples that should not be pooled
  4. Independent variables are correlated with error terms (violation)