Topic 3: Multiple Regression Model Flashcards
The OLS estimates are chosen to minimize what MLR model with k variables?
the sum of [(y_i) - (beta^0) - (beta^1xi1) - (beta^2xi2)-…-(beta^kxik)]^2
In MLR, holding all other intendent variables equal except x, the change in y = ?
change in y^ = beta1 change in x^1
What does it mean to say that a OLS estimator is unbiased?
That means that the expected value of the estimates is equal to the population coefficient.
What is an important assumption for the unbiasedness of estimators?
the zero conditional mean assumption is what makes the expected value of the estimates to be equal to the parameter
What is the bias in a simple regression from leaving out an important variable, beta2?
beta2 * the sum of [(xi1 - xbar)xi2] / the sum of [(xi1 - xbar)^2]
When the correlation between x1 and 2x is greater than zero, and x2 is omitted from the estimated equation, when is the bias positive and negative?
When corr(x1,x2)>0 and beta2>0 the bias is positive (the expected value is greater than the parameter) and when corr(x1,x2)>0 and beta2<0 the bias is negative.
When the correlation between x1 and 2x is less than zero, and x2 is omitted from the estimated equation, when is the bias positive and negative?
When corr(x1,x2)<0 and beta2>0 the bias is negative (the expected value is less than the parameter) and when corr(x1,x2)>0 and beta2<0 the bias is positive.
How to remember the bias in beta~1, when x2 is omitted?
When the correlation and the beta2 have the same sign the bias is positive, when they have different signs, the bias is negative.
Stata: How to leave how certain observations from a regression?
reg y x1 x2…. if var =
What is multicollinearity?
Correlation among independent variables in a MLR model. An R-squared of 1 shows perfect collinearity, but there is no standard amount under 1 that always is too much.
How do you interpret a the effect of a model on y where two variables are correlated (but not perfectly) with the variable of interest?
Combine the coefficients of the two variables in the regression to get the effect of the variable of interest on y.
Why is the data in an experiment better than observational data when it comes to MLR assumptions?
In the experiment, it is given that the treatment is randomly assigned, thus the zero conditional mean assumption will hold.
Would omitting a variable cause bias in your estimators?
Yes.
Would including an irrelevant (uncorrelated) variable cause biased in your estimators?
No.
Even though including uncorrelated variables does not cause bias, why would you not want to include such variables?
Because they could have undesirable effects on your variance.