Introduction to Multiple Regression Flashcards
Are there always omitted variables in a regression model
Yes because reality is interconnected and not entirely causal
When are omitted variables a problem for regression functions
When they are not unbiased aka they fail to fulfill the E(u|X=x) = 0 requirement
What is omitted variable bias
Bias in the OLM estimator that is a result of some variable being omitted. For the bias to uccur the omitted variable Z must be at least a little correlated with X and also be a determinant of Y
If room temperature is a determinant of test scores but is not correlated at all with the amount of teachers in the class will the admission of room temperature lead to a omitted variable bias in the regression model of grades based on teachers in class
No, both the requirements of correlation and determinants must be fulfilled by the third factor for its omission to result in a omitted variable bias.
What is P a sign fore in statistics
correlation
Is there a formulaic way to determine if the slope is biased due to an omitted variable
If there is correlation between the independent variable X and the residual u
What is reverse causality
when effects causes the cause because of an expected effect. For example when patients choose a treatment because they believe it will cure them
How do you avoid omitted variable bias
You ither use a randomized and controlled sample that compensates for the bias or you use multiple regression aka include the omitted variable in the regression function
How is the standard error of the regression SER effected by multiple regression
the denominator it is divided by is subtracted by 1 in addition to the number of independent variables.
R² always increases when you add another regressor
Yes because the method gets more to work with, that is why the adjusted -R² penalizes you for using another regressor.
How do you calculate adjusted -R² from R²
You multiply the regression with the sample size subtracted by one divided by the sample size subtracted by one and the amount of regressors
What is perfect multicollinearity
When one of the regressors is an
exact linear function of the other regressors
Explain why the dummy variable trap leads to perfect multicillinearity
Becouse if the variables are muturally exclusive and exhaustive you can re write one variable as a linear function of the other.
What is the dummy variable trap
That when you have a group of dummy varables that are mutually exclusive and exhaustive aka a few classes buch each datapoint falls withing one of them they achive perfect multicollinerity if you add a constant to the regression.
How do you escape the dummy variable trap
Omit one of the variables