Reg review Flashcards
what is the standard error?
The estimated standard deviation of the sampling distribution of the slope parameter, which tells us how precise our estimate is.
what is the p-value?
The smallest significance level at which we would reject the null hypothesis.
What is a confidence interval?
Over repeated sampling, we would expect 95% of confidence intervals constructed in this manner to contain the true population parameter.
What does it mean for a coefficient to be unbiased?
If an estimator is unbiased, then the mean of the sampling distribution of the estimate should be centered on the true population parameter (E(beta1_hat) = beta1_population)
Definition of beta1 coefficient?
Beta1_hat is the slope of y with respect to x when all other regressors are held constant, or fixed, a one-unit change in X is associated with a beta1_hat change in Y, holding all else constant.
What is an endogenous regressor?
correlated with the error term, or correlated with Y through the error term
What is an exogenous regressor?
uncorrelated with error and has a direct impact on Y, should be included to avoid OVB.
What is consistency?
Betahat is a consistent estimator of betapop, as N approaches infinity, betahat converges in probability to betapop (a.k.a. asymptotic unbiasedness, large samples property)
What is the law of large numbers?
Our estimates of the population mean and variance will converge in probability to the true population parameters.
What is the CLT?
As N approaches infinity, the sampling distribution will be normally distributed.
What are the Guass-markov assumptions for MLR and which are needed for unbiasedness and consistency? Which are needed to be BLUE?
Gauss Markov Assumptions for MLR: (1 – 4 for unbiasedness, 1 – 5 for BLUEs (best linear unbiased estimators))
- Linear in parameters
- Random sampling (independent and identically distributed random variables)
- No perfect collinearity (between any of the predictors, r < 1 between all regressors)
- Zero conditional mean (the expected value of U conditional on all Xs is equal to zero)
- Homoskedasticity (the variance of U conditional on all Xs is equal to the variance of U, sigma-squared)
Probability of type 1 error?
alpha
Probability of type 2 error?
beta
What is r-squared?
Proportion of the sample variation in Y that is explained by X
Factors affecting sampling variances of OLS slope estimators
1) Error Variance: Take more stuff out of the error (make σ2 smaller); add more explanatory variables. As error variance in pop decreases, Var((β_j ) ̂) gets smaller.
2) Total Sample Variation: It is easier to estimate how xj affects y if we see more variation in xj (increase SSTj); increase SSTj by increasing the sample size.
3) As Rj2 gets bigger so does Var(b1). If xj is unrelated to all other independent variables, it is easier to estimate its ceteris paribus effect on y.
What happens when you have heteroskedasticity?
Variance formulas for OLS invalid, does not affect beta coefficients; cannot perform F/t-tests.