3- Classical linear regression Flashcards
What does the error term (u) capture in a regression?
All the effects on the dependent variable not in the explanatory variables
What are the betas in a regression?
The gradients associated with each of the theoretical parameters
What is always the first independent variable in a regression?
A constant β₀, it represents the y-intercept i.e. value of dependent variable when there’s no input
What is the population regression line?
The sum of the independent variables; the relationship that holds between x and y on average (without error term) y = β₀ + Xβ₁
Once x and y values are collected how is a line of best fit determined?
Minimising the sum of each error term squared
What is the error term for each observation plotted?
The vertical distance between the population regression line
What do the independent variable subscripts denote (xᵢₖ)?
The first letter is the number of observations and the second letter is the number of parameters
What are the 6 assumptions of the linear regression model?
-Linearity
-Identification condition
-Exogeneity
-Homoskedasticity
-Normality
-X can be fixed or random
Explain the linearity assumption
Linearity in parameters means the betas have index 1 so are not exponential
Explain the identification condition
Number of observations must be at least as great as the number of parameters
Explain the exogeneity assumption
Expected value of any u on X is zero E(uᵢ|X)=0 meaning no observation conveys any information about u
Explain the homoskedasticity assumption
Variance of the error term is constant across observations
Explain the normality assumption
Error terms are normally distributed
What are the 3 main properties of the OLS estimator?
-Unbiased
-Given variance covariance matrix
-Estimator is the best in that it has minimum variance
What does it mean that the OLS estimator is unbiased?
The expected value of all estimated betas will give their true value E(^βols)= β