lecture 3 Flashcards
what is the genera; form a regression equation
y = a + bX + u
what dpoes X represent in an regression equation
X is the independent or explanatory variable
what dpoes y represent in an regression equation
Y is the dependent or explained variable
what does u represent in a regression equ
u is a random error or disturbance
what does ‘a and b)’ represent in a regression equation?
α and β are parameters which characterise the relationship between
Y and X. The parameters are not observable directly.
what are the 2 interpretations of the regression model
- X values are chose by the investigators e.g. process of experimentation
- X and Y are jointly distributed random variables with cov(x,y)=/= 0
whats is the method of least squares?
calculating an estiamte of an unknown value using observable data
method is to choose estimates of the parameters
which minimises the residual sum of squares.
min(𝑅𝑆𝑆)=∑(𝑌_𝑖−𝛼hat−𝛽hat𝑋)^2
this is OLS
what is the estimate for Bhat?
‘sum of all’(X-Xbar)(Y-Ybar)/’sum of all(x-xbar)
covarience of X and Y over varience of X
what is the estimate for a hat?
Ybar - bhat*Xbar
what is the 2 stage procedure to calculate OLS
1.Calculate the slope coefficient as the ratio of the sample
covariance of X and Y to the sample variance of X
- Calculate the intercept using the property that the regression
line passes through the sample means of the data.
what are ‘ a and b’?
population parameters
what are ‘ahat and bhat’?
population parameters based on
sample data.
what are estimators
random variables because they are
constructed from the random variables Y and (possibly) X.
The population parameters are not random variables. They
are unknown/unobservable parameters which we must estimate
using the data available.
what is maximum likelehood?
alternative way to generate estimates of unknown parameters
The errors are assumed to be independent, identically distributed
(iid),normal random variables
take logs of the likelihood function
The method of maximum likelihood involves choosing
estimates of the population parameters which maximise the
log-likelihood function.
compare Maximum likelihood and OLS
1st 2 FOC are identical to OLS equations
This is different from the formula normally used for the variance
of a least squares regression because it does not adjust for the
loss of degrees of freedom when estimating the other regression
parameters.
The maximum likelihood estimator of the error variance will be
biased in small samples. However, the bias will tend to zero as the
sample size becomes large