lecture 3 Flashcards

1
Q

points on a regression line

A

all data points have a residual, we want them to average to zero, does OLS do this?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

what do OLS estimators minimize

A

the sum of squared residuals
where: uhat i = Y i - Bhat i - Bhat1 Xi

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

minimizing the sum of squared residuals gave us

A

two first order conditions (FOC) (SUM n, i=1)
1. uhat i = 0
2. uhat i X i = 0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

OLS estimators

A

Bhat 1 = sXY/S^2 X
Bhat o = Ybar - bhat 1 Xbar

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

what does necessary conditions mean for FOC

A

they must be satisfied when we’ve estimated our estimate line via OLS

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

SUM n, i=1 uhat i = 0

A

1st first order condition
- states that the sum of the residuals must be equal to zero

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

SUM n, i=1 uhati X i = 0

A

2nd first order condition

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

OLS sample covariance =

A

0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

based on OLS first order conditions we have decided

A

u hat bar = 0
S u caret X = 0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

first order conditions are not

A

assumptions; they follow from the choice made to use OLS to estimate the regression line

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

graphical representations

A

y = give us the average residuals to sum to zero so the line is through the middle, not higher or lower
u caret = when put onto a residual line we can see that it is not the optimal height for a regression line

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

OLS regression lines are

A

average residual = 0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

what is the height of the regression line

A

such that the average residuals sum to zero

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

what can be said about the OLS estimator of the intercept

A

Bcaret o + Bcaret 1 X bar = Ybar
OLS sets the height of the regression line at Xbar to Ybar: ensuring average residual will be equal to 0
- plug in mean for X, it should equal Ybar if the average residual is 0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

OLS regressor must go through what

A

Xbar, Ybar
- slope doesn’t matter, it just needs to hit this point

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

even when the average residual is 0 what else is needed

A

covariance needs to be 0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

what is the 2 step process for OLS

A

satisfy 2 FOC

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

how do the 2 FOC create the best fit line

A
  1. the first FOC of OLS sets the height of the regression at Xbar to be Ybar, ensuring the average residual is zero
  2. adding the second FOC ensures the slope of the regression results in zero covariance between the X and the residuals
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

goal of OLS

A

estimate a population regression line using a sample

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

OLS

A

estimation method for population regression line

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

implications of choosing OLS

A

1 . OLS must go through (Xbar, Ybar) ensuring residual is zero
2. the OLS slope ensures there is zero covariance between X and u caret

22
Q

what variance do we prefer for an estimator

A

the one with more variance
/4 better than /2

23
Q

least squares assumptions: line written as

A

there is a population regression line:
Yi = Bo + B1Xi + ui

24
Q

what can be observed in a population regression line

A

Y and X
Bo, B1, ui are not observed

25
least square assumptions #1: zero conditional mean
critical assumption about the error term = E(ui|Xi)
26
what does the zero conditional mean assumption assume
zero covariance and no correlation between u and X
27
in the zero conditional mean assumption, u and X are -------
independent
28
if X and u are correlated then
zero conditional mean assumption cannot be satisfied
29
what does zero conditional mean assumption do for our line
ensures the true regression line is located in the center of the data
30
what do we say about u as a function of X in assumption 1
u is not a function of X
31
two components of the zero conditional mean assumption (about u0
1. expected value of the error term is a constant - it does not depend on x 2. this constant is zero
32
what graph would OLS not perform well
- with the zero conditional mean assumption slope and intercept surround the true line - data surrounds the true line, center of data is the truth
33
with zero conditional mean assumption what can we say about true regression line and population conditional mean
true regression line is located at the population conditional mean
34
what do assumptions 1-3 imply/do
they allow for us to certain of factors concerning OLS estimators
35
least squares assumption #2: sample is i.i.d.
independently and identically distributed for the joint distribution
36
joint distribution
Joint: probability that the random variables simultaneously take on certain values, say x and y EX: rain, no rain, long commute, short commute, makes 4 separate options All sum to 1
37
independent and identically distributed
independent: one observation does not tell you about another, randomly selected identically: drawn from the same population
38
how to get i.i.d. data
simple random sample from a large population
39
critical value for 95%
1.96
40
critical value for 90%
1.64
41
least squares assumption #: X and Y have non-zero finite moments
- rules out having very large outliers - meant to rule out difficult to work with variables
42
using these 3 assumptions what 4 things can we say about the OLS estimators; properties we know to be true when sufficient conditions are satisfied
1. OLS estimators are unbiased 2. OLS estimators are consistent 3. we can derive OLS variance estimators 4. OLS estimators are asymptotically normal (only FOCs always hold)
43
1. OLS estimators are unbiased
E(Bcaret o) = Bo - intercept E(Bcaret o) = B1 - slope this means that the expected values for these are what the OLS estimators turn out to be
44
2. OLS estimators are consistent
law of large numbers for Bcaret o and Bcaret 1
45
law of large numbers
When a large number of random variables with the same mean are averaged together, the large values tend to balance the small values, and their sample average is close to their common mean; if you have enough data you can get extremely close
46
4. OLS estimators are asymptotically normal
approximation works better when there is more data - essentially saying n is sufficiently large that the approximation is good enough
47
4. OLS estimators are asymptotically normal: infinitely large meaning
normal distribution
48
properties of OLS that hold ALL the time
1. FOC: u caret bar = 0; OLS regression line passes through (Xbar, Ybar) 2. FOC: s ucaret X = 0; residuals are uncorrelated with regressor
49
given that 1. FOC holds
then 2. FOC holds
50
why isn't E(u|X) on the list of properties of OLS that hold all the time
statement is about errors in population not residuals
51