L1 Flashcards

1
Q

Slope of regression line = ?

A

Expected effect on Y of a unit change in X

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

How does an OLS estimator work?

A

It minimises the avg square of differences between the actual Yi values and the predicted ones

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the interpretation of beta1(hat)?

A

For each extra unit of X, Y on avg. is beta1(hat) higher/lower

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the interpretation of beta0(hat)?

A

It is the intercept (ie. Y when X=0)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Give an example when an interpretation of Beta0(hat) would not make sense?

A

If looking at test scores to student-teacher ratio, by extrapolating outside the data the line predicts a very high score, even though having 0-students to a teacher would not actually give a score!

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What does R^2 measure?

A

The fraction of the variance of Y that is explained by X (ranges from 0 (none) to 1 (complete explanation))

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What does it mean if R^2 is 0 or 1 for explained sum of squares?

A

If 0, means ESS=0

If 1, means TSS=ESS

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

When does R^2=r^2?

A

When there is only one X variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is r^2?

A

The square of the correlation coefficient between X and Y

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What does the standard error of the regression measure (SER)?

A

The magnitude of a typical regression residual in the units of Y

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What does the root mean squared error measure?

A

Same as the SER; in large samples the measures are very close

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Briefly, what are the least squares assumptions?

A

1) Error and X variable(s) are independent: E(u|X=x)=0
2) Observations are i.i.d. (Xi and Yi)
3) Large outliers in X and/or Y are rare
(not the same as the OLS assumptions!)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

When do we often encounter non i.i.d. sampling?

A

TS data and Panel data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is the sampling uncertainty of an OLS estimator?

A

When a different sample of data might lead to different beta1 estimate

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

How is the variance of Beta1(hat) related to the sample size?

A

Inversely proportional; as sample size increases, variance decreases

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

See: consistency, approximations and CLT

A

now

17
Q

How does the variance of X affect the variance of Beta1(hat)?

A

Larger variance of X, smaller variance of beta1(hat) (see notes)

if there is more variance in X, there is more information in the data therefore the regression is plotted with more confidence therefore less variance in coefficient

18
Q

State the CLT?

A

Suppose you have a set of i.i.d. observations with expected value 0 and variance=(sigma)^2, then when n is LARGE, 1/n(sum of obs.) is approximately distributed by N(0, Variance/n)

19
Q

Explain LSA 1: Error and X variable(s) are independent: E(u|X=x)=0?

A

For any given value of X, the mean of the error is zero!

Consider an ideal randomised controlled experiment: X is randomly assigned to people by a computer tf since X is randomly assigned, all other characteristics (present in u) are randomised so u and X are independent!

If an IRCE isn’t in action must consider if LSA 1 actually holds!

20
Q

Explain LSA 2: Observations are i.i.d. (Xi and Yi)? (ie. explain each i?)

A

Arises automatically if the entity is sampled by simple random sampling!

entities selected from same pop. so (Xi,Yi) are identically distributed for all i, and since they are selected at random the values (Xi,Yi) are also independently distributed tf i.i.d.

21
Q

Explain why a larger variance in the set of observations, X, leads to a better estimate of beta?

A

Eqn. in notes tells us that the variance of X is inversely proportional to the variance of beta(hat1)

if there is more variance in the X, this gives more information in the data therefore the regression will be plotted with more confidence and tf less variance in beta(hat1)