3/4- Linear regression with one regressor Flashcards

1
Q

What 3 main things does linear regression allow us to do?

A
  • Quantify different relationships (through correlations)
  • Estimate causal effects under assumptions
  • Carry out a predictive analysis
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What does an error term summarise?

A

All the determinants of y besides x

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the causal/treatment effect?

A

E(y|x=1) - E(y|x=0) = β1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Why will an experiment not give us the true value of β1 unless we can sample the entire population?

A

Randomisation ensures E(y|x=1) = E(y|x=0) in samples as n goes to infinity. But, in any sample, the sample average value of u may differ between small and large classes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What are the 2 main properties of ^β1 as an estimator of β1?

A
  • ^β1 is an unbiased estimator of β1, if we did the experiment repeatedly on different samples, the average value of ^β1 would equal β1: E(^β1)=β1
  • ^β1 is a consistent estimator of β1, as the sample size goes to infinity ^β1 gets closer to β1
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is the selection bias?

A

The difference between E(u|x=1) - E(u|x=0)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What does the slope of the population regression line (β1) give us?

A

The average change in Y for a unit change in x

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What do beta coefficients with hats (^β) refer to, and what do those without (β) refer to?

A
  • ^β refers to a sample

- β refers to population parameters

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the line of best fit?

A

The best fit line is the closest to all the sample points, so to find it we have to minimize the distance between the line and the data points

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What does the intercept (β0) of the OLS estimator tell us?

A

The predicted value when X = 0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What does the Total Sum of Squares (TSS) tell us?

A

TSS measures the variation in the dependent variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What does the Explained Sum of Squares (ESS) tell us?

A

ESS represents variation explained by the regression

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What does the Residual Sum of Squares (RSS) tell us?

A

RSS represents variation not explained by the regression

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What does R^2 measure?

A

The fraction of the variance of Y that is explained by the regression. It measures how well the regression line fits the data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

How does TSS relate to ESS and RSS?

A

TSS = ESS + RSS

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What are the range of values R^2 can take?

A

Between 0 and 1, 0 implies no fit and 1 implies perfect fit

17
Q

How does the interpretation of β1 change when regression is in semi-logarithmic form?

A

A one unit change in x produces a β1*100 percent change in Y

18
Q

How can you calculate the error term/residual (u)?

A

Find the difference between the actual results for a given parameter and the result the model would predict given that parameter

19
Q

What is a key feature of OLS residuals?

A

They all add up to 0

20
Q

What is the relationship between the predicted/fitted value of y and the residual?

A

The sample covariance between them is 0

21
Q

What is an advantage of using a semi-logarithmic regression?

A

If we just had the dependent variable then the value of an extra unit of the independent variable would be constant

22
Q

What is a constant elasticity model?

A

When both the dependent and independent variable are logarithmic

23
Q

What is the relationship between y and β1 in the constant elasticity model?

A

a 1% increase in y leads to a β1% increase in x

24
Q

What does the expected value of the OLS estimator tell us?

A

Tells us if the estimator gets it right on average, it is the average outcome across all possible random samples

25
Q

What does the standard error of the OLS estimator tell us?

A

Tells us how reliable our results are

26
Q

What are the 4 main assumptions for the unbiasedness of SLR?

A
  • Linear in parameters: can have log(y) but not log(β)
  • Random sampling
  • Sample variation in the explanatory variable: sample outcomes on xi are not all the same value
  • Zero conditional mean: E(u|x)=0
27
Q

What happens to the parameters if the OLS is unbiased? i.e all 4 assumptions hold

A

E(^β1)=β1