Binary Dep for kids Flashcards

1
Q

What three models do we have?

A

Logit, Probit, LPM

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

How do we interpret Binary dependent variable regression

A

interpreted as a conditional probability function

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is conditional probability

A

Conditional probability is the probability of one thing being true given that another thing is true

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Difference between Probit & Logit and LPM

A
  • Probit, Logit allows for non-linear relationship between dependent and regressors
  • Probit, Logit will be between 0 to 1
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

-

A

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

R2 Interpretation

A

No meaningful interpretation

- regression line never able to fit the data perfectly because y is binary and regressors are continious

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

R2 relies on ___ which make it unusable

A

linear relationshipt between X and Y

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What measures the fit of the model

A

PseudoR2 measureas the fit using the likelihood function

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is a good PseudoR2 value

A

Rule of thumb is between 0,2 and 0,4

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

PsuedoR2 is also called

A

McFadden

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Standard errors in LPM are always

A

Heteroscedastic, so we use robust standard errors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

.

A

.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

When Y is the binary variable -> explain the regression

A

The population regression function shows the probability that Y = 1 given the value of the regressors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Why is it called LPM

A

Because the probability that Y = 1 is a linear function of the regressors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is Probit and Logit regressions

A

They are regression models that are nonlinear when Y is used as a binary variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Difference between LPM and Probit & Logit

A

Probit & Logit regressions ensure that the predicted probability will be between 0 and 1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Probit Regression uses …..

A

Cumulative Distribution function

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What is cumulative distribution function

A

It is the probability that the variable takes a value less than or equal to X

19
Q

What does Probit and Logit regression allow for that LPM doesnt

A

Probit and Logit models allows for non-linear relationship between regressors and dependent variable.

20
Q

Logit Model uses _________

A

Logistic cumulative distribution function

21
Q

Logit and Probit Models are appropriate when attempting to model ___

A

a dichotomus dependent variable, e.g. yes/no, agree/disagree, like/dislike.

22
Q

How does the Probit and Logit model look like? Shape

A

S-Shape, y is between 0 and 1

23
Q

Y axis shows

A

We can think of the y-axis as originally having a value 0 to 1. But this value get transformed into the value of log(p/(1.p)). So if p (or y) was 0,5, the new value on this axis is log(0,5/(1-0,5))=0.

when p = 1, we get log(1) - log(0). This equals positive infinity, we now got both positive and negative infinity

24
Q

where infinity come from

A

when p = 1, we get log(1) - log(0). This equals positive infinity, we now got both positive and negative infinity

25
Q

.

A

.

26
Q

What is the z value

A

Rule of thumb: Z should be over 2 and p under 0,05 for H0 to be rejected

estimated intercept divided by the standard error

  • the number of standard deviations the estimated intercept is away from 0 on a standard normal curve (Wald test)
27
Q

Why cant we use Least Squares method

A

Intuitively we want to draw the best line with least squares as in OLS simple regressions, but our residuals go to infinity, so cant use Least Squares

28
Q

Maximum Likelihood Intuitive of the mean

A
  1. Imagine that you have a line of observed values.
  2. Then imagine that you test every point on that line for where you get the highest likelihood of observing the data
  3. when all areas are checked you pick the one that maximizes the likelihood
29
Q

likelihood in statistics means

A

trying to find the optimal value for the mean or std for a distribution

30
Q

How do wee find the best regression line

A

maximum likelihood

31
Q

if p-value is < 0,05

A

there is a statistically significant association between the response variable and dependent

32
Q

.

A

.

33
Q

Consistency means:

A

Increased sample will make the Beta converge to the real beta

34
Q

Unbiasedness means

A

The expected value of the beta will on average be correct.

Not an overestimate or underestimate

35
Q

What are the assumptions on the parameters

A
  • Consitency: increase sample = converge to true population value
  • Unbiased: Expected B will equal true B
36
Q

What do we use instead of R2

A

PseudoR2 (Mcfdden)

37
Q

is there a reason to use LPM over Probit, Lobit?

A
  • more easy to interpret

- it can be discussed if there are not extreme prop values

38
Q

how to interpret probit coeff

A

A positive coefficient means that an increase in the predictor leads to an increase in the predicted probability.

A negative coefficient means that an increase in the predictor leads to a decrease in the predicted probability

39
Q

What are the tests for parameters

A

z - test: one parameter

likelihood ratio test: several parameters

40
Q

What are the interpretation for the marginal effects in the three models

A

LPM assumes a that the distribution

41
Q

Marginal effects of Probit and Logit

A

Use Probability Distribution Function to find it

42
Q

When can LPM be used

A

The basic insight is that the linear probability model can be used whenever the relationship between probability and log odds is approximately linear over the range of modeled probabilities.

43
Q

rule of thumb for when to use LPM versus logit

A
  • if the probabilities are extreme, like yes/no, close to 0 or 1, logit is better
  • if they are more moderate, like between 0,20 and 0,8, LPM can be used —then the linear and logistic models fit about equally well, and the linear model should be favored for its ease of interpretation.
44
Q

LPM is bad with

A

very large or very small probabilities