Regression Part-7 Heteroskedasticity Flashcards

1
Q

What is homoskedasticity?

A

Random errors appearing in the PRF have equal variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Why does hetero come into picture?

A
  1. Error learning models (num of errors decrease) 2. outliers 3. Model misspecification 4. Skewness 5. Incorrect transformation 6. data collection improvement
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is pure hetero?

A

When error term is a function of error term. Eg. of comparing spending changes in CA vs RI is affected by population size which acts as the proportionality factor

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is impure hetero?

A

Caused by model misspecification - omitted variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

WHat is a BLUE?

A

Best - min variance; Linear function of observed variable; Unbiased Estimator–> Expected value of OLS estimator is equal to the true value of parameters

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What does the gauss markov theorem say?

A

If linear model satisfies the CLRM assumptions then OLS estimators are BLUE. Expected value of disturbance is zero, Expected value of disturbance squared is a constant equal to varinace.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

If hetero exists then do we have BLUE for our model?

A

No, since they wont have minimum variance. We need to find BLUE

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Why cant we use OLS estimators in case of hetero?

A

Since OLS gives equal weights to all observations despite the difference in variability in populations we cant use the estimators. It becomes LUE since its not efficient for small and large samples. GLS will give BLUE

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Why use GLS?

A

GLS is OLS on the transformed variables that satisfy the standard least-squares assumptions. In GLS the weight assigned to each observation is inversely proportional to its πœŽπœŽπ‘–π‘– - i.e., the observations coming
from a population with larger πœŽπœŽπ‘–π‘– will get relatively smaller weight and those from a population with smaller πœŽπœŽπ‘–π‘–
will get relatively larger weight in minimizing sum of error squares.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What assumption are we making on residuals?

A

although residuals squared (sample) are not the same thing as the square of random error (population), they can be used as proxies, especially if the sample size is sufficiently large.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What are informal tests?

A

Plot residual squared with predicted Y and X to see if theres a pattern

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is the limitation of park test?

A

The error term introduced using the equation could bring heteroskedasticity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is the Park Test?

A

Park formalizes the graphical method by suggesting that πœŽπ‘–
2 is some function of the explanatory variable 𝑋𝑖

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is the glejser test?

A

We regress the absolute value of 𝑒𝑖 (estimated sample residual) on the 𝑋 variable related to the heteroscedastic variance 𝜎i squared.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Heteroscedasticity is likely to be more seen in cross sectional or time series data?

A

cross-sectional data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly