Week 10 Flashcards

1
Q

What are the 5 regression assumptions?

A

Ass 1: For fixed values of the independent variables, means of error is 0
Ass 2: Variance and standard deviation of independent variables always constant (no fan shape shit) - only homoscedasticity
Ass 3: For any values of independent variables, the dependent variables and errors are normally distributed (bell curved)
Ass 4: Errors are probabilistically independent thus provide no information on the values of other errors - However, problematic assumptions with time series data due to autocorrelation
Ass 5: No independent variable can be EXACT linear combination of any other independent variables due to Exact Multicollinearity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is Exact Multicollinearity?

A

When at least one of the independent variables is redundant and not needed

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is Multicollinearity?

A

When independent variables are highly correlated but not exactly which causes estimates to be unreliable and difficult to interpret

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the inferences about the regression coefficients?

A

The challenge of deciding which independent variables to include since including more will increase R Square
Check Adjusted R Square (Higher) and Standard Error of Estimates (Lower)
Relevancy and data availability will also came in play

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is Parsimony Principle?

A

Explaining the most with the least, using the least independent variables to explain the dependent

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

How do we estimate parameters?

A
  1. Point of estimates -> use the statistics as the estimate parameters (different sample, different point of estimates)
  2. Find confidence interval with t-value
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What are the two different standard error of estimate?

A

One is on top of the output, in the “Regression Statistics”
This measure the error when using the regression equation to predict the dependent variable

One is at the bottom of the output, in the variables tables
This measure the accuracy of each independent variable coefficient

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

How do we calculate confidence interval?

A

b +- t-multiple x Sb

Where:
b is the coefficient of intercept/independent variables
t-multiple is the number you get from t-tables, see from the “degrees of freedom - 1” and confidence seek (0.025 for 95%)
Sb is the standard of error of intercept/independent variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is t-values

A

Ratio of estimated coefficient to its standard error, above 4 / 1 is significant

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the important hypothesis test for a regression coefficient?

A

You want to decide if a particular independent variable belongs in the regression equation.
If the p-value < 0.05, reject null (significant)
If the p-value > 0.05, eliminate variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is ANOVA Test

A

Analysis of Variance, explain variation in dependent variables
The larger the F-ratio is, the better since it’s more explainable
If ANOVA (Significant F) < 0.05, reject null (significant), more than 1 variable can explain the changes in dependent variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is multicollinearity?

A

When there is a fairly strong linear relationship among a set of explanatory variables -> highly correlated, prolematic
Usual signs:
“wrong” signs show that they are not significant
t-values too small
p-values too large
impossible to predict and causes unreliable estimate

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

How to test for multicollinearity?

A

VIF and R Square
VIF (Variance Inflation Factor) -> 1/(1-R Square)
If:
VIF = 1 -> No Correlation
VIF > 4 -> Some Problem
VIF > 10 -> Serious multicollinearity problem

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

How to decide to include/exclude variables

A
  1. t-values and p-values associated, if p-values > 0.05, exclude
  2. Check t-values < 1, exclude
  3. low t-values & high p-values, exclude
  4. economic or physical theory (common sense)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

When can we predict revenues?

A

When the R Square is large and Se (Standard of Error) is small

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What are the 2 types of prediction?

A

Point Prediction: Substitute the given values of the X’s into the est. regression equation
Prediction Interval: T-multiple of the relevant standard error on either side of the Point Prediction