Week 6 Flashcards

1
Q

Homoscedasticity

A

Simply put, homoscedasticity means “having the same scatter.” For it to exist in a set of data, the points must be about the same distance from the line, as shown in the picture above. The opposite is heteroscedasticity (“different scatter”), where points are at widely varying distances from the regression line.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

The volatility of the error term ?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

OLS variance

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Difference between error term and residuals

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

The variance of the OLS slope and intercept are defined as

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

The T distribution (also called Student’s T Distribution)

A

is a family of distributions that look almost identical to the normal distribution curve, only a bit shorter and fatter. The t distribution is used instead of the normal distribution when you have small samples (for more on this, see: t-score vs. z-score). The larger the sample size, the more the t distribution looks like the normal distribution.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

If the normality of a and b holds, then the standardized OLS estimates are distributed as:

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Hypothesis testing on OLS coefficients — we can test two different types of null hypothesis:

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Hypothesis testing table:

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Significance level

A

It should be clear from the discussion so far that whether we reject or do not reject the null hypothesis depends critically on α, the level of significance or the probability of committing a Type I error, the probability of rejecting the
true hypothesis.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Why is an alpha level of .05 commonly used?

A

Seeing as the alpha level is the probability of making a Type I error, it seems to make sense that we make this area as tiny as possible. For example, if we set the alpha level at 10% then there is large chance that we might incorrectly reject the null hypothesis, while an alpha level of 1% would make the area tiny. So why not use a tiny area instead of the standard 5%?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Type I error

A

Supporting the alternate hypothesis when the null hypothesis is true.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Type II error

A

Not supporting the alternate hypothesis when the alternate hypothesis is true.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Alpha Levels / Significance Levels: Type I and Type II errors example

A

In an example of a courtroom, let’s say that the null hypothesis is that a man is innocent and the alternate hypothesis is that he is guilty. if you convict an innocent man (Type I error), you support the alternate hypothesis (that he is guilty). A type II error would be letting a guilty man go free.

An alpha level is the probability of a type I error, or you reject the null hypothesis when it is true. A related term, beta, is the opposite; the probability of rejecting the alternate hypothesis when it is true.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Null hypothesis

A

Null hypothesis, H0: The world is flat.
Alternate hypothesis: The world is round.
Several scientists, including Copernicus, set out to disprove the null hypothesis. This eventually led to the rejection of the null and the acceptance of the alternate. Most people accepted it — the ones that didn’t created the Flat Earth Society!. What would have happened if Copernicus had not disproved the it and merely proved the alternate? No one would have listened to him. In order to change people’s thinking, he first had to prove that their thinking was wrong.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Solving the arbitraireness of choosing the alpha level

A
17
Q

The p-value as a measure of evidence against Ho

A
18
Q

Multicollinearity

A

Multicollinearity generally occurs when there are high correlations between two or more predictor variables. In other words, one predictor variable can be used to predict the other. This creates redundant information, skewing the results in a regression model. Examples of correlated predictor variables (also called multicollinear predictors) are: a person’s height and weight, age and sales price of a car, or years of education and annual income.

19
Q
A
20
Q

Decision rule

A
21
Q

F Test

A

a catch-all term for any test that uses the F-distribution. In most cases, when people talk about the F-Test, what they are actually talking about is The F-Test to Compare Two Variances. However, the f-statistic is used in a variety of tests including regression analysis, the Chow test and the Scheffe Test (a post-hoc ANOVA test).

22
Q

Goodness of fit test

A

used to test if sample data fits a distribution from a certain population (i.e. a population with a normal distribution or one with a Weibull distribution). In other words, it tells you if your sample data represents the data you would expect to find in the actual population.

23
Q

Adjusted R-squared

A
24
Q
A