For the Midterm Flashcards

1
Q

What is the variable Y called?

A

The response variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What are the X variables called?

A

The predictor, or explanatory variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is multiple linear regression?

A

It relates one numerical characteristic, the response variable, to one or more predictor or explanatory variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

In multiple linear regression, what do we typically assume about epsilon?

A

That the epsilon are independent, homoscedastic (same variability) and, if the sample size is small to moderate, that they are approximately normal in distribution.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What are regression methods used for?

A

1) Identify and characterize the relationships between the response and predictor/explanatory variables.
2) Estimate or predict the value of the response variable for combinations of the predictor/explanatory variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is the objective of time series analysis?

A

To identify patterns and trends, and to predict future observations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What does UCLM mean?

A

Upper confidence limit for regression line

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What does LCLM mean?

A

Lower Confidence Limit for Regression Line

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What does UCL mean?

A

Upper prediction limit

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What does LCL mean?

A

Lower prediction limit

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the simple linear regression model?

A

Y_i = B_o + B_i*x_i + epsilon

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

How do we denote the ith residual?

A

e_i

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

How are residuals found?

A

y_i - y_hat_i

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is the prediction interval?

A

The interval a new value is likely to be in

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is the confidence interval?

A

The interval we’re confident a value we already have is in.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What sort of hypothesis test do we run on the slope in multiple linear regression?

A

We test whether it’s equal to 0 or not! It we reject H_o, then the slope is not 0.

17
Q

What does the analysis of variance do?

A

Tests to see if there exist two slopes that equal each other, and if they equal zero.

18
Q

What does the SSE tell us?

A

How much of the model’s variance is due to random error.

19
Q

What does SSE denote?

A

Sum of Squares due to Error

20
Q

What does SSR denote?

A

Sum of Squares due to the Model (or our regression)

21
Q

What does the SSR tell us?

A

How much of our model’s variance is understood by the model.

22
Q

What does R-square tell us?

A

The proportion of our model’s variance that is understood by the model. (the closer to 1, the more accurate our model)

23
Q

Why do we care about partial regressions?

A

They help us to see the relationship between Y and a single X!

24
Q

What do we use partial plots to determine?

A

If there is a linear relationship between Y and each individual X (predictor variable)

25
Q

What is collinearity?

A

When one or more predictor variables are close to being a linear combination of the other predictor variables.

26
Q

What are symptoms of collinearity?

A
  • The regression coefficients have unlogical “signs” (- or +)
  • The regression coefficients are huge in magnitude, and have even larger standard errots
  • The individual coefficients are nonsignificant, but when grouped with other coefficients are significant.
27
Q

How can we diagnose collinearity?

A
  1. A correlation matrix (look for variables that seem to have a linear relationship.
  2. Regress each predictor variable on all the other predictor variables.
    • Look for high values of R-square
    • A Variance Inflation Factor ≥ 10 suggests collinearity.
  3. Look at the condition indices (large values and jumps in the values indicate collinearity.)
28
Q

How can we fix collinearity?

A
  • Backward elimination
  • Forward Selection
  • “Stepwise”?? Selection
29
Q

How does backward elimination work?

A

It takes out variables whose p-values are greater than alpha (usually .05)

30
Q

How does forward selection work?

A

It adds variables that 1) have the highest correlation with the current model, and 2) have a p-value less than alpha.