Chapter 5 Flashcards

1
Q

Linear Regression and Linear Correlation

A

analyze the relationship between two quantitative variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Linear Regression

A

Used to analyze the relationship between two quantitative variables when one variable responds to the other, Explanatory and Response

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Linear Correlation

A

Used to analyze the relationship between two quantitative variables to determine whether a change in one variable is associated with a change in the other variable.
* So, the two variables are co-related.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Direction

A

Positive correlation: as one variable increases, the other also increases (r has + sign)
Negative correlation: as one variable increases, the other decreases (r has ─ sign)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Form

A

May be a straight-line relationship (linear) or curved (Here we only deal with linear)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Strength

A

The magnitude of r indicates the strength of the linear relationship between the two
variables.
r close to -1 or 1 indicates a strong linear relationship
r close to 0 indicates no relationship or a weak linear relation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Lurking Variable (=Extraneous Variable)

A

Variable that is hidden or not measured and that may cause a change in the measured variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Confounding Variable

A

Variables that have confusing effects on other variables, making it difficult to determine which might be the causal variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Regression equation

A

the equation of the regression line

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Analysis of Residuals

A

Regression tries to minimize the errors due to deviations not explained by the regression equation (least squares criterion)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Residual = error (e)

A

vertical distance from the regression line to a data point (may be + or –)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Residual Sum of Squares = Error Sum of Squares (SSE)

A

the variation in the observed values of the response variable that is not explained by the regression

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Least-squares criterion

A

Tries to minimize the Residual or Error Sum of Squares (SSE) in order to get the “best fit” line

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Predictor variable (= explanatory variable)

A

= x-variable, which can be used to make predictions about the other variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Response variable

A

y-variable, whose values respond to changes in the predictor variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Interpolation

A

using the regression equation to make predictions about the response variable, within the range of the observed values of x

17
Q

Extrapolation

A

using the regression equation to make predictions about the response variable, outside the range of the observed values of x

18
Q

The coefficient of determination (R2) = [correlation coefficient]2

A

the fraction or percentage of variation in the observed values of the response variable
that is accounted for by the regression analysis
Always: 0 ≤ R2 ≤ 1
OR 0% ≤ R2 ≤ 100%

19
Q

Population regression line (linearity) (Conditions) for Regression Inferences

A

The relationship between the two variables must be approximately linear. In other words, there are constants β0 and β1 such that, for each value x of the predictor variable, the conditional mean of the response variable is β0 + β1x.

20
Q

Equal standard deviations (homoscedasticity)
(Conditions) for Regression Inferences

A

The standard deviations of y- values must be approximately the same for all values of x

21
Q

Normal populations
(Conditions) for Regression Inferences

A

For each value of x, the corresponding y-values must be normally distributed)

22
Q

No Serious Outliers
(Conditions) for Regression Inferences

A

Significant outliers can drastically change the regression model

23
Q

Independent observations
(Conditions) for Regression Inferences

A

The observations of the response variable are independent of one another. This implies that the observations of the predictor variable not need to be independent.

23
Q

Independent observations:

A

The observations of the response variable are independent of one another. This implies that the observations of the predictor variable not need to be independent.