W5: RQ for Predictions 2 Flashcards

1
Q

What is a full regression equation involving 2 IVs

A

Yi = a + b1X1i + b2X2i + ei

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is a model regression equation involving 2 IVs

A

Y^i = a + b1X1i + b2X2i

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is an intercept in a regression equation. What is it signified by

A

a. Expected score on DV when all IVs = 0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is a partial regression coefficient in a regression equation. What is it signified by

A

b1, b2. Expected change in IV variable for each unit change in an IV, holding constant scores on all other IV

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the Sum of Squares in a regression equation. What is SS Total Indicated by

A

SStotal = SSreg + SSres

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is the aim of OLS. Use sum of squares to explain

A

Maximize SSreg; Minimize SSres

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is R^2.

A

Effect size measuring strength of prediction.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Formula of R^2. And what does it mean? What is another name for R^2

A

R^2 = SSreg/SStotal.

R&2 is the proportion of SStotal account for by SSreg.

Another name: Coefficient of Determination

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the range of R^2

A

0 to 1. Closer to 1 = Stronger

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

To calculate a confidence on R^2, what are the 4 things we require

A
  1. ) Estimated R^2 value
  2. ) Numerator df (no. of IVs)
  3. ) Denominator df (n-IVs-1)
  4. ) Desired confidence level
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Typically what is the biasness/consistency of R^2

A

IT is often biased, but consistent

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

How is an adjusted R^2 better than R^2

A

It is usually less biased, but we should always report both values

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

How do we make meaningful comparison between IVs in a multiple regression

A
  1. ) Transform regression coefficient to STANDARDIZED PARTIAL regression coefficient (z-scores)
  2. ) Semi-Partial Correlations as effect size esimate
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Making meaningful comparison between IVs in a multiple regression: Transform regression coefficient to STANDARDIZED PARTIAL regression coefficient (z-scores)… When interpreting coefficients, what is the difference. When is this method useful?

A

Coefficients are interpreted in SD units

Example: One SD increase in variable X will increase in 0.5 SD decrease in variable Y.

Only useful when IV has an arbitrary scaling

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Making meaningful comparison between IVs in a multiple regression: Transform regression coefficient to STANDARDIZED PARTIAL regression coefficient (z-scores)… What is the intercept

A

Always 0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Making meaningful comparison between IVs in a multiple regression: Semi-Partial Correlations. Overview.

A

Correlation between DV and EACH focal IV, when effects of other IVs have been removed from focal IV.

17
Q

Making meaningful comparison between IVs in a multiple regression: Semi-Partial Correlations. What is the SQUARED semipartial correlation

A

Indicates proportion of variation in DV uniquely explained by each IV.

18
Q

What are the 4 statistical assumptions underlying the linear regression model

A
  1. ) Independence of Observations
  2. ) Linearity
  3. ) Constant variance of residuals/ homoscedastiscity
  4. ) Normality of Residual Scores

ILHN I love Honey Nougats

19
Q

Statistical Assumption Linear Regression: Independence of Observations. How do we meet this

A

Met usually as long as

  1. ) Scores are not duplicated for bigger sample
  2. ) Responses on one variable does not determine person’s response to another variable
20
Q

Statistical Assumption Linear Regression: Linearity. How do we meet this. There are 4 ways

A
  1. ) Scatterplot Matrix
  2. ) Scatterplot of Residual Scores
  3. ) Marginal Model Plots
  4. ) Marginal Conditional Plots
21
Q

Statistical Assumption Linear Regression: Linearity. Scatterplot Matrix

A

Defined by variables being assigned to rows (Y-Axis) or Column (X-axis)
Examine scatter plots where DV is on the Y-Axis.

Non-Linearity would be a U shape.

22
Q

Statistical Assumption Linear Regression: Linearity. Scatterplot of Residual Scores

A

Scatterplot of Residual Scores against

(1) Observed IV scores
(2) Predicted scores on DV

23
Q

Statistical Assumption Linear Regression: Linearity. Marignal Model Plot

A

Marginal model plots of scores on the DV (Y) against scores on each IV and on predict scores (on X)

24
Q

Statistical Assumption Linear Regression: Linearity. Marignal Conditional Plot. Why is it especiallyg good

A

It shows partial regression line after other IVs are partialled out.

25
Q

Statistical Assumption Linear Regression: Homoscedasticity, What are the 2 ways

A

1,) Residual Plots (Similar to linearity)

2.) Breusch-Pagan Test

26
Q

Statistical Assumption Linear Regression: Homoscedasticity, Residual Plots

A

Even distribution of residuals

27
Q

Statistical Assumption Linear Regression: Homoscedasticity, Breusch-Pagan Test Overview

A

Null-Hypothesis Test that assumes the variance of residuals are constant/homoscedastic

28
Q

Statistical Assumption Linear Regression: Homoscedasticity, Breusch-Pagan Test. What happens when the P value is small

A

We have reason to reject the assumption of constant variance for the residuals

29
Q

Statistical Assumption Linear Regression: Normality of Residuals. What are the 3 ways

A
  1. ) qqplot
  2. ) Histogram
  3. ) boxplot
30
Q

Statistical Assumption Linear Regression: Normality of Residuals. qqplot. What does it contain

A

Middle line: Strict normality
Side Lines: confidence

See residuals inside confidence

31
Q

Statistical Assumption Linear Regression: Normality of Residuals. qqplot. What must we look out for

A
  1. ) Outliers: Vary large studentized residual value in absolute terms (High standard deviation)
  2. ) influential Case: If removed from analysis, results in regression coefficient changing notably in value
32
Q

Statistical Assumption Linear Regression: Normality of Residuals. qqplot. How do we find influential cases

A

Cook’s D as a measure of overly influential values.

> 1: problem.