Linear and Multiple Regression Flashcards

1
Q

What are regression’s used for?

A

To predict an outcome from a predictor variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the equation of a straight line?

A

y=mx+c

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What does b0 represent?

A

The intercept

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What does b1 represent?

A

The slope

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What does the regression line aim to do?

A

Ensure the line of best fit produces the smallest amount of residuals.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What does the R^2 value tell us?

A

How good the model/regression is - the amount of variance explained by the model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What does the F ratio tell us about a linear regressionn model?

A

If it is significant.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What does the F ratio tell us about a linear regression model?

A

If the regression predicts a significant amount of the variation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

How do you find the difference between the observed and predicted Y scores? (SS residual)

A

Predicted Y scores using the equation minus the actual Y scores, then square the values to stop them cancelling each other out.
∑(Y’− Y)^2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

How do you find the total variance of Y scores in the data set? (SS Total)

A

Each data point minus the mean for all Y data points, added together.
∑(Y - M)2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What does the SS Residual tell us?

A

An estimate of the amount of variation that is NOT predicted by our regression

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

How do you work out the SS regression?

A

SS Total - SS Residual

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What does the SS Regression tell us?

A

An estimate of the amount of variance explained by the regression of the model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What affects the estimate of the SS regression?

A

Sample size and the amount of total variation in the sample.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

How do you work out the R^2 value using SS values?

A

SS regression divided by the SS total

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is another representation of SS regression?

A

SS m

16
Q

What is the F Ratio?

A

The ratio between two variances (predicted and error)

17
Q

How to work out the F ratio?

A

SS regression is divided by the degrees of freedom.

18
Q

What does the p-value mean?

A

It tells us whether the result is significant so we know whether to accept or reject the null hypothesis.

19
Q

What assumptions have to be met for a simple regression? (6)

A
  • Outcome must be continuous
  • Predictors must not have zero variance
  • All values of the outcome should come from a different person/item
  • The relationship we model in reality must be linear
  • Homoscedasticity
  • Residuals must be normally distributed
20
Q

What are the 2 types of multiple regression?

A
  • Forced Entry Regression

- Hierarchical Regression

21
Q

What is the adjusted R^2?

A

R^2 will be an overestimate of the real R^2 In the population so it is adjusted down to allow for the overestimation of R^2.

22
Q

What are the unstandardised (b) and standardised (beta) coefficients used for?

A

Unstandardised - used within any equation.

Standardised - allows us to make comparisons across the predictors.

23
Q

What is the difference between Forced Entry and Hierarchical regression?

A

In forced entry, all the predictors are entered into the analysis at once. In hierarchical regression, some variables are controlled for.

24
Q

How do we compare models in hierarchical regression?

A

We check to see if more variance is explained in the second model compared to the first.

25
Q

If the beta value is positive for a dummy variable, what does this mean?

A

The category coded as ‘1’ is higher (scores higher) than the category coded as ‘0’.

26
Q

What is multicollinearity?

A

When predictors are highly correlated with each other.

27
Q

What numbers cause alarm in the collinearity VIF column?

A

Anything close to 10. (Anything over 5 is quite problematic).

28
Q

How much of the sample is allowed to be over 2SD away from the mean?

A

About 5%.

29
Q

What should the Cook’s distance be?

A

All below 1.