Lecture 12 Flashcards

1
Q

how many predictors and response variable are in simple/univariate linear regression?

A

one predictor and one variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

how many predictors and response variable are in multiple linear regression?

A

more than one predictor variable and one response variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

why is it called multiple linear regression?

A

we have p>1 predictors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

why the model is called regression?

A

we are modelling a response variable (y) as a function predictor variables (x1,…xp)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

what is the relationship between Xj and Y in multiple linear regression model?

A

the relationship is linear

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

is collinearity beneficial amongst explanatory variables?

A

no, it can complicate or prevent the identification of optimal set of explanatory variables for a statistical model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

what could be the source of multicollinearity?

A

if two variables have high variability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

what is another way of detecting multicollinearity?

A

to estimate the multiple regression and then examine the output carefully

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

how VIF value is linked to multicollinearity?

A

• Calculate VIF as an indicator of multicollinearity. The larger the value of VIFj, the more “troublesome” or collinear the variable Xj. As a rule of thumb, if the VIF of a variable exceeds 10, that variable is said to be highly collinear

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

what degrees of freedom refer to?

A

the number of values involved in the calculations that have the freedom to vary

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the coefficient of determination in multiple regression?

A

The coefficient of multiple determination (R2) measures the proportion of variation in the dependent variable that can be predicted from the set of independent variables in a multiple regression equation. When the regression equation fits the data well, R2 will be large (i.e., close to 1); and vice versa

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

describe forward model

A

‘forward’ - it is done in the opposite way. Instead of doing the whole list, here, it will work with first variable and add the second one. Before adding the next variable, the significance will be checked. Insignificant variables will be dropped.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

describe backward model

A

‘backward’ - the model will take all the list of the variables , check which one is insignificant and drop the one which is most insignificant. It will do this one by one. We will end up with the list of significant variables only

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

what AIC stands for

A

Akaike information criteria

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

what BIC stands for

A

Bayesian information criteria

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

why the multivariate regression model is called multivariate?

A

because we have m>1 response variables

17
Q

why the multivariate regression model is multiple?

A

because we have p>1 predictors. if we have p=1, then the model is multivariate simple linear regression model