Lecture 6 - Binary Logistic Regression Flashcards
Logistic Regression is…?
….a modified form of the linear regression framework we have learned about so far. Logistic regression modifies the output from the linear regression model to transform it from a linear line to a sigmoidal curve, which has bounds of 0-1. Normal linear regression does not have bounds and in theory can run to infinity in either direction.
The point of logistic regression is…?
…to model a criterion variable which is binary i.e. one that can only take on one of two values. Just like in linear regression, the predictor variables can be continuous or categorical, and multiple predictors are fine.
The output from a logistic regression is usually interpreted as…?
…providing a probabilistic ‘guess’, based on the predictor variable scores, as to whether the criterion variable will be a 0 or a 1. So, if for a given person the output of the model is 0.9, this can be considered a 90% guess that that individual has a ‘1’ on the criterion variable.
In the SPSS output, where is the difference in -2LL between the model and model without predictors?
In the Omnibus table, under chi-square.
In the SPSS output, where is the equivalent of SSE-left, but for -2LL, between the model and model without predictors?
In the Model Summary table, under -2 log-likelihood.
What is the Likelihood ratio?
The difference in -2LL between the model and model without predictors.
What is the name of the equivalent for Likelihood ratio in linear regression?
SSE-reduced.
Using the SPSS output from Logistic regression. How can you calculate the -2LL of the model with no predictors (the equivalent of SSE-total)?
Add the likelihood ratio from the Omnibus table with the -2LL (SEE-left equivalent) from the Model Summary table.
What is the -2LL?
-2 Log Likelihood. It is -2 multiplied with the Log likelihood.
Why do we use -2LL and not only the Log Likelihood?
Because the -2LL has a known distribution, the chi-square distribution, making it possible to compare two values of -2LL and calculate a p-value for the difference between these two values.
How do you calculate Hosmer & Lemeshow’s r^2?
(Likelihood ratio) / (-2LL of the model with no predictors )
What does the Likelihood ratio divided by -2LL of the model with no predictors calculate?
Hosmer & Lemeshow’s r^2
What does Hosmer and Lemmeshows r^2 mean?
What proportion of -2LL we reduced by including predictors in the model
In the “Variables in the Equations” table, what does the “B” column denote?
The b0(bottom) and b1(top) of the model.
What is the “Wald” number/value and how is calculated in logistic regression?
Wald is a measure that can be thought of as the equivalent to the t-value in linear regression. It compares a predictor’s b1 to the predictor’s SE (being the “b1” of the no-predictor model. This generates a p-value. Wald is calculated by:
(b1 for the predictor / SE for the predictor) ^2 = Wald