MODULE 2 S2.2 Flashcards
Linear Regression, Ridge and Lasso
A constraint where each feature should have as little effect on the outcome as possible (which translates to having a small slope), while still predicting well.
Regularization
The ‘slope’ parameter is also called _______ or coefficients.
weight
Regression where the coefficients (w) are chosen not only so that they can predict well on the training data, but also to fit an additional constraint.
Ridge Regression
T/F Linear Regression is also known as Ordinal Least Squares.
FALSE (Ordinary)
It is the sum of the squared differences between the predictions and the true values.
Mean Squared Error
It is a linear model for classification problems.
Logistic Regression
Logistic Regression happens by fitting a logistic function, also known as the _______________
sigmoid function
An alternative to Ridge for regularizing linear regression.
Lasso
It is also referred to as w, weights, coefficients.
Slope
T/F In Linear Regression, the final output of the model is numeric value (numeric predictions).
TRUE
The algorithm used for solving regression problems.
Linear Regression
It makes a prediction using a linear function of the input features.
Linear Model
T/F In ridge regression, a higher alpha means a more restricted model, so we expect the entries of coef_ to have smaller magnitude for a high value of alpha than for a low value of alpha.
TRUE
T/F When comparing training set and test set scores, we find that we predict very accurately on the training set, but the R2 on the test set is much worse. This is a sign of underfitting.
FALSE
T/F Ridge regression is a linear regression model that controls complexity to avoid overfitting.
TRUE
T/F Dichotomous classes means Yes or No.
TRUE
T/F In Ridge regression is α (alpha) is lesser, the penalty becomes larger.
FALSE
What regularization is lasso regression?
L1 regularization
What does the variable b represent?
Intercept / Offset
When comparing training set and test set scores, we find that we predict very accurately on the training set, but the R2 on the test set is much worse. This is a sign of:
Overfitting
Linear Regression is also known as _________________
Ordinary Least Squares (OLS)
It generates a formula to create a best-fit line to predict unknown values.
Linear Model
It means explicitly restricting a model to avoid overfitting.
Regularization
It generates a probability between 0 and 1.
Logistic Regression
They are called linear because they assume there is a __________ relationship between the __________ variable and each of its ________________.
linear
outcome
predictors
T/F The ‘offset’ parameter is also called slope.
FALSE (intercept / b)
It is one of the most commonly used alternatives to standard linear regression.
Ridge Regression
Types of Linear Models
Linear Regression
Logistic Regression
Linear regression finds the parameters ___ and ___ that minimize the ______________ error between predictions and the true regression targets, ___ , on the training set.
w
b
mean squared
y
A model that is easy to analyze and understand.
Lasso
It is the simplest and most classic linear method for regression.
Linear Regression
What regularization is ridge regression?
L2 regularization
T/F Ridge is generally preferred over Lasso, but if you want a model that is easy to analyze and understand then use Lasso.
TRUE
T/F If α is larger, the penalty becomes larger and w should be smaller.
TRUE