MODULE 2 S2.3.1 Flashcards
Linear Model for Classification (Logistic Regression)
T/F In Naïve Bayes, training and testing are fast and easy to understand and process.
TRUE
3 Kinds of Naïve Bayes Classifier in Scikit-learn
GaussianNB - continuous data
BernoulliNB - binary data, text data
MultinomialNB - integer count data, text data
Classification algorithms in linear models
- How well weights and intercepts fir the training data
- Cost function or loss function
- Regularization methods can be used
Logistic Regression is a ______________ algorithm, not a regression algorithm.
classification
Default C value:
1
T/F Using a higher value of C stresses the importance that each individual data point be classified correctly.
TRUE
______________ and ____________ are a text-like used to count sparse data.
BerniuolliNB
MultinomialNB
T/F In Naïve Bayes, Larger alpha decreases the complexity of the model but does not change the performance.
TRUE
______________ is a high-dimensional dataset.
GaussianNB
It is a family of classifiers that are quite similar to the linear models.
Naïve Bayes Classifier
T/F Logistic Regression is also a regression algorithm.
FALSE
A regression where the dependent variable is in an ordered state (i.e., ordinal). The dependent variable (y) specifies an order with two or more categories or levels.
Ordinal Logistic Regression
Linear SVM is implemented in svm.LinearSVC.
SVC stands for
Support Vector Classifier
It predicts the relationship between the independent and binary dependent variables.
Binary Logistic Regression
For LogisticRegression and LinearSVC the trade-off parameter that determines the strength of the regularization is called ______.
C
A regression type that has more than two possible outcomes.
Multinomial Logistic Regression
It is a statistical technique used to develop predictive models with categorical dependent variables having dichotomous or binary outcomes
Binary Logistic Regression
T/F The reason that Naïve Bayes are so efficient is that they learn parameters by looking at each feature individually and collect simple per-class statistics from each feature.
TRUE
The two most common linear classification algorithms
Logistic Regression
Linear Support Vector Machines (Linear SVMs)
In Naïve Bayes, generalization performance is ___________________.
slightly slower
A categorical dependent variable has two or more discrete outcomes in a __________ regression type.
Multinomial
Multinomial Logistic Regression
T/F Naïve Bayes works well with sparse high-dimensional datasets and is not parameter sensitive.
TRUE
Higher values of C correspond to _________ regularization.
less
In Naïve Bayes, training speed is ____________ than linear classifier.
faster
T/F Using low values of C will cause the algorithms to try to adjust to the “majority” of data points.
TRUE