MCQ Flashcards
Which statement is true about 1-vs-1 classification?
It needs as many models as there are class pairs, and each model predicts which class an observation is more likely to belong to
Which statement is true about 1-vs-all classification?
It needs as many models as there are classes, and each models predicts the probability of an observation to belong to a class
In a binary classification setting, accuracy is…
The ratio of number of correct predictions to the number of observations
An advantage of using the bootstrap method over other model selection methods is….
It uses fewer data points for training and so each model is more accurate
The classical estimate of the model error obtained using the bootstrap method is
It overestimates the real error, because each model is trained on average on 63.2% of the data set
What are typical assumptions of the error term ε, when the relation between inputs and
output is written as y = f(X) + ε (where y is the output, X is the input, and f is the
real relationship between input and output)?
It’s expected value is zero
Which one of the following about the F1 score is true?
It is close to zero for models bad at discriminating the positive class
Given data with 3 numeric features, 2 categorical features and 1 numeric label, how many subset of features should I consider, if I do feature selection via an exhaustive search of all possible combinations?
32
What is the synonym of “feature”?
Independent variable
What is true about forward stepwise feature selection?
It starts with considering no features, just predicting mean label
The gradient descent method
Moves against the direction of gradient in every iteration
In the hold out validation method
We divide our data into large training set and small test set
After selecting a winning model in the hold out validation method, it’s a good idea to
Retraining the training+test data before using it in production
The estimate of the error of the model we get using the hold out validation method
Is an overestimation of the true model error because we train on fewer data points
The hyperparameters of a model
Are chosen by the user usually by hyper parameter tuning