Quiz 5 Flashcards
True or False
In Support Vector Machines, we maximize (║w║^2)/2 subject to the margin constraints.
False
True or False
In kernelized SVMs, the kernel matrix K has to be positive definite.
True
True or False
If two random variables are independent, then they have to be uncorrelated.
True
True or False
Isocontours of Gaussian distributions have axes whose lengths are proportional to the eigenvalues of the
covariance matrix.
False
True or False
Cross validation will guarantee that our model does not overt.
False
True or False
In logistic regression, the Hessian of the (non regularized) log likelihood is positive denite.
False
Given a binary classification scenario with Gaussian class conditionals and equal prior probabilities, the
optimal decision boundary will be linear.
False
True or False
The hyperparameters in the regularized logistic regression model are η (learning rate) and λ (regularization
term).
False
The Bayes risk for a decision problem is zero when…
the class distributions P(X|Y ) do not overlap and the the prior probability for one class is 1.
Gaussian discriminant analysis…
models P(Y = y|X) as a logistic function, is an example of a generative model and can be used to classify points without ever computing an exponential.
Ridge regression…
reduces variance at the expense of higher bias.
Logistic regression…
minimizes a convex cost function and can be used with a polynomial kernel.
In least-squares linear regression, imposing a Gaussian prior on the weights is equivalent to…
L2 regularization
In terms of the bias-variance trade-off, which of the following is/are substantially more harmful to the test error than the training error?
Bias
Loss
Variance
Risk
Variance
In Gaussian discriminant analysis, if two classes come from Gaussian distributions that have different means, may or may not have different covariance matrices, and may or may not have different priors, what are some of the possible decision boundary shapes?
1) hyperplane
2) a nonlinear quadric surface (quadric = the isosurface of a quadratic function)
3) the empty set (the classifier always returns the same class)