Classification Flashcards

1
Q

What are the two stages of classification?

A

Inference stage: use training data to learn a model for P(Ck|X).
Decision stage: use posterior probabilities to make optimal class assignement.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What are three approaches to solving classification problems?

A

Generative, Discriminative and Discriminant-based models.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are generative classification models?

A

These are approaches to solving classification problems consisting in modeling a distribution (P(X|Ck) or P(X, Ck)) and deduce P(Ck|X) from it. The class can then be determined using decision theory.
N.B.: It is called “generative” because we can use the learnt distribution (P(X|Ck) or P(X, Ck)) to generate synthetic data in the input space.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What are discriminative classification models?

A

These are approaches to solving classification problems consisting in directly model P(Ck|X) and then use decision theory tp determine the class.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What are discriminant-based classification models?

A

These are approaches to solving classification problems consisting in finding a function f(X) (called “discriminant function”) which maps X to class labels.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What are lineary separable classification problems?

A

These are classification problems that can be solved by a linear model, e.g. whose classes can be separated by hyperplans.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is the 1-of-k coding scheme (a.k.a. “one-hot encoding”)?

A

It is the encoding of the target variable of a classification problem into a vector of length [number of classes] which components are all 0 except the one corresponding to the class the datapoint belong to.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What are discriminant functions and linear discriminants?

A

Discriminant functions are functions mapping an input vector X to a class Ck.
Linear discriminants are discriminant functions which decision surfaces are hyperplanes.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the geometrical interpretation of y(x) = w.T * x + w0 as a linear discriminant function?

A

-The decision surface is orthogonal to w.
-w0 can be considered as a threshold and controls the distance of the decision surface to the origin (which is equal to - w0/|w|).
-The signed orthogonal distance to the decision surface of a datapoint X is y(x)/|w|.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is a One-vs-the-rest classifier?

A

It’s a multi-class classifier consisting in a set of k-1 classifiers, each of which is solving a two class classification problem by separating points that are in Ck from points that are not.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is a One-vs-one classifier?

A

It’s a multi-class classifier consisting in a set of k(k-1)/2 discriminant functions, each of which discriminates one of every possible pair of classes. Majority voting is then used for classification.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is a k-class discriminant?

A

It’s a discriminant comprising k linear functions of the form yi(x) = wi.T * x + wi0.
x belongs to the class k such that yk(x) is the maximum among all yi(x) values.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly