SVM Flashcards

1
Q

Idea behind SVM?
Best Hyper-Plane

A

To maximize the gap between plane and point so that classification happens as widely as possible.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Why SVM?

A
  • Linear & Non-Linear
  • Classification, Regression & Outlier Detection (One Class SVM)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is One Class SVM?

A

Anomaly Detection technique

Separates majority class with minority using kernel trick.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

When SVM? (Real world examples)

A

Text classification, Image classification, Spam Detection, Handwriting identification, Gene expression analysis, Face detection and Anomaly detection

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is Hyperplane?

A

In a p-dimensional space, a hyperplane is a sub-space of p-1 dimensions.

It is the plane that separates the support vectors.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is Support Vector?

A

Points touching the decision boundary (classifier).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Maximum Margin Hyperplane/Hard Margin Hyperplane

A

Hyperplane whose distance from points on either side is maximum.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Advantage of SVM

A
  1. Robust to Outliers.
  2. Can also be used for Non Linear Classification
  3. Can also be used for Regression
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Why is the coefficient of the equation of the line of hyperplane on either side is 1?

A

Just to simplify the calculation because even if we take the values as K, we can divide the whole equation by K and get the eqn. with coefficient 1.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Soft Margin classifier / Support vector classifier

A

When the data is not perfectly separable, misclassifications are allowed and a penalty term is added allowing for a trade-off between a wider margin and few misclassifications.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

SVM Error

A

Margin Error + Classification Error
i.e., Hinge Loss + Regularization Term (Penalty)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is “C” in Hinge Loss?

A

C is a hyper parameter which strikes a balance between making the gap as big as possible and reducing mistakes for misclassification.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is Kernel in SVM?

A

Kernel is a mathematical function which is used to map lower dimensional feature space into higher dimensional feature space.

Examples:- Linear, Polynomial, Radial Basis Function (RBF) and sigmoid.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What happens with increase or decrease in “C” in SVM?

A

With increasing in C, the margin widens and with decrease in C, it become less tolerant of violations, and so the margin narrows.

C controls the bias-variance trade-off.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is the advantage of using a kernel rather than simply enlarging the feature space using functions of the original features?

A

Computational - kernel do not forms extra column for enlarged features, rather it transforms the existing features.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

SVM for more than two classes

A
  1. One vs one - Kc2 SVM - voting count
  2. One vs All - k SVM
17
Q

Is the concept of kernel only applicable to SVM?

A

No, We can apply nonlinear kernels in other classification technique as well. However, for historical reason, the use of kernel is widespread in context of SVM