Lecture 6 Flashcards

1
Q

What causes irregular boundries?

A

Irregular distribution, imbalanced training sizes, and outliers

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What causes misclassifications?

A

unoptimized decision boundries

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Support Vectors

A

a subset of vectors that support or determine the boundry

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the goal of support vector machines?

A

to learn a boundary that leads to the largest margin (buffer) from points on both sides

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What points of the data set have an influence on the decision boundry (when using an svm)?

A

Only the support vectors. any point that isn’t a support vector has no influence and can be messed with while having no affect on the decision boundary.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What data points do SVMs use to compute predictions?

A

Only the support vectors, not the whole training set

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What dimension is a decision boundary for a dataset with 2 features?

A

It is the third dimension and creates a 2D plane

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Where is the decision function equal to zero?

A

On the decision boundary

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

When are inputs labeled as ‘undefined’?

A

When they lie between the margins i.e. between -1 and 1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the goal of decision functions for SVMs?

A

to maximize the margin between the data points and the hyperplane

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

How can optimal values of w and b be found?

A

through optimization via projective gradient descent

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What do your graph and results look like when parameters w and b are optimized?

A

The algorithm correctly classifies the training examples and the margin is maximized

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

In a margin-based classifier, what happens to the margin when the weight vector w gets smaller?

A

The margin gets larger

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

In hard margin SVMs, do we minimize ‘1/2 abs(w)^2’ or ‘abs(w)’ and why?

A

1/2 abs(w)^2 because abs(w) is not differentiable at w = 0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What do you do when the data is not linearly separable?

A

Introduce slack variables and allow “error” in classification

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What does the data have to look like for you to not be able to use a hard margin SVM?

A

A margin can’t cleaning split the data without leaving, for example, a blue piece on the red side. In other words, if blue pieces separate a red piece from the other red pieces.

17
Q

What are the two conflicting elements of soft margin SVMs?

A
  1. making the slack variables as small as possible to reduce the margin violations and
  2. making w^T * w as small as possible to increase the margin
18
Q

What does a kernel function map?

A

It maps the low dimensional data to high dimensional space

19
Q

What type of function is the kernel function?

A

a similarity function

20
Q

What’s the relationship between the weight vector w and the margin?

A

The smaller the weight vector w, the larger the margin

21
Q

When using RBF kernel in SVM what does a high Gamma value signify?

A

The model would consider only the points close to the hyperplane for modeling.

22
Q

Gamma parameter

A

The gamma parameter in SVM tuning signifies the influence of points either near or far away from the hyperplane. For a low gamma, the model will be too constrained and include all points of the training dataset, without really capturing the shape. For a higher gamma, the model will capture the shape of the dataset well.

23
Q

Disadvantages of SVMs

A

It is not suited to larger datasets as the training time with SVM’s can be high

It is less effective on noisier datasets with overlapping classes

It was originally designed as a 2-class classifier