Support Vector Machines Flashcards
The basic idea
SVMs create a feature space, which represent the features of an object. Let’s say we want to classify images as dog vs cat.
Our features are weight, ear size, and nose type - these are our axis labels.
We then map all of our data points onto this vector space and see if we can draw a plane (line) that successfully separates dogs from cats.
Margin Classifier
This is where you draw your line / plane to classify your points.
You can use a:
maximal margin classifier (essentially the midpoint), sensitive to outliers - which can render this margin classifier pretty useless
soft margins - allow for a certain amount of misclassifications
Soft margin classifiers
AKA support vector classifier
How to we find the optimal one?
We use cross validation to determine how many misclassifications and observations to allow inside the soft margin
Observations on the edge and within the soft margin are called support vectors
Unique attributes for SVMs
They can handle outliers via using support vector classifiers / soft margins
They allow for misclassifications and can handle overlapping classifications
What to do if your soft margin has a ton of overlap?
SUPPORT VECTOR MACHINES!!
Support vector machines use some sort of y = f(x) to map your same data into a higher dimensional space. From there it sees if it can draw a line in that higher space that can successfully classify your points
Kernel Trick
The kernel trick reduces the amount of computation required for support vector machines by avoiding the math that transforms the data from low to high dimensions.