Classification Flashcards
Describe the steps of Image classification
1) Feature extraction
2) Feature description
3) Classification
How can we classify images without labels
Use unsupervised clustering techniques
What do we mean by semi-supervised learning?
We only have labels for a part of the dataset
What is MNIST
A dataset of 70 000 28x28 pixel handwritten digits
Describe the typical pre-processing of digit recognition in general images
- Detect the digits in the large image
- Normalize the size of the digit, for example, to 28x,28 pixels
- Normalize the location, place mass center in the middle
- “Slant” Make the orientation canonical
Describe K-Nearest Neighbour algorithm
Classify by taking a majority vote among the K-nearest neigbhours
Name some distance measurements
L2 (Euclidean), L1(Manhatten)…
Name some advantages and disadvantages of K-Nearest Neighbour
- It works reasonably well
- No training required
- Nonlinear decision boundaries
- Multi-class
- All training data must be stored in memory
- Long evaluation time
Why shouldn’t we use the test set to tune hyperparameters? What should we do instead?
Tuning hyperparameters with the test set overfits the model to the test set. Use a validation set or cross-validation.
What’s the advantage and disadvantage of using cross-validation vs. a dedicated validation set
Cross-validation makes more data available for training but is computational more expensive.
What’s the formula of a general linear classifier
w*x + b = y
how can we use the formula of a linear classifier, wx + b = c to assign a class to the data?
in class (+1) if c>0, else not in class (-1)
What is considered the best hyperplane for SVMs?
The hyperplane that maximalizes the margin, the combined distance from the closest points in both classes to the hyperplane.
What points do we need to determine the hyperplane in an SVM?
Only the points closest to the hyperplane
Derive the formula for making the margin equal to 2 in an SVM.
𝒘 ∙ 𝒙𝟏 + 𝑏 = -1
𝒘 ∙ (𝒙𝟏 + 𝑚𝒏) + 𝑏 = 1.
m 𝒘 ∙ 𝒏 = 2
𝒏 = 𝒘/||𝒘||