Week 4 - Segmentation and Clustering p1,2 Flashcards
What is the goal of Grouping
Gather features that belong together for efficiency of further processing
Obtain an intermediate representation that compactly describes key parts
What is the Gestalt principle
That grouping is key to visual perception
Elements in a collection have properties that result from relationships
(negative space between features)
What are the gestalt factors
not grouped ( line of identical dots, spaced evenly)
Proximity
Similarity
Common fate (Elements that move in the same direction belong together)
Common region
Symmetry
Closure
Continuity
Parallelism
How do gestalt principles work with computer vision
It is difficult to know how best to apply these ideas via algorithms
What is Top-down segmentation
pixels belong together because they are from the same object
What is bottom-up segmentation
pixels belong together because they look ‘similar’
We define this similarity: colour, texture, gradient
How do we measure segmentation success
It is very hard to measure the success of segmentation; it depends on the application
Compare to human segmentation or to “ground truth”
- But often there is no objective definition segmentation
- Not even agreement between humans
What are superpixels
Compact and perceptually meaningful groups of pixels in an image
The idea is to not even try to compute a “correct” segmentation
Lets be content with over-segmentation (more superpixels less big regions)
- there will be the lines that follow the correct segmentation
How do we segment via clustering
Using a pixel Intensity histogram
Often we will get curves (peaks) at different respective grey levels (objects)
We use clustering to determine the main intensities
How do we use sum of squared differences in clustering
Σclusters i Σpoints p in cluster i ||p - ci||²
We want to minimise SSD between all points p and their nearest centre ci
What is the chicken and egg problem in clustering
Know the cluster centre → can allocate correct points to groups
know the group memberships → can get the centres of each cluster
What is k-means clustering
1) Pick random k cluster centres
2) determine group members using distance to centre
3) calculate average of each group and update ci
Will always converge to some local minimum but not necessarily the global minimum of SSD||p - ci||²
What is Feature Space
A very important factor
Depending on what we chose, we will group pixels in different ways
For example,
Intensity similarity in 1D (grayscale)
Or 3D: (r,g,b)
Or 5D: (r,g,b,x,y)
This encodes both similarity and proximity
What are the pros of K means
Simple, fast to compute
Converges to local minimum of within-cluster squared error
What are the cons of k means
Have to chose a k
Sensitive to initial centres
Sensitive to outliers
Detects spherical clusters only
What is a ‘hard assignment’
Assigning each data point to exactly one cluster/group
It is something we want to avoid