Denisty Estimation + Mean shift, Markov Random Fields,Feature and corner detection Flashcards

1
Q

Limitations and beyond Kmeans

A

Limitations
1-Needs k to be provided
2-Non Optimal
3-Forces symetry

Beyond:
1-Create Density Function
2-Look For modes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Creating density function for segmentatiton

A

Starting point: set of samples
Approac: Density estimation

Density function:computionally complex

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Mean Shift

A

-Find stationary peaks without computing distribution function.
-models distribution using continuos nonparametric model

Avoid density function? Estimate gradient instead.

Meanshift steepest ascend method

Meanshfit vector same direction as gradient.

Procedure:
-Compute mean shift vector
-move kernel window
-stop if gradient close to 0
-prube mode by perturbation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Meanshift properties and optimization

A

Properties:
1-Automatic convergence speed
2-Steps are smaller towards mode.
3-Convergence depends on step size

Optimization
1-Divide space into windows and run parallelly

Pros
1.Doesnt assume cluster spherical
2.1 parameter window size
3- Find a variable nr of modes
4-robust to outliers

Cons
1-depends on window size -> hard to find the best, if bad modes are merged
2-computionally expensive
3-doesnt scale well with dimension of feature space

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Mean shift Clustering

A

Clustering criterion
1. Define attraction basin as the region for which trajectories lead to same mode.

  1. All data points in same basin = cluster

To perserve discontinuities we have a joint domain spatial+color

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Segmentation and optimization

A

1 additional approach -> segmentation is a per-pixel labelling task.

Labelling::
-Define a set of labels
-Define and energy function
-Assign a label to minimize the energy

Energy function:
1-Data term
Depends on specific image features at pixel location
Evaluates how good label fits pixel
Defines match criterion based on image feature vector.

2.Smoothing term
Depends on labels of neighbors.
Normally 4 neighbors considered
Favors same label at adjacen pixels.

Balance:
1-To detect strong boundaries: dont penalize with smoothing term
2-Weak boundaries like noise shouldnt affect labelling so smoothing term shouldnt be too weak.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Markov Random FIelds

A

The energy function evaluated over the whole
image is defined as:
𝐸 𝑓 = ෍
π‘βˆˆΞ©
𝐸data 𝑝, 𝑓𝑝 + ෍
π‘žβˆˆπ΄(𝑝)
𝐸smooth(𝑓𝑝, π‘“π‘ž)
* This model is called Markov Random Field (MRF)
* The model considers local interactions between
adjacent pixels
-Interactions are modeled considering a graph
Goal: select the labelling that minimizes the
energy function

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Optimizing Random Markov Fields using Belief Propagation

A

MRF-> minimization problem
Belief Propagation:algorithm to solve it

Based on message passing between neighboring pixels
Message has info about labels
Run several iterations

A message is a function mapping
– The ordered discrete set of labels – 𝐿
– Into an array of length π‘š having, for
each position, the message value
* Non-negative real values
– Message value for label 𝑙 sent from 𝑝
to π‘ž at time 𝑑
– Function of the label 𝑙 ∈ 𝐿

The reply of node 𝑝 depends on
– The data term at 𝑝 for label β„Ž (minimized)
– The smoothing term if the labels 𝑙 and β„Ž are different
– The information brought by the neighbors of 𝑝
(excluding π‘ž) in the previous iteration (time 𝑑 βˆ’ 1)
– Node π‘ž is excluded from the computation
* It cannot influcence the opinion of 𝑝 about π‘ž itself

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Features

A

Feature is a meaningful part of image.
Output: a set of points + description of the region (AKA
signature, fingerprint, …)

The ideal keypoints shall be
– Stable and repeatable
– Invariant to transformations (e.g., rotations)
– Insensitive to illumination changes
– Accurate

The ideal descriptor shall be
– Robust to occlusion and clutter
– Robust to noise, blur, compression, discretization
– Discriminative
– Stable over changes in viewing angle and illumination
– Computationally efficient (many features per image)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Harris Corners to detect salient points(corners)

A

Idea: Consider image and shifted version
if uniform region:two similar patches.
if salient point; different patches.

Corner = large difference.

Harris.
- Conisder a patch and a displacement.
- Similarity is measured with autocorrelation.

Autocorrelation matrix properties.
-real
-symetric
-orthogonal eigenvectors
-points to max data spread eigenvectors

Studying the eigenvalues we get information
about the type of patch
– If both eigenvalues are small: uniform region
– Only one large eigenvalue: edge
– Two large eigenvalues: corner

The Harris corner detector is
– Invariant to brightness offset:
– Invariant to shift and rotations
– Not invariant to scaling

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Beyond Harris Corners

A

Susan and Usan
-Analyze circular window around point
-no derivatives
-edge and corner detector
-robust to noise

USAN: Comparte between nucleus and pixels. The portion with intensity difference from nucleus given a threshold.

Susan i smalles usan.

Harris focuses on specific point.
OTHERS on BLOBS(diferent properties than surrounding).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

MSER

A

Connected areas with uniform intensity surrounded by contrasting background (blobs)
MSER DETECTOR IS SAME AS BLOB DETECTOR.

ALG:
Apply thresholds
compute connected binary regions
compute statistics
analyze how presisten is blob.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Scale Space Concept

A

Goal: analyze image at multi scales
Tool: scale space 0> original signal to family of derived signals.

-fine details are gradually suppressed.

Requirements:
Linearity
Spatial shift invariance
no new local extrema created
scale invariance

Gaussian smoothing does it

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Edge detection in scale space

A

Strong edges found at all scales.
At fine scales,accurate but many.
At coarse scales, inacurate but strong.
Combine them.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

SIFT Feature in detail
Keypoint detection
Descriptor calculation

Invariant to: translation, rotation, scale and other
transforms

Application
Matching and
registration
* Object recognition
* Query by example

A

-image content mapped into local fature cordinates

SIFT features:
– Local – robust to occlusions
– Distinctive – distinguish objects in large databases
– Dense – many features can be found even on
small objects
– Efficiency

Algorithm
* Scale-space extrema detection
* Keypoint localization
* Orientation measurement
* Descriptor calculation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

SIFT 1.scale space

A

-organized in octaves (from one to other sigma of gaussian is doubled)
(last image in one octave same smoothing as first one in next octave)

To detect extrema:
1-Layer substraction.
evaluate differences between images with same smooting but different smoothing intensity. (difference of gaussians)
Scale space + substraction of consecutive layers

2)Laplacian of Gausians ->smoothing + derivative filter.
removes noise
zero crossing
isotropic

17
Q

SIFT 2.Keypoint localization

A

-search for maxima and minima of DoG
-compares with 8 neighbors in current, previous and next scale level.
-done in all scales ->scale independent.

Subpixel accuarcy = interpolation

More scale levels considered
– More points
– More unstable
– Best value: s=3

Discard edge points – require that both
eigenvalues of 𝐻 are large
* Means: require that both
curvatures are high
* Similar to the discussion about
Harris corners
* Similar considerations about
eigenvalues ratios

18
Q

SIFT 3.Orientation measurement

A

-each keypoint has its scale.
-select that gaussian for that scale. (scale invariance)
-compute magnitude and orientation

Rotation invariance: the dominant
local direction shall be measured
36 bins of 10Β° for histogram. Look for 2 peaks. Fit parabolas.

19
Q

SIFT 4. Descriptor calculation

A

SIFT evaluates a descriptor for each keypoint
– In the image 𝐿 corresponding to the keypoint scale
– Considering coordinates that are rotated based on
the measured keypoint orientation.

Compute gradient
magnitude and direction

Values weighted with a
gaussian function
centered on the
keypoint

Evaluate histogram of
each region
– 8 bins (45Β° each)
* Save the values

  • Additional details are considered by the SIFT
    algorithm
  • Enhanced invariance to illumination: feature
    vector normalized to unit length
  • Enhanced robustness to large gradient
    magnitudes: threshold to 0.2 on all the
    components
    – Nonlinear illumination effect – e.g., camera saturation
    – Further normalization to 1 after thresholding
20
Q

SIFT robustness

A

Noise:random rotations
Affine transfor: Random scale change + 2% noise

21
Q

Beyond SIFT

A

PCA
-reduces dimensionality
-max variance of projected data
-min mean squared dist

Principal component #1 points in the direction
of largest variance
* Each other principal component
– Is orthogonal to the previous ones
– Points in the direction of largest variance of the
residual subspace

22
Q

BEYOND SIFT 2.
PCA SIFT
SURF

A

PCA SIFT.
1-3 same as sift
4->refer to 41 41 patch normalized to a canonical direction

SIFT: weighted histograms
* PCA-SIFT: concatenate horizontal and vertical
gradients (39x39) into a long vector
– Length: 2 directions x 39x39 = 3042
* Normalize the vector to unit norm
* Reduce the vector using PCA
– E.g., from 3042 to 36.

SURF
Speed-up
computations by
fast approximation
of hessian matrix
and descriptors
using integral
images
Integral image 𝐼Σ(π‘₯, 𝑦)
– Each pixel represents
the sum of all pixels in
the rectangle between
A box filter is composed of black and white rectangles
– White: positive weight
– Black: negative weight
* Box filters exploit the integral image
– Calculation is constant-time irrespective of the filter size.
Keypoints: compute the Hessian matrix in scale
space, find the maximum of the determinant

  • Faster than SIFT
  • Less robust to illumination and viewpoint
    changes WRT SIFT
23
Q

BEYOND SIFT 3. GLOH
BRIEF
ORB

A

Gradient Location-Orientation Histogram
* Obtained (again) modifying the 4th step of the
SIFT algorithm
* Computes SIFT descriptor using a log-polar
location grid.

PCA used for reducing dimensionality to 128

BRIEF
Based on:
– Gaussian smoothing
– Pairs of pixels compared inside a window
– Build a vector with comparison output
* Vectors compared using the Hamming distance (XOR)
Random pattern provides the best performance
– Best solution: isotropic gaussian distribution

  • Oriented FAST and Rotated
    BRIEF
  • FAST corner detector
  • Add rotation invariance to
    BRIEF
  • Orientation assignment
    based on the intensity
    centroid WRT the central
    pixel