Jon's Topics Flashcards

1
Q

Robust, Repeatable, Invariance and Constraints

A

Robust and Repeatable computer vision is achieved through engineered Invariance and applied Constraints. (We wants Robust and Repeatable, we achieve using Invariance and Constraints)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Feature Space

A

Abstract space defined by a feature extraction process. Transforming raw data into feature of some fixed number of elements (dimensionality of the space).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Distance Measures

A

Euclidean (L2) Distance: 2rt[SUMi=1->n (q_i-p_i)^2]

L1 Distance: SUMi=1->n ( | p_i-q_i | ). Where p,q = points

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Supervised ML: Classification

A

Binary - Linear: Learn a hyperplane that separates two classes with a minimum error. Non-Linear: Can learn curved lines to fit data, lose generality by overfitting.
Multiclass - KNN

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Unsupervised ML: Clustering

A

Aims to group data without any prior knowledge. K-Means: Guaranteed to find optimal results, doesn’t perform consistently. Trial and error to find best k value.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Image Features (4 types)

A

Global, Grid-based, Region-based, Local

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Global Features

A

Extracted contents from entire image, outputs single Feature Vector. Image Histogram is common. Joint Colour Histogram usually considers all pixel colour components together.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Segmentation

A

Used for segmenting image into regions for Region based features. Using: Thresholding - Basic, Otsu, and Local or Adaptive Thresholding. K-Means

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Otsu Thresholding

A

Otsu: Assumes image contains two classes of pixels (fore/background). Calculates optimum threshold to separate two classes with minimal intra-class variance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Local or Adaptive Thresholding

A

Computes a different threshold for each pixel, based on values of neighbouring pixels. E.g. Mean adaptive thresh sets pixel to BG if it’s value is less than mean of it’s neighbours plus an offset.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Segmentation using K-Means

A

Simple method: Cluster colour vector of all pixels, assign each pixel based on closest cluster centroid. Could spatially encode FV to help base segments on relative position of pixels, to avoid loner different region pixels.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Connected Components

A

Type of segment in which all pixels are reachable to each other through a path of spatially adjacent pixels. Commonly 4-connectivity or 8-connectivity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Connected Component Labelling

A

Takes binary image and produces a set of connected components. Uses 2 pass algorithm

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Simple Scalar Features

A
Area: #pixels in component
Perimeter: Length around component
Inner Border: Inner pixels forming edge
Outer Border: Outer pixels of edge
Compactness: Measure how tightly packed pixels are
Dispersion: How spread out shape is
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Moments

A

Describes Distribution of pixels in a shape. Standard 2D Cartesian Moments can be used as shape descriptors. Different shapes have different moments. Not invariant to scaling, translation, and rotation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Central Moments + Normalised Central Moments + Hu Moments

A

Central: Translation invariant as moments computed about centroid of component.
Normalised Central: Translation and Scale Invariant.
Hu: 7 scale, rotation and invariant moments.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Chain Codes

A

Simple way of encoding boundary of an object. Walk around boundary and encode direction on each step as a number. Cyclically shift code so it forms smallest integer value. Invariant to start point.

18
Q

Chain Codes Properties

A

Make rotation invariant by encoding differences in direction. Make scale invariant by resampling component to fixed size. Perimeter for 8-con chain code = #(Even numbers in CC) + Rt2 * #(Odd numbers in CC). Not good for shape matching due to noise, resampling and no good distance metric.

19
Q

Fourier Descriptor

A

Encode shape info by decomposing boundary into small set of frequency components. 1: Define representation of the curve (boundary). 2: Expand representation using Fourier Theory.

20
Q

Region Adjacency Graph

A

Build graph from set of connected components. Each node = component. Nodes connected if they share a border. Invariant to distortion, not to occlusion. Good for creating markers like QR Codes.

21
Q

Point Distribution Models

A

Learning a low dimensional parametric model of how the shape of an object can vary. Shape represented by number of fixed 2D points at distinguishable locations on the object. Needs training set of images with same points marked.

22
Q

Procrustes Analysis

A

Choose a reference shape, superimpose (align) all instances to current reference shape. Compute mean shape of the current set of images. If Euclidean distance between mean shape and reference shape is above a threshold, set reference shape to mean shape and repeat.

23
Q

Local Interest Point

A

Good Local Interest Point is invariant to global & local lighting change, invariant to difference in camera’s location. Detect using corner or blob detection.

24
Q

Corner Detector: Harris + Stephens

A

Search for corners through a small window. Corner found if shift in window by a small amount results in large change in intensity.

25
Q

Structure Tensor

A

Encodes the image derivatives in the x, y and xy directions. Is the second moment matrix. Symmetric matrix with M=[{Ew(Ix(xi,yi)^2), Ew(Ix(xi,yi)Iy(xi,yi)}, {Ew(Ix(xi,yi)Iy(xi,yi), Ew(Iy(xi,yi)^2)}]

26
Q

Laplacian Of Gaussian

A

Second Derivative of Gaussian. By finding local minima or maximum, you get a blob detector

27
Q

Narrow-baseline stereo, Wide-baseline stereo

A

Narrow: 2 images are very similar - local features have only been moved by a few pixels, applicable for video tracking. Wide: 2 images are not similar at all, applicable for generic matching tasks.

28
Q

Robust Local description

A

Narrow baseline: Robust to rotation & lighting not so important. Descriptiveness can be lower as search is over a smaller area.
Wide baseline: Need to be robust to intensity change, invariant to rotation. Highly Descriptive to avoid mismatches! Robust to small localisation errors of interest points.

29
Q

Matching by correlation

A

Use pixel values to describe a local region around an interest point. Basic template matching can be used to perform matching. Works well for regions with small differences, better in narrow baseline.

30
Q

Local Intensity Histogram

A

Represent local region with a histogram of pixel values describing the region around each interest point, as opposed to comparing raw pixel values. Might not be distinctive. Only rotation invariant if circular window is used. Not invariant to illumination changes. Sensitive to interest point localisation

31
Q

Localisation Sensitivity

A

Small changes in interest point. Overcome by weighting. Pixels at edge of sampling patch have less effect, those nearer have more effect. Commonly use a Gaussian weighting.

32
Q

Local Gradient Histogram

A

Encode gradient directions within a window instead of a raw pixel values in a histogram. Compute gradient orientations and magnitudes using partial derivatives of the image. Gradient Direction ‘Theta’, magnitude ‘m’

33
Q

Local Gradient Histogram properties

A

Gradient Magnitudes are invariant to brightness change. Not rotation invariant -> Find dominant orientation and cyclically shift so it is in first bin.

34
Q

SIFT Feature

A

Like Local Gradient Histogram, incorporates spatial binning. Creates multiple gradient histograms about the interest point and appends them to one feature vector. Usually spatial 4x4 grid of histograms with 8 orientations = 128-dimensional feature. Discriminative and robust!

35
Q

Feature Distinctiveness

A

Local feature too distinctive won’t match subtle variations due to noise. If it’s not distinctive enough it will match everything.

36
Q

Geometric Mapping

A

A transform function mapping x,y co-ords of points in one image to another. Used to constrain matches between two images

37
Q

Affine transform

A

Transform that retains parallel lines, but allows translation,scaling,aspect ratio,rotation,skewing. Usually: y=Ax+b where A is a 2x2 transform matrix encoding scale,rotation&scew, b=vector encoding translation. 6 degrees of freedom.

38
Q

Planar Homography

A

Adds homogeneous co-ords to deal with perspective (i.e. an object moving further away gets smaller).

39
Q

Recovering geometric mapping transform matrix

A

Solve set of simultaneous equations. 4 point matches for Homograph, 3 for affine.

40
Q

RANSAC

A

Robust Estimation algorithm to robustly estimate inliers of the model. Assume: M data items required to estimate model T. N data items in total.

41
Q

RANSAC Pseudocode

A
  1. Select M data items at random
  2. Estimate model T
  3. Calculate K = how many N data items fit T within a tolerance tol.
  4. If K is large enough, aaccept T, or computer least squares estimates using all inliers. Exit with success
  5. Repeat steps 1-4 N iteration times.
  6. Fail if no good T fit of data.