Segmentation and Clustering p2 - Lecture 8 - Week 4 Flashcards

1
Q

How are clusters modelled in Gaussian Mixture Models

A

As Gaussians, not just by their mean

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What are the parameters to the univariate Normal (Gaussian) Distribution?

A

The mean and the variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are the parameters to the multivariate Normal Distribution?

A

a vector containing the mean position
A symmetric “positive definite” covariance matrix

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What are the three types of covariance

A

Spherical, diagonal and full

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is a generative model?

A

Instead of treating the data as a bunch of points, assume that they are all generated by sampling a continuous function

This is defined by a vector of parameters theta

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

How can we model data with multiple clusters with multiple gaussians?

A

By starting with parameters describing each cluster
Mean muc, variance sigmac, “size” pic
This models a probability distribution of the cluster of:
p(x) = Sumc(pic * N(x ; muc, sigmac))

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is expectation Maximisation (EM)?

A

Goal:
FInal blob parameters theta that maximise the likelihood function:
P(data|theta) = PI(x)(x|theta)

Approach:
1. E-Step: Given current guess of blobs, compute ownership of each point
2. M-step: given ownership probabilities, update blobs to maximise likelihood function
3. Repeat until convergence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is expectation maximisation (EM) useful for?

A

Any clustering problem
Any model estimation problem
Missing data problems
Finding outliers
Segmentation problems
- Based on colour
- Based on motion
- Foreground/background seperation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Pros and cons of mixture of gaussians / expectation maximisation (EM)?

A

Pros:
- Probabilistic interpretation
- Soft assignments between data points and clusters
- Generative model, can predict novel data points
- Relatively compact storage

Cons:
- Local minima
- Initialisation
- Often a good idea to start with some k-means iterations
- Need to know number of components (number of clusters)
- Need to choose a generative model
- Numerical problems are often a nuisance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly