L2 - Types Of Learning Flashcards

(15 cards)

1
Q

Supervised learning can be used to map from input data to an output, what is the difference between Classification and Regression in this instance ?

A

In Classification the output is discrete values (e.g., “spam” or “not spam”, “cat” or “dog”).
In regression the goal is to predict a continuous numerical value. The output is a set of real numbers (e.g., price, temperature, age).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is Supervised Learning?

A

Training algorithms using labeled datasets, where each input is paired with the correct output. This allows the algorithm to learn the mapping from inputs to outputs, enabling it to make predictions or decisions when presented with new, unseen data.
Examples:
* Classification
* Regression

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is unsupervised learning?

A

Training algorithms using labeled datasets, unsupervised learning algorithms are tasked with finding patterns and relationships within the data without any prior knowledge of the data’s meaning.
Examples:
* Clustering
* Density Estimation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Give three examples of classification algorithms

A
  1. Spam Filtering
  2. Object Detection
  3. Weather Prediction
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is Clustering?

A

Grouping Data into clusters depending on similarity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is Classification?

A

Assigning classes to Data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is Regression?

A

mapping input to output by a set of real numbers.
Eg The relation between hand and eye data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Give examples of Regression algorithm use

A
  1. Stock market trend and prediction
  2. Machinery lifetime estimate
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Name the four classification methods covered by the module.

A
  1. Support Vector Machines
  2. Discriminant Analysis
  3. Naive Bayes
  4. Nearest Neighbour
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

name the five Regression methods covered by the module.

A
  1. Linear Regression GLM
  2. SVR GPR
  3. Ensemble methods
    **4. Decision trees
  4. Neural Networks**
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Name the five Clustering methods covered by the course.

A
  1. K-means, Kmedoids, Fuzzy C-means
  2. Hierarchical
  3. Gaussian mixture
  4. Neural networks
  5. Hidden markov model
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is K-means clustering?

A

A method of dividing observations into a given number of groups by determining the location of k group centers.
Then assigning each observation to a group according to which center the it is closest to.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

How do you determin k group centers in k-means clustering?

A
  1. Start with random centre locations
  2. Assign the data to the closest centres
  3. For each group, calculate the groups actual average centre and update the k centre to this location
  4. Repeat until the centres are fixed

5.(optional) repeat for new starting centres. For each solution calculate the total distance between data and their closest centre. The optimal solution will have the lowest total distance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Hierarchical Clustering

A

Hierarchical clustering is a method of cluster analysis that seeks to build a hierarchy of clusters. It’s commonly used in data mining and statistics to group similar objects into clusters based on their characteristics.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Gaussian Mixture Models

A

A Gaussian Mixture Model (GMM) is a probabilistic model used to represent a dataset as a mixture of several Gaussian (normal) distributions.

Instead of assigning each data point to a single cluster (like K-Means), GMM assumes that:

The data is generated from a mixture of several Gaussian distributions.
Each data point has a probability of belonging to each cluster.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly