Machine Learning Flashcards

1
Q

Inductive learning

A

generalize from a given set of (training) examples so that accurate predictions can be made about future examples; learn an unknown function

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

how to represent a “thing” in machine learning

A

x: example or instance of a specific object; represented by a feature vector; each dimension - feature

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

feature vector representation

A

extract a feature vector x, that describes all attribute relevant for an object; each x is a list of (attribute, value) pairs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

types of features

A

numerical features - discrete or continuous
categorical features - no intrinsic ordering
ordinal features - similar to categorical but clear ordering

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

point in feature vector representation

A

each example can be interpreted as a point in a D-dimensional feature space, where D is the number of features/attributes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Training set

A

A training set is a collection of examples (instances), which is the input to the learning process; assume instances are independent and identically distributed. training set = experience given to learning algorithm

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

idd

A

independent and identically distributed

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Unsupervised learning

A

training set = x1, …xn; no “teacher” to show how examples should be handled; tasks: clustering, discovery, novelty detection; dimensionality reduction

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

goal of clustering

A

group training samples into clusters such that examples in the same cluster are similar, and examples in different clusters are different

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Clustering methods

A

Hierarchical Agglomerative Clustering
K-means Clustering
Mean Shift Clustering

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Hierachical Clustering General Idea

A

initially every point is in its own cluster
find the pair of clusters that are the closest
merge the two into a single cluster
repeat
end result: binary tree

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

How to measure closeness between 2 clusters (hierarchical clustering)

A

Single linkage, complete-linkage, average linkage

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

single-linkage

A

the shortest distance from any member of 1 cluster to any member of another cluster

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

complete linkage

A

the largest distance from any member of 1 cluster to any member of another cluster

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

average linkage

A

the average distance between all pairs of members, one from each cluster

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How to measure the distance between a pair of examples?

A

Euclidean, manhattan/city block, hamming

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Dendrogram

A

binary tree resulting from hierarchical clustering; the tree can be cut at any level to produce different numbers of clusters

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What factors affect the outcome of hierarchical clustering?

A

features used, range of values for each feature, linkage method, distance metric used, weight of each feature

19
Q

K-Means Clustering

A

Specify the desired number of clusters and use an iterative algorithm to find them

20
Q

K-means clustering general idea

A

if have cluster centers, for each point, choose closest center, if have points, choose a cluster center to be the mean/centroid of the points in the cluster; repeat until convergence

21
Q

K-means algorithm

A

input: x…xn
select k cluster centers c1, …ck
fore each point x, determine it’s cluster by finding closest cluster center; update cluster centers as centroid (mean)

22
Q

Distortion

A

Sum of squared distances of each data point to its cluster center; optimal clustering minimizes distortion (over all possible possible cluster locations/assignments)

23
Q

How to pick k in k-means clustering

A

pick number of k to minimize distortion - pick one close to elbow of the distortion curve

24
Q

Does K-means always terminate?

A

Yes (finite number of ways of partioning a finite number of points into k groups)

25
Q

Is K-means guarunteed to find “optimal clustering”?

A

No, but each iteration will reduce the error/distortion of the clustering

26
Q

Which local optimum k-means goes to is determined by:

A

solely by the starting cluster centers

27
Q

Ways to pick good starting cluster centers

A
  1. run multiple times with different starting random cluster centers
  2. pick starting points based on farthest apart from each other
28
Q

How to pick number of clusters

A

difficult problem, heuristics depend on number of points/dimensions

29
Q

mean shift clustering

A

choose search window size and initial location; compute centroid; center search window at mean; repeat

30
Q

mean shift clustering objective:

A

find the densest region

31
Q

Supervised learning

A

learns a function H: x->y in some function family H, such that h(x) predictions the true label y on future data (classification if discrete, regression if continuous)

32
Q

what is a label (supervised learning)?

A

the desired prediction on an instance x

33
Q

what is a class?

A

discrete label

34
Q

concept learning

A

determine from a given set of examples if a given example is or is not an instance of the concept/class/category

35
Q

positive/negative example (concept learning)

A

if an example is or isn’t an instance of concept/class/category

36
Q

Supervised Concept Learning by Induction

A

Given a training set of positive and negative examples of a concept, construct a description that accurately classifies whether future examples are positive or negative

37
Q

Nearest Neighbor Classification

A

Save each training example in feature space; classify a new example by giving it the same classification as its nearest neighbor

38
Q

How to pick k for nearest neighbor?

A

split data into training/tuning sets
classify turning set with different k values
pick k that produces least turning-set error

39
Q

When doesn’t k-NN generalize well?

A

If the examples in each class are not well clustered

40
Q

How is distribution from training set important?

A

If train on set from different distribution from future data, will not be as accurate

41
Q

Inductive bias

A

Any knowledge created by generalization from specific facts cannot be proven true, only false; used when one function is chosen over another

42
Q

Inductive biases commonly used in ML?

A

restricted hypothesis space bias

preference bias

43
Q

Restricted hypothesis space bias

A

allow only certain types of h’s, not aribitrary ones

44
Q

preference bias

A

define a metric for comparing h’s so as to determine whether one is better than another