Class 6 Flashcards

1
Q

computational learning theory

A

lies at the intersection of AI, stats, and theoretical CS

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

sample complexity

A

number of required examples to get to probably approximately correct

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

approximately correct

A

a hypothesis that is consistent with low error rate after a large set of training examples

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

linear functions

A

“fitting a straight line”

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

linear regression

A

task of finding the best fitting line

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

weight space

A

all of the possible settings for the weights

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

alpha

A

step size, also called learning rate

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

epoch

A

step that covers all the training examples

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

decision boundary

A

line that separates two classes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

linear separator

A

linear decision boundary

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

logistic regression

A

process of fitting the weights to a model to minimize loss

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

parametric model

A

learning model that summarizes data with a set of parameters of fixed size (independent of the number of training examples)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

nonparametric model

A

learning model that cannot be characterized by a bounded set of parameters – this method retains all data points as part of the model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

table lookup

A

simplest instance based learning model – all training examples put into table, doesn’t generalize well

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

curse of dimensionality

A

nearest neighbors works well in low dimensions with plenty of data – at higher dimensions it doesn’t work well

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

locality sensitive hash

A

get around randomness and exact matching issues found in hash tables

17
Q

maximum margin separator

A

a decision boundary with the largest possible distance to example points – helps them generalize well

18
Q

kernel trick

A

SVMs can embed data into a higher dimensional space using this trick

19
Q

ensemble learning

A

selects a collection of hypotheses and combines their predictions by averaging, voting, or some other means of ML

20
Q

bagging

A

short for bootstrap aggregating

21
Q

random forest model

A

form of decision tree bagging in which extra steps to make ensemble of K trees more diverse to reduce variance

22
Q

boosting

A

most popular ensemble method, used a weighted training set, all examples start with equal weight, weight increased if training example fails, able to overcome any amount of bias in base model, approximates Bayesian learning

23
Q

gradient boosting

A

form of boosting that uses gradient descent as opposed to weighted examples

24
Q

semisupervised learning

A

typeof learning that involves giving a few labeled examples and using them to mine more information from a large collection of unlabeled examples

25
Q

crowdsourcing

A

using paid workers or unpaid volunteers operating over the internet