Loss Flashcards

1
Q

Holder continuity

A

Holder continuity implies (uniform) continuity

Where smaller α means the function can be rougher

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Lipschitz continuity

A

Holder continuity with α = 1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What does a neural network do fundamentally

A

It approximates a function

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Objective of deep learning

A

Find a function that accomplishes given task by turning inputs into outputs in an, in some sense, optimal way

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

General loss function

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Auto encoder

A

Used in unsupervised learning

Dim reduction; has I = O and aims to reconstruct its input x where data must pass through a hidden layer with d units for d &laquo_space;I

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Relate auto encoders to PCA

A

Auto encoders are a non linear extension of PCA

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Risk

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Mini batch risk

A

As architecture and activation functions have already been specified, the mini batch risk is not fully determined by the parameter vector

Here #B is size of B

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Compare squared loss to absolute loss

A

Squared loss targets mean

Abs loss targets median

Squared loss significantly amplifies loss of a prediction far from actual value causing outliers to disproportionately effect training

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

How to manage limitation of squared loss

A

To minimise the effect of outliers but retain quadratic behaviour of loss in a δ neighbourhood (for small δ>0) we use Huber loss

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Alternative to Huber loss

A

Log cosh loss

Like Huber: behaves quadratically in region of true value and almost linearly far away

Unlike Huber: twice differentiable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Weighted sum of 1D losses

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Categorical cross entropy

A

Used to train multi class classifies (as opposed to binary)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Kulback Leibler divergence

A

Measure of the discrepancy between the empirical distribution of labels and the distribution predicted by the network

A measure of how 1 P dist is different to another

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Piecewise affine

A