Class 11 Flashcards

1
Q

one hot vector

A

assign a value to every word, could be used to describe every word

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

word embedding

A

slightly more sophisticated approach of assigning a low dimensional vector, learned automatically from the data, do good job of representing words in isolation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

character level model

A

alternative to word embeddings – input is a sequence of characters, each encoded as a one hot vector – model has to learn how characters come together to form words

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

language model

A

probability distribution over sequences of words, need to create one that has sufficient context

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

machine translation

A

has the goal of translating a source language to a target language

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

sequence to sequence model

A

neural network architecture created by using an RNN in conjunction with an LSTM, most commonly used for machine translation (MT) but can also be used to generate a text caption for an image or summarization – 3 shortcomings: nearby context bias, fixed context size limit, slower sequential parsing

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

decoding

A

process of generating target words from source words

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

transformer architecture

A

architecture that uses a self attention mechanism that can model long distance context without a sequential dependency

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

multiheaded attention

A

can be used to address the problem of too much self attention – breaks sentences into pieces and applies the attention model to the pieces

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

positional embedding

A

technique used by a transformer to capture the ordering of words

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

transformer encoders

A

used to text classification tasks

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

transformer decoders

A

used for text classification tasks but uses a version of self attention where each word can only attend to words before it

How well did you know this?
1
Not at all
2
3
4
5
Perfectly