Word Senses and WordNet Flashcards

1
Q

What is a word sense?

A

It is a discrete representation for one meaning of a word.

E.g. a mouse can be a mouse controlling a computer system, or it can be the animal

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

How is a sense defined?

A

They are defined through its relationship to other senses

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is a synonym?

A

A synonym is two different words that have a near identical meaning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is an antonym?

A

An antonym is two words that have almost completely opposite meanings

e.g. long and short, big and little, up and down

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is a hyponym?

A

A hyponym is a more specific lemma (subclass)

e.g a car is a vehicle, a dog is an animal, a mango is a fruit

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is a hypernym?

A

It is a more general lemma (superclass or superordinate)

e.g. A vehicle can be a car, an animal can be a dog, a fruit can be a mango

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is a meroynm?

A

It is a lemma which is part of something (part-whole or part of)

e.g. a wheel is part of a car, a leg is part of a chair

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is a holonym?

A

A holonym is a lemma which something can belong to

e.g. a car has wheels, a chair has legs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is WordNet 3.0?

A

It is a lexical database containing nearly 120,000 nouns, 11,000 verbs, 22,000 adjectives and 4,000 adverbs.

On average, there are 1.23 senses per noun and 2.16 senses per verb

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is a synset?

A

A synset is a set of synonyms

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What does supersense mean?

A

Supersense is a semantic category.

There are 26 categories for noun, 15 for verb, 2 for adjective and 1 for adverb

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What can synsets have?

A

They can have relationships to each other, as seen in the image

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is Word Sense Disambiguation?

A

It is a way of, given a sentence, selecting the correct word sense for all the words

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What training data is used with Word Sense Disambiguation?

A

Datasets with word sense annotations such as SensEval and SemEval that use WordNet annotations

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What type of problem is word sense disambiguation defined as?

A

It is defined as a sequence classification problem

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is the input and output for the word sense disambiguation problem?

A

The input is a sentence, a sequence of words

The output is the word sense labels, which are a sequence of tags that are WordNet sense annotations

17
Q

What does the image show?

A

It shows how Word Sense Disambiguation works in practice. We have an input sentence, and then the classifier chooses the correct word sense based on the context of the rest of the sentence.

18
Q

What baselines are used in WordNet?

A

One is to take the statistically most frequent sense in the corpus. as we are picking the most frequent one. With this, we can use one sense per discourse (remember the previous sense and use that) - on average this works well but remember, this is just the baseline.

19
Q

What embeddings are used with a WordSense Disambiguation classifier?

A

Pre-trained contextual BERT embeddings

20
Q

Explain what the image shows

A

We have our input sentence, which is then encoded using the encoder (BERT embeddings). The decoder then creates word embeddings for each of the words and the word embeddings are aggregated together to form sense embeddings. For any given word, we want to plot where that word falls in the latent space and find the closest sense to that.

21
Q

What happens during the training phase of WSD?

A

For each word sense we get, we average across the entire corpus each token embedding (context) for the particular word to get a sense embedding.

22
Q

What happens during the test phase of WSD?

A

When we encounter unseen words, we look up the word embedding for BERT, plot the word embedding in the vector space and use a cosine distance measure to find the closest sense embedding to the new word.

23
Q

What happens when we encounter a sense that is not in the corpus of WordNet?

A

One option is to choose the statistically most frequent sense in the corpus (not good)

The other option is to impute the sense using WordNet taxonomy (averaging sense embeddings of children to compute embeddings for hypernyms)