Chapter 13 Word Associating Mining Flashcards

1
Q

paradigmatic relation

A

Word wa and wb have a paradigmatic relation if they can be substituted for each other.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

syntagmatic relation

A

the two words that have this relation can be combined with each other in a grammatical sentence - meaning that these two words are semantically related

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

In general, there are two type of word relations

A

one is called paradigmatic

the other is called syntagmatic

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Examining context is a general way of discovering paradigmatic words.

A

similar left context
similar right context
similar general context

How similar are context (“cat”) and context (“dog”)?
How similar are context (“cat”) and context (“computer”)?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

we ask the question, what words tend to occur to the left of eats? What words tend to occur to the right of eats?

A

This is the intuition we would like to capture. In other words, if we see eats occur in the sentence, that should increase the chance that meat would also occur.
This is syntagmatic relation.

How helpful is the occurrence of “eats” for predicting occurrence of “meat”?
How helpful is the occurrence of “eats” for predicting occurrence of “text”?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

how to discover paradigmatic relations

A

the general idea of discovering paradigmatic relations is to compute the similarity of context of two words. For

By viewing context in the vector space model, we convert the problem of paradigmatic relation discovery into the problem of computing the vectors and their similarity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

EOW (expected overlap of words)

A

Fig 13.5

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

BM25 Weighting

A

In order to achieve this desired weighting, we will use BM25 weighting, which is of course based on the BM25 retrieval function. It is able to solve the above two problems by sublinearly transforming the count of wi in d1 and including the IDF weighting heuristic in the similarity measure.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly