Information Extraction Flashcards

1
Q

What is Information Extraction?

A

It is turning unstructured text into structured data, such as a relational database or set of extracted tuples

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is Relation Extraction (RE)?

A

Relation Extractions finds semantic relations among text entities, such as parent-child, part-whole or geospatial relations

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What can be used to encode relational informtaion?

A

Knowledge-graphs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is Event Extraction?

A

It finds event in which entities participate

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is Temporal Extraction?

A

It is finding times, dates and durations

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is Knowledge Base Population (KBP)?

A

It is the task of populating knowledge bases from unstructured text using extracted information

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What datasets exist to perform relational extraction?

A

The ACE relation extraction dataset

Wikipedia info boxes (DBpedia and Wikidata)

WordNet

TACRED dataset

SemEval

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What are some relations in RE?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is pattern-based RE?

A

It is pattern-based relation extraction. It uses hand-crafted lexico-syntactic patterns that were to be followed. They were tailored to specific domain lexicons, so would only work in that domain. It had high precision, but low recall and were expensive to make

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is Supervised RE?

A

It is supervised relation extraction with an annotated corpus, which we pass to a model which we want to learn the patterns in the corpus to be able to perform extraction. It works by finding pairs for named entities and would classify the relation within the sentence for that pair

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the input and output for a supervised RE model?

A

The input X is a feature set for the entity pair, and the output Y is a prediction of the relation for the provided pair.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What type of classifier can be used in a RE model?

A

It can be LogReg, RF, RNN but in this course we use a Transformer model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Why do we use a transformer model with RE?

A

This is because the self-attention works well for this type of problem and can learn what parts to focus on

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What are some techniques to improve RE models?

A

Replace SUBJ and OBJ with NER tags to avoid overfitting to lexical terms

Use RoBERTa or SPANbert pre-trained word embeddings instead of vanilla BERT, as pre-training is done using single sentences rather than sentence pairs with a separator between them

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Why do NEs tend to overfit in deep learning RE models?

A

RE labelled datasets are too small, so there will not be enough examples of every possible NE phrase, so the model is likely to overfit the dataset samples of the ones in the training set

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is a semi-supervised RE method?

A

We can use a semi-supervised RE approach with bootstrapping

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What is bootstrapping?

A

It is where we have a high-quality hand crafted small set of seed tuples for the relations in the form (relation, e1, e2)

The algorithm will find sentences which contain instances of the seed tuples, extract patterns and find new seed tuples

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What happens when we run many iterations of boot strapping?

A

We get semantic drift, so the seed tuples start looking for other things

19
Q

What are some methods to reduce semantic drift?

A

Apply a confidence threshold to the extraction patterns to improve quality of tuples

Limit the dependency graph walk for new tuples

20
Q

What is the distant supervision method for RE?

A

It is the use of a knowledge-base such as DBpedia as a source of seed tuples (r, e1, e2)

21
Q

What does using a knowledge base avoid?

A

It avoids the semantic drift problems of the bootstrapping approach

22
Q

What are the steps taken with Distant Supervision?

A

Start with a text corpus, use a NER tagger, and then match entities within the database, work out the relation to get the relation term for the seed tuple. We can then add matches to the training set as a feature set with the occurrence frequency

23
Q

What is an issue with using distant supervision?

A

It generates a very large training set for supervised RE, meaning that it is very noisy and will have low precision

24
Q

What can be done to reduce noise during distant supervision RE?

A

A GAN or incremental training approaches can be used

25
Q

What is unsupervised RE?

A

It extracts relations with no training data.

26
Q

What can Unsupervised RE also be called?

A

Open Information Extraction or OpenIE

27
Q

What are the steps in unsupervised RE?

A

You take the text corpus, apply a POS tagger, identify verb-based POS patterns. This performs a syntax and lexical constrained walk of token in a sentence to explore different options to connect arguments to a relation in order to build a tuple.

28
Q

What are unsupervised RE methods good at?

A

They provide an unlimited number of relations and entity types

29
Q

What is an issue with using unsupervised RE?

A

The extracted phrases are not semantically grounded to a database entry making it harder for applications to use the relations

30
Q

How can we evaluate RE models?

A

Supervised RE - P/R/F1

Semi-supervised and Unsupervised RE - Random sample + human inspection to compute precision or P@Yield or P@R which is precision at different levels of recall

31
Q

What can temporal extraction be?

A

It can be absolute (easy) or relative (contextual to a reference point in text)

Rule based systems look for lexical triggers encoded as temporal expressions

32
Q

What are some approaches for Temporal Extraction?

A

A sequence labelling approach using BIO tags with feature templates encoding feature sets for a supervised classifier

33
Q

What does temporal normalisation do?

A

It maps temporal expressions to a point in time or quantified duration

34
Q

What do temporal normalisation approaches tend to be?

A

They tend to be rule-based

35
Q

What are fully qualified expressions and what are temporal anchors?

A

Fully qualified expressions are ones that appear in full in text, such as a full date.

Temporal anchors are followed by a relative temporal expression. These are words such as yesterday, tomorrow, the weekend.

36
Q

What do we aim to do with temporal anchors?

A

We try and make time statements based on the anchors.

For example, if we see yesterday, we know that means -1 day from the date of the post.

For example, through the weekend would mean the coming weekend after the start

37
Q

What do events in English often correspond to?

A

They correspond to verbs often, but not always (e.g. nouns can introduce an event)

38
Q

How do we perform event extraction?

A

We use sequence labelling approaches using BIO tags and feature templates that encode features for a supervised classifier

39
Q

What do we do with the extracted relations?

A

We can populate a knowledge base.

40
Q

How do we populate a knowledge base?

A

We start with a partial KB and a large corpus of text, apply our KB technique to a large corpus to generate a fully populated KB with some noise as these approaches are not perfect.

41
Q

What are the two approaches to filling a KB?

A

Slot filling and entity linking

42
Q

What is slot filling?

A

We complete all the known information that we can about a particular entity. The input is a bit of text and some NER tags, apply the supervised RE to compute some relation tuples. Having done this, we use the output of those tuples to see how they are connected to form a knowledge graph by computing 0-hop and 1-hop slots.

43
Q

Explain what the image shows

A

In the image, we have some text. We have a query entity for Mike Penner, which is getting the knowledge for Mike Penner. The table in the bottom right shows the relation tuples that have been extracted from the text. A 0-hop slot looks for the immediate connection, so Mike Penner has a spouse, which is Lisa Dillman. A 1-hop slot takes the object of the 0 hop, and use it as the subject for the 1-hop. So here, we want to see if there is a title for Lisa Dillman, which there is. This can create an n-deep knowledge graph

44
Q

What is entity linking?

A

It takes text entities (names of things that are entities), and connects them to unique knowledge graph entities. It is typically done by using an existing KB in order to work out the links better.