NER Flashcards

1
Q

What does NER stand for?

A

Named Entity Recognition

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is a Named Entity?

A

It is anything with a proper name

People, Locations, Organisations, Events, Dates, Time, Money

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is Named Entity Recognition?

A

It is the task of labelling a text span with types of named entities

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What are some popular NER tags?

A

People (PER)

Organisations (ORG)

Locations (LOC)

Geo-Political Entity (GPE)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What are NER approached based on?

A

They are based on tagsets

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is a popular NER tagset? How many types does it define?

A

The Automatic Contact Extraction (ACE) tagset is a very popular that defines 7 types.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What are some challenges of NER?

A

It works with spans, so working out how big a phrase needs to be labelled for a NE

Named entity type ambiguity, where a NE can be different types depending on the context

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the key method used for NER?

A

We treat it as a sequence labelling problem so we use BIO tagging

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is BIO tagging?

A

It is a common approach for sequence labelling requiring span-recognition

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What does BIO stand for?

A

Begin, Inside, Outside

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the idea of BIO tagging?

A

We assign a tag to each word in our sequence, and each tag may represent the beginning, the middle or the end of something

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Explain what the image shows

A

It shows different types of tagging methods. It shows that IO is difficult to comprehend, as it is difficult to understand where one NE begins and ends. BIO tagging builds on this by using begin labels, which shows where NEs begin. BIOES takes this even further.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What model is used to learn text according to the BIO scheme to identify NEs?

A

Conditional Random Fields (CRFs)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What type of features are useful for NER?

A

Non-word features such as captilisation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Why is the Hidden Markov Model (HMM) not a good model for NER?

A

As they are generative, is is hard to add feature patterns

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What type of model is CRF?

A

It is a discriminative sequence model based on a log-linear model. It is widely used for this type of sequence labelling problem.

17
Q

What is the input and output for a CRF model, and their lengths?

A

The input is a sequence of words, the output is a sequence of BIO tags. The length of the input will always be the same as the length of the output

18
Q

What does CRF want to find?

A

It wants to find the most probable sequence given a set of all possible sequences for a set of words.

19
Q

What is the equation for CRF?

A

Y hat is the most probable sequence, which is found using argmax. We compute the probability for each possible sequence we have given the input words and let argmax take the most probable

20
Q

What does CRF define?

A

A function F, which takes an input of a sequence of words and a sequence of BIO tags up to that point.

21
Q

What needs to be created in order to performed NER?

A

A set of K features

22
Q

What does each feature have in a CRF model?

A

Each feature has a corresponding weight

23
Q

What is the global feature vector in a CRF model?

A

It is the sum of all local features

24
Q

What are the local features in a CRF model?

A

Local features are features at a particular word index in the sentence. For each index position in the sentence, we use the local feature function to compute the features for that particular position. This can be summed to give a global feature vector for the entire sequence.

25
Q

What makes a linear chain CRF gain its name?

A

The input is the set of words X, the index we are dealing with, the tag that is being predicted and the previous tag. It can only look one tag in the past, a full CRF would be able to look further behind, but it is much more computationally expensive.

26
Q

How are local feature values created?

A

They are populated using a manually designed feature template for each word in the sequence.

27
Q

Where can features get their information from?

A

They can get them from anywhere in the word sequence X, as we can define the context window to be whatever we want it to be.

28
Q

What value types can a feature set contain?

A

It can contain many different types (numeric, text, boolean)

29
Q

What are some common types of features?

A

Words, POS tags, Word shape type, Word prefix or suffix, Match to a lexicon (e.g. list of names) or gazetteer (e.g. list of cities)

30
Q

What happens to the feature values over a sentence?

A

They are summed over the entire sentence X, meaning that there are always K features regardless of the sentence length.

For example, if we have 7 local features, and a sentence that is 5 words long, we can get the local features for each of the words, so each word in the sequence is now represented by the local feature vector. To get just one feature vector, we then sum each of these feature vectors to get just one feature vector that has 7 features representing the entire sentence.

31
Q

Explain what equation shows.

A

The equation is for the CRF model. It takes the maximum of the sum for each token position, the sum of the weighted local feature vector. For every feature type, for every position in the sequence, we sum it. This selects the most likely tag sequence for an input X.

Simply put, we compute the global feature vector for the input sentence. We then weight each of the features in the global feature vector differently and we want to find the best solution with argmax

32
Q

How do we evaluate NER taggers?

A

We can use micro or macro Precision, Recall and F1 scores

33
Q

What do we apply the evaluation metrics to?

A

The entities as a whole rather than single words or tags as the NEs can be of different lengths

34
Q

What needs to be avoided with NE?

A

Bias needs to be avoided with entities, e.g. is New York a good match for New York City, or do want exactly New York City as the match?

35
Q

What happens to the O tags in NE?

A

The O tag matches will be removed since the majority of the corpus is an O tag, and this will bias results reported using mean scores to the O class performance.

35
Q

What happens to the O tags in NE?

A

The O tag matches will be removed since the majority of the corpus is an O tag, and this will bias results reported using mean scores to the O class performance.

36
Q

What are some alternatives to the CRF model?

A

Word Representations (CBOW, skip-gram, word2vec, GloVE, BERT)

Character Representations (LSTM, GRU, CNN)

Model (CNN, LSTM, GRU)

Decoder (softmax, CRF)

37
Q

How is CRF different to HMM?

A

In HMM, we have to use Bayes theorem and the likelihood P(X|Y) to calculate P(Y|X).

In CRF, we can compute P(Y|X) directly, training the CRF to discriminate among the best possible tag sequences