W3 L1 Information extraction Flashcards

1
Q

why information extraction

A

we need it to get structure information (facts we want) from an unstructured/potentially messy input

ie where was einstein born
-> we need to extract this information from the many relevant documents

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

are terms like albert einstien a challenge for information extraction

A

yes, words that should be treated as a single unit are a challenge, some words are part of larger groups and should be grouped together

ie albert einstien is together not two seperate words

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

does the order of words matter in sentneces and information extraction

A

yes words occupy specific positions in a sentence for a reason and relate to each other in particular ways

the order of words matter, its not random

was einstine born albert where is garbage

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

what are parts of speech and pos tagging

A

parts of speech are nouns verbs adj and adverbs

part of speech tagging is identifying and tagging these parts of speech

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

why is pos tagging hard

A

there are many different intepretations for the same words

ie
we can fish
-> we go fishing
-> we put fish in cans

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

what is a markov chain

A

a model that defines the chance of a sequence of random variables/states happening

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

what is in a markov chain

A

states, transitions
give us markov chains
these markov chains have transition probabilities

the starting of these transition probabilities is the start distribution

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

what is the assumtion we are making in markov chain

A

they are memoryless

in order to predict the weather for tomorrow you only need to consider the weather today

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

what is a first oder markov chain

A

only the current state matters in determining the future

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

recall what these notations are

S, A, pi

A

S = s1…sN set of N states

A = a1….aNN transition probability matrix
the sum of the rows must equal 1

pi = pi1…piN initial probability distribution
the sum of pi = 1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

how can we apply markov models to language

A

we use markov chains to calculate the probability of certain words appearing given the previous word

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

how can we use hidden markov models for pos tagging

A

if we recorded some observations that depended on the hidden events we can try to identify whta the hidden events were that were underneath, causing our surface observations

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

in pos tagging what would be the observations and the hidden states

A

observations = words in sentences

hidden states = part of speech

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

what elements does a hidden markov model have

A

states, transition matrix, initial probabiltiy distribution

AND

O = o1…oT sequence of T observations

B = bi(ot) sequence of observation likihoods (emission probabilities)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly