w3 POS tagging Flashcards

1
Q

what is information extraction IE

A

extracting information from documents based on a unstructured/messy input

ie search on google where was einstien born

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

what are the challenges in information extraction

A

words ocupy specific positions relative to each other, order of the words matter so its changing one term in the inquiry can change the search results

also words sometimes are meant to be in groups

ie albert einstien
but computer will see that as two words

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

what is pos tagging

A

breaking down sentences into parts of speech, tagging words as nouns verbs adj, etc

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

why is pos tagging hard

A

there are a lot of interpretations of sentences

nouns and verbs can change spending on intention

we can fish = we know how to pish
we can fish = we put fish in cans

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

what is a markov chain?

A

a model that defines the chance of sequences of random variables (states), these states are predefined

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

what is a transition probability in a markov chain

A

how likey a state is to transition to another state

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

what is a start distubution

A

the chance of starting at a certain state

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

what are the components of the markov chain

A

states (S), trainsitions (A) which give us markov chains

these chains have transition probabilities

and the starting probability spread is the start distribution (symbol is pi)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

what is a powerful point about markov chains

A

they are memoryless,

current probabilities are not impacted by things that happened in the past

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

is pos real

A

i dont think so, scientists randomly assigned and categorized these words but that means the concept of pos tags is made up

its a theoretical construct

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

in a markov chain

when it says

a1 = [0.6 0.1 0.3] what does that mean

A

this means

the probablity of going between diff states ie

P(s1 -> s1) = 0.6
P(s1 -> s2) = 0.1
P(s1 -> s3) = 0.3

the chance of state 1 changing to state 2 is 10%

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

what are different order markkov chains

A

the order of a markov chain tells you how many past expereinces it uses to predict a future

ie a first order markov chain only uses the current state to make predictions

p(s=a|s1…..s(i-1)) = P(s = a |s (i-1)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

what is the ∑j Aij = 1 ∀i

A

all of the changes to my state will summ up to one

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

what does it mean when the summation over all the inital probabilities doesnt equal one

A

while it usually equals one, if it doesnt that means theres something there that you havent captured

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

if there is a zero transition probability between two words what does that mean

A

there is no chance the words will show up next to each other

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

how would u calcuate the probabiltiy of

s1 -> s2 -> s3 -> s2

given markov chain/ transiton probabilities

A

If the transition matrix is:

P = 0.1 0.4 0.5
0.2 0.3 0.5
0.3 0.7 0.0

s1 = 1 s2 = 2 s3 = 3

then

0.4⋅0.5⋅0.7=0.14

because

P(s1 -> s2) = 0.4 * ….. etc etc

17
Q

how do hidden markov models relate to pos tagging

A

you are going to have some inner/hidden transitions assoicated with the pos

you are itnerested in deetecting the underlying sequence of hidden states which is what u will assign as ur pos

18
Q

what is the goal of hmm and pos

A

Given some input observable sequence of words x one to x n and some tag set, which is basically what this part of speech things are,

you want to come up with some output sequence where y I corresponds to the part of speech associated with x I.

the words are the obserable observation and
the parts of speech tags are hidden

19
Q

what are observations and emission probabilties

A

O = o1, o2 … oN a sequence of N observatiosn

B = bi * oN a sequence of N observation liklihoods/emission probabiltiies

the chance of an observation popping up bc of a state

20
Q

what is conditional probabiltiy formula

If I have
3 green shirts, of which I like 2 and
8 blue shirts, of which I like 6
What is the probability I like the shirt I am wearing given that I
am wearing a blue shirt?

A

P(A|B)

6/8

of blue shirts i like/ total # of blue shirts