Syntactic Analysis: POS-Tagging Flashcards

1
Q

POS Tagging

A
  • Tags what part of speech each word is
  • Uses POS Taggers like WordNet, Penn Bank, Brown Corpus, etc.
  • Words that have more than one POS occur more often
  • Eric Brill’s algorithm it has a set of rules for walking through a sentence and assigning labels. As it steps through the words, it applies the rules and it may change previous word taggings.
    1) Assign the most common POS word initially
    2) List of rules that can change the initial tag based on the words before and after the word. Set of a few hundred rules.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Using POS tags

A

-Default tagger: assigns the same tag to each token; good for establishing a baseline

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Type vs. Token

A

Unique vocabulary word, you can have more than one of the same word in tokens. Types are unique tokens.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Hidden Markov models (HMMs)

A
  • Been around since the 1980s
  • Statistical approach (stochastic)
  • Baum-Welch algorithm to generate HMM from training data
  • Viterbi algorithm - determines the most likely sequence of “states” of the HMM, most likely sequence of POS tags.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Stochastic Taggers (aka statistical models)

A
  • Stochastic incorporate frequency and/or probabiility
  • POS probability based on the preceding word(s)
  • Hidden Markov models (HMMs)
  • POS with probabilities for each word, multiply them all together, use the highest scoring model of all HMMs, then tag sentence based on that model
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Dynamic programming

A

Fibonacci Series
Recursively find the answer
Fib(0) = 0, Fib(1) = 1
Transforms to a matrix, one end and go across the matrix depending on which one is most likely
Next tag depends on values of the next two tags

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Machine learning taggers

A
  • N-gram trainer
  • Naive Bayes
  • Neural network
  • SWVM
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Machine learning taggers

A
  • N-gram trainer
  • Naive Bayes - nbt.tag(tokens) - 93.1% accurate
  • Neural network
  • SVM
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Training data

A
  • Penn Treebank: Over 7 million POS-tagged text made up of IBM training manuals, WSJ articles, transcribed telephone conversations
  • You can build your own, e.g., if you are trying to POS tag tweets
  • Could just tag based no the most common POS for that word (90% accuracy)
  • Human consensus among English teach tagging WSJ articles is 96%, probably not possible to get to 100%
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Process

A
  • Gather corpus
  • Sentence tokenizer
  • Word tokenizer
  • POS tagger
  • Higher Level Analysis
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Combine taggers

A
  • We can combine taggers through a voting engine. This uses a panel of different taggers (NB, HMM, Brill, etc.)
  • Every different POS tagger gets a vote and majority wins
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

POS Tagging Use Cases

A
  • Narrow down synset lookups in WordNet, also helps with synonyms/hyponyms to substitute related search terms
  • Rudimentary Named Entity Recognition or Extraction (get names of proper people, places, things)
  • Potentially more useful word clouds (e.g., eliminate verbs to make a word cloud less messy)
  • More robust sentiment analysis, mixed sentiment review of movies
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Tropoynym

A

More specific cases of another verb. Troponym for speaking would be verbalizer, whimper, etc.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly