W4 L2 Flashcards
what is lexical semantics
a branch of nlp that deals with word sense
but what is a sense bearing unit
does a word always have the same meaning
no! the same spelling of a word can have different meanings
homonymy
when two words are spelt the same but have different meaning
i went to the bank
we sat on the bank
what is word net
the largest most popular database of lexical relations
what is a homophone
same pronouncation with different spelling
to too two
these are hard for speech models
homographs
same spelling but different pronounciaiton
drum bass
fish bass
hard for text to speech agents
what is polysemy
when words are extended/ the meaning is transfered from words
bank
blood bank
what is metonymy
a subtype of polysemy
ie shakespear
author: shakepear wrote hamlet
works of author: i studied shakespear at school
turkey animal vs turkey meat
what synonymy
words that have the similar meaning
ie sofa and couch
antonymy
words that have opposite meanings
big small
what is hypo and nypernym
hypo is a sub term
and hyper is a super term
socretes is a man
all men are mortal
socreties is moral
hypo to hyper
but not hyper to hyp
not all men are socreties
what are semeantic fields
allow you to combine various terms that are related to the same domain
ie flight booking, plane, price, meal
-> domain of air travel
what is a concept
and abstract idea representign the fundemental characteristics of what it represents
what is classical concept theory
aristotle
concepts have a definitional structure, a list of features and all memebrs of this class must have these features
what is prototype concept therory
properties of concepts are not definitional
memebers tend to posses them but not strictly required
ie memebrs tend to look similar but they dont need to be super restrictive requriements
what is the theory concept theory
catagorization by concept
as new evidence comes, new members join and definitions change
definitions are in relation to each other
what is the big bag of words method
machine learning models reqiure numbers so we can use classification to turn words lemmas synsets and concepts into features
this ignores word order so its called bag of words method
what do we need to apply to features to classififcation
feature selection, feature weighting, normalization
what is feature selection
selecting the most important of the features
what is feature weighting and normalisiation
weighing the most important features after feature selection (tfidf)
what is the power of bayes formula
bayes formula allowss us to flip the coniditions
P(class|content) =
P(content)
what is the P(class | content) in bayes formula
posteriro probabilitiy
this is the postierior probabilitiy we are interested in in bayes formula that we need to use other formulas to get
what is P(class) in bayes formula
prior probabilitiy
our belife about the class distrbution before we see more evidence
what is P(content |class) in bayes formula
the likelihood
it shows how likely it is to see this exact combinations of features