WordNet Flashcards
Polysemous
having multiple meanings
Word sense
A discrete representation of one aspect of the meaning of a word.
thesaurus
Database that represents word senses.
Antonymy
When two words / lemmas have two opposite meanings.
WSD (abbreviation)
word sense disambiguation
word sense disambiguation
The task of determining which sense of a word is being used in a particular context.
How does WordNet represent different senses of the same word?
with superscript:
mouse1, mouse2 etc.
embedding
a point in semantic space
glosses
textual definitions for each sense, in dictionaries or thesauruses.
Why can a gloss still be useful, even though it is often circular and relies on real-world knowledge that we as humans have?
It is just a sentence, and we can compute sentence embeddings that show something about the meaning of the sense.
Zeugma
A conjunction of different readings of a word, for example ‘Air France serves breakfast and the USA’.
Synonymy
When two senses of two different words / lemmas are identical.
Reversives
A group of antonyms that describe change or movement in opposite directions.
Hyponymy
When a word/sense is more specific, denoting a subclass of another word/sense.
Hypernymy
When a word/sense describes a less specific or superclass of another word/sense.
Superordinate ==
Hypernym
Other word for hypernym
Superordinate
What does ‘A IS-A B’ mean?
A is an instance of B, or B contains A.
Meronymy
A part-whole relation.
Holynymy
A whole-part relation. Example: car is a holonym of wheel.
Structured polysemy
When the senses of a word are related semantically.
Metonymy
A type of polysemy relation. When the use of one aspect of a concept or entity is used to refer to other aspects of the entity or the entity itself.
Synset
The set of near-synonyms for a WordNet sense.
Supersenses
The lexicographic categories in which senses are grouped in WordNet.
Lexical sample tasks
Situations where we just need to disambiguate a small number of words.
All-words task
A problem in which we have to disambiguate all the words in some text.
Semantic concordance
A corpus in which each open-class word in each sentence is labeledwith its word sense from a specific dictionary or thesaurus.
Most frequent sense baseline
Choose the most frequent sense for each word from the senses in a labeled corpus. Can be quite accurate.
One sense per discourse baseline
Not generally used as a baseline, but holds better for coarse-grained senses.
How does the 1-nearest-neighbor WSD algorithm produce contextual sense embeddings (vs<\sub>) for sense s?
It averages the n contextual representations vi for each of the n tokens of the sense s.
How does the 1-nearest-neighbor WSD algorithm produce its output?
It chooses the sense whose embedding has the highest cosine similarity with the target word in context.
What class of WSD algorithms does not require labeled data?
knowledge-based algorithms
Lesk algorithm
A WSD algorithm that chooses the sense whose dictionary/gloss/definition shares the most words with the target word’s neighborhood.
WiC (abbreviation)
word-in-context
word-in-context task
Given two sentences with a target word in different contexts, the system must decide whether the words are used in the same sense or in different senses.
Retrofitting / counterfitting
Methods that learn a second mapping after embeddings have been trained, shifting the embeddings such that synonyms are closer to each other and antonyms are further apart.
WSI (abbreviation)
word sense induction
Word sense induction
An unsupervised approach to WSD
Agglomerative clustering
Type of clustering where each of the N training instances is initially assigned to its own cluster. New clusters are formed bottom-up by merging similar clusters.