Sequence Models - NLPs Flashcards
What are word embeddings
Features of a word, say ‘gender’,’royalty’ etc etc
Learning word embeddings need a huge text corpus. True/False?
True…sometimes, billions of words.
How do we use are analogies in NLP? Man:Woman=King:??
Minimise with each word in the vocab (e_king - e_man + e_woman)
What is cosine similarity?
Explain
What is negative sampling?
It is trying to find the next word “juice” in a “….orange juice. “Orange” in the context and “Juice” is the positive match. We will also have 5 negative matches.
Rule based systems have _____ precision and _____ recall (high/log)
high precision and low recall.
What is Semantic Slot Filling in NLP?
We have slots (like “Find flights between A to B”), which we try to fill using rule based approaches.
What are the steps in Semantic Slot Filling?
- Feature Engineering 2. Probabilistic Graphical Model
What are the 4 levels of NLP Pyramid
Morphology (words) -> Syntax -> Semantics (meanings)-> Pragmatics
What is co-reference in a text?
Which reference the same entity in a text
What are constituency Trees?
The various parts of speech organized as a tree in a sentence.
- Why should we ignore high frequency n-grams
- Why should we ignore low frequency n-grams
High frequency n-grams lead to overfitting
Low frequency n-grams are like stop-words, which do not add any value