Sequence Models - NLPs Flashcards

1
Q

What are word embeddings

A

Features of a word, say ‘gender’,’royalty’ etc etc

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Learning word embeddings need a huge text corpus. True/False?

A

True…sometimes, billions of words.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

How do we use are analogies in NLP? Man:Woman=King:??

A

Minimise with each word in the vocab (e_king - e_man + e_woman)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is cosine similarity?

A

Explain

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is negative sampling?

A

It is trying to find the next word “juice” in a “….orange juice. “Orange” in the context and “Juice” is the positive match. We will also have 5 negative matches.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Rule based systems have _____ precision and _____ recall (high/log)

A

high precision and low recall.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is Semantic Slot Filling in NLP?

A

We have slots (like “Find flights between A to B”), which we try to fill using rule based approaches.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What are the steps in Semantic Slot Filling?

A
  1. Feature Engineering 2. Probabilistic Graphical Model
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What are the 4 levels of NLP Pyramid

A

Morphology (words) -> Syntax -> Semantics (meanings)-> Pragmatics

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is co-reference in a text?

A

Which reference the same entity in a text

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What are constituency Trees?

A

The various parts of speech organized as a tree in a sentence.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q
  1. Why should we ignore high frequency n-grams
  2. Why should we ignore low frequency n-grams
A

High frequency n-grams lead to overfitting

Low frequency n-grams are like stop-words, which do not add any value

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q
A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly