Distributional Semantics Flashcards

1
Q

How do we learn new words?

A
  • look in a dictionary
  • from experience of usage
  • similar words from the past
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the distributional hypothesis?

A

Similar context suggests similar meanings

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

In distributional semantics, we want to find f, where f is

A

a function that takes in and transforms and compresses contexts to produce a vector that encompasses the meaning of a word
meaning(w) = f(c1, c2, c3, c4)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

How do we find function, f?

A

use co-occurrence vectors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

what is a cooccurrence vector?

A

collect a corpus of documents or sentences

apply basic preprocessing like lower case

count how many times word u appears with word v

the meaning of u is vector [(count(u,v1), count(u,v2)…]

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

what are the benefits of cooccurrence vectors (3)

A
  • meaning of a word is vector so we can compute similarities like the cosine similarities
  • can visualise word meanings
  • can directly use these vectors as input to machine learning models
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

what are the disadvantages of cooccurrence vectors

A

distributional semantics beyond words

cant capture all aspects of semantics

How well did you know this?
1
Not at all
2
3
4
5
Perfectly