Language Flashcards

1
Q

spans all tasks where the AI gets human language as input

A

Natural Language Processing

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

examples of tasks in Natural Language Processing

A

• automatic summarization
• information extraction
• language identification
• machine translation
• named entity recognition
• speech recognition
• text classification
• word sense disambiguation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

sentence structure

A

Syntax

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

meaning of words or sentences

A

Semantics

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

system of rules for generating sentences in a language

A

Formal Grammar

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

text is abstracted from its meaning to represent the structure of the sentence using formal grammar

A

Context-Free Grammar

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

a sequence of n items from a sample of text.

A

n-gram

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

a contiguous sequence of n characters from a sample of text

A

character n-gram

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

a contiguous sequence of n words from a sample of text

A

word n-gram

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

a contiguous sequence of 1 item from a sample of text

A

unigram

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

a contiguous sequence of 2 item from a sample of text

A

bigram

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

a contiguous sequence of 3 item from a sample of text

A

trigrams

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

task of splitting a sequence of characters into pieces (tokens)

A

Tokenization

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

the task of splitting a sequence of characters into words

A

word tokenization

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

the task of splitting a sequence of characters into sentences

A

sentence tokenization

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How to generate text using a Markov Model

A

Markov Models

17
Q

a model that represents text as an unordered collection of words.

A

Bag-of-words Model

18
Q

adding a value α to each value in our distribution to smooth the data

A

additive smoothing

19
Q

adds 1 to each value in our distribution, pretending that all values have been observed at least once.

A

Laplace Smoothing

20
Q

task of finding relevant documents in response to a user query.

A

Information retrieval

21
Q

models for discovering the topics for a set of documents

A

topic modeling

22
Q

counting how many times a term appears in a document.

A

term frequency

23
Q

words that have little meaning on their own, but are used to grammatically connect other words

A

function words

24
Q

am, by, do, is, which, with, yet,

A

function words

25
Q

words that carry meaning independently

A

content words

26
Q

algorithm, category, computer, …

A

content words

27
Q

measure of how common or rare a word is across documents in a corpus

A

Inverse Document Frequency

28
Q

ranking of what words are important in a document by multiplying term frequency (TF) by inverse document frequency (IDF)

29
Q

task of extracting knowledge from documents

A

Information Extraction

30
Q

each word is represented with a vector that consists of as many values as we have words

A

One-Hot Representation

31
Q

meaning is distributed across multiple values in a vector.

A

Distributed Representation

32
Q

algorithm for generating distributed representations of words

33
Q

a neural network architecture for predicting context given a target word

A

Skip-Gram Architecture