Large Language models Flashcards

You may prefer our related Brainscape-certified flashcards:
1
Q

How do neural networks aim to emulate the brain’s manner of computation?

A

it uses lots of processors like neurons, talking to each other.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

what do Classifier Networks do?

A

they learn to map input onto categories through training examples

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

how do we build a classifier?

A

we train it on many training examples

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

supervised learning

A

each training example is an input paired with the correct category

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

what are neural networks similar to?

A

a probability distribution or model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

when training a network inputs cause what to happen?

A

it generates activity in all its output units. As more activity in the outputs increases the chance of getting the correct output

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

How could you interpret the outputs of a neural network?

A

as a probability distribution as all possible outcomes can sum to 1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

How does a network for language process work?

A

by taking a sequence of words as its input/ prompt and learning to predict the next word

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

what is self-supervised learning?

A

a network prepends it doesn’t know the next word and then checks if its guess was right.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

what does a language processing network need to predict a good distribution?

A

good word representations that can handle long prompts

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

How should we represent words in a neural network?

A

One-Hot Encoding
A sparse vector where each word in the vocabulary is represented by a unique vector with a single high (1) value and all other values low (0).

“cat” 10000
“dog” 10001

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Contextualized Word Embeddings

A

Word representations that take the surrounding context into account, providing different embeddings for the same word in different contexts.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

what are embeddings?

A

distributed word representations that are ‘stuck’ into good places in an n-dimensional space.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

how are LLM’s input word sequences processed?

A

by an encoder network

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

how are LLM’s output word sequences processed?

A

by a decoder network

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

what are transformers

A

LLMs that have learned to attend to the right input words.

17
Q

what is the role of reinforcement learning in creating chatbots like GPT

A

It is used to teach the model to respond appropriately.

18
Q

How do Large Language Models Select the Next Word?

A
  • Statistical probabilities learned from the training data.
  • Context derived from the preceding text.