Week 5 - (part 1) LLMs And Gen AI Flashcards

1
Q

To train the model, what sort of data is being used?

A

Common crawl, books, wikipedia etc.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is Generative Pretraining?

A

Based on the algorithm, the model is trying to predict the next output based on probability of co-occurrence.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Generative Language Models do sampling + classification over tokens. Do you have to classify all the words in the dictionary?

A

No. Only the sub-words. This gives us a smaller class size to predict. This estimate is done for efficiency reasons. If you have all the words, too many classes and a lot of words not used.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Why don’t we use character classification?

A

If you use characters, the class size is too small and it takes very long.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

How do they decide to group the sub-words?

A

Through frequency and statistics

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Instruction fine tuning and human feedback

A

Instruction fine tuning - Step 1: a labeller demonstrates the desired output behaviour by telling the machine what is the correct answer given a set of inputs

Human feedback - Step 2: A labeller ranks the outputs from best to worst

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Pros and cons of feedback

A

[+] train and adapt to come up with better answers
[+] overly compliant
[-] may not be safe

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Why does the model hallucinate?

A

The model is always told there is one right answer. It will predict to its best ability based on probability. Probability is baked into its weight and parameter.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

How to mitigate true preferential learning

A

Ask the model to output sub word tokens equivalent to “i dont know” and ask the model some questions with no answer. You are manipulating the probabilities.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Why can’t we just use LLMs for everything and rely on probability?

A

Efficiency Low
Lack of updatability
Issue of provenance
Effective
Good at synthesising information

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What are the benefits of retrieval based NLP

A

Efficient, updatable, provenance, effective, synthesis

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Do auto regressive LMS simply predict the next token?

A

Yes, that’s all they do (TO a certain extent). They predict scores over the entire vocabulary at each step. We then use those scores to compel them to predict some other token or other. They also present data in their internal and output representations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

How is generative pre-training different from predictive text?

A

Generative pre training involves looking through more things to produce more useful outputs. The goal is to learn USEFUL knowledge.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Emergent abilities - what AI chatbots can do that autocorrect cannot?

A

Play chess - an internal model of chess rules and strategy must be programmed in the first place

How well did you know this?
1
Not at all
2
3
4
5
Perfectly