Transformers:Self attention Flashcards

1
Q

The self attention operation takes n inputs and how many outputs?

A

n

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

If the keys, queries, and values are generated from the same sequence, what type of attention so we have?

A

Self attention

https://medium.com/@angelina.yang/whats-the-difference-between-attention-and-self-attention-in-transformer-models-2846665880b6

How well did you know this?
1
Not at all
2
3
4
5
Perfectly