Seq To Seq Models Flashcards
1
Q
In any arbitrary decoder step of the decoder [in an encoder decoder model that I guess is assumed to not have attention] is the input at that step a word or a sequence of words (eg the entire output sequence up to this step)?
A
I’m pretty sure it’s just one word (all the other info I’m assuming is held in the hidden state?]