Unit 3 - Recurrent Neural Networks Flashcards

1
Q

What is recurrent neural network?

A

Variation to feed forward network

Output from previous step are fed as input to the current step

Talk about traditional neural networks

Less complexity due to using same parameters for each input

Uses a hidden state that stores all the information about the previous steps or a sequence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Example of RNN:

A

To predict the next word in the sentence, you need to know the previous words of the sentence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Types of RNN:

A

One to one
One to many
Many to one
Many to many

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is one to one RNN?

A

Single input and single output

Fixed input and output sizes so works like a traditional neural network

Example: image classification

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is one to many RNN?

A

Gives multiple outputs for a single input

Fixed input size and gives sequence of data outputs

Example: music generation, image captioning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is many to one RNN?

A

Single output is required for multiple input units or sequence of them

Example: sentiment analysis

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is many to many RNN with equal unit size?

A

Used to generate sequence of output data from a sequence of input units

In this case, both input output units are same in number

Example, name-entity recognition

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is many to many RNN with unequal unit size?

A

Which net sequence of output data from sequence of input data

This case inputs, an outputs have different number of units

Example: machine translation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is LSTM?

A

Variation of RNN

Critical components - memory cell and gates

Contents of memory formulated by forget and input gates

If both gates closed, no change in memory contents

No vanishing gradient problem

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Three neural networks used in LSTM:

A
  1. Forget network - info not needed is removed
  2. Remember network - adds the info from input if needed to the state
  3. Select network - presents the version of internal state as the output
  4. State - stores the formulated input
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

LSTM Gates:

A
  1. Forget gate - input combined with previous output to generate a fraction 0 and 1 and then multiplies with previous state.
  2. Input gate - which new info enters state of LSTM. Output of input gate(fraction 0 to 1) multiplied with tan h block and added to previous state.
  3. Output gate - exposes the current memory state
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is encoder decoder architecture?

A

Maps input domain to output domain in two stage network:

  1. Encoder - compresses input into latent space representation.
  2. Decoder - reconstructs from the latent representation to predict the output
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is sequence to sequence model?

A

Map fixed length input to fixed length output. Length of input and output may differ.

English to Chinese example

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

How Seq to Seq model works?

A

Uses encoder-decoder architecture

Consists of 3 parts:
1. Encoder - has several units of LSTM or GRU cells where each accepts an element of input sequence. Collects info from that element and forwards it .

  1. Intermediate (encoder) vector - final hidden state from the encoder part of the model. Encapsulates info of all the input elements.
  2. Decoder - like encoder, has several recurrent units. Predicts output from each unit at each time step. Takes hidden state from previous step and produces output
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Applications of Seq to Seq model:

A
  1. Machine translation
  2. Question and answer
  3. Video captioning
  4. Speech recognition
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is a recursive neural network?

A

Ability to model hierarchical structure in the training datasets

Applying same set of weights recursively over tree-like structures

Generalise the RNN from chain-like struct to tree-like struct.

17
Q

Types of Recursive Neural Network:

A
  1. Recursive autoencoder - learns to reconstruct input. In NLP, reconstructs the contexts. In supervised learning, learns likelihood of certain labels
  2. Recursive Neural Tensor Network -