Handout #8 - Recurrent Neural Networks Flashcards

1
Q

Explain what can be used to train a RNN

A

Back-Propogation Through Time (BPTT)

  1. Unroll the network to expand it into a standard feedforward network and then apply back-propogation as per usual.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
1
Q

What’s the problem with BPTT

A

The unrolled network can grow very large and might be hard to fit into the GPU memory.

Process is seq. -> can’t be parallelised.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What’s the problem with the Simple RNN layer

A

RNN can grow very deep -> gradient descent can vanish (or explode) very quickly.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What time of data is RNN’s used for?

A

It’s used on sequential data -> any data with tie series (e.g. audio signal, stock market, machine translation)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Is an RNN a feedforward network?

A

Not, it’s cyclic

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Explain why LSTM is useful

A

Deals with the exploding and vanishing gradient problem (when unrolling the network).

LSTM has three gates; forget gate, input gate and output gate.

  1. Forget gate; forget irrelevant information
  2. Input gate; add/update new information
  3. Output gate; pass updated information
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Explain GRU

A

A simpler alternative to LSTM -> faster to train.

Instead of linear combination (w1u1 + w2u2), the gating mechanism is based on a multiplication of both inputs.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is the critical issue with RNN

A

They aren’t suitable for transfer learning.

Can’t do stuff in parallel.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly