Lecture 10 - LSTM - CNN -GAN - BatchNorm Flashcards
LSTM stands for
Long Short-Term Memory
LSTM is used for
Image captioning, stock market prediction, machine translation and text classification (for sequence processing)
What are the Components of LSTM?
Memory cell, input gate, forget gate, output gate
What is the function of the input gate?
Control what information gets stored in long-term state/memory cell
Note: Before the input gate, there is one more gate which controls the flow into the input gate
What is the function of the forget gate?
Regulate how much of information in long-term state is persisted across time instants
What is the function of the output gate?
Controls how much information to output from cell at a time instant
GRU cell is a simplified version of ___ cell
GRU cell is a simplified version of LSTM cell
GRUs combine which two LSTM gates to decide what information should be committed to long-term memory?
Forget and Input gates
For real - Use this website to learn RNNs/LSTM etc
http://colah.github.io/posts/2015-08-Understanding-LSTMs/
What is BatchNorm primarily used for?
Stabilising data - Also makes ANN faster and more stable through normalization (by recentering and rescaling)
Very useful for deep learning - Makes hidden layers very efficient
BatchNorm is tricky to use in
RNN (Recurrent Neural Networks)
What are adversarial examples?
Adversarial examples are inputs to machine learning models that an attacker has intentionally designed to cause the model to make a mistake; they’re like optical illusions for machines.
https://openai.com/blog/adversarial-example-research/
What are the two types of adversarial image attacks?
Untargeted adversarial attacks: cannot control output label of adversarial image
Targeted adversarial attacks: can control output label of image
Name some ML models that have had adversarial attacks
DNN, Clustering, Naive Bayes, Decision tree, multilayer perceptron, SVM
Adversarial attacks impact model performance by
Injecting adversarial/malicious data into training datasets that can cause decreased performance. Model failure is known as poisoning