Week 2 (Neural Nets) Flashcards
What is a perceptron (and what activation function does it use)
What is a MLP (and how is the activation function different)
What things need to be defined for an MLP
How does the NN forward pass step work?
What is the NN backward pass step?
How is the error calculated for a neural network and how is this used to calculate the gradient
How is the gradient calculated for hidden layers (ie back propagation)
Answer this (neural network gradient):
What does a_j, z_j and h represent in neural networks
What is the typical delta_k for a neural network, ie the gradient on the last layer. Then generalise, what is the formula for delta_j
What is the backpropagation formula for delta_j
What is the general backgropagation playbook
How can back propagation be made more efficient
Store the previous gradients (ie gradients closer to the output layer) as they are reused on earlier layers.
How does gradient descent work
What are some methods to reduce overfitting of neural networks
Dropout, early stopping, regularisation