Neural Networks Flashcards
What is a feedforward network?
A network where the inputs are put through multiple layers, every layer doing calculations before sending the input to the next layer
How do backpropagation work?
- You run one feedforward cycle
- compare expected output to the output (y-hat)
- calculate the error between them
- Run backprop = see which neurons are doing a bad job predicting and decrease their weights and vice versa for good neurons
- repeat
What is overfitting?
Your model is performing well on the training data but poorly on test data
What is underfitting?
Your model is performing poorly on train data
When do you want to use L1 regularization?
L1 is better when you want to find the best features with fewer weights - it is more sparse
When do you want to use L2 regularization?
L2 is better for training neural models since it tries to optimize and keep all weights
Why use dropout?
Because dropout prevents some neurons to become overly decisive in the decision-making.
What is Stochastic Gradient Descent?
SGD allows you to divide the train data in each epoch into batches. You run each batch through the network, update the weights (backprop) and then repeat until all batches has run - you then run the next epoch