Module 5 Flashcards
Binary classification
Only two possible classes
Multi-class classification
Multiple classes
Multi-label classification
Each input can belong to more than one class
Batching
- combining vectors of several data points into one matrix
- improves speed
- reduces noise
Backpropagation
Propagating the gradients backwards through the network layers
Low learning rate
Model will take forever to converge
Learning rate too high
Keep stepping over the optimal values
Learning rate decay
Reducing the learning rate by a factor
Weight initialisation
- zeros
- draw randomly from N(0,1)
Network capacity
Correlation between the capacity of a neural network and its ability to overfit
Network underfitting
Increase the number of neurons/layers
Network overfitting
Lowering the number of neurons/layers
Regularisation
Adding some information/constraints to stop the model from overfitting
L1 regularisation
Adding the absolute weights to the loss function
L2 regularisation
Adding the squared weights to the loss function