C2W1 Practical Aspects of Deep Learning Flashcards
Data should be divided into
Train (99%) / dev (0.5%) / test (0.5%)
Bias/Variance
High bias - bad on train - algorithms fails to lear train data
High variance - good on train, bad on test.
How to reduce bias
Bigger network / train longer
How to reduce variance
More data / regularization
Regularization L1 / L2
Just reduces weight of some weights, input parameter Alpha
Dropout regularization
Drop random neurons during training on each sample. Do division of the remaining to adjust values. Don’t drop them when evaluating model.
Other regularizations methods
Twist the image, early stopping (when dev error is optimal)
Normalizing Inputs
Center the data with respect to the plot center (subtract the mean), x - mu
Reduce (normalize) the variance: x / sigma. Sigma is not an input parameter
Weight Initialization for Deep Networks
There are different techniques, for example Xaviar initialization if activation function is tahn