Introduction to Deep Learning Flashcards
In a single perceptron(neuron), what is the formula for calculating a single output?
y = g(W0 + XW)
1. Dot product (XW)
2. Add bias (W0)
3. Apply non-linearity (g)
What are the 3 layers of NN?
Input layer, hidden layers, output layer.
The input layer receives data, hidden layers process it, and the output layer produces the final predictions.
How to determine how well the model perform?
__ Function
Using Loss Function
What are the 2 types of loss functions?
- Binary __ loss
- Mean __
- Binary cross entropy loss (Classification)
- Mean Squared Error (MSE) (Regression)
What is the problem of training NN?
Finding the vector __. The __ determine the performance, because they…
Finding the vector weights(W). The weights determine the performance, because they directly influence the output and thus the loss.
What is backpropagation in simple sense?
Backpropa uses the __ rule to compute the gradient of the loss …
Backpropagation uses the chain rule to compute the gradient of the loss with respect to the weights (W). This allows us to update the weights in the direction that minimizes the loss.
In summary, what is forwardpropagation?
FP refers to the computation of outputs from the inputs in NN.
What is the optimization method we can use during backpropagation?
__ (GD)
The method is gradient descent.
How does GD works?
It sets the ___ __ (LR) …
It sets the learning rates by designing an adaptive LR that adapts to the landscape.
What are 2 tips to train NN?
- S…
- B…
Stochastic GD (SGD) and batching
Batching is better than SGD. Why?
It uses ___ batch of data to compute.
Instead of using single training point, batching uses small batch of data to compute. It is more accurate and faster to train.
How to deal with overfitting in NN?
Re___
Regularization
What is the 2 regularization technique?
- D…
- E… S…
- Dropout
- Early Stopping
How does Dropout work?
- Randonly set some ___ to 0.
- Drop __% of __ in layers.
- Randomly set some activations to 0.
- Drop 50% of activation in layers
How does Early Stopping work?
Stop the model from ___ before ___
Stop the model from training before overfitting.