Week 10 Flashcards
What is the Loss Function?
loss(hw) = sum ((yj - hw(xi))^2)
What is the Perceptron Learning Rule?
wi = wi + axi(y - hw(x))
How are the weights changed for the Perceptron Learning Rule?
If y = 1 but hw(x) = 0:
Make wTx larger so hw(x) outputs 1.
wi increased when xi is positive.
If y = 0 but hw(x) = 1, do opposite.
What is an epoch?
The time after which we update the weights of the Perceptron Learning Rule.
One example.
What are some issues in Linear Classification?
Hard threshold.
Function is not differentiable - weights learning could be unpredictable.
What is Under/Overfitting?
Underfitting: Model not doing well on seen data.
Overfitting: Model trained too specific on training data (not good for generalisation).
What is Regularisation?
Achieve generalisation in ML.
Require loss and regularisation paramater lamda.
What is Drop Out?
Encourages network to adapt, not memorise.
Drop a certain number of weights.
What are the benefits of Stochastic gradient descent?
Quicker to converge in practice.
Helps to avoid overfitting.