Multi-Layer Neural Network Flashcards
What is the issue with a perceptron?
It cannot separate anything that is not linearly separable
What is a multi-layer perceptron?
> Contains multiple neurons
> Contains multiple layers
- Input layer
- Hidden layer
- Output layer
What is the simplest example of something that is not linearly separable?
XOR gate
Theoretically, how many layers is required for a universal aproximator?
2 layers
What is the issue with just 2 layers?
In theory only 2 layers are enough to aproximate any function but in practice the hiddle layer needs to be so large that it is computationally impossible so instead we can add mutiiple hidden layers
What is the equation for mean squared error?
E(x) = 1/2 ∑ (yn - tn )^2
Why do we use a sigmoid function?
For multiple layers we cannot use step functions because it is not differentiable. A sigmoid function aproximates a step function and is differentiable
What is the equation for the sigmoid?
f(x) = 1/(1 + eβx)
β = Changes the steepness of the sigmoid function
What is the derivative of the sigmoid function?
σ’ = (βe-βx) / (1 + e-βx)2
What is the sigmoid activation function?
y(wTx) = 1 / (1 + e-βw^T x)