Neural Networks Flashcards

1
Q

What is a perceptron

A

an artificial neuron which takes in many input signals and produces a single binary output signal (0 or 1)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

explain the role of bias terms in a NN

A
  • bias terms add a level of flexibility and adaptability to the model.
  • they “shift” the activation function, providing every neuron with a trainable constant value, in addition to the inputs
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Perceptron convergence theorem

A

If the data is linearly separable and therefore a set of weights exist that are consistent with the data, then the Perceptron algorithm will eventually converge to a consistent set of weights

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Perceptron cycling theorem

A

If the data is not linearly separable, the Perceptron algorithm will eventually repeat a set of weights and threshold at the end of some epoch and therefore enter an infinite loop

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the main limitation of a Single Perceptron, also give a solution.

A

Can only handle linearly separable problems
Multi-layer perceptron

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is impressive about a MLP with just one hidden layer

A

the single hidden layer is enough to represent an approximation of any function to an arbitrary degree of accuracy

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Why multilayer NN over a single layer?

A
  • need for a wide architecture in shallow NNs
  • shallow networks have a higher affinity for overfitting
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the purpose of the cost function in a NN

A

Also known as the loss function, it quantifies the inconsistency between predicted values and the corresponding correct values

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Explain the role of activation functions in NN

A

They play a crucial role by introducing non-linearities to the model, which are essential for enabling NN to learn complex patterns in the data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

If you were to simply use the identity function (f(x) = x) as an activation function, what is the
class of functions that you will be restricted to, in terms of learning?

A

linear functions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Explain the difference between two common activation functions of your choice

A

Sigmoid vs TanH
1. Output Range:
- Sigmoid: (0,1): used for binary classification
- tanh: (-1, 1): suitable for zero-centred data
2. Symmetry:
- Sigmoid is asymmetric, biased towards positive values
- tanh is symmetric around the origin (0, 0)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Advantages of NNs

A
  • Can learn and model non-linear and complex relationships.
  • Works well when training data is noisy or inaccurate.
  • Fast performance once a network is trained
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Disadvantages of NNs

A
  • Often require a large number of training examples.
  • Training time can be very long.
  • Network is like a “black box”. A human cannot look inside and easily understand the model or interpret the outputs
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Give an example of a simple function that can’t be learned by a single perceptron

A

XOR, IOR (non-linear functions)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Briefly explain the role of the gradient descent algorithm in training neural networks

A

It plays a fundamental role in training NNs. It aims to minimise the error of the NN by adjusting the model parameters (weights and bias) based on the gradient of the cost function with respect to these parameters

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Give an intuitive answer as to why adding hidden neurons in a NN increases the set of functions that can be learned by the network

A

Adding hidden nodes increases the expressive power of a NN by enhancing its ability to detect and combine features in a non-linear fashion. This increased capacity allows the network to learn a broader and more diverse set of functions, making it better suited for capturing complex relationships in the data.

17
Q

True or False: The back-propagation algorithm learns a globally optimal neural network with hidden layers.
Also, provide justification for your answer.

A

False. Backpropagation aims to find a set of weights and biases that minimise the loss function, but it may converge to a local minimum instead of a global optimum. This is because of the complexity of the optimisation landscape, and other factors such as initialisation and architecture