Past Papers Flashcards

1
Q

What are hyper parameters?

A

Set before training

Be sure to specify that they DEFINE network’s architecture

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

For an L1 regularised neural network, write down how the regularisation term changes the way the parameters are modified during back prop

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

For an L1 regularised neural network, write the loss function expanded around θ* and show what the minimum is for θi

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

For an L2 regularised neural network:

Write down how regularisation term changes the way the parameters update

Expand around the min of the loss function

Write down an expression for the ith component of the minimum

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

When doing questions with L1 remember to

A

Mention that we are introducing sparsity to the solution -> some parameters will go to zero if they are not significant

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Describe what early stopping does

A

At each iteration of early stopping, we check how the validation or test set errors behave. After p (patience) consecutive iterations where the test error gets worse, the algorithm terminates

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

GPT4o definition of Universal Approximation Property

A

A feedforward NN with a single hidden layer containing a finite number of neurons can approximate any continuous function cation on a compact subset of Rd, given an appropriate activation function

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the capacity of an infinitely wide neural network with a single hidden layer?

A

By the UAP, this network can approximate any function and therefore has infinite capacity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

When asked to compare loss functions?

A

Remember to check wether the functions are bounded & wether they are differentiable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

When discussing MSE?

A

Remember to mention that it is preferred for regression

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

How does L2 regularisation change the way the parameters are updated using backprop?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

When finishing values that minimise MSE for linear regression

A

Mention linear independence of system of derivatives with respect to β

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

How many times are parameters updated ?

A

N * (1 - validation_split) = num of training samples (X)

X / batch size = num of batches (B)

B * epochs = number of parameter updates

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Total # of parameters in a NN

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

UAP conditions on g

A

Maps R to R
Measurable
Non polynomial
Bounded on any finite interval
The closure of the set of all discontinuities of g in R has zero lebesgue measure

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

UAP property 1

A
17
Q

UAP property 2

A