True/False Flashcards

1
Q

The mean-field theory for Hopfield network yields the exact value for the critical storage capacity.

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

That the energy cannot increase under the deterministic Hopfield dynamics is a consequence of the fact that the weights are symmetric.

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

The stochastic update rule for the Hopfield network is different from the Metropolis algorithm.

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

All stored patterns are local minima of the energy function.

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

The detailed balance condition is a necessary condition for the Markov-Chan Monte-Carlo algorithm to converge.

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

That the energy cannot increase under the deterministic Hopfield dynamics is valid only if the thresholds are put to zero.

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

For a given 𝛼, the one-step error probability for the deterministic Hopfield network is lower when the diagonal weights are set to zero.

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

In the limit of N → ∞ the order parameter m𝜇 can have more than one component of order unity, the other components are small.

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

The stochastic update rule for the Hopfield network is identical to the Metropolis algorithm.

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

The detailed balance condition is a necessary condition for the Markov-Chain Monte-Carlo algorithm to converge.

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

That the energy cannot increase under the deterministic Hopfield dynamics is a consequence of the fact that the weights are symmetric.

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

The mean-field theory for the Hopfield network yields the exact value for the critical storage capacity.

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

All stored patterns are local minima of the energy function.

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Not all local minima of the energy function of the Hopfield network correspond to stored patterns.

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

The stochastic update rule of the Hopfield network is different from the Metropolis algorithm.

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

That the energy cannot increase under the deterministic Hopfield dynamics is a consequence of the fact that the diagonal weights are set to zero.

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

That the energy cannot increase under the deterministic Hopfield dynamics holds also when the thresholds are zero.

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

The detailed condition is a necessary condition for the Markov-Chain Monte-Carlo algorithm to converge.

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

A perceptron that solves the parity problem with N inputs contains at least N^2 hidden neurons.

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Increasing the number of hidden neurons in the network increases the risk of overfitting.

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

Two hidden layers are necessary to approximate any real valued-function with N inputs and one output in terms of a perceptron.

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

Using stochastic gradient decent in backpropagation assures that the energy either decreases or stays constant.

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

In minimisation with a Lagrange multiplier, the function multiplying the Lagrange multiplier can also assume negative values.

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

Some of the functions with 5 Boolean valued inputs and one Boolean valued output are linearly separable.

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

Different layers of a deep network learn at different speeds because their effects on the output are different.

26
Q

The weights in a perceptron are symmetric.

27
Q

L1-regularisation reduces small weights more than L2-regularisation.

28
Q

Weight decay helps against overfitting.

29
Q

Increasing the number of hidden neurons in the network increases the risk of overfitting.

30
Q

Two hidden layers are necessary to approximate any real valued-function with N inputs and one output in terms of a perceptron.

31
Q

Pruning increases the risk of overfitting.

32
Q

Using stochastic gradient decent in backpropagation assures that the energy either decreases or stays constant.

33
Q

In minimisation with Lagrange multiplier, the function the Lagrange multiplier must be equal to or larger than zero.

34
Q

Back-propagation is a form of unsupervised learning.

35
Q

To make use of back-propagation, it is necessary to know how the target outputs of input patterns in the training set.

36
Q

“Early stopping” in back-propagation helps to avoid being stuck in local minima of energy.

37
Q

“Early stopping” in back-propagation is a way to avoid overfitting.

38
Q

Using stochastic path through weight space in back-propagation helps to avoid being stuck in local minima of energy.

39
Q

Using stochastic path through weight space in back-propagation prevents overfitting.

40
Q

Using stochastic path through weight space in back-propagation assures that the energy either decreases or stays constant.

41
Q

There are 2^(2^n) functions with n Boolean valued inputs and one Boolean valued output.

42
Q

None of the functions with 5 Boolean valued inputs and one Boolean valued output are linearly separable.

43
Q

There are precisely 24 functions with 3 Boolean valued inputs and one Boolean valued output (equal to zero ore one) where exactly three of the possible inputs maps to zero.

44
Q

Oja’s learning is a form of unsupervised learning.

45
Q

Then dimension of the output space of a Kohonen network must be equal to the dimension of the input space.

46
Q

The number of neurons in the input layer of a perceptron is equal to the number of input patterns.

47
Q

You need access to the state of all neurons in a multilayer perceptron when updating all weights through backpropagation.

48
Q

Consider the Hopfield network. If a pattern is stable it must be an eigenvector of the weight matrix.

49
Q

If you store two orthogonal patterns in a Hopfield network, they must always turn out unstable.

50
Q

Kohonen algorithm learns convex distributions better than concave ones.

51
Q

The number of N-dimensional Boolean functions is 2^N.

A

False. it is (the number of choices)^(the number of inputs)

52
Q

The weight matrices in a perceptron are symmetric.

53
Q

Using g(b)=b as activation function and putting all thresholds to zero in a multilayer perceptron allows you to solve some linearly inseparable problems.

54
Q

You need at least four radial basis functions for the XOR-problem to be linearly separable in the space spanned by the radial spaces.

55
Q

Consider p>2 patterns uniformly distributed on a circle. None of the eigenvalues of the covariance matrix of the patterns is zero.

56
Q

Assume that the weight vector in Oja’s rule corresponds to a stable steady state after a given iteration. The weight vector may change in the next iteration.

57
Q

If your Kohonen network is supposed to learn the distribution P(xi), it is important to generate the patterns xi^(my) before you start training the network.

58
Q

All one-dimensional Boolean problems are linearly separable.

59
Q

In Kohonen’s algorithm. the neurons have fixed positions in the output space.

60
Q

Some elements of the covariance matrix are variances.