10 - The Algorithm that Put Paid to a Persistent Myth Flashcards

1
Q

What did Minsky and Papert prove about single-layer perceptrons?

A

They proved that single-layer perceptrons could not solve the XOR problem

This proof is often cited as a turning point in neural network research.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Who is Geoffrey Hinton?

A

A key figure behind the modern deep learning revolution

Hinton became interested in neural networks in the mid-1960s.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What influenced Hinton’s interest in how brains learn?

A

A mathematician friend exploring how memories are stored in the brain

This led Hinton to study the mind and neural networks.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What did Hinton study at university?

A

Physics and physiology

However, he found the curriculum insufficient regarding understanding the brain.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What book deeply influenced Hinton?

A

The Organization of Behavior by Donald Hebb

This book impacted Hinton’s thinking on neural networks and learning.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What was Hinton’s doctoral focus?

A

Solving constrained optimization problems using neural networks

Hinton believed multi-layer networks could eventually learn.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What was the key limitation of single-layer perceptrons according to Minsky and Papert?

A

They could not solve the XOR problem, which is a specific instance of a broader class of problems

This limitation led to skepticism about neural networks for some time.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is back-propagation?

A

A method for training multi-layer neural networks by propagating error corrections back through the network

Introduced by Rosenblatt in his work on neural networks.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What issue arises when initializing all weights in a neural network to zero?

A

All neurons produce the same output, leading to symmetry and ineffective learning

This problem prevents the network from detecting different features.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What did Rosenblatt suggest for updating weights in a neural network?

A

A stochastic process that introduces randomness to weight updates

This approach aimed to break symmetry in the network.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What was Hinton’s belief about the nature of neurons in neural networks?

A

Neurons had to be stochastic to ensure different learning outcomes

This belief was based on Rosenblatt’s argument about non-deterministic procedures.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What was Hinton’s experience in academia post-Ph.D.?

A

He faced rejection in the UK and eventually found a position in the US

This move was significant for his career in neural networks.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is the gradient descent method?

A

A technique to minimize error by updating weights in the opposite direction of the error gradient

Used in training neural networks to find optimal weight values.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is a major challenge with the error function in neural networks?

A

It is not convex and can have multiple local minima

This complexity makes finding the global minimum more difficult.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What phenomenon can occur with hill climbing algorithms?

A

The mesa phenomenon, where the algorithm gets stuck in flat regions of the error space

This can impede finding better solutions in optimization tasks.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is the hill-climbing technique?

A

A method where performance must improve to a local optimum where no small change in controls yields improvement.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What phenomenon can hill climbing encounter according to Minsky and Selfridge?

A

The mesa phenomenon.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What is the mesa phenomenon?

A

A situation where small tweaks to parameters do not improve performance or lead to large performance changes.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

What was Minsky and Papert’s view of multi-layer neural networks?

A

They had a dismal view, suggesting a deliberate sabotage of research into neural networks.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Who independently developed methods relevant to the backpropagation algorithm in 1960-61?

A

Henry J. Kelley and Arthur E. Bryson.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

What contribution did Stuart Dreyfus make in 1962?

A

He derived formulas based on the chain rule to augment the Kelley-Bryson method.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

Who demonstrated techniques for using stochastic gradient descent in 1967?

A

Shun’ichi Amari.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

What did Seppo Linnainmaa develop in 1970?

A

The code for efficient backpropagation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

What was the title of Paul Werbos’s 1974 Ph.D. thesis?

A

Beyond Regression: New Tools for Prediction and Analysis in the Behavioral Sciences.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

Who developed the modern version of the backpropagation algorithm in the early 1980s?

A

Rumelhart, Hinton, and Williams.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

What does the equation y = wx + b represent?

A

The output of a neuron given a weight w, bias b, and scalar input x.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

What is the delta rule used for?

A

Finding the weight and bias in a neuron.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
28
Q

What is the formula for calculating the error in the delta rule?

A

e = y - yhat.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
29
Q

What does loss represent in the context of the delta rule?

A

loss = (y - yhat)².

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
30
Q

What does MSE stand for?

A

Mean Squared Error.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
31
Q

In the context of the delta rule, what is the learning rate denoted by?

A

α (alpha).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
32
Q

What happens to the weight and bias during the update process?

A

They are adjusted by a small fraction of the gradient.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
33
Q

What is a hyperplane in machine learning?

A

A line that separates different classes in a dataset.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
34
Q

What is the XOR problem in neural networks?

A

A classification problem where data points cannot be separated by a single linear line.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
35
Q

What is required to solve the XOR problem?

A

At least two layers of neurons.

36
Q

What do the neurons in the first layer of a neural network for XOR do?

A

They find two lines to separate the data.

37
Q

What is the output of a neuron that takes in two inputs x1 and x2?

A

y = w1x1 + w2x2 + b.

38
Q

What is the purpose of the second layer in a neural network for XOR?

A

To create a weighted sum of the outputs of the first layer’s neurons.

39
Q

What is the output of a simple linear neuron?

A

y = w1 * x1 + w2 * x2 + b

40
Q

What is the role of the activation function in a neuron?

A

Transforms the weighted sum input into an output.

41
Q

What is a threshold function?

A

A function that outputs 1 if z > 0 and 0 otherwise.

42
Q

True or False: The threshold function is differentiable everywhere.

43
Q

What is the sigmoid function used for?

A

To create a smooth, differentiable activation function.

44
Q

As z tends to infinity, what does the sigmoid function approach?

45
Q

As z tends to minus infinity, what does the sigmoid function approach?

46
Q

What is the structure of a simple neural network for the XOR problem?

A

Three layers: input layer, hidden layer with two neurons, output layer with one neuron.

47
Q

Fill in the blank: The output neuron takes a weighted sum of the outputs of the two hidden neurons and passes that through a _______.

A

sigmoid activation function

48
Q

What is the loss function defined as in this context?

A

L = e^2, where e is the error (y - yhat).

49
Q

What technique was developed for calculating partial derivatives in neural networks?

A

Backpropagation

50
Q

Who were the key researchers in developing the backpropagation algorithm?

A

Werbos, Rumelhart, Hinton, Williams

51
Q

What does backpropagation allow us to compute?

A

The gradients of the loss function with respect to weights and biases.

52
Q

What is required for the chain rule to be applied in backpropagation?

A

Every operation must be differentiable.

53
Q

What is the significance of breaking symmetry in neural networks?

A

To ensure that neurons learn different features and do not produce the same output.

54
Q

How can symmetry be broken during the initialization of weights?

A

By setting initial weights to small random values.

55
Q

What is the role of the output layer in a neural network for classifying digits?

A

It has one neuron for each digit class, firing the corresponding neuron for the detected digit.

56
Q

What is a multi-layer perceptron?

A

A fully connected deep neural network.

57
Q

What does a fully connected neural network mean?

A

Each neuron in a layer receives inputs from all neurons in the previous layer.

58
Q

Fill in the blank: The first layer in a neural network for image recognition has _______ neurons, one for each pixel.

59
Q

What type of activation function was initially used in binary threshold neurons?

A

Threshold activation function

60
Q

What is the main advantage of using a sigmoid function over a threshold function?

A

It is differentiable everywhere.

61
Q

What does the backpropagation algorithm enable networks to learn?

A

Interesting representations of data.

62
Q

True or False: Neural networks require predefined features from the data.

63
Q

What happens to the output of neurons in a well-trained network for digit recognition?

A

The correct digit neuron fires significantly more than others.

64
Q

What is the primary challenge mentioned regarding complex networks with many layers?

A

Calculating partial derivatives becomes unrealistic.

65
Q

What is the main advantage of neural networks over traditional algorithms like support vector machines?

A

Neural networks can learn to represent features internally without needing predefined features.

66
Q

What are the features needed to separate circles from triangles in a two-dimensional dataset?

A

Nonlinear features such as [x1, x2, x1x2].

67
Q

What is the role of hidden units in a neural network according to Rumelhart, Hinton, and Williams?

A

Hidden units represent important features of the task domain.

68
Q

What distinguishes backpropagation from earlier methods like the perceptron-convergence procedure?

A

Backpropagation allows for the creation of useful new features automatically.

69
Q

Who are the authors of the influential paper on backpropagation?

A

Rumelhart, Hinton, and Williams.

70
Q

What significant event occurred in 1987 related to Rumelhart’s career?

A

Rumelhart moved to Stanford University.

71
Q

What is the sigmoid function represented by?

A

u = 1 + e^{-z}

72
Q

What must the output of each neuron in a neural network layer pass through?

A

An activation function.

73
Q

Fill in the blank: The output of layer 1 after activation can be expressed as _______.

A

a1 = σ(z1)

74
Q

What is the formula for calculating the error in a neural network?

A

e = (y - ŷ)

75
Q

What is the loss function represented by in the context of a neural network?

76
Q

What is the purpose of calculating the gradient of the loss function?

A

To update the weights and biases in the neural network.

77
Q

True or False: More hidden layer neurons always result in a rougher decision boundary.

78
Q

What type of functions can be used as activation functions in neural networks?

A

Any differentiable function.

79
Q

What does the term ‘gradient’ refer to in the context of neural networks?

A

The rate of change of the loss function with respect to weights and biases.

80
Q

What is the output of the final layer in a simple neural network with one output neuron?

A

ŷ = σ(z4)

81
Q

What condition must be satisfied for the activation function in a neural network?

A

It must be differentiable.

82
Q

What is the significance of learning features automatically in neural networks?

A

It allows for more effective and flexible modeling of complex data.

83
Q

Fill in the blank: The ability to create useful new features distinguishes _______ from earlier methods.

A

backpropagation

84
Q

What does the term ‘delta rule’ refer to in neural networks?

A

A method used to update the weights based on the gradient.

85
Q

Who independently developed an algorithm achieving similar results to backpropagation?

A

Yann LeCun.

86
Q

What illness did Rumelhart suffer from before his retirement?

A

Pick’s disease.