Module 13: Neural Networks Flashcards

1
Q

Which of the following are true about linear classifiers? Please check all that apply.

They can be used for both regression and classification.
All sets of data points are linearly separable.
When using the perceptron learning rule, the weights are updated when the actual output does not match the hypothesis output.
The learning rule must be applied to one example at a time.

A

They can be used for both regression and classification.

When using the perceptron learning rule, the weights are updated when the actual output does not match the hypothesis output.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Which of the following best characterizes the difference between parametric and nonparametric models?

A

A parametric model has a fixed size on the number of parameters.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

T/F
For linearly separable data, there exists only one decision boundary that separates the classes.

A

False

There could be multiple boundaries, but their performance may vary.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Y/N

A

Is it possible that the assignment of observations to clusters do not change between successive iterations in K-Means?

When approaching the converge point, the state may not change.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Relative to single-layer perceptrons, neural networks gain their power from

A

stacking of layers

Multi-layer architecture of neural network provides enough parameters to get better performance on the task.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

In neural networks, nonlinear activation functions: Please check all that apply.

  • Make it possible to do the gradient calculation in backpropagation, as opposed to using step function which isn’t differentiable.
  • Help to learn nonlinear decision boundaries.
  • Are applied only to the output units.
  • Always output values between 0 and 1.
A

Make it possible to do the gradient calculation in backpropagation, as opposed to using step function which isn’t differentiable.
Help to learn nonlinear decision boundaries.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

T/F
When performing k-means clustering, each observation always starts in its own cluster, and then pairs of clusters are merged in each iteration.

A

False

This is only true of agglomerative clustering.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly