Activation+ Flashcards

Exam2

1
Q

Should a linear function be used as the activation function in the hidden layers of a neural network?

A

No, because this is no different than linear regression

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Can it be used at the outputlayer?

A

Yes if it is in a regression where the output can be positive or negative

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What activation function is recommended for the hidden layer of neural networks because it speeds up training time?

A

ReLU (Rectified Linear Unit)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Can the sigmoid function be used in the hidden layer of a neural network?

A

Yes, but it converges slower than ReLU

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What activation function should be used at the output layer of a model that predicts home prices?

A

ReLU

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What activation function should be used at the output layer of a model that predicts temperatures?

A

Linear

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What activation function should be used at the output layer of a model that recognizes images of 15 animals?

A

SoftMax

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What activation function should be used at the output layer of a binary classification problem?

A

Sigmoid

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

The SoftMax output of a 4 class classification problem outputs [.1,.1,.1,.7] - what class is the prediction?

A

The class with the highest probability is considered the predicted class, which is 0.7

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What role do activation functions play in creating the decision boundary of classification problems?

A

Allows you to find non-linear decision boundaries

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Is [.1,.1,.1,.1] possible output of the SoftMax function with 4 classes?

A

YES, it is possible because this scenario indicates that the model has assigned an equal probability of 0.1 to each of the four classes, suggesting uncertainty or lack of confidence in making a specific prediction

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Explain what the Adam optimizer does

A

The Adam Optimizer algorithm adapts the learning rate for each parameter, w. This leads to faster convergence and better performance compared to traditional optimization techniques with fixed learning rates

How well did you know this?
1
Not at all
2
3
4
5
Perfectly