lesson_3_flashcards

1
Q

What is regularization in deep learning?

A

Techniques to prevent overfitting by penalizing large weights or encouraging sparsity, such as L1, L2, or dropout regularization.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

How does dropout regularization work?

A

At each iteration, nodes are randomly ‘dropped’ (deactivated) with a probability (p), preventing over-reliance on specific features.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is batch normalization?

A

A method to normalize layer outputs during training, improving gradient flow and learning stability.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is data augmentation?

A

Techniques like flipping, rotating, or adding noise to expand datasets artificially, improving model generalization.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Why is weight initialization important?

A

Proper initialization ensures effective gradient flow, avoiding vanishing or exploding gradients, and helps models converge faster.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is Xavier initialization?

A

A method to maintain consistent variance of activations across layers by scaling weights based on the number of input and output nodes.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What are optimizers in deep learning?

A

Algorithms like SGD, Adam, and RMSProp used to adjust model parameters to minimize the loss function.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is momentum in optimization?

A

A technique to smooth updates by incorporating past gradients, improving convergence speed and stability.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the vanishing gradient problem?

A

A phenomenon where gradients become very small as they propagate back through layers, slowing or halting learning.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What are the key considerations for designing a neural network architecture?

A

Understanding the data, selecting appropriate layers, ensuring gradient flow, and leveraging domain-specific insights.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the purpose of normalization in deep learning?

A

To standardize input or layer data, ensuring balanced and effective gradient flow throughout the network.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is data preprocessing?

A

Preparing raw data for training through techniques like normalization, scaling, or encoding.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

How does ReLU improve gradient flow?

A

By providing non-saturating gradients for positive inputs, ensuring gradients remain large enough to drive learning.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is the importance of hyperparameter tuning?

A

Adjusting settings like learning rate, batch size, or regularization strength can significantly impact model performance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is the difference between overfitting and underfitting?

A

Overfitting occurs when a model performs well on training data but poorly on unseen data, while underfitting fails to capture patterns in training data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly