5 Flashcards

1
Q

What is the intuition behind L2 loss function?

A

Measures the squared difference between the predicted value and the actual value, encouraging the model to minimize large errors.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the intuition behind L1 loss function?

A

Measures the absolute difference between the predicted and actual values. This makes it more robust to outliers because it does not square the differences, which would amplify large errors.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the intuition behind Smooth L1 loss function?

A

Combination of L1 and L2 loss, and it behaves differently depending on the size of the error. When the error is small, it behaves like L2 loss (squared differences), and when the error is large, it behaves like L1 loss (absolute differences). This makes it more robust to outliers than L2 while not being as harsh as L1 for small errors.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the intuition behind Negative Log Likelihood loss function?

A

Calculates the likelihood of the predicted class being correct. For a given class, if the model is confident and correct, the loss will be small. If the model is wrong, the loss increases sharply.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the intuition behind Cross Entropy loss function?

A

Compares the predicted probability distribution of classes with the actual distribution (one-hot encoded labels). It is particularly effective when used with softmax activations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is the intuition behind Binary Cross-Entropy loss function?

A

Measures the difference between two probability distributions, one being the predicted probability for class 1 and the actual label for the class

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is the intuition behind Cosine Embeddings Loss function?

A

Calculates the cosine of the angle between the predicted and actual vectors. If they are similar (i.e., their angle is small), the cosine similarity will be close to 1, and the loss will be small.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Why do we need to augment the data?

A

Improving the performance of machine learning models by increasing data diversity, improving generalization, preventing overfitting, and making models more robust to variations in real-world data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

“Black box” issues

A

Lack of Interpretability (not transparent processes)
Visualizing Filters
Visualizing Activations
Maximally Activating Patches
Occlusion Experiments (masking part of the picture)
Saliency Maps

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Advantages and disadvantages of Deep Learning network

A

Advantages: easy to implement, tailored approach, performance boost, can help with overfitting

Disadvantages: time consuming, trial and error, relational gaps, can lead to overfitting

How well did you know this?
1
Not at all
2
3
4
5
Perfectly