Segment/Depth Anything Flashcards

1
Q

How the focal loss helps in an imbalanced classification

A

Focal loss is takes Cross Entropy loss and adds a modulating factor that helps to bring the loss further down when an example is classified well. That makes it focus more on hard examples.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Why is focal loss good for segmentation

A

In segmentation we have class a heavy class imbalance because most of the pixels are negative and only a few are positive for the specific class we want.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the task proposal for segmenting anything

A

Return a valid segmentation mask given any prompt

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What can be ambiguous in a point prompt for a segmentation

A

A point can be the whole object, a part of the object or the sub-part of the object.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

In the training of the model, how did the vit evolved?

A

They started with vit-base and moved to vit-huge in the 2nd stage.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Write the equation for focal loss & cross-entropy loss

A

FL=((1-p_t)^gamma)*CE
CE = -log(p_t)
p_t = p if y==1 else 1-p

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

List depth anythin main contributions

A

Train on both labeled and unlabeled data to create a strong small student network.
Unlabeled training with strong perturbations: gaussian blur, color jitter & CutMix.
Semantic priors loss for encoder comparison.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Explain the idea behind semantic prior method

A

Create a loss from the student model encoder to a frozen big encoder to enforce producing similar feature extraction.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Explain in words the equation of the semantic prior method

A

take a cosine similarity of each feature value and divide it by the amount of features (take 1 minus for loss)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

How to combat semantic prior when pixels can be in different depths but of the same object

A

Create an alpha threshold. and when the cosine would be greater then the pixel would not be considered in the loss.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

List the strong perturbations they used

A

colorJitter
GaussianBlur
CutMix

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Explain the steps to create the CutMix loss.

A

We take 2 unlabeled images
We take the teacher model and predict on both images.
We initialise a mask
We use the student model to predict on the combined images with the mask.
We take the distance of the prediction between: Student network multiplied by mask. Teacher network multiplied by the mask.
The distance is the same depth loss distance.
Do the same for outside the mask.
Combine the results

How well did you know this?
1
Not at all
2
3
4
5
Perfectly