College 4 Flashcards
Name and define 3 properties of convolution.
- sparse connectivity: convolution kernel is much smaller than the input (less connections)
- Parameter sharing:
kernel coefficients are identical for each input location - Equivariant representations:
convolution value covaries with input value (I you shift your input (image) your output is going to be the same with the same shift)
If you have 2d convolution, with an 32x32 input and a 3x3 filter how many parameters do you have to learn? And what will be the size of the feature map
- 9 (3x3)
2. 28 x 28
If you apply 6 filters in convolution how many feature maps do you get?
6
What is the filter size of a 2d convolution for an input with N channels?
(3,3,N)
1,1,2,4
5,6,7,8
3,2,1,0
1,2,3,4
What is the result of applying 2d max pooling with a 2x2 filter and stride: 2?
6,8
3,4
If you have a convolution over a 7 by 7 input
Filter size: 3x3
stride: 1
What is the output size?
5x5
If you have a convolution over a 7 by 7 input
Filter size: 3x3
stride: 2
What is the output size?
3x3
What can be an advantage of maxpooling?
more robustness (to little shifts in the input) / better generalisation
What can be an advantage of increasing strides?
efficiency / space reduction
If you have a convolution over a 8 by 8 input Filter size: 3x3 stride: 3 padding: 2 What is the output size?
4x4
How do you calculate the width and height of the output size
width_out = ((Width_input - filter_width + 2 x padding) / Stride) + 1
height_out = ((Height_input - filter_height + 2 x padding) / Stride) + 1
If you have a convolution over a 5 by 5 input Filter size: 3x3 stride: 2 padding: 1 What is the output size?
3x3
If you have a convolution over a 64 by 64 by 3 input
Filter size: 4x4x3
filters: 32
stride:2
What is the number of feature maps?
What is the output width and height?
What is the number of parameters of the convolutional layer?
output width: 31 output height: 31 number of feature maps: 32 number of parameters: 1568 (4x4x3 x32 + 1 , 1 for bias)
Define: transposed convolution
A specific transformation is not always useful, so a more robust way to upsample is to learn som filters that allow going from a feature map to a larger one.
Apply transposed convolution:
input =
0, 1
2, 3
kernel =
0, 1
2, 3
0, 0, 1
0, 4, 6
4, 12, 9