DL-05a - Convolutional neural network Flashcards

1
Q

DL-05a - Convolutional neural network

How do you perform convolution on discrete data?

A

Slide a filter (kernel) over the input data matrix, applying element-wise multiplication, and sum the results to create a new output matrix.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

DL-05a - Convolutional neural network

Describe how to perform convolution for 1 output element.

A

(See image)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

DL-05a - Convolutional neural network

Describe what comes out of after one pass. (See image)

A

A WxHx1 tensor (1 channel) at a time. Can have multiple output channels.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

DL-05a - Convolutional neural network

What is padding?

A

The process of adding extra pixels or data around the input image or feature map to maintain the spatial dimensions during convolution.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

DL-05a - Convolutional neural network

What is stride?

A

The number of units by which the filter slides over the input during the convolution process. I.e. jumping factor.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

DL-05a - Convolutional neural network

Why are fully connected layers not sufficient for image data?

A

When flattening the image, you lose information about spatial correlation (adjacency information).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

DL-05a - Convolutional neural network

Why is the number of parameters in a CNN lower than in a fully connected network?

A

The number of parameters in a CNN is lower because it uses shared weights and local connectivity, reducing redundancy and computational complexity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

DL-05a - Convolutional neural network

How many parameters are in a convolutional layer?

A

n(wh*ch) + n

w, h = x size
ch = number of channels
W = Kernel of size w * h * ch
b = bias term
Use n kernels per input, add n biases -> n(Wx + b) -> n((wh) + 1) -> nw*h + n

(See image)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

DL-05a - Convolutional neural network

Is a smaller or larger kernel preferred?

A

Depends.
- Large captures global features
- Small captures local features

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

DL-05a - Convolutional neural network

What kernel size captures global features?

A

Large kernels.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

DL-05a - Convolutional neural network

What kernel size captures local features?

A

Small kernels.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

DL-05a - Convolutional neural network

What does the pooling layer do? (3)

A
  • Downsample spatial dimensions.
  • Reduces number of params
  • Helps control overfitting
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

DL-05a - Convolutional neural network

What are the common pooling layer types? (2)

A
  • Max pooling
  • Average pooling
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

DL-05a - Convolutional neural network

Is average or max pooling better?

A

Max pooling has been shown to work better in practice.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

DL-05a - Convolutional neural network

What is an alternative to pooling?

A

Using conv layers with stride >1.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly