Week 1: principles of deep learning in artificial networks Flashcards

1
Q

What is a deep network?

A

A learning network that transforms or extracts features using multiple nonlinear processing units arranged in multiple layers with hierarchical organisation and different levels of representation and abstraction

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

How do complex outcomes emerge in deep networks?

A

From interactions between many simple steps

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is a representation in a deep network?

A

The information the computer is given. Each representation is built from an earlier representation that can transfer the features and extract complex features from simpler features

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What are the 4 operations in a linear-nonlinear layer

A

Filter/convolve
Threshold/rectification
Pool
Normalise

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the purpose of the filter/convolve operation?

A

To determine how well each group of nearby pixels matches each of a group of filters

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

How does the filter operation work?

A

The input is a pixel map

  • convolution step looks for a pattern in a group of neighbouring pixels that corresponds to the convolution filter
  • if source pixels follow the filter pattern the results is a high value, if input area is all same brightness result will be 0, if source pixels are opposite to the filter the result will be negative
  • output is feature maps
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is the purpose of the threshold/rectify operation?

A

Introduce nonlinearity by setting negative activations of units to zero (and maybe set a maximum activation)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the goal of the ReLU activation function?

A

To only activate the output feature map if its value reaches a certain threshold

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the purpose of the pooling operation?

A

Downsample the units to improve computational efficiency

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

How does the pooling operation work?

A

Typically takes the maximum of a square of 2x2 neighbouring units of the feature map

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the purpose of the normalise operation?

A

Rescale responses of each feature map to have mean 0 and standard deviation 1 so each feature map contributes similarly to classification

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Why is normalisation necessary?

A

The range could be different between feature maps, weighting some more than others

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What tasks are deep learning useful for?

A

Useful for achieving tasks that are difficult to describe formally. Tasks that are difficult for computers but intuitive for humans

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is the final layer of the network

A

The final fully-connected layer links pattern of most abstracted, top-level features to required response
-the last feature map is flattened into a line of independent units where each is connected to all the others

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What does the softmax function do?

A

Determines the probability of the desired response

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How does the softmax function work?

A

Each input image has a score reflecting the match bewegen the top layer’s activation pattern. The score is converted to a probability that this input image falls into each category

17
Q

How are convolution filters and weights related?

A

The convolution filters form the weights of connections between the nodes in a neural network

18
Q

How does a network learn the weights of the connections?

A

Using back-propagation error

19
Q

What does back-propagation error do?

A

Adjusts the filter structure - the link between layers

20
Q

What is the basis for back-propagation error?

A

The match or conflict of expected and actual outputs

21
Q

Why do filters generally have a single set of weights for all positions in the feature map?

A
  • if a feature is useful to compute at one position, it is probably useful at another position
  • the filter values are weights that need to be learned, it is computationally demanding if the set is too large
  • the convolution operation is a fast matrix function, if filters are not fixed the convolution operation cannot be used
22
Q

What is a definition of machine learning?

A

A computer program is said to learn from experience E with respect to some class of tasks T and performance measure P if its performance at tasks in T, as measured by P, improves with experience E