Deep Learning Flashcards

1
Q

What is an activation function? Why is it important?

A

The activation function is the function the neuron applies to the weighted sum of its inputs. It is important because it is how you introduce nonlinearity, since the function can be anything.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is a step function?

A

A function that produces binary output based on some threshold.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is a perceptron?

A

A perceptron is a neuron with the step function z > 0.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What does MLP stand for?

A

Multi-layer perceptron.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is a multi-layer perceptron?

A

It’s when you chain together multiple layers of logic gates built from perceptrons, to represent complex logical functions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is “universality”?

A

If a network has universality, then it can approximate any continuous function arbitrarily with just one hidden layer, given enough units.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

The definition of universality talks about one hidden layer. If that’s the case, why would you want multiple layers?

A

You can represent things more compactly if you use multiple layers.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What does “differentiable” mean?

A

A function is differentiable if you can find the slope of its tangent at any point.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is ReLU?

A

ReLU is an activation function that is 0 for x < 0 and x for x > 0.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Why do we like to use ReLU instead of step function?

A

The step function has no useful gradient since it is just flat everywhere. ReLU does have a useful gradient.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is risk? What is empirical risk?

A

Risk is the actual expected loss over the entire real distribution of data. We don’t know the real distribution of data, we just have our training set, so we only have empirical risk (the expected loss over the training set distribution).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What’s the point of the VC Dimension?

A

The VC dimension proves that the generalization bound is finite, even for an infinite hypothesis set. Generalization bound is risk <= empiricalRisk + error. And VC dimension says that error is finitely bounded.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is the objective function?

A

It’s the loss function that we want to minimize.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is the vanishing gradient problem?

A

It’s when the gradient goes to zero – which often happens because you’re multiplying a lot of tiny weights together.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is it called when the gradient goes to zero?

A

The vanishing gradient problem.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is “over-saturation”?

A

It’s when the activation function is in a flat region – such as < 0 for the ReLU – so the gradient is zero, which causes the vanishing gradient problem.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Give three solutions to the vanishing gradient problem

A

Any of these four: Better initialization, Better activation function, Regularization to rescale the gradient, Gradient clipping

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What does it mean if a model is “non-identifiable”?

A

It means that the model can get the same minimum loss with multiple different settings of the weights.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Why is it a problem is a model is non-identifiable?

A

If we were using weights to draw conclusions about the model, having two different sets of weights both claim to be optimal would make that difficult.

20
Q

What does it mean if a function is “convex”?

A

It means that it has a single minimum, and a bowl shape going towards that minimum. Specifically, it means that if you pick any two points and draw a line between them, that line won’t intersect the function.

21
Q

Are deep neural networks convex?

A

Absolutely not.

22
Q

Give three examples of problems caused by nonconvexity

A

Any of: Local minima, Saddle points, Cliff structures, Asymptotes

23
Q

Why don’t you want to just initialize all the weights to zero?

A

For ReLU, this yields zero gradients. In general, this means every hidden unit with the same activation function would have the same output.

24
Q

What is the most common way to initialize?

A

Sampling from a distribution

25
Q

What is “warm start”?

A

It’s where you train the model on a different objective just to get some initialization weights, and then switch to the actual objective.

26
Q

What is dropout regularization?

A

Randomly remove hidden units by hardcoding their activation to 0, randomly choosing in each iteration of gradient descent. This simulates a bunch of different models.

27
Q

What is ensemble regularization?

A

Train a bunch of different models and average their predictions.

28
Q

What is variational noise and why would you do it?

A

It’s when you add noise from a Gaussian distribution during the forward pass part of training. It’s a way to regularize.

29
Q

What is a convolution?

A

It’s a window-based operation where you combine two vectors to extract spatial or contextual information.

30
Q

What is the vector you use to do the convolution called?

A

The kernel

31
Q

What is the kernel, in the context of CNNs?

A

It’s the vector you use to define and apply the convolution to your input

32
Q

What does CNN stand for?

A

Convolutional Neural Network

33
Q

What is a Convolutional Neural Network?

A

It’s a neural network where you have a convolutional layer.

34
Q

Why would you want to use a CNN?

A

It’s able to find local structure, and interactions between nearby components.

35
Q

What is “pooling”?

A

It’s where you combine the outputs of a convolutional layer, with something like average or L2 norm, to zoom out on the convolution region.

36
Q

What does RNN stand for?

A

Recurrent Neural Network

37
Q

What is unique about a Recurrent Neural Network?

A

It captures information about a sequence of values, by adding a loop to the hidden unit.

38
Q

What is the “transition function” in the context of an RNN?

A

The transition function is the term for the weights, because they are constant and don’t vary with time.

39
Q

What is the “unrolled network”?

A

It’s an RNN where you’ve expanded the loop into a bunch of iterations side by side – so the network at time 1, then feeding into itself at time 2, and then time 3. It ends up looking like a ladder.

40
Q

What does LSTM stand for?

A

Long short-term memory

41
Q

What is unique about an LSTM?

A

It has memory units that hang onto information over some duration.

42
Q

What does GAN stand for?

A

Generative Adversarial Network

43
Q

What are the two parts of a GAN?

A

First, a Generator creates fake examples to try and fool the discriminator. Then the Discriminator learns to discern real examples from fake generated ones. And the two are learned together.

44
Q

What is an “adversarial example”?

A

It’s one that has been minimally changed from a real example, but that causes the network to produce a very different prediction.

45
Q

What is “dataset augmentation”?

A

It’s when you add a bunch of new manually-concocted examples to your training set to try and help the model generalize.

46
Q

What are example of how people do dataset augmentation?

A

They might crop the data, or rotate it, or add noise, or erase random bits.