Deep Learning Flashcards

1
Q

What is the heart of deep learning?

A

Neural networks which are composed of neurons.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

How is a neuron defined?

A
  • Input = a vector of numeric inputs
  • Output = a scalar
  • Parameters
    1) A vector of weights, one for each input plus a bias term (b)
    2) An activation function f
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is a single-neuron neural network?

A

A perceptron

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

How to train a neural network?

A

With the perceptron algorithm, within which each iteration is termed an “epoch”.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What are some common non-linear activation functions?

A
  • (Logistic) sigmoid
  • Hyperbolic tan (“tanh”)
  • Rectified Linear Unit
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

How to get neural network to be equivalent to logistic regression?

A

A neural network with a single neuron and a sigmoid activation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is the power of neural nets?

A

Stacking multiple neurons together in different ways.

  • Layers of parallel neurons of varying sizes
  • Feeding layers into hidden layers of varying sizes

e. g. a “fully-connected feed-forward neural network” takes the following form:
- the INPUT LAYER is made up of the individual features
- each HIDDEN LAYER is made up of an arbitrary number of neurons, each of which is connected to all neurons in the preceding layer, and all neurons in the following layer
- the OUTPUT LAYER combines the inputs from the preceding layer into the output

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What would be the Output Layer Activation Function in a neural net if you wanted to do multiclass classification?

A

SOFTMAX

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What would be the Output Layer Activation Function in a neural net if you wanted to do regression?

A

identify function or Sigmoid or tanh

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the requirement for activation functions in Neural Nets?

A

They must be non-linear otherwise it’ll just collapse down to a linear regression model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Describe the Universal Approximation Theorem, what’s so good about it?

A

A feed-forward neural network with a single hidden layer (and finite neurons) is able to approximate any continuous function.

It’s good because it’s possible for a feed forward neural net with non-linear activation functions to learn any continuous basis function DYNAMICALLY, unlike SVMS eg, where the kernel is a hyperparameter

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

How to train a Neural Net with Hidden Layers?

A

Train neural nets with BACK PROPAGATION

  • Compute errors at the output layer each weight using partial differentiation
  • Propagate those errors back to each of the input layers
  • STILL HAS A LEARNING RATE
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

CONS of Neural Nets?

A

Prone to chronic overfitting
- Due to large number of parameters

*Regularisation is critical

Or,
EARLY STOPPING, stop training when performance peaks on the dev. data

OR,
DROPOUT

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Theoretical Properties of Neural Networks

A
  • Can be applied to either classfification or regression
  • Parametric
  • Batch
  • Relies on continuous features
  • Assuming at least one hidden layer
  • Complex to train, but produce relatively compact models
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Why is a perceptron (which uses a sigmoid activation function) equivalent to logistic regression?

A
  • A perceptron has a weight associated with each input
  • The output is acquired by applying the activation function f(x) = 1 / (1 + e^-x) to the linear combination of inputs, which simplifies to f(sumof(wi * ai)) - this is the same as logistic regression function
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What makes a feed forward neural network more interesting?

A

Basically, each node in the first layer is a logistic regression model.

However, these values are themselves the input to another layer, so that we are effectively progressively stacking lots of logistic regression models.

17
Q

Why is a neural network suitable for deep learning?

A

Hypothetically, the weights across the network describe some useful properties of the model.
In effect, we hope to engineer the necessary features to solve our problem, based only on the simples inputs.

18
Q

What is significant about the representation that we attempt to learn?

A

The representation is simultaneously useful for solving the problem and un-interpretable by humans.

Can discover properties of the data that were previously unknown to be useful.