04 neural networks* Flashcards

1
Q

artificial neurons

A

input signals = independent variables
triggers when activated = activation functions
nerve impulse = output value

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

perceptron

A

input layer = linear regression
perceptron = calculate weighted sum on input, bias factor
produces output value

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

activation functions

A

threshold function
if x >= 0, output 1
if x < 0, output 0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

problems with perception

A

non linear curves, eg. XOR problem`

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

feed forward neural network

A

one or more nodes in hidden layer
- node in one layer is linked to every single node on the second layer
- more than one activation functions can be used
- simulate thinking process of multiple input signals with different combinations

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

neuron

A

non input node (hidden node) and output node
can be used with:
- sigmoid
- rectifier linear unit RELU
- hyperbolic tangent tanh

to account for non-linear

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

sigmoid

A

1/ (1 + pow(e, -x))
0 to 1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

hyperbolic tangent

A

(pow(e, x) - pow(e, -x)) / (pow(e, x) + pow(e, -x))
-1 to 1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

rectifier linear unit

A

max(0, x)
0 to infinite

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

training artificial neural network

A
  1. initialise weights
  2. set epochs (how many rounds)
    - forward propagation
    - predict results y
    - compute errors (mean squared or cross entropy)
    - back propagation (update weights)
  3. repeat
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

back propagation

A
  • update weights to minimise loss/cost/error
  • done with gradient descent
  • plot sum of square errors (y) and intercept (x) (form a happy curve)
  • by measuring the gradient, we can find out the point where error is the lowest, therefore selecting that parameter (gradient = 0 at the lowest point of the curve)
  • each step to measure the next gradient is done using “learn rate”
  • if learning rate too small, slow traing
  • if learning rate too high, miss optimal point
  • max step configured to usually be 1000
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

overfitting

A

common in deep learning when high numbers of hidden nodes and layers
solutions:
- complexity
- dropout
- regularisation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

tuning dataset

A

split data into training and testing set

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

machine learning vs deep learning

A

ML
input -> feature engineering -> classification -> output

Deep learning
input -> feature extraction + classification -> output

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

padding

A

add 0s to the edge so that output and input is the same size

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

why should we use pooling

A
  1. reduce number of parameters and computation
  2. avoid overfitting
17
Q

recurrent neuron networks ** how to draw

A

h(t) represent memory
y(t) represent output

output h is combined with new iteration to produce new output

18
Q

long short term memory ** explain and rmb how to draw

A

unique kind of RNN

x(t) represent inputs
c(t) represent
h(t) represent

  1. first layer -> forget gate outputs f(t)
    prev output + current input + activation function
    combine with c(t-1),
    if c(t-1) = 0, discard,
    if c(t-1) = 1, keep
  2. second layer -> input gate layer
    inputs are
    outputs are i(t) and ~c(t)
  3. candidate cell
    update c(t) to a new value
    c(t) = f(t) * c(t-1) + i(t) * ~c(t)
  4. output layer