ANN Lecture 1 & 2 - Introduction and Multi Layer Perceptron Flashcards

1
Q

Marr’s Level of Analysis

A

Computational Level:
What is the aim of the process?
Which problem is solved?
[Psychology]

Algorithmic Level:
Which algorithm is used?
How is the problem solved?
[Cognitive Science]

Implementational Level:
How is the algorithm realized
or implemented?
[Neuroscience]

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Explicit Knowledge

A

Can be formalized and can be implemented in an algorithm, so a computer will probably be better than a human.
e.g.: chess, integrating complex formulas, navigation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Implicit Knowledge

A

Can’t be formalized. It is knowledge that a human has and can use, but cannot access verbally.
e.g.: walking, juggling, all other motor skills, object/speech recognition etc.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Perceptron
and
XOR-Problem

A

Output =
Step-Activation-Function(Weighted input-vector + Bias)

A perceptron seperates a space into two parts since the perceptron is a hyperplane. But XOR is is not linearly seperable, so a perceptron can not learn it. A MLP (Multi-Layer-Perceptron) though is able to solve the XOR-Problem.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Multi-Layer-Perceptron

A
  • Feed-forward neural network.
  • > Input-Layer -> Hidden Layers -> Output Layer
  • Forward step
  • Back propagation step
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Tasks for a MLP

A

Regression:
Finding the relation ship between two continuous variables.

Classification:
Predicting a label for a sample.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

One Hot Encoding

A

Vector-length of one-hot encoded vector is equal to the number of labels. Label can be interpreted as a discrete probability distribution. (Single one, remainder zeros)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Output Function for Classification

A

Softmax Function:
Y_i = (e^d_i)/sum_j(e^d_j)
Takes an Vector as input and turns it into a discrete probability distribution.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Loss Function for regression

A
  • Sum squared error (approximate the real valued outputs as good as possible)
    Loss = 1/2 * SUM_i(T_i - Y_i)^2
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Loss Function for Classification

A

Cross Entropy (Similarity matters more than the distance between output and target)

Loss = - SUM_i(T_i * log(Y_i))

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

MLP - Forward Step

A
Drive = Weight Matrix x Input-Vector + Bias-Vector
Activation = Sigmoid-Function(Drive)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Activation-Functions

A

Heavisided Step Function:
(One or Zero)
Logistic Function:
(Soft transition between One and Zero, sigmoid)
TanH-Function:
(Soft transition between One and minus One, sigmoid)
ReLu-Function:
(Zero up to Zero, Linear slope from zero)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Supervised Learning

A
  • We have input data with corresponding targets (labels)
  • We are trying to find the underlying mapping of targets to the input
  • This mapping should explain all our data and generalized to unseen data.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly