L2_NCC_Perceptron Flashcards

1
Q

Cognitive functions have to be investigated on 3 levels?

A

• Computational Level
What does a cognitive function do?
• Algorithmic Level
What is the functional organization within a cognitive module?
• Implementational Level
What is the physical/physiological realization of this algorithm?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Explain the 3 Steps of Nearest Centroid Classifier (NCC)

A
  1. Calculate the class means of the Prototypes w1 and w2
  2. find the linear classification boundary: w.Tx-ß = 0
    with w = w1 - w2 and ß =1/2 * (w1.T
    w1 -w2.T*w2)
  3. Classify the class label sign f(x)= +1 / -1
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Basic Artificial Neural Networks model (4 point)

A
  • Input nodes xi receive information
  • Inputs are multiplied with a weighting factor wi and summed up
  • Integrated input is mapped through some (non-linear) function f (·)
  • f(x)= +1 if x is preferred stimulus / -1 if x is any other stimulus
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Goal of the The Perceptron Learning Algorithm

A

Binary classification of multivariate data x ∈ RD

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Input of the The Perceptron Learning Algorithm

A

Learning rate η and N tupels (xn,yn) where

  • xn ∈ RD is the D-dimensional data
  • yn ∈ {−1, +1} is the corresponding label
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Output of the The Perceptron Learning Algorithm

A

Weight vector w ∈ RD such that w.T*xN =

≥0 if yn = +1 / < 0 if yn = −1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is a good w?

A

We need an error function
that tells us how good w is.
Then we chose w such that the error function is minimized.
Perceptron error EP is a function of the weights w, where M denotes the index set of all misclassified data xm

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Explain the Perceptron Learning Algorithm (2 steps)

A
  1. Initialize wOLD (randomly, 1/n, …)
  2. While there are misclassified data points
    Pick a random misclassified data point xm
    Descent in direction of the gradient at single data point xM:
    Em(w) = −w.⊤ * xM * yM
    ∇Em (w) = −xm ym

wNEW ← wOLD-η∇Em(wOLD) = wOLD + η * xM * yM

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

[Novikoff, 1962; Rosenblatt, 1962] about The Perceptron Learning Algorithm

A

If there is a solution, the perceptron algorithm will find it in a finite number of steps

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Convergence on non-linearably-separable sets:

A

wnew ← wold + ηxmym
t→∞
• Proven for variable learning rate η(t), with η(t) → 0
• Best convergence speed is achieved for η(t) ∼ 1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Problems with Nearest Centroid Classification

A

Non-linear Data

Correlated Data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

High accuracy but ”lousy” weight vector - why?

A

Misconception: signal channels (here: pixels) with large classifier weights are strongly related to the class label

The purpose of the weight vector is two-fold: amplify the signal of interest, while at the same time surpress signals of no interest.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly