Chapter 4, W5: Elements of Connectionist Cognitive Science Flashcards
Cartesian rationalism
The notion that the products of thought were rational conclusions drawn from the rule-governed manipulation of pre-existing ideas.
- The Cartesian view that thinking is equivalent to performing mental logic—that it is a mental discourse of computation or calculation —has inspired the logicism that serves as the foundation of the classical approach.
Empiricism
The view that the source of ideas was experienced; i.e., the source of all ideas is experience.
- Locke was one of the pioneers of empiricism, a reaction against Carteisn philosophy
- Locke argued for experience over innateness, for nurture over nature
Connectionist Elements
- Element 1: Association
- Element 2: Decision
Artificial Neural Network
The basic medium of connectionism is a type of model called an artificial neural network (AFN), or a parallel distributed processing (PDP) network. AFNs are:
- “neuronally inspired” networks
- built from simple processors (artificial neurons)
- learn from experience
- operate in parallel
Artificial neural networks are exposed to environmental stimulation—activation of their input units—which results in changes to connection weights.
Laws of Association (Arsitole)
Fundamental to connectionism’s empiricism is the key idea
of association: different ideas can be linked together, so that if one arises, then the association between them causes the other to arise as well.
- Contiguity or habit
- similarity
- contrast
The behaviour of a processor in an artificial neural network, which is analogous to a neuron, can be characterized as follows:
- the processor computes the total signal (its net input) being sent to it by other processors in the network.
- the unit uses an activation function to convert its net input into internal activity (usually a continuous number between 0 and 1) on the basis of this computed signal.
- the unit converts its internal activity into an output signal and sends this signal on to other processors.
A network uses parallel processing because many, if
not all, of its units, will perform their operations simultaneously.
The signal sent by one processor to another is a number that is transmitted through a weighted connection, which is analogous to a synapse.
Weight
The weight is a number that defines the nature and strength of the connection. For example, inhibitory connections have negative weights, and excitatory connections have positive weights. Strong connections have strong weights (i.e., the absolute value of the weight is large), while weak connections have near-zero weights.
Tabula Rasa
Tabula rasa, or the blank slate: the notion of a mind being blank in the absence of experience. Modern connectionist networks can be described as endorsing the notion of the blank slate.
This is because prior to learning, the pattern of connections in modern networks has no pre-existing structure. The networks either start literally as blank slates, with all connection weights being equal to zero or they start with all connection weights being assigned small, randomly selected values
Interpret the Figure: Modifiable Connections
Figure 4-1 is James’ biological account of association:
- Illustrates two ideas, A and B, each represented as a pattern of activity in its own set of neurons. A is represented by activity in neurons a, b, c, d, and e; B is represented by activity in neurons l, m, n, o, and p. The assumption is that A represents an experience that occurred immediately before B. When B occurs, activating its neurons, residual activity in the neurons representing A permits the two patterns to be associated by the law of habit. That is, the “tracts” connecting the neurons (the “modifiable connections” in Figure 4-1) have their strengths modified.
How does this figure reveal three properties that are common to modern connectionist networks.
- First, the system is parallel: more than one neuron can be operating at the same time.
- Second, the system is convergent: the activity of one of the output neurons depends upon receiving or summing the signals sent by multiple input neurons.
- Third, the system is distributed: the association between A and B is the set of states of the many “tracts” illustrated in Figure 4-1; there is not just a single associative link
Hebb Rule
- Connectionists use the Hebb Rule: “neurons that fire together wire together”
When an axon of cell A is near enough to excite cell B and repeatedly or persistently takes part in firing it, some growth process or metabolic change takes place in one or both cells such that A’s efficiency, as one of the cells firing B, is increased.
Standard Pattern Associator (“Distributed Memory”)
The standard pattern associator, which is structurally identical to Figure 4-1 is a memory capable of learning associations between pairs of input patterns or learning to associate an input pattern with a categorizing response.
- The standard pattern associator is empiricist in the sense that its knowledge is acquired by experience.
- Memory = begins as a blank slate (i.e., all of the connections between processors start with weights equal to zero.)
- During a learning phase: pairs of to-be-associated patterns simultaneously activate the input and output units in Figure 4-1
- With each presented pair, all of the connection weights— the strength of each connection between an input and an output processor—are modified by adding a value to them. Value follows Hebbs Rule.
The standard pattern associator is called a distributed memory because its knowledge is stored throughout all the connections in the network, and because this one set of connections can store several different associations.
Classical Conditioning
Classical conditioning is an example of the associative law of contiguity at work.
- The conditioned stimulus is not associated with the unconditioned response.
- After repeated pairings with an unconditioned stimulus (contiguity), the conditioned stimulus becomes associated with the desired response.
Emergence
where the properties of a whole (i.e., a complex idea) are more than the sum of the properties of the parts (i.e., a set of associated simple ideas).
Emergent properties are often defined as properties that are not found in any component of a system but are still features of the system as a whole.
- emergence results from nonlinearity
How is this system considered a linear system?
If a system is linear, then its whole behaviour is exactly equal to the sum of the behaviours of its parts. Output unit activity is exactly equal to net input.