Cog models/ neural basis of speech perception Flashcards

1
Q

Motor theory of speech perception
What are the 2 components of the theory?

A
  1. Speech perception is the result of a specialised speech module that operates separately from the mechanisms involved in perceiving non-speech sounds and is uniquely human (“speech is special”)- speech is somehow special

Evidence: Speech (but not other sounds) are perceived categorically

  1. The objects of speech perception are intended articulatory events rather acoustic events

Evidence: Speech sounds are highly variable (articulations less so). Lip movements are more stable than acoustics

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Perceiving intended articulatory events

A

If you hear a sound- comes through ears, relayed to brain. This is acoustic info

Auditory cortex recognising acoustic cues

According to Lieberman because acoustics is so variable, he’s saying that the sound you’re getting, the brain is trying to latch onto this to perceive the underlying gestures to perceive these sounds. Rather than the acoustic consequences of the speech sounds.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What method can we use which might offer support for speech perception?

A

fMRI

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Evidence for the motor theory of speech perception
- What does passive listening to meaningless monosyllables activate?
- what other areas are also activated?

A

Passive listening to meaningless monosyllables activates auditory cortex (region that is normally related to sound processing)

But motor (activate if you make a movement) and premotor (activate if you plan t make a movement) areas are also activated

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is TMS and what does it involve?

A

Transcranial Magnetic Stimulation

Use a coil that makes magnetic pulses over specific brain regions. This will disrupt activity in the underlying brain regions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Evidence for the motor theory of speech perception
Procedure and result

A

Procedure:
Asking pps to do speech discrimination task (hearing pairs of speech sounds and have to discriminate between sounds- are they different or the same). Also have control task where they have to do a visual task (nothing to do with sound or speech)

Results:
TMS over premotor areas interferes with phoneme discrimination in noise but not colour discrimination

Have a selective disruption of speech when TMS is applied over motor areas but not when it’s applied over control.

Evidence that motor areas are causally involved in speech perception

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Evidence against the motor theory of speech perception

A

Categorical perception can also be demonstrated for non-speech sounds
- So not the result of a specialised speech module

With training, chinchillas (rodents) shows the same phoneme boundary for a /da/-/ta/ continuum as humans

Steep change in identification function when you’re near phoneme boundary
- So not unique to humans

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Motor theory of speech perception
Two components of the theory
Are the theories correct?

A

X - This aspect of the motor theory isn’t quite right- its not a result of the mechanism

  1. Speech perception is the result of a specialised speech module that operates separately from the mechanisms involved in perceiving non-speech sounds and is uniquely human (“speech is special”)
    - Evidence: Speech (but not other sounds) are perceived categorically

? - more evidence of this aspect- it does seem that processing gestures is involved in speech perception. It’s about how extreme you take this on.

  1. The objects of speech perception are intended articulatory events rather acoustic events
    - Evidence: Speech sounds are highly variable (articulations less so)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Brain basis of speech perception: Classic model

A
  • Superior temporal gyrus for speech perception (Wernicke’s area- posterior section of superior temporal gyrus)
  • Inferior frontal gyrus for speech production (Broca’s area)
  • Left hemisphere dominant
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What are the two streams for speech processing that are engaged in a task-dependent manner

A

Dorsal and Ventral

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Brain basis of speech perception:
Dorsal stream

A

Dorsal stream for mapping speech sounds onto articulatory representations
- activated for tasks focussing on perception of speech sounds (e.g. phoneme discrimination)
- Left hemisphere dominant
- Broca’s area part of the dorsal stream but also involved in perception (not just production)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Brain basis of speech perception:
Ventral stream

A

Ventral stream for mapping speech sounds onto lexical representations
- Activated for tasks focussing on comprehension (e.g. word recognition)
- Bilateral (left and right hemispheres)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Brain basis of speech perception:
What does it explain?

A

Explains why some aphasics can’t tell apart phonemes but can recognise words (and vice versa)

This model can explain counterintuitive finding- damage to dorsal stream but ventral stream is intact. For other patients, they will have ventral stream damages.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Evidence for ventral stream processing

A
  1. Williams, Nestor and Hodges
    Example of semantic task: Pyramids and Palm Trees test. Present picture of pyramid. Choose between picture of palm tree or foil (other type of tree).

Anterior temporal damage associated with semantic impairment (have trouble choosing correct picture)

  1. Bates et al.
    Inferior temporal damage associated with comprehension deficits
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Evidence for dorsal stream processing

A
  1. Wilson et al.
    Listening to syllables activates motor and premotor areas
  2. Meister et al.
    TMS ver premotor areas interferes with phoneme discrimination in noise but not colour discrimination
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Cohort model:
Explain the process of recognising spoken words

A

You have a set of word representations in your mind

You have a memory trace if what different words should sound like in your head

As you’re hearing speech (not even the whole word) the speech input you’re getting activates words signal in your lexicon that you’ve heard so far.

If it starts with c- it will activate all words starting with this

As time goes on, it will result in a more selective activation

cohert- set of words that respond to speech signal

With cath- many words before can be ruled out

This will play out until moments of time in uniqueness point

17
Q

Cohort model:
Uniqueness point (UP) and key features

A

Uniqueness point: Time-point in the speech signal when only one word becomes consistent with the speech input

Word is recognised at the UP, even before the whole word has been heard (‘optimal efficiency’)- instead of waiting for the whole word to be heard, the mind will only use what it needs to recognise the word. It will recognise cathedral at this point in time even before hearing the whole word.

Key features:
- Words are activated immediately upon minimal input
- Multiple words are activated
- Words compete for recognition (‘lexical competition’)

18
Q

Cohort model: Evidence from shadowing task

A

Listeners presented with speech and they have to shadow what they’re hearing as soon as they can

Histogram showing response latencies

  • Average response latency is approx. 250 ms
  • Average duration of words is 375 ms
  • So implies that listeners recognise words even before they have heard the ends of the words
  • Consistent with Cohort model

Average duration of words 375 ms. Also point out that it would take around 50-100 ms to make response.

19
Q

What does learning new words slow down?

A

Recognition of existing words

20
Q

Cohort model: Shortcoming

A
  • The Cohort model is still very influential in the field
  • However, it is an example of a ‘verbal model’, which makes it difficult to evaluate
  • A better way of being sure what a model or theory predicts, is to implement it as a computer program (‘computational model’)

Average duration of words 375 ms. Also point out that it would take around 50-100 ms to make response.

21
Q

TRACE model of speech perception

Within-layer inhibitory connections vs Bi-directional excitatory connections

A

Within-layer inhibitory connections- influencing lexical connections- if you have cat activated, this will inhibit other words like bad and bat.

Bi-directional excitatory connections- link between different levels of processing - link between the word stage and phoneme stage

22
Q

TRACE model of speech perception

A

Different brain areas and stages of processing

Can peak in each of these stages- have neuron like inputs (one responding to b, k ect and a word part of the model)

If you have activation of acoustic features that are consistent with book for example this activation will travel up and excite the neuron thats consistent with codes for b and this will activate words that start with b

In TRACE, also have activations going the other way- if you have activation of a particular word such as bad, for these top down connections, this will enhance activity for phonemes that comprise bad such as ‘b’ and ‘a’

23
Q

TRACE model: Evidence from eyetracking

A

Camera monitors where pp’s are looking on the screen

They hear a spoken instruction eg. take the beaker and put it on the black circle

This method of eye tracking is a way to see the time course of speech processing

By using eye tracking, you can look at the more intermediate fine grained processes that lead to the final behavioural outcome

24
Q

Eye tracking data and TRACE model results

A

Initially they’re not looking at any object but as more of the spoken word is revealed, they start to look at the beaker.

Looking at words that start with the same phoneme or that rhythme/ are unrelated

TRACE- Present same stimuli as human pp’s to model and look at activation in word level and compare data to human pp’s data.

The activations and trace closely respond with eye tracking data suggesting that TRACE is doing well at modelling the dynamics of spoken word recognition.

25
Q

TRACE model: Explaining context effects

A

Lexical context influences how you perceive ambiguous speech sounds.

26
Q

TRACE’s explanation of Ganong effect

A
  • Word-level activity ‘feeds back’ to influence lower-level phoneme activity
  • although this explanation has been strongly challenged

For phonemes and because they’re ambiguous they might start off equally activated

  • ift through top down connection will bias activity fot that g phoneme unit
  • iss will bias activity in the k unit

This is what affects perception of speech contexts, they are influenced by higher level word context