Exam 3 Flashcards
what is the external part of the ear called?
the pinna
what does the outer ear consist of?
pinna, ear canal, and the eardrum
what does the middle ear consist of?
from ear drum to the oval window: contains three small bones malleus, incus, and stapes
passage through the middle ear does what to the sound?
amplifies it
what does the inner ear consist of?
semicircular canal and cochlea
what happens in the cochlea?
mechanical sound waves are converted to electrical nerve impulses
for an unwound cochlea, there is a thicker and thinner end which is which? (Apical or Basal end)
thicker is the Apical end, thinner is the Basal end
for an unwound cochlea which frequencies do the Apical and Basal ends move more for?
the Apical end moves more for lower frequencies because thicker = lower resonant frequency
the Basal end moves more for high frequencies because thin = higher resonant frequency
the basilar membrane has a thick and thin end which are which (Apical and Basal)
Apical is thick and Basal is thin
the basilar membrane is tonotopically organized - what does that mean?
different locations on the membrane correspond to different frequencies
Denes and Pinson 1993 : 90
- shows how far the basilar membrane is pushed out of place by different frequencies
What was the conclusion found?
lower frequencies (25 Hz) are higher farther (30 mm) from the stapes than higher frequencies (1600 Hz –> 17 mm)
explain how hair cells work - what is their role?
they are attached to the basilar membrane, the hair cell fires if movement of the basilar membrane pushes the cell out of position sufficiently
what is the response curve of a hair cell?
shows the lowest intensity at which a pure tone at a given frequency triggers a firing of the cell - the low point shows the freq. the hair cell responds to most readily - the closer to the apical end (thick) the lower the resonant frequency
Moore 1997 : 33
- shows the response curves for different hair cells
What does the lowest point show?
the lowest point is the characteristic freq. where it will fire at the lowest amplitude
what is the most important factor of a hair cell?
the location of it - they are all the same otherwise
what causes hair loss at certain frequencies?
the hair cells are pushed too far and sheared off
the outer hair cells are different from inner how?
when the outer hair cells fire they change length to push back on the basilar membrane and amplify the signal
Denes and Pinson 1993 : 95
- shows human’s hearing range
What does this show about our speech sounds as humans?
What is the peak sensitivity?
speech sounds evolved to be where our hearing is particularly good - the peak sensitivity is between 1000 and 10,000 Hz
tonotopically organized signals from the ear are passed to the brain through what?
the auditory nerve, through various bodies in the brainstem and to the cerebrum (uppermost and outermost part of the brain)
signals from the right ear are passed to where? what is this called?
the left hemisphere of the brain - decussation
where is the auditory cortex located and what does it border?
in each hemisphere of the temporal lobe of the cerebral cortex on the superior temporal gyrus (STG) - borders the lateral (Sylvian) fissure
what is the primary auditory cortex?
entryway into the cerebral cortex for signals from the ears
how is the primary auditory cortex organized?
tonotopically - different locations correspond to different frequency bands
the frequency-based locations in the primary auditory cortex correspond to what?
frequency-sensitive locations on the basilar membrane
damage to the primary auditory cortex could cause what?
aphasia
Bear et al. 2007
- both hemispheres of the brain have an auditory cortex
But what?
but one is dominant to speech processing - left for 93% of people (96% of right-handed, 70% left handed)
what is dichotic listening?
speech materials are processed in the opposite hemisphere of the ear it receives it from, therefore there is often a right-ear processing advantage for speech but NOT for non-speech sounds like music or humming
what is Wernicke’s Area and where is it?
middle region of the STG that if injured causes problems with perception and comprehension (Wernicke’s Aphasia)
where is Wernicke’s Area in relation to the auditory cortex?
posterior, the auditory cortex is the “bottom” part of the STG
when and who discovered Wernicke’s area?
1874, German neurologist Karl Wernicke - it was early evidence for brain area specialization
True or false:
electrical stimulation of Wernicke’s area interferes with identification of speech sounds, discrimination between speech sounds, and comprehension of speech
true
what are combination-sensitive neurons?
in the STG, respond to particular patterns of frequency and amplitude - they fire only if there is an activation of a particular combination of primary cells
Mesgarani, Cheung, Johnson and Chang (2014) - study of 6 adults whose skulls were opened for epilepsy surgery, electrodes were placed on the surface of the left STG (electrocorticography).
True or false: this study was testing to see which parts of the brain activated when the patient was producing speech and when they were not.
False - the study was to see which parts of the brain were active when speech was playing, but inactive during silence
Mesgarani, Cheung, Johnson and Chang (2014) - study of 6 adults whose skulls were opened for epilepsy surgery, electrodes were placed on the surface of the left STG (electrocorticography).
True or false: the patients passively listened to 500 samples of SAE sentences
True
Mesgarani, Cheung, Johnson and Chang (2014) - study of 6 adults whose skulls were opened for epilepsy surgery, electrodes were placed on the surface of the left STG (electrocorticography).
True or false: researchers found that when passively listening to speech, the STG was activated constantly, but was not in silence
False - different groups of neurons in the STG activated for different classes of sounds
Mesgarani, Cheung, Johnson and Chang (2014) - study of 6 adults whose skulls were opened for epilepsy surgery, electrodes were placed on the surface of the left STG (electrocorticography).
True or false:
e1 responded to the sibilant fricatives /s, ʃ , z/
False - e1 responded to the plosives /b, d, g, p, t, k/
Mesgarani, Cheung, Johnson and Chang (2014) - study of 6 adults whose skulls were opened for epilepsy surgery, electrodes were placed on the surface of the left STG (electrocorticography).
True or false:
e2 responded to the sibilant fricatives /s, ʃ , z/
true
Mesgarani, Cheung, Johnson and Chang (2014) - study of 6 adults whose skulls were opened for epilepsy surgery, electrodes were placed on the surface of the left STG (electrocorticography).
True or false:
e3 responded to the low-back vocoids (vowels and glides) /ɑ, aʊ/
true
Mesgarani, Cheung, Johnson and Chang (2014) - study of 6 adults whose skulls were opened for epilepsy surgery, electrodes were placed on the surface of the left STG (electrocorticography).
True or false:
e4 responded to the plosives /b, d, g, p, t, k/
False - e4 responded to the high-front vocoids /i, j/
Mesgarani, Cheung, Johnson and Chang (2014) - study of 6 adults whose skulls were opened for epilepsy surgery, electrodes were placed on the surface of the left STG (electrocorticography).
True or false:
e5 responded to nasals /m, n, ŋ/
true
Mesgarani, Cheung, Johnson and Chang (2014)
What is PSI?
phoneme selectivity index, represents the number of other phonemes statistically distinguishable from that phoneme in the response of a specific electrode
Mesgarani, Cheung, Johnson and Chang (2014)
what does a PSI = 0 mean?
that electrode does NOT distinguish between that phoneme and any others
Mesgarani, Cheung, Johnson and Chang (2014)
true or false:
PSI = 32 means the electrode can detect 32 phonemes
false - it means the electrode is maximally selective, the phoneme is distinguishable from all other phonemes in the response of that electrode
Mesgarani, Cheung, Johnson and Chang (2014)
true or false:
neurons sensitive to a particular acoustic combination are located near neurons sensitive to similar combinations and are therefore tonotopically organized
false - they are organized by phonetic category
Mesgarani, Cheung, Johnson and Chang (2014)
true or false:
from the basilar membrane to the primary auditory cortex, sound is represented tonotopically in the form of time-varying frequency spectrum, corresponding to a spectrogram
true
what is a category?
a set of entities or events that all elicit an equivalent response
categories are essential to learning and cognition - why?
we can only generalize particular experiences to general knowledge through the use of categories
true or false:
speech categories are the same across people and situations
false - they vary greatly from speaker to speaker and context to context; each person has a broad range of phonetic events they pull from to decode a word or sound
true or false:
an acoustic continuum is a series of items that differ gradiently for a series of acoustic properties
false - only one acoustic property not multiple
true or false:
an F1 continuum would be a series of items that have the same F1 but are different in other aspects
false - the items would differ ONLY in F1
true or false:
in a F1 continuum, the difference in F1 of each of the items in the series is the same as the preceding member
true
Lisker & Abramson 1970
- varied VOT in word-initial stops, using speech synthesis from -150 to 150 ms in 10 ms steps for each place of articulation (bilabial, apical, velar) - subjects who spoke Thai, Spanish, and English were asked to identify the initial consonant of the stimulus among a choice of sounds in their language
true or false:
the space where the lines for either sound meets is called the perceptual/identification boundary
true
Lisker & Abramson 1970
- varied VOT in word-initial stops, using speech synthesis from -150 to 150 ms in 10 ms steps for each place of articulation (bilabial, apical, velar) - subjects who spoke Thai, Spanish, and English were asked to identify the initial consonant of the stimulus among a choice of sounds in their language
true or false:
the study found that at low VOT, English speakers identified the stop as voiceless 100% of the time
false - at low VOT the subjects identified the sounds as VOICED 100% of the time
Lisker & Abramson 1970
- varied VOT in word-initial stops, using speech synthesis from -150 to 150 ms in 10 ms steps for each place of articulation (bilabial, apical, velar) - subjects who spoke Thai, Spanish, and English were asked to identify the initial consonant of the stimulus among a choice of sounds in their language
true or false:
at high VOT the English subjects identified the stop as voiceless 100% of the time
true
Lisker & Abramson 1970
- varied VOT in word-initial stops, using speech synthesis from -150 to 150 ms in 10 ms steps for each place of articulation (bilabial, apical, velar) - subjects who spoke Thai, Spanish, and English were asked to identify the initial consonant of the stimulus among a choice of sounds in their language
true or false:
the perceptual / identification boundary is where subjects were able to tell the stops apart 100% of the time
false - the boundary is where they identified the stimulus as voiced 50% of the time and voiceless 50% of the time
true or false:
Lisker and Abramson (1964) found that further forward places of articulation are associated with greater VOT values
false - places of articulation that are further back are associated with greater VOT values
Lisker & Abramson 1970
- varied VOT in word-initial stops, using speech synthesis from -150 to 150 ms in 10 ms steps for each place of articulation (bilabial, apical, velar) - subjects who spoke Thai, Spanish, and English were asked to identify the initial consonant of the stimulus among a choice of sounds in their language
true or false:
a conclusion drawn from this study is that the identification boundary is at a lower VOT for alveolars and velars than for bilabials
false - because alveolar and velar sounds are further back in the mouth = greater (higher) VOT
what is categorical perception?
listeners ignore the differences of sounds on the same side of the perceptual boundary and only discriminate sounds that lie on opposite sides
Lisker & Abramson 1970
- varied VOT in word-initial stops, using speech synthesis from -150 to 150 ms in 10 ms steps for each place of articulation (bilabial, apical, velar) - subjects who spoke Thai, Spanish, and English were asked to identify the initial consonant of the stimulus among a choice of sounds in their language
true or false:
this study found that speakers differentiate sounds within each side of the perceptual boundary
false - they ignore the differences of those on the same side and only discriminate sounds that lie on opposite sides of the boundary
Liberman et al 1957
- synthesized a series of stop-vowel syllables that were alike in steady-state values of F1 and F2 - they only differed in the onset value of the initial F2 transition from way above F2 steady-state to way below (hand drawn looked like eyebrows) - subjects were asked to identify as b, d, or g
true or false:
when F2 pointed down, subjects identified the consonant as d
false - F2 pointing down was identified as b
Liberman et al 1957
- synthesized a series of stop-vowel syllables that were alike in steady-state values of F1 and F2 - they only differed in the onset value of the initial F2 transition from way above F2 steady-state to way below (hand drawn looked like eyebrows) - subjects were asked to identify as b, d, or g
true or false:
when F2 was flat, subjects identified the consonant as g
false - F2 was flat it was identified as d
Liberman et al 1957
- synthesized a series of stop-vowel syllables that were alike in steady-state values of F1 and F2 - they only differed in the onset value of the initial F2 transition from way above F2 steady-state to way below (hand drawn looked like eyebrows) - subjects were asked to identify as b, d, or g
true or false:
when F2 pointed up, subjects identified the consonant as b
false - F2 pointed up was identified as g
Liberman et al 1957
- synthesized a series of stop-vowel syllables that were alike in steady-state values of F1 and F2 - they only differed in the onset value of the initial F2 transition from way above F2 steady-state to way below (hand drawn looked like eyebrows) - subjects were asked to identify as b, d, or g
there is only one ambiguous stop in between which two stimulus?
3 (almost always b) and #5 (almost always d) - boundary between b and d
Liberman et al 1957
discrimination experiment:
- synthesized a series of stop-vowel syllables that were alike in steady-state values of F1 and F2 - they only differed in the onset value of the initial F2 transition from way above F2 steady-state to way below (hand drawn looked like eyebrows) - subjects listened to a series of 3 syllables (b, d, or g) together (e.g. ABX) where A and B are different and X is either identical to A or B
true or false:
if two of the syllables were within the same category (same side) subjects found it hard to discriminate between them
true
why might humans be more sensitive to acoustic cues that distinguish categories and insensitive to those within the categories?
because the acoustic differences within categories do NOT help with our goal of identifying what sound is being produced
Miyawaki et al. 1975
- synthesized syllables with a sonorant consonant followed by [ɑ], the only difference was the frequency of F3 in the consonant (r, l) - subjects (SAE and Japanese) heard each in random order and asked to determine if they were l or r (law or raw).
true or false:
stimuli with a low F3 in the consonant was identified as “l” nearly 100% of the time
false - low F3 were identified as “r” nearly 100% of the time
Miyawaki et al. 1975
- synthesized syllables with a sonorant consonant followed by [ɑ], the only difference was the frequency of F3 in the consonant (r, l) - subjects (SAE and Japanese) heard each in random order and asked to determine if they were l or r (law or raw).
true or false:
stimuli with a high F3 in the consonant were identified as “l” nearly 100% of the time
true
Miyawaki et al. 1975
- synthesized syllables with a sonorant consonant followed by [ɑ], the only difference was the frequency of F3 in the consonant (r, l) - subjects (SAE and Japanese) heard each in random order and asked to determine if they were l or r (law or raw).
there was one stimulus that could not be clearly assigned by the subjects - what does this mean and which one was it?
7, it was the identification boundary between l and r
Miyawaki et al. 1975
- synthesized syllables with a sonorant consonant followed by [ɑ], the only difference was the frequency of F3 in the consonant (r, l) - subjects (SAE and Japanese) heard each in random order and asked to determine if they were l or r (law or raw).
what were the three main findings of this study?
- SAE speakers did well distinguishing the sounds on opposite sides of the boundary
- SAE speakers were guessing/leaving it to chance when discriminating within the categories
- Japanese speakers, having no contrast between the sounds in Japanese, could not distinguish the sounds
are vowels similar in discrimination to consonants? why or why not?
no they aren’t, there is a perceptual boundary but it is not a peak in discriminability like consonants, it is gradable - people can discriminate within vowel categories as well as between them
what is one hypothesis as to why consonants have a perceptual boundary and vowels don’t?
categorical perception may be limited to rapid, dynamic acoustic properties, like the VOT and F2 formant transitions between consonants and vowels, but vowels have steady-state formant patterns that stay the same for what in speech is a long time
what is speaker normalization?
the listener’s ability to handle/understand the differences among speakers that are unlike what they have heard before
what are the 3 main ways speaker’s voices differ and which one is the MAIN way?
- MOST IMPORTANTLY differ in the formant frequencies
- they differ in f0 (higher or lower pitch) depending on the length of their vocal chords
- voice quality as measured in open quotient or spectral tilt
true or false:
only F1 is higher in women than men
false - F1 and F2 are higher in women than men
men are generally larger than women, and women are larger than children - what does this mean in terms of their voices?
men have longer vocal tracts than women who have longer ones than children therefore men have the lowest resonant frequencies, then women’s, then children - however that does not mean that all large people have the deepest voices
true or false:
the difference between men and women lies mainly in the length of the pharynx
true
true or false:
Peter Ladefoged has formant values for his vowels that are close to those of SAE but are not the same vowel and are therefore easily confused
false - though the formant values are close for one vowel said by him and a different one said by a SAE speaker, they do NOT get confused for each other - distinction is NOT ONLY in formant values (speaker normalization)
what is one of the biggest problems when developing automatic speech recognition software?
computers cannot, as easily or as well as humans, perform speaker normalization when encountering a new voice unlike what they’ve heard before
Ladefoged and Broadbent 1957
- synthesized 4 syllables differing only in F1 and F2, in isolation the syllables were identified as bit, bet, bat, and but - they also synthesized (via F1 and F2) the syllables of a carrier sentence “Please say what this word is” before the next word which subjects had to identify.
true or false:
with the “normal” carrier sentence test word A (F1: 375 Hz) was identified as “bat”
false - with the “normal” carrier sentence test word A (375 Hz) was identified as “bit”
Ladefoged and Broadbent 1957
- synthesized 4 syllables differing only in F1 and F2, in isolation the syllables were identified as bit, bet, bat, and but - they also synthesized (via F1 and F2) the syllables of a carrier sentence “Please say what this word is” before the next word which subjects had to identify.
true or false:
with the “normal” carrier sentence test word B (F1: 450) was identified as “bet”
true
Ladefoged and Broadbent 1957
- synthesized 4 syllables differing only in F1 and F2, in isolation the syllables were identified as bit, bet, bat, and but - they also synthesized (via F1 and F2) the syllables of a carrier sentence “Please say what this word is” before the next word which subjects had to identify.
true or false:
with the “normal” carrier sentence test word C (F1: 575 Hz) was identified as “but”
false - with the “normal” carrier sentence test word C (F1: 575 Hz) was identified as “bat”
Ladefoged and Broadbent 1957
- synthesized 4 syllables differing only in F1 and F2, in isolation the syllables were identified as bit, bet, bat, and but - they also synthesized (via F1 and F2) the syllables of a carrier sentence “Please say what this word is” before the next word which subjects had to identify.
true or false:
with the “normal” carrier sentence test word D (F1: 600 Hz, F2: 1300 Hz) was identified as “bat”
false - with the “normal” carrier sentence test word D (F1: 600 Hz, F2: 1300 Hz) was identified as “but”
Ladefoged and Broadbent 1957
- synthesized 4 syllables differing only in F1 and F2, in isolation the syllables were identified as bit, bet, bat, and but - they also synthesized (via F1 and F2) the syllables of a carrier sentence “Please say what this word is” before the next word which subjects had to identify.
true or false:
when F1 was lowered in the carrier sentence, test word A (375 Hz) started to be identified as “bet”
true - in the low F1 context, the value of 375 Hz counted as high in comparison so the vowel was judged to be low
Ladefoged and Broadbent 1957
- synthesized 4 syllables differing only in F1 and F2, in isolation the syllables were identified as bit, bet, bat, and but - they also synthesized (via F1 and F2) the syllables of a carrier sentence “Please say what this word is” before the next word which subjects had to identify.
true or false:
when F1 was raised in the carrier sentence, test word B (450 Hz) started to be identified as “bat”
false - “bet” began to be identified as “bit” because with the context of high F1 values in the carrier, 450 Hz counted as low in comparison so the vowel was judged to be high
Ladefoged and Broadbent 1957
- synthesized 4 syllables differing only in F1 and F2, in isolation the syllables were identified as bit, bet, bat, and but - they also synthesized (via F1 and F2) the syllables of a carrier sentence “Please say what this word is” before the next word which subjects had to identify.
true or false:
when F1 was raised in the carrier sentence, test word C (575 Hz) started to be identified as “bit”
false - “bat” started to be identified as “bet” because compared to the high F1 in the carrier, 575 Hz was not that high so the vowel was judged to be mid rather than low
Ladefoged and Broadbent 1957
- synthesized 4 syllables differing only in F1 and F2, in isolation the syllables were identified as bit, bet, bat, and but - they also synthesized (via F1 and F2) the syllables of a carrier sentence “Please say what this word is” before the next word which subjects had to identify.
true or false:
when F2 was lowered in the carrier sentence, test word D (F1: 600 Hz, F2: 1300 Hz) started to be identified as “but”
false - “but” started to be identified as “bat” because compared to the low F2 values, 1300 Hz was not all that low, and was judged to be front
Ladefoged and Broadbent 1957
- synthesized 4 syllables differing only in F1 and F2, in isolation the syllables were identified as bit, bet, bat, and but - they also synthesized (via F1 and F2) the syllables of a carrier sentence “Please say what this word is” before the next word which subjects had to identify.
what was the conclusion found by this study?
listeners notice where the formants are in vowels from a new speaker and adapt their model of the vowel space to fit the new voice - their expectations change as they learn where the new speakers vowels are which can happen in a matter of seconds - intelligent problem solving NOT passive
Mullenix et al. 1989
- one group of listeners identified lists of words in noise produced by a single speaker, while another group heard the same words produced by multiple speakers
true or false:
the group hearing a single speaker in noise identified the words slower and less accurately that those hearing multiple speakers
false - the group hearing a single speaker in noise were faster and more accurate because over the short period of time they were able to learn more about the single voice and improve their processing of their speech
Nygaard and Pisoni 1998
- subjects listen to samples from 10 different speakers over 10 days, they learned voices well enough to match a new sample to them, and were presented with a word that needed to identify in noise.
true or false:
subjects made fewer errors identifying words in noise if it was produced by one of the voices they were already familiar with
true
what is priming?
previous exposure to one stimulus (the prime) improves processing performance (accuracy and speed) on the task with a later stimulus (the target)
Nygaard and Pisoni 1998
- subjects listen to samples from 10 different speakers over 10 days, they learned voices well enough to match a new sample to them, and were presented with a word that needed to identify in noise.
how does priming help explain the results this study?
the priming was greater when the prime and the target were produced by the same voice which implies that the voice was part of the memory representation for the prime
Goldinger 1996, 1998
- exposed subjects to words produced by different speakers in a study session, they were tested in various tasks involving those words in a test session (e.g. have you heard this word before in the study session?)
what were the results of this experiment?
the subjects were quicker and more accurate in making this judgement if they heard the word produced by the same speaker who produced it in the study session
Goldinger 1996, 1998
- exposed subjects to words produced by different speakers in a study session, they were tested in various tasks involving those words in a test session - they were asked to identify sounds in the word, discriminate sounds in the word, and repeat the word as quickly as possible (shadowing)
what were the results of these tasks?
they all were done faster and more accurately if they had previously heard the test word produced by the same voice as in the test
Goldinger 1996, 1998
- exposed subjects to words produced by different speakers in a study session, they were tested in various tasks involving those words in a test session - they were asked to identify sounds in the word, discriminate sounds in the word, and repeat the word as quickly as possible (shadowing)
true or false:
this experiment shows that activating both word and voice at the same time is less effective than just activating the word
false - it is more effective to activate both the word and the voice because they both activate memory of that word and that voice
what is an exemplar?
the memory representation of a category like the word “cat” consists of every instance of that word one has ever encountered organized by recency, speaker, context, etc.
true or false:
if you heard a word recently from a speaker, you can more quickly process that word in a new instance from the same speaker
true - speaker normalization is partially responsible for this
what is the perceptual challenge of coarticulation?
there is a different version of every phoneme for every preceding sound and for every following sound - the differences can be as large as those between categories
what is the main problem with speech recognition programs?
it is hard for them to account for coarticulation so a speech sound or spoken word from one context won’t match a sample of the same sound or word from another context
true or false:
to counteract coarticulation problems, listeners remember vowels with the sound that precedes it, not by itself
false - they remember the sound that precedes it and the sound that follows it
Lindblom and Studdert-Kennedy 1967
- series of high vowels synthesized, varying just in the freq. of F2, from clear I to with high F2 to clear ʊ with low F2 - the vowels were spliced into three environments: isolation, w__w, j__j
what would be the expectation for the F2 of the vowel between j__j?
F2 would be high, like it is in j
Lindblom and Studdert-Kennedy 1967
- series of high vowels synthesized, varying just in the freq. of F2, from clear I to with high F2 to clear ʊ with low F2 - the vowels were spliced into three environments: isolation, w__w, j__j
what would be the expectation for the F2 of the vowel between w__w?
F2 of the vowel would be low, like w
Lindblom and Studdert-Kennedy 1967
- series of high vowels synthesized, varying just in the freq. of F2, from clear I to with high F2 to clear ʊ with low F2 - the vowels were spliced into three environments: isolation, w__w, j__j - 10 speakers of SAE listened to the words and identified them as containing I or ʊ
true or false:
when the vowel had a high F2 it was identified as ʊ no matter the context
false - it was identified as I no matter the context
Lindblom and Studdert-Kennedy 1967
- series of high vowels synthesized, varying just in the freq. of F2, from clear I to with high F2 to clear ʊ with low F2 - the vowels were spliced into three environments: isolation, w__w, j__j - 10 speakers of SAE listened to the words and identified them as containing I or ʊ
true or false:
when the vowel had a low F2 it was identified as ʊ no matter the context
true
Lindblom and Studdert-Kennedy 1967
- series of high vowels synthesized, varying just in the freq. of F2, from clear I to with high F2 to clear ʊ with low F2 - the vowels were spliced into three environments: isolation, w__w, j__j - 10 speakers of SAE listened to the words and identified them as containing I or ʊ
true or false:
when the vowel has intermediate F2 values, it was identified as ʊ more often in the low F2 w__w environment than in isolation
false - it was identified as I
Lindblom and Studdert-Kennedy 1967
- series of high vowels synthesized, varying just in the freq. of F2, from clear I to with high F2 to clear ʊ with low F2 - the vowels were spliced into three environments: isolation, w__w, j__j - 10 speakers of SAE listened to the words and identified them as containing I or ʊ
true or false:
when the vowel has intermediate F2 values, it was identified as I less often in the high F2 j__j environment than in isolation
true
Lindblom and Studdert-Kennedy 1967
- series of high vowels synthesized, varying just in the freq. of F2, from clear I to with high F2 to clear ʊ with low F2 - the vowels were spliced into three environments: isolation, w__w, j__j - 10 speakers of SAE listened to the words and identified them as containing I or ʊ
how did the identification boundary shift for I and ʊ?
shifted toward lower values of F2 in the low F2 context and toward higher F2 values in the higher F2 context