Word Recognition Flashcards
COHORT’s assumptions (per TRACE)
- Uses first sound or first CV to determine words in initial candidate set (“cohort”).
- Eliminates words from cohort as successive phonemes arrive.
- Phoneme-to-word inhibition
- Words can also be eliminated based on semantic content, but initial cohort is determined by acoustics.
- Word recognition occurs when there is a single item ileft n candidate set.
- Word recognition can influence phoneme identification after the word has been recognized.
COHORT person
Marslen-Wilson, circa 1980
Good evidence for COHORT
In gating, the uniqueness point usually matches the participant’s acceptance/certainty point.
Confirms COHORT’s prediction that words are recognized when just one item remains in cohort.
COHORT’s problems (per TRACE)
- Cannot cope with distortion or underspecified onsets
- No way to recover items that were removed from cohort
- Want to reject “present” for pleasant but accept “blacelet” for bracelet.
- Assumes listeners always knows where a word’s onset is.
COHORT’s three stages of word recognition
- Access (when the bottom-up perceptual input first activates lexical representations)
- Selection (narrowing down the activation candidate set)
- Integration (retrieve and integrate semantic/syntactic details)
COHORT
- accesses all items consistent with input and
- evaluates the multiple words in parallel,
- using information as it becomes available.
COHORT’s segmentation strategy
Segmentation is implicit.
Utterance onset marks onset of first word.
Offset of each word marks onset of next word.
Controversial features of TRACE
Spatializing time, so lots of units are duplicated.
Assuming interactive effects between layers.
Activation in TRACE
It’s continuous, based on how acoustic/phonetic features map onto lexical representations.
Supports partial activation rhymes because they are bottom-up matches.
Interactivity in TRACE
- Top-down connections from lexical items to phonemes.
- The top-down connections from phonemes to feature detectors are usually disabled.
Competition in TRACE
Temporally overlapping units in phonemic and lexical layers inhibit one one another.
How TRACE models time
Units in phoneme and lexical layers are repeated every few time slices.
It spatializes time.
Coarticulation in TRACE
Input phoneme’s features are spread over 11 steps, but the centers of adjacent input phonemes are 6 steps apart.
TRACE’S acoustic features
- Acute
- Burst
- Consonantal
- Diffuse
- Power
- Vocalic
- Voiced
Each with nine levels of activation, each with a feature detector at every timestep.
So there would be Voice0, Voice1, …, Voice8 feature detector units at each step.
Shortlist’s main idea
- At each phoneme time-step, a shortlist of matching words is generated
- The words in the shortlists that overlap each other compete with each other via lateral inhibition
- Separates lexical access (shortlist formation based on match scores) from competition (overlapping words across the lists have to compete with each other)
Shortlist’s scoring system
- 1 point for each matching phoneme
- -3 points for each mismatching phoneme
- Strong mismatch penalty will keep mostly onset-matching items in the shortlist.
- Rhymes will only appear in the list
- when shortlist is sparse and
- when there have been multiple matching phonemes to overcome initial mismatch.
- There is a shortlist at each phoneme timestep, consisting of words with top match scores.
Simple recurrent networks
- Learning network
- Layers for input units, hidden units, context units, output units
- Context units are exact copy of last time steps hidden units
- Hidden units combine information from input and previous state.
- Interactive in the sense that the context interacts with the input units.
- Recurrence is self feedback
Overtraining in SRNs
- If you train an SRN until error asymptotes, it will not show rhyme effects
- If you train until each target reaches from recognition threshold, rhyme effects will remain intact.
- Adults learning novel neighborhoods look like these SRNs (Magnuson et al. 2003)
Localist representations
One unit for each word.
Competing units compete as their activation changes over time.
Distributed representations
- All items are represented by a shared set of units.
- Competition shows up in the blend of hidden representations.
- Makes predictions priming effects of ambiguous inputs activates a blend of competitors.
Distributed Cohort Model
SRN but two output layers: phonological and lexical semantics
Distributed representation of phonological and semantic features in the hidden units
Cross-Modal Semantic Priming
- Castle-candy-sweet priming
- Castle activates its cohort (including candy) which in turn activate semantic cohorts (including sweet).
- Works for cohorts but not for rhymes
- Cohorts don’t get much inhibition during start of word
- When rhymes finally have perceptual support, the onset-cohort are strongly inhibiting them
Naming / Repetition / Shadowing
- Play a wordlike stimulus. Listener repeats word.
- Measure accuracy and response latency.
- Word frequency and neighborhood density influence outcome measures.
- Unlike real recognition
- bc it’s post-perceptual
- bc listener may pay close attention to the speech sounds (without semantic processing).
Lexical Decision
- Play a wordlike stimulus. Listener decides whether stimulus is a word or not.
- Measure accuracy and response latency.
- Sensitive to
- word frequency (more false rejects of less common words)
- neighborhood density (probably, slower to make decision when stimulus overlaps with a lot of words).
- Unlike real recognition bc
- it’s a post-perceptual judgment about stimulus.
- Also a listener can make decision w/o knowing meaning of word (I can recognize tamp but I don’t know what it means).
Gating
- Early gates have a wide variety of guess (many words activated)
- Uniquess point: When there is only one completion of a word
- Recognition point: When correct guessing occurs
- Recognition can precede uniqueness, especially for higher frequeny words