Very helpful PPT: Models of word recognition

PSY 369: Psycholinguistics
Language Comprehension
Word recognition & speech recognition
Lexical access


How do we retrieve the linguistic
information from Long-term memory?





How is the information organized/stored?
What factors are involved in retrieving
information from the lexicon?
Models of lexical access
Lexical access


How do we retrieve the linguistic
information from Long-term memory?





How is the information organized/stored?
What factors are involved in retrieving
information from the lexicon?
Models of lexical access
Models of lexical access


Serial comparison models




Parallel comparison models





Search model (Forster, 1976, 1979, 1987, 1989)
Logogen model (Morton, 1969)
Cohort model (Marslen-Wilson, 1987, 1990)

Connectionist models


Interactive Activation Model (McClelland and
Rumelhart, 1981)
Search model (e.g., Forster, 1976)


Access of the lexicon is considered autonomous, independent of
other systems involved in processing language




A complete perceptual representation of the perceived stimulus is
constructed
The representation is compared with representations in access files


Three access files:






Orthographic
Phonological
Syntactic/semantic (for language production)

Access files are organized in a series of bins (first syllable
or letters)



Position within the bins is organized by lexical frequency
Access files have “pointers” to meaning information in semantic
memory
Search model (e.g., Forster, 1976)
Visual input
Pointers
Decreasing frequency

Entries in order of

Access
codes

Auditory input

/kat/

cat

Mental lexicon mat

cat

mouse
Search model (e.g., Forster, 1976)
Search model (Forster, 1976, 1979, 1987, 1989)







Frequency effects
 Bin organization
Access
codes
Repetition priming effects
 Temporary reordering of bins in
response to recent encounter
Semantic priming effects
 Accounted for by cross
referencing in the lexicon
Context effects
Mental lexicon
 Search is considered to be
autonomous, un affected by
context (so context effects are
“post-access”)
Decreasing frequency



Entries in order of



Visual input

Auditory input
Pointer
s

/kat/

cat

mat

cat

mouse
Logogen model (Morton 1969)


The lexical entry for each word comes with a logogen


Logogens specify word‟s attributes




e.g., semantic, orthographic, phonological

Activated in two ways





cat

By sensory input
By contextual information

Access (recognition) when reach threshold


Different thresholds depending on different factors




e.g., frequency

Access makes information associated with word
available

dog

cap
Logogen model (Morton 1969)
Auditory
stimuli
Auditory
analysis
Context
system

Visual
stimuli
Visual
analysis

Semantic
Attributes

Logogen
system
Available Responses

Output
buffer
Responses

cat

dog

cap
Think of a logogen as being like a
„strength-o-meter‟ at a fairground
When the bell rings, the
logogen has „fired‟
‘cat’
[kæt]

• What makes the logogen fire?
– seeing/hearing the word
• What happens once the logogen has fired?
– access to lexical entry!
‘cat’
[kæt]

• So how does this help
us to explain the
frequency effect?
– High frequency
words have a lower
threshold for firing

–e.g., cat vs. cot

‘cot’
[kot]

Low
freq
takes
longer
‘doctor’
[doktə]

• Spreading activation
from doctor lowers the
threshold for nurse to
fire
– So nurse take less
time to fire

doctor

‘nurse’
[nə:s]

Spreading
activation
network
doctor nurse

nurse
Interactive Activation Model (IAM)
Proposed to account for
Word Superiority effect

McClelland and Rumelhart, (1981)
The Word-Superiority Effect (Reicher, 1969)

+

Until the participant hits some start key
The Word-Superiority Effect (Reicher, 1969)

COURSE

Presented briefly … say 25 ms
The Word-Superiority Effect (Reicher, 1969)
U
&&&&&
A
Mask presented with alternatives above and below
the target letter … participants must pick one as the
letter they believe was presented in that position.
The Word-Superiority Effect (Reicher, 1969)
+

+
E

KLANE
E
&
T

Letter only
Say 60%

+
PLANE

E
&&&&&
T

E
&&&&&
T

Letter in Nonword
Say 65%

Letter in Word
Say 80%

Why is identification better when a letter is presented in a word?
Also goes by the name: Interactive Activation and Competition Model (IAC)

Interactive Activation Model (IAM)
Previous models posed a
bottom-up flow of
information (from features
to letters to words).
IAM also poses a topdown flow of information
Nodes:
• (visual) feature
McClelland and Rumelhart, (1981)
• (positional) letter
• word detectors
• Inhibitory and excitatory connections between them.
Interactive Activation Model (IAM)


Inhibitory connections within
levels




If the first letter of a word is “a”, it
isn‟t “b” or “c” or …

Inhibitory and excitatory
connections between levels
(bottom-up and top-down)




If the first letter is “a” the word
could be “apple” or “ant” or
…., but not “book” or “church”
or……
If there is growing evidence that
the word is “apple” that evidence
confirms that the first letter is
“a”, and not “b”…..

McClelland and Rumelhart, (1981)
IAM & the word superiority effect


The model processes at
the word and letter levels
simultaneously




Letters in words benefit from
bottom-up and top-down
activation
But letters alone receive only
bottom-up activation.
McClelland and Rumelhart, (1981)
Cohort model (Marslen-Wilson & Welch, 1978)


Specifically for auditory word recognition
(covered in chapter 9 of textbook)


Speakers can recognize a word very rapidly




Recognition point (uniqueness point)




Usually within 200-250 msec
Point at which a word is unambiguously different from
other words and can be recognized (strong emphasis on
word onsets)

Three stages of word recognition
1) activate a set of possible candidates
2) narrow the search to one candidate
3) integrate single candidate into semantic and
syntactic context
Cohort model


Prior context: “I took the car for a …”

/s/

/sp/

…
soap
spinach
psychologist
spin
spit
sun
spank
…

spinach
spin
spit
spank
…

time

/spi/

spinach
spin
spit
…

/spin/

spin
Comparing the models


Each model can account for major findings
(e.g., frequency, semantic priming, context), but they
do so in different ways.







Search model is serial, bottom-up, and autonomous
Logogen is parallel and interactive (information flows up and
down)
AIM is both bottom-up and top-down, uses facilitation and
inhibition
Cohort is bottom-up but parallel initially, but then interactive
at a later stage
Different signals
Visual word recognition

Speech Perception

Where are you going




Some parallel input
Orthography








Letters

Clear delineation
Difficult to learn

Serial input
Phonetics/Phonology





Acoustic features

Usually no delineation
“Easy” to learn
Different signals
Visual word recognition

Speech Perception

Where are you going




Some parallel input
Orthography








Letters

Clear delineation
Difficult to learn

Serial input
Phonetics/Phonology





Acoustic features

Usually no delineation
“Easy” to learn
Speech perception


Articulatory phonetics


Production based




Place and manner of articulation

Acoustic phonetics


Based on the acoustic signal


Formants, transitions, co-articulation, etc.
Speech production to perception



Acoustic cues are extracted and stored in sensory
memory and then mapped onto linguistic information


Air is pushed into the larynx across the vocal cords and into the
mouth nose, different types of sounds are produced.




The different qualities of the sounds are represented in formants

The formants and other features are mapped onto phonemes
Acoustic features


Spectrogram




Frequency



Time on the x-axis
Frequency (pressure
under which the air is
pushed) on the y-axis
Amplitude is
represented by the
darkness of the lines

time
Acoustic features


Acoustic features


Formants - bands of resonant frequencies






Formant transitions - up or down movement of formants
Steady states - flat formant patterns

Bursts - sudden release of air
Voice onset time (VOT) - when the voicing begins
relative to the onset of the phoneme
Formants in a wide-band spectrogram

<-- Formant transitions -->

<-- F 3
<-- F 2

<-- F1


Formants - bands of resonant frequencies



Formant transitions - up or down movement of formants
Steady states - flat formant patterns
Formants in a wide-band spectrogram

Burst -->
<-- Formant transitions -->

<-- F1


Bursts – sudden release of air

<-- F 3
<-- F 2
Voice-Onset Time (VOT)

bit
5 ms

pit
40 ms
Categorical Perception


Categorical Perception is the perception of different
sensory phenomena as being qualitatively, or
categorically, different.



Liberman et al (1957)


Used the speech synthesizer to create a series of syllables panning
categories /b/, /d/, /g/ (followed by /a/)





Was done by manipulating the F2 formant

Stimuli formed a physical continuum
Result, people didn‟t “hear” a continuum, instead classified them into
three categories
Categorical Perception


Liberman et al (1957)

1. Set up a continuum of sounds between two categories
/ba/

1

-

...

3

…

/da/

5

…

7
Categorical Perception


Liberman et al (1957)

2. Run an identification experiment
Sharp phoneme boundary

100



% /ba/

Our perception of
phonemes is categorical
rather than continuous.

0

1

...

3

…

5 …

7
Hard Problems in Speech Perception
Wave form



Linearity (parallel transmission): Acoustic features
often spread themselves out over other sounds
 Where does show start and money end?



Demo's and info
Hard Problems in Speech Perception
Wave form



Invariance:
 One phoneme should have a one waveform




But, the /i/ („ee‟) in „money‟ and „me‟ are different

There aren‟t invariant cues for phonetic segments


Although the search continues


Demo's and info
Hard Problems in Speech Perception
Wave form



Co-articulation: the influence of the articulation
(pronunciation) of one phoneme on that of another
phoneme.


Essentially, producing more than one speech sound at once



Demo's and info
Hard Problems in Speech Perception


Trading relations
 Most phonetic distinctions have more than one
acoustic cue as a result of the particular
articulatory gesture that gives the distinction.




slit–split – the /p/ relies on silence and rising
formant, different mixtures of these can result in the
same perception

Perception must establish some "trade-off"
between the different cues.
Hard Problems in Speech Perception


The McGurk effect: McGurk and MacDonald (1976)
• Showed people a video where the audio and the
video don‟t match
• Think “dubbed movie”




McGurk effect
McGurk effect2

• visual /ga/ with auditory /ba/ often hear /da/


Implications
• phoneme perception is an active process
• influenced by both audio and visual information
Motor theory of speech perception


A. Liberman (and others, initially proposed in late 50‟s)




Direct translation of acoustic speech into articulatorally
defined categories
Holds that speech perception and motor control involved
linked (or the same) neural processes


Theory held that categorical perception was a direct reflection of
articulatory organization






Categories with discrete gestures (e.g., consonants) will be perceived
categorically
Categories with continuous gestures (e.g., vowels) will be perceived
continuously

There is a speech perception module that operates
independently of general auditory perception
Frontal slices showing differential activation elicited during lip and tongue movements
(Left), syllable articulation including [p] and [t] (Center), and listening to syllables including
[p] and [t] (Right)

Pulvermüller F et al. PNAS 2006;103:7865-7870

©2006 by National Academy of Sciences
Motor theory of speech perception


Some problems for MT




Categorical perception found in non-speech sounds
(e.g., music)
Categorical perception for speech sounds in non-humans


Chinchillas can be trained to show categorical perception of /t/ and /d/
consonant-vowel syllables (Kuhl & Miller, 1975)
Other theories of speech perception


Direct Realist Theory (C. Fowler and others)






Similar to Motor theory, articulation representations are
key, but here they are directly perceived
Perceiving speech is part of a more general perception
of gestures that involves the motor system

General Auditory Approach (e.g., Diehl, Massaro)


Do not invoke special mechanisms for speech
perception, instead rely on more general mechanisms of
audition and perception



For nice reviews see:

Diehl, Lotto, & Holt (2003)

Galantucci, Fowler, Turvey (2006)
Top-down effects on Speech Perception



Phoneme restoration effect
Sentence context effects
Phoneme restoration effect
Listen to a sentence which contained a word from
which a phoneme was deleted and replaced with
another noise (e.g., a cough)
The state governors met with their respective legi*latures
convening in the capital city.
* /s/ deleted and replaced with a cough

Click here for a demo and additional information
Phoneme restoration effect
Typical results:
Participants heard the word normally, despite the
missing phoneme

Usually failed to identify which phoneme was missing

Interpretation
We can use top-down knowledge to “fill in” the
missing information
Phoneme restoration effect
Further experiments (Warren and Warren, 1970):
What if the missing phoneme was ambiguous
The *eel was on the axle.
The *eel was on the shoe.
The *eel was on the orange.
The *eel was on the table.

Results:
Participants heard the contextually appropriate word
normally, despite the missing phoneme
Phoneme restoration effect


Possible loci of phoneme restoration effects


Perceptual loci of effect:




Lexical or sentential context influences the way in
which the word is initially perceived.

Post-perceptual loci of effect:


Lexical or sentential context influences decisions
about the nature of the missing phoneme information.
Beyond the segment
Shillcock (1990): hear a sentence, make a lexical decision to
a word that pops up on computer screen (cross-modal
priming)
Hear:
The scientist made a new discovery last year.

NUDIST
Cross-modal priming
Shillcock (1990): hear a sentence, make a lexical decision to
a word that pops up on computer screen (cross-modal
priming)
Hear:

The scientist made a novel discovery last year.

NUDIST
Cross-modal priming
Shillcock (1990): hear a sentence, make a lexical decision to
a word that pops up on computer screen (cross-modal
priming)
Hear:

The scientist made a novel discovery last year.
The scientist made a new discovery last year. faster
Cross-modal priming
Shillcock (1990): hear a sentence, make a lexical decision to
a word that pops up on computer screen (cross-modal
priming)
Hear:

The scientist made a novel discovery last year.
The scientist made a new discovery last year. faster

NUDIST gets primed by segmentation error
Although no conscious report of hearing “nudist”
Beyond the segment


Prosody and intonation


English:
 Speech is divided into phrases.
 Word stress is meaningful in English.
 Stressed syllables are aligned in a fairly regular
rhythm, while unstressed syllables take very little time.
 Every phrase has a focus.
 An extended flat or low-rising intonation at the end of a
phrase can indicate that a speaker intends to continue
to speak.
 A falling intonation sounds more final.
Beyond the segment


Prosodic factors (supra segmentals)


Stress




Rate




Emphasis on syllables in sentences
Speed of articulation

Intonation


Use of pitch to signify different meanings across
sentences
Beyond the segment


Stress effects


On meaning




“black bird” versus “blackbird”

Top-down effects on perception


Better anticipation of upcoming segments when syllable
is stressed
Beyond the segment


Rate effects


How fast you speak has an impact on the speech
sounds




Faster talking - shorter vowels, shorter VOT

Normalization


Taking speed and speaker information into account
 Rate normalization
 Speaker normalization
1 von 58

Recomendados

Theories of Speech Perception von
Theories of Speech PerceptionTheories of Speech Perception
Theories of Speech PerceptionAsma Agha Mashkoor
18.4K views36 Folien
Speech production von
Speech productionSpeech production
Speech productiongcuf
11.3K views33 Folien
Neurolinguistics: Brain and Language von
Neurolinguistics: Brain and Language Neurolinguistics: Brain and Language
Neurolinguistics: Brain and Language Mohammed Mallah
10.9K views34 Folien
Key issues in 2nd language acquisition von
Key issues in 2nd language acquisitionKey issues in 2nd language acquisition
Key issues in 2nd language acquisitionSamir1370
7.4K views15 Folien
Language Acquisition: Lecture 2 Phonological Development von
Language Acquisition: Lecture 2 Phonological DevelopmentLanguage Acquisition: Lecture 2 Phonological Development
Language Acquisition: Lecture 2 Phonological Developmentsuascolleges
7.2K views16 Folien
Behaviorism ,Introduction to language Learning Theories & Behaviorist Theory von
Behaviorism ,Introduction to language Learning Theories & Behaviorist TheoryBehaviorism ,Introduction to language Learning Theories & Behaviorist Theory
Behaviorism ,Introduction to language Learning Theories & Behaviorist TheoryNaqvisailya
4.5K views41 Folien

Más contenido relacionado

Was ist angesagt?

Language and Human's Brain von
Language and Human's  Brain Language and Human's  Brain
Language and Human's Brain Irma Fitriani
19.1K views37 Folien
Discourse analysis new von
Discourse analysis newDiscourse analysis new
Discourse analysis newHarry Subagyo
21.7K views19 Folien
Language Acquisition von
Language AcquisitionLanguage Acquisition
Language AcquisitionMilton Gomez
38.3K views13 Folien
Language and the brain von
Language and the brainLanguage and the brain
Language and the brainyaseen zebary
4.8K views17 Folien
Introduction to psycholinguistics von
Introduction to psycholinguisticsIntroduction to psycholinguistics
Introduction to psycholinguisticsLusya Liann
20.6K views18 Folien
Visual Word Recognition. The Journey from Features to Meaning von
Visual Word Recognition. The Journey from Features to MeaningVisual Word Recognition. The Journey from Features to Meaning
Visual Word Recognition. The Journey from Features to Meaningfawzia
3.5K views24 Folien

Was ist angesagt?(20)

Language and Human's Brain von Irma Fitriani
Language and Human's  Brain Language and Human's  Brain
Language and Human's Brain
Irma Fitriani19.1K views
Discourse analysis new von Harry Subagyo
Discourse analysis newDiscourse analysis new
Discourse analysis new
Harry Subagyo21.7K views
Language Acquisition von Milton Gomez
Language AcquisitionLanguage Acquisition
Language Acquisition
Milton Gomez38.3K views
Introduction to psycholinguistics von Lusya Liann
Introduction to psycholinguisticsIntroduction to psycholinguistics
Introduction to psycholinguistics
Lusya Liann20.6K views
Visual Word Recognition. The Journey from Features to Meaning von fawzia
Visual Word Recognition. The Journey from Features to MeaningVisual Word Recognition. The Journey from Features to Meaning
Visual Word Recognition. The Journey from Features to Meaning
fawzia3.5K views
Introduction to psycholinguistics von IvanClans
Introduction to psycholinguisticsIntroduction to psycholinguistics
Introduction to psycholinguistics
IvanClans494 views
Production and Comprehension Process of Language von Riska Daenangsari
Production and Comprehension Process of LanguageProduction and Comprehension Process of Language
Production and Comprehension Process of Language
Riska Daenangsari4.9K views
Language Processing in Brain von Hafsa Awan
Language Processing in Brain Language Processing in Brain
Language Processing in Brain
Hafsa Awan2.5K views
Assimilation and Dissimilation von nirmeennimmu
Assimilation and DissimilationAssimilation and Dissimilation
Assimilation and Dissimilation
nirmeennimmu1.3K views
Multilingualism and bilingualism von Farooq Niazi
Multilingualism and bilingualismMultilingualism and bilingualism
Multilingualism and bilingualism
Farooq Niazi15.2K views
The Linguistics of Second Language Acquisition von kashmasardar
The Linguistics of Second Language AcquisitionThe Linguistics of Second Language Acquisition
The Linguistics of Second Language Acquisition
kashmasardar711 views
Critical discourse analysis of the ideology of media presented through news von Mazhar Ranjha
Critical discourse analysis of the ideology of media presented through newsCritical discourse analysis of the ideology of media presented through news
Critical discourse analysis of the ideology of media presented through news
Mazhar Ranjha16.3K views
LANGUAGE PRODUCTION IN PSYCOLINGUISTIC von Anisa Asharie
LANGUAGE PRODUCTION IN PSYCOLINGUISTICLANGUAGE PRODUCTION IN PSYCOLINGUISTIC
LANGUAGE PRODUCTION IN PSYCOLINGUISTIC
Anisa Asharie9.8K views

Similar a Very helpful PPT: Models of word recognition

Comprehension and Language (handout version) von
Comprehension and Language (handout version)Comprehension and Language (handout version)
Comprehension and Language (handout version)Al Alva
322 views17 Folien
An Introduction To Speech Sciences (Acoustic Analysis Of Speech) von
An Introduction To Speech Sciences (Acoustic Analysis Of Speech)An Introduction To Speech Sciences (Acoustic Analysis Of Speech)
An Introduction To Speech Sciences (Acoustic Analysis Of Speech)Jeff Nelson
5 views10 Folien
L2 Thinking von
L2 ThinkingL2 Thinking
L2 ThinkingMelissa Foster
816 views48 Folien
Theories of speech perception.pptx von
Theories of speech perception.pptxTheories of speech perception.pptx
Theories of speech perception.pptxsherin444916
84 views41 Folien
Lecture phonetics von
Lecture phoneticsLecture phonetics
Lecture phoneticsNara Erjanovna
8K views67 Folien
speech recognition and removal of disfluencies von
speech recognition and removal of disfluenciesspeech recognition and removal of disfluencies
speech recognition and removal of disfluenciesAnkit Sharma
641 views31 Folien

Similar a Very helpful PPT: Models of word recognition(20)

Comprehension and Language (handout version) von Al Alva
Comprehension and Language (handout version)Comprehension and Language (handout version)
Comprehension and Language (handout version)
Al Alva322 views
An Introduction To Speech Sciences (Acoustic Analysis Of Speech) von Jeff Nelson
An Introduction To Speech Sciences (Acoustic Analysis Of Speech)An Introduction To Speech Sciences (Acoustic Analysis Of Speech)
An Introduction To Speech Sciences (Acoustic Analysis Of Speech)
Jeff Nelson5 views
Theories of speech perception.pptx von sherin444916
Theories of speech perception.pptxTheories of speech perception.pptx
Theories of speech perception.pptx
sherin44491684 views
speech recognition and removal of disfluencies von Ankit Sharma
speech recognition and removal of disfluenciesspeech recognition and removal of disfluencies
speech recognition and removal of disfluencies
Ankit Sharma641 views
How do we generate spoken words This issue is a fasci-natin.docx von wellesleyterresa
How do we generate spoken words This issue is a fasci-natin.docxHow do we generate spoken words This issue is a fasci-natin.docx
How do we generate spoken words This issue is a fasci-natin.docx
Mindmap2 von anbray723
Mindmap2Mindmap2
Mindmap2
anbray7231.4K views
Chapters 1-3 CommunicationsIntroductionCommuni.docx von spoonerneddy
Chapters 1-3  CommunicationsIntroductionCommuni.docxChapters 1-3  CommunicationsIntroductionCommuni.docx
Chapters 1-3 CommunicationsIntroductionCommuni.docx
spoonerneddy14 views
speech production in psycholinguistics von Aseel K. Mahmood
speech production in psycholinguistics speech production in psycholinguistics
speech production in psycholinguistics
Aseel K. Mahmood33K views
Automatic speech recognition von Manthan Gandhi
Automatic speech recognitionAutomatic speech recognition
Automatic speech recognition
Manthan Gandhi1.1K views

Último

Papal.pdf von
Papal.pdfPapal.pdf
Papal.pdfMariaKenney3
76 views24 Folien
The Picture Of A Photograph von
The Picture Of A PhotographThe Picture Of A Photograph
The Picture Of A PhotographEvelyn Donaldson
38 views81 Folien
Presentation_NC_Future now 2006.pdf von
Presentation_NC_Future now 2006.pdfPresentation_NC_Future now 2006.pdf
Presentation_NC_Future now 2006.pdfLora
38 views74 Folien
JRN 362 - Lecture Twenty-Two von
JRN 362 - Lecture Twenty-TwoJRN 362 - Lecture Twenty-Two
JRN 362 - Lecture Twenty-TwoRich Hanley
39 views157 Folien
ANGULARJS.pdf von
ANGULARJS.pdfANGULARJS.pdf
ANGULARJS.pdfArthyR3
54 views10 Folien

Último(20)

Presentation_NC_Future now 2006.pdf von Lora
Presentation_NC_Future now 2006.pdfPresentation_NC_Future now 2006.pdf
Presentation_NC_Future now 2006.pdf
Lora 38 views
JRN 362 - Lecture Twenty-Two von Rich Hanley
JRN 362 - Lecture Twenty-TwoJRN 362 - Lecture Twenty-Two
JRN 362 - Lecture Twenty-Two
Rich Hanley39 views
ANGULARJS.pdf von ArthyR3
ANGULARJS.pdfANGULARJS.pdf
ANGULARJS.pdf
ArthyR354 views
Guidelines & Identification of Early Sepsis DR. NN CHAVAN 02122023.pptx von Niranjan Chavan
Guidelines & Identification of Early Sepsis DR. NN CHAVAN 02122023.pptxGuidelines & Identification of Early Sepsis DR. NN CHAVAN 02122023.pptx
Guidelines & Identification of Early Sepsis DR. NN CHAVAN 02122023.pptx
Niranjan Chavan43 views
Ask The Expert! Nonprofit Website Tools, Tips, and Technology.pdf von TechSoup
 Ask The Expert! Nonprofit Website Tools, Tips, and Technology.pdf Ask The Expert! Nonprofit Website Tools, Tips, and Technology.pdf
Ask The Expert! Nonprofit Website Tools, Tips, and Technology.pdf
TechSoup 67 views
JRN 362 - Lecture Twenty-Three (Epilogue) von Rich Hanley
JRN 362 - Lecture Twenty-Three (Epilogue)JRN 362 - Lecture Twenty-Three (Epilogue)
JRN 362 - Lecture Twenty-Three (Epilogue)
Rich Hanley44 views
Creative Restart 2023: Christophe Wechsler - From the Inside Out: Cultivating... von Taste
Creative Restart 2023: Christophe Wechsler - From the Inside Out: Cultivating...Creative Restart 2023: Christophe Wechsler - From the Inside Out: Cultivating...
Creative Restart 2023: Christophe Wechsler - From the Inside Out: Cultivating...
Taste39 views
INT-244 Topic 6b Confucianism von S Meyer
INT-244 Topic 6b ConfucianismINT-244 Topic 6b Confucianism
INT-244 Topic 6b Confucianism
S Meyer51 views
Peripheral artery diseases by Dr. Garvit.pptx von garvitnanecha
Peripheral artery diseases by Dr. Garvit.pptxPeripheral artery diseases by Dr. Garvit.pptx
Peripheral artery diseases by Dr. Garvit.pptx
garvitnanecha135 views
OOPs - JAVA Quick Reference.pdf von ArthyR3
OOPs - JAVA Quick Reference.pdfOOPs - JAVA Quick Reference.pdf
OOPs - JAVA Quick Reference.pdf
ArthyR376 views
NodeJS and ExpressJS.pdf von ArthyR3
NodeJS and ExpressJS.pdfNodeJS and ExpressJS.pdf
NodeJS and ExpressJS.pdf
ArthyR353 views

Very helpful PPT: Models of word recognition

  • 1. PSY 369: Psycholinguistics Language Comprehension Word recognition & speech recognition
  • 2. Lexical access  How do we retrieve the linguistic information from Long-term memory?    How is the information organized/stored? What factors are involved in retrieving information from the lexicon? Models of lexical access
  • 3. Lexical access  How do we retrieve the linguistic information from Long-term memory?    How is the information organized/stored? What factors are involved in retrieving information from the lexicon? Models of lexical access
  • 4. Models of lexical access  Serial comparison models   Parallel comparison models    Search model (Forster, 1976, 1979, 1987, 1989) Logogen model (Morton, 1969) Cohort model (Marslen-Wilson, 1987, 1990) Connectionist models  Interactive Activation Model (McClelland and Rumelhart, 1981)
  • 5. Search model (e.g., Forster, 1976)  Access of the lexicon is considered autonomous, independent of other systems involved in processing language   A complete perceptual representation of the perceived stimulus is constructed The representation is compared with representations in access files  Three access files:     Orthographic Phonological Syntactic/semantic (for language production) Access files are organized in a series of bins (first syllable or letters)   Position within the bins is organized by lexical frequency Access files have “pointers” to meaning information in semantic memory
  • 6. Search model (e.g., Forster, 1976) Visual input Pointers Decreasing frequency Entries in order of Access codes Auditory input /kat/ cat Mental lexicon mat cat mouse
  • 7. Search model (e.g., Forster, 1976) Search model (Forster, 1976, 1979, 1987, 1989)    Frequency effects  Bin organization Access codes Repetition priming effects  Temporary reordering of bins in response to recent encounter Semantic priming effects  Accounted for by cross referencing in the lexicon Context effects Mental lexicon  Search is considered to be autonomous, un affected by context (so context effects are “post-access”) Decreasing frequency  Entries in order of  Visual input Auditory input Pointer s /kat/ cat mat cat mouse
  • 8. Logogen model (Morton 1969)  The lexical entry for each word comes with a logogen  Logogens specify word‟s attributes   e.g., semantic, orthographic, phonological Activated in two ways    cat By sensory input By contextual information Access (recognition) when reach threshold  Different thresholds depending on different factors   e.g., frequency Access makes information associated with word available dog cap
  • 9. Logogen model (Morton 1969) Auditory stimuli Auditory analysis Context system Visual stimuli Visual analysis Semantic Attributes Logogen system Available Responses Output buffer Responses cat dog cap
  • 10. Think of a logogen as being like a „strength-o-meter‟ at a fairground When the bell rings, the logogen has „fired‟
  • 11. ‘cat’ [kæt] • What makes the logogen fire? – seeing/hearing the word • What happens once the logogen has fired? – access to lexical entry!
  • 12. ‘cat’ [kæt] • So how does this help us to explain the frequency effect? – High frequency words have a lower threshold for firing –e.g., cat vs. cot ‘cot’ [kot] Low freq takes longer
  • 13. ‘doctor’ [doktə] • Spreading activation from doctor lowers the threshold for nurse to fire – So nurse take less time to fire doctor ‘nurse’ [nə:s] Spreading activation network doctor nurse nurse
  • 14. Interactive Activation Model (IAM) Proposed to account for Word Superiority effect McClelland and Rumelhart, (1981)
  • 15. The Word-Superiority Effect (Reicher, 1969) + Until the participant hits some start key
  • 16. The Word-Superiority Effect (Reicher, 1969) COURSE Presented briefly … say 25 ms
  • 17. The Word-Superiority Effect (Reicher, 1969) U &&&&& A Mask presented with alternatives above and below the target letter … participants must pick one as the letter they believe was presented in that position.
  • 18. The Word-Superiority Effect (Reicher, 1969) + + E KLANE E & T Letter only Say 60% + PLANE E &&&&& T E &&&&& T Letter in Nonword Say 65% Letter in Word Say 80% Why is identification better when a letter is presented in a word?
  • 19. Also goes by the name: Interactive Activation and Competition Model (IAC) Interactive Activation Model (IAM) Previous models posed a bottom-up flow of information (from features to letters to words). IAM also poses a topdown flow of information Nodes: • (visual) feature McClelland and Rumelhart, (1981) • (positional) letter • word detectors • Inhibitory and excitatory connections between them.
  • 20. Interactive Activation Model (IAM)  Inhibitory connections within levels   If the first letter of a word is “a”, it isn‟t “b” or “c” or … Inhibitory and excitatory connections between levels (bottom-up and top-down)   If the first letter is “a” the word could be “apple” or “ant” or …., but not “book” or “church” or…… If there is growing evidence that the word is “apple” that evidence confirms that the first letter is “a”, and not “b”….. McClelland and Rumelhart, (1981)
  • 21. IAM & the word superiority effect  The model processes at the word and letter levels simultaneously   Letters in words benefit from bottom-up and top-down activation But letters alone receive only bottom-up activation. McClelland and Rumelhart, (1981)
  • 22. Cohort model (Marslen-Wilson & Welch, 1978)  Specifically for auditory word recognition (covered in chapter 9 of textbook)  Speakers can recognize a word very rapidly   Recognition point (uniqueness point)   Usually within 200-250 msec Point at which a word is unambiguously different from other words and can be recognized (strong emphasis on word onsets) Three stages of word recognition 1) activate a set of possible candidates 2) narrow the search to one candidate 3) integrate single candidate into semantic and syntactic context
  • 23. Cohort model  Prior context: “I took the car for a …” /s/ /sp/ … soap spinach psychologist spin spit sun spank … spinach spin spit spank … time /spi/ spinach spin spit … /spin/ spin
  • 24. Comparing the models  Each model can account for major findings (e.g., frequency, semantic priming, context), but they do so in different ways.     Search model is serial, bottom-up, and autonomous Logogen is parallel and interactive (information flows up and down) AIM is both bottom-up and top-down, uses facilitation and inhibition Cohort is bottom-up but parallel initially, but then interactive at a later stage
  • 25. Different signals Visual word recognition Speech Perception Where are you going   Some parallel input Orthography      Letters Clear delineation Difficult to learn Serial input Phonetics/Phonology    Acoustic features Usually no delineation “Easy” to learn
  • 26. Different signals Visual word recognition Speech Perception Where are you going   Some parallel input Orthography      Letters Clear delineation Difficult to learn Serial input Phonetics/Phonology    Acoustic features Usually no delineation “Easy” to learn
  • 27. Speech perception  Articulatory phonetics  Production based   Place and manner of articulation Acoustic phonetics  Based on the acoustic signal  Formants, transitions, co-articulation, etc.
  • 28. Speech production to perception  Acoustic cues are extracted and stored in sensory memory and then mapped onto linguistic information  Air is pushed into the larynx across the vocal cords and into the mouth nose, different types of sounds are produced.   The different qualities of the sounds are represented in formants The formants and other features are mapped onto phonemes
  • 29. Acoustic features  Spectrogram   Frequency  Time on the x-axis Frequency (pressure under which the air is pushed) on the y-axis Amplitude is represented by the darkness of the lines time
  • 30. Acoustic features  Acoustic features  Formants - bands of resonant frequencies     Formant transitions - up or down movement of formants Steady states - flat formant patterns Bursts - sudden release of air Voice onset time (VOT) - when the voicing begins relative to the onset of the phoneme
  • 31. Formants in a wide-band spectrogram <-- Formant transitions --> <-- F 3 <-- F 2 <-- F1  Formants - bands of resonant frequencies   Formant transitions - up or down movement of formants Steady states - flat formant patterns
  • 32. Formants in a wide-band spectrogram Burst --> <-- Formant transitions --> <-- F1  Bursts – sudden release of air <-- F 3 <-- F 2
  • 34. Categorical Perception  Categorical Perception is the perception of different sensory phenomena as being qualitatively, or categorically, different.  Liberman et al (1957)  Used the speech synthesizer to create a series of syllables panning categories /b/, /d/, /g/ (followed by /a/)    Was done by manipulating the F2 formant Stimuli formed a physical continuum Result, people didn‟t “hear” a continuum, instead classified them into three categories
  • 35. Categorical Perception  Liberman et al (1957) 1. Set up a continuum of sounds between two categories /ba/ 1 - ... 3 … /da/ 5 … 7
  • 36. Categorical Perception  Liberman et al (1957) 2. Run an identification experiment Sharp phoneme boundary 100  % /ba/ Our perception of phonemes is categorical rather than continuous. 0 1 ... 3 … 5 … 7
  • 37. Hard Problems in Speech Perception Wave form  Linearity (parallel transmission): Acoustic features often spread themselves out over other sounds  Where does show start and money end?  Demo's and info
  • 38. Hard Problems in Speech Perception Wave form  Invariance:  One phoneme should have a one waveform   But, the /i/ („ee‟) in „money‟ and „me‟ are different There aren‟t invariant cues for phonetic segments  Although the search continues  Demo's and info
  • 39. Hard Problems in Speech Perception Wave form  Co-articulation: the influence of the articulation (pronunciation) of one phoneme on that of another phoneme.  Essentially, producing more than one speech sound at once  Demo's and info
  • 40. Hard Problems in Speech Perception  Trading relations  Most phonetic distinctions have more than one acoustic cue as a result of the particular articulatory gesture that gives the distinction.   slit–split – the /p/ relies on silence and rising formant, different mixtures of these can result in the same perception Perception must establish some "trade-off" between the different cues.
  • 41. Hard Problems in Speech Perception  The McGurk effect: McGurk and MacDonald (1976) • Showed people a video where the audio and the video don‟t match • Think “dubbed movie”   McGurk effect McGurk effect2 • visual /ga/ with auditory /ba/ often hear /da/  Implications • phoneme perception is an active process • influenced by both audio and visual information
  • 42. Motor theory of speech perception  A. Liberman (and others, initially proposed in late 50‟s)   Direct translation of acoustic speech into articulatorally defined categories Holds that speech perception and motor control involved linked (or the same) neural processes  Theory held that categorical perception was a direct reflection of articulatory organization    Categories with discrete gestures (e.g., consonants) will be perceived categorically Categories with continuous gestures (e.g., vowels) will be perceived continuously There is a speech perception module that operates independently of general auditory perception
  • 43. Frontal slices showing differential activation elicited during lip and tongue movements (Left), syllable articulation including [p] and [t] (Center), and listening to syllables including [p] and [t] (Right) Pulvermüller F et al. PNAS 2006;103:7865-7870 ©2006 by National Academy of Sciences
  • 44. Motor theory of speech perception  Some problems for MT   Categorical perception found in non-speech sounds (e.g., music) Categorical perception for speech sounds in non-humans  Chinchillas can be trained to show categorical perception of /t/ and /d/ consonant-vowel syllables (Kuhl & Miller, 1975)
  • 45. Other theories of speech perception  Direct Realist Theory (C. Fowler and others)    Similar to Motor theory, articulation representations are key, but here they are directly perceived Perceiving speech is part of a more general perception of gestures that involves the motor system General Auditory Approach (e.g., Diehl, Massaro)  Do not invoke special mechanisms for speech perception, instead rely on more general mechanisms of audition and perception  For nice reviews see:  Diehl, Lotto, & Holt (2003)  Galantucci, Fowler, Turvey (2006)
  • 46. Top-down effects on Speech Perception   Phoneme restoration effect Sentence context effects
  • 47. Phoneme restoration effect Listen to a sentence which contained a word from which a phoneme was deleted and replaced with another noise (e.g., a cough) The state governors met with their respective legi*latures convening in the capital city. * /s/ deleted and replaced with a cough Click here for a demo and additional information
  • 48. Phoneme restoration effect Typical results: Participants heard the word normally, despite the missing phoneme Usually failed to identify which phoneme was missing Interpretation We can use top-down knowledge to “fill in” the missing information
  • 49. Phoneme restoration effect Further experiments (Warren and Warren, 1970): What if the missing phoneme was ambiguous The *eel was on the axle. The *eel was on the shoe. The *eel was on the orange. The *eel was on the table. Results: Participants heard the contextually appropriate word normally, despite the missing phoneme
  • 50. Phoneme restoration effect  Possible loci of phoneme restoration effects  Perceptual loci of effect:   Lexical or sentential context influences the way in which the word is initially perceived. Post-perceptual loci of effect:  Lexical or sentential context influences decisions about the nature of the missing phoneme information.
  • 51. Beyond the segment Shillcock (1990): hear a sentence, make a lexical decision to a word that pops up on computer screen (cross-modal priming) Hear: The scientist made a new discovery last year. NUDIST
  • 52. Cross-modal priming Shillcock (1990): hear a sentence, make a lexical decision to a word that pops up on computer screen (cross-modal priming) Hear: The scientist made a novel discovery last year. NUDIST
  • 53. Cross-modal priming Shillcock (1990): hear a sentence, make a lexical decision to a word that pops up on computer screen (cross-modal priming) Hear: The scientist made a novel discovery last year. The scientist made a new discovery last year. faster
  • 54. Cross-modal priming Shillcock (1990): hear a sentence, make a lexical decision to a word that pops up on computer screen (cross-modal priming) Hear: The scientist made a novel discovery last year. The scientist made a new discovery last year. faster NUDIST gets primed by segmentation error Although no conscious report of hearing “nudist”
  • 55. Beyond the segment  Prosody and intonation  English:  Speech is divided into phrases.  Word stress is meaningful in English.  Stressed syllables are aligned in a fairly regular rhythm, while unstressed syllables take very little time.  Every phrase has a focus.  An extended flat or low-rising intonation at the end of a phrase can indicate that a speaker intends to continue to speak.  A falling intonation sounds more final.
  • 56. Beyond the segment  Prosodic factors (supra segmentals)  Stress   Rate   Emphasis on syllables in sentences Speed of articulation Intonation  Use of pitch to signify different meanings across sentences
  • 57. Beyond the segment  Stress effects  On meaning   “black bird” versus “blackbird” Top-down effects on perception  Better anticipation of upcoming segments when syllable is stressed
  • 58. Beyond the segment  Rate effects  How fast you speak has an impact on the speech sounds   Faster talking - shorter vowels, shorter VOT Normalization  Taking speed and speaker information into account  Rate normalization  Speaker normalization