Language Comprehension Speech Perception Meaning Representation.

Slides:



Advertisements
Similar presentations
Chapter 12: Speech and Music Perception
Advertisements

Human Speech Recognition Julia Hirschberg CS4706 (thanks to John-Paul Hosum for some slides)
Language and Cognition Colombo, June 2011 Day 8 Aphasia: disorders of comprehension.
Chapter 12 Speech Perception. Animals use sound to communicate in many ways Bird calls Bird calls Whale calls Whale calls Baboons shrieks Baboons shrieks.
The Perception of Speech. Speech is for rapid communication Speech is composed of units of sound called phonemes –examples of phonemes: /ba/ in bat, /pa/
Language Comprehension Speech Perception Semantic Processing & Naming Deficits.
Cognition, 8e by Margaret W. MatlinChapter 2 Cognition, 8e Chapter 2 Perceptual Processes I: Visual and Auditory Recognition.
The Neuroscience of Language. What is language? What is it for? Rapid efficient communication – (as such, other kinds of communication might be called.
Ling 240: Language and Mind Acquisition of Phonology.
Speech perception 2 Perceptual organization of speech.
Writing Workshop Here are some typical writing style issues which people have trouble with.
Development of Speech Perception. Issues in the development of speech perception Are the mechanisms peculiar to speech perception evident in young infants?
The Perception of Speech. Speech is for rapid communication Speech is composed of units of sound called phonemes –examples of phonemes: /ba/ in bat, /pa/
Auditory Word Recognition
Speech Perception Overview of Questions Can computers perceive speech as well as humans? Does each word that we hear have a unique pattern associated.
Exam 1 Monday, Tuesday, Wednesday next week WebCT testing centre Covers everything up to and including hearing (i.e. this lecture)
Language. Using Language What is language for? Using Language What is language for? – Rapid, efficient communication To accomplish this goal, what needs.
Cognitive Processes PSY 334 Chapter 2 – Perception April 9, 2003.
What is Cognitive Science? … is the interdisciplinary study of mind and intelligence, embracing philosophy, psychology, artificial intelligence, neuroscience,
Visual Cognition II Object Perception. Theories of Object Recognition Template matching models Feature matching Models Recognition-by-components Configural.
Language. Processing Stages Language makes it possible for us to inform each other about the world. A sentence (string of words) is perceived. That string.
WHAT, WHERE, & HOW SYSTEMS AGNOSIAS!. What, Where, & How Systems.
PSY 369: Psycholinguistics
SPEECH PERCEPTION The Speech Stimulus Perceiving Phonemes Top-Down Processing Is Speech Special?
Visual Cognition II Object Perception. Theories of Object Recognition Template matching models Feature matching Models Recognition-by-components Configural.
What is Phonetics? Short answer: The study of speech sounds in all their aspects. Phonetics is about describing speech. (Note: phonetics ¹ phonics) Phonetic.
Sound and Speech. The vocal tract Figures from Graddol et al.
Reading. Reading Research Processes involved in reading –Orthography (the spelling of words) –Phonology (the sound of words) –Word meaning –Syntax –Higher-level.
Reading & Speech Perception Connectionist Approach E.g., Seidenberg and McClelland (1989) and Plaut (1996). Central to these models is the absence of.
Phonetics, day 2 Oct 3, 2008 Phonetics 1.Experimental a. production b. perception 2. Surveys/Interviews.
Language Comprehension Speech Perception Naming Deficits.
Memory II Are there multiple LTM memory systems? How do you learn a new skill? How do you learn a new fact? How about learning about an event? Is there.
PSY 369: Psycholinguistics Language Comprehension: Visual perception.
An aside: peripheral drift illusion illusion of motion is strongest when reading text (such as this) while viewing the image in your periphery. Blinking.
The Perception of Speech
Cognitive Processes PSY 334 Chapter 2 – Perception.
Cognitive Psychology Chapter 3. Visual Consciousness Transduction of the visible spectrum (400 nm to 700 nm) of electromagnetic radiation. Crossing.
Audio Scene Analysis and Music Cognitive Elements of Music Listening
Speech Perception 4/6/00 Acoustic-Perceptual Invariance in Speech Perceptual Constancy or Perceptual Invariance: –Perpetual constancy is necessary, however,
? Speech Perception []
1 Speech Perception 3/30/00. 2 Speech Perception How do we perceive speech? –Multifaceted process –Not fully understood –Models & theories attempt to.
Speech Perception1 Fricatives and Affricates We will be looking at acoustic cues in terms of … –Manner –Place –voicing.
PSY 369: Psycholinguistics Language Comprehension Word recognition & speech recognition.
What is Language? Education 388 Lecture 3 January 23, 2008 Kenji Hakuta, Professor.
Speech Perception 4/4/00.
Cognitive Processes PSY 334
PSY 369: Psycholinguistics Language Comprehension Speech recognition.
Sensation & Perception
Acoustic Continua and Phonetic Categories Frequency - Tones.
1 Cross-language evidence for three factors in speech perception Sandra Anacleto uOttawa.
Cognitive Processes PSY 334 Chapter 5 – Meaning-Based Knowledge Representation.
Chapter 13: Speech Perception. The Acoustic Signal Produced by air that is pushed up from the lungs through the vocal cords and into the vocal tract Vowels.
Solve this maze at your leisure.
Language Perception.
Motor Theory + Signal Detection Theory
WebCT You will find a link to WebCT under the “Current Students” heading on It is your responsibility to know how to work WebCT!
Heilmann, J., Ellis Weismer, S., Evans, J. and C. Hollar. (2005). Utility of the MacArthur–Bates Communicative Development Inventory in Identifying Language.
Audio Scene Analysis and Music Cognitive Elements of Music Listening Kevin D. Donohue Databeam Professor Electrical and Computer Engineering University.
Solve this maze at your leisure. Start at phil’s house. At first, you can only make right turns through the maze. Each time you cross the red zigzag sign.
Long Term Memory LONG TERM MEMORY (LTM)  Variety of information stored in LTM:  The capital of Turkey  How to drive a car.
VISUAL WORD RECOGNITION. What is Word Recognition? Features, letters & word interactions Interactive Activation Model Lexical and Sublexical Approach.
Perception & Pattern Recognition 1 Perception Pattern Recognition Theories of Pattern Recognition Bottom-up vs. Top-Down Processing & Pattern Recognition.
Cognitive Processes PSY 334
Speech Perception Models
Cognitive Processes PSY 334
Copyright © American Speech-Language-Hearing Association
What is Phonetics? Short answer: The study of speech sounds in all their aspects. Phonetics is about describing speech. (Note: phonetics ¹ phonics) Phonetic.
Language Comprehension
Cognitive Processes PSY 334
Topic: Language perception
Presentation transcript:

Language Comprehension Speech Perception Meaning Representation

Speech Comprehension and Production

Speech Perception The first step in comprehending spoken language is to identify the words being spoken, performed in multiple stages: 1. Phonemes are detected (/b/, /e/, /t/, /e/, /r/, ) 2. Phonemes are combined into syllables (/be/ /ter/) 3. Syllables are combined into words (“better”) 4. Word meaning retrieved from memory

Spectrogram: I owe you a yo-yo

Speech perception: two problems Words are not neatly segmented (e.g., by pauses) Lack of phoneme invariance Coarticulation = consecutive speech sounds blend into each other due to mechanical constraints on articulators Speaker differences; pitch affected by age and sex; different dialects, talking speeds etc.

How Do Listeners Deal with Variability in Acoustic Input? Use of visual cues: –McGurk effect Use of semantic cues: –Phonemic restoration Categorical perception: continuous changes in input are mapped on to discrete percepts

Phonemic restoration Auditory presentation Perception Legislature legislature Legi_laturelegi lature Legi*lature legislature It was found that the *eel was on the axle. wheel It was found that the *eel was on the shoe. heel It was found that the *eel was on the orange. peel It was found that the *eel was on the table. meal Warren, R. M. (1970). Perceptual restorations of missing speech sounds. Science, 167,

McGurk Effect Perception of auditory event affected by visual processing Harry McGurk and John MacDonald in "Hearing lips and seeing voices", Nature 264, (1976). Demo 1 AVI: MOV: Demo 2 MOV:

McGurk Effect McGurk effect in video: –lip movements = “ga” –speech sound = “ba” –speech perception = “da” (for 98% of adults) Demonstrates parallel & interactive processing: speech perception is based on multiple sources of information, e.g. lip movements, auditory information. Brain makes reasonable assumption that both sources are informative and “fuses” the information.

Categorical Perception Categorical perception: high level cognitive processes (i.e., categorization) can influence perceptual processes Categorization Perception of Sounds/Images categorical perception

Differences among items that fall into different categories are exaggerated, and differences among items that fall into the same category are minimized. (from Rob Goldstone, Indiana University)

Examples from “LAKE” to “RAKE” – from /da/ to /ga/ Good /ga/Good /da/

Identification: Discontinuity at Boundary % of /ga/ response 100% 0% 50% Token

Pairwise discrimination Good /ga/Good /da/ Discriminate these pairs (straddle the category boundary) Discriminate these pairs

Pairwise Discrimination (same/different) % Correct Discrimination _22_33_44_55_66_77_8 Pair of stimuli

What Happened? Physical World Perceptual Representation

Categorical Perception depends on language In one language a difference in sound may make a difference between words; in another, it might not Example –The Japanese language does not distinguish between /l/ and /r/ –These sounds belong to the same category for Japanese listeners –They find it very hard to discriminate between them (Massaro, 1994)

Dental Stop Retroflex Stop Hindi Non-English Contrasts Uvular Velar Salish (Native North American—Canadian—language)

Models of Spoken Word Identification The Cohort Model –Marslen-Wilson & Welsh, 1978 –Revised, Marslen-Wilson, 1989 The TRACE Model –Similar to the Interactive Activation model –McClelland & Elman, 1986

Online word recognition: the cohort model

Recognizing Spoken Words: The Cohort Model All candidates considered in parallel Candidates eliminated as more evidence becomes available in the speech input Uniqueness point occurs when only one candidate remains

Evidence for activation of spurious words If we recognize words by recognizing a cohort of possibilities, then the cohort words should excert some influence Shillcock (1990). Test for semantic priming of cohort words: “He picked up the trombone” “trom” “bone” Semantic priming for the word “rib”

TRACE model Similar to interactive activation model but applied to speech recognition Connections between levels are bi-directional and excitatory  top-down effects Connections within levels are inhibitory producing competition between alternatives (McClelland & Elman, 1986)

TRACE Model (McClelland & Elman, 1986)

Human Eye Tracking Data ‘Pick up the beaker’ Allopenna, Magnuson & Tanenhaus (1998) Eye tracking device to measure where subjects are looking

Human Eye Tracking Data Human eye tracking data highly similar to TRACE predictions Allopenna, Magnuson & Tanenhaus (1998)

Representing Meaning

Mental representation of meaning as a network of interconnected features Evidence comes from patients with category-specific impairments –more difficulty activating semantic representation for some categories than for others

Category Specific Semantic Deficits Warrington and Shallice (1984) reported a patient called JBR who following an acute lesion to the left temporal lobe (as a result of herpes encephalitis) had a selective deficit when asked to name pictures from just one semantic category – living things. By contrast JBR was able to name non-living objects very well including those with low frequency names such as ‘accordion’ that were matched for the number of letters in the name and the visual complexity of the object. Other patients have shown opposite pattern

Summary of patient data Living Nonliving Animal Fruit Artefacts x x√ (Warrington & Shallice,1984). √ √ x (Sheridan & Humphreys, 1993). x √ √ (Hart & Gordon, 1992). √ x x(Hillis & Caramazza, 1991). √ x √ (Hart, Berndt, Caramazza, 1985).

Representing Meaning

Implications Different types of objects depend on different types of encoding  perceptual information  functional information

Sensory-Functional Approach Category specific effects on recognition result from a correlated factor such as the ratio of visual versus functional features of an object –living more visual and nonliving more functional. Farah & McClelland (1991) report a dictionary study showing the ratio of visual to functional features for living things and nonliving things: –living things was 7.7:1 and nonliving was 1.4:1.

A neural network model of category- specific impairments A single system with functional and visual features. Model was trained to discriminate 20 living and nonliving things Two main layers: semantic and input. Semantic has 3:1 ratio of visual and functional properties Objects have visual and functional property codes –7.7:1 for living things –1.4:1 for nonliving things Farah and McClelland (1991)

Simulating the Effects of Brain Damage by “lesioning” the model Farah and McClelland (1991) Visual Lesions: selective impairment of living things Functional Lesions: selective impairment of non-living things