Outline.! Neural representation of speech sounds. " Basic intro " Sounds and categories " How do we perceive sounds? " Is speech sounds special?

Similar documents
Language Speech. Speech is the preferred modality for language.

Does Wernicke's Aphasia necessitate pure word deafness? Or the other way around? Or can they be independent? Or is that completely uncertain yet?

Chapter 11: Sound, The Auditory System, and Pitch Perception

MULTI-CHANNEL COMMUNICATION

group by pitch: similar frequencies tend to be grouped together - attributed to a common source.

ID# Exam 2 PS 325, Fall 2003

Hearing. istockphoto/thinkstock

ID# Final Exam PS325, Fall 1997

PSY 215 Lecture 10 Topic: Hearing Chapter 7, pages

5. Which word refers to making

Hearing. PSYCHOLOGY (8th Edition, in Modules) David Myers. Module 14. Hearing. Hearing

THE EAR Dr. Lily V. Hughes, Audiologist

Speech Generation and Perception

Speech (Sound) Processing

11 Music and Speech Perception

Topics in Linguistic Theory: Laboratory Phonology Spring 2007

17.4 Sound and Hearing

FAILURES OF OBJECT RECOGNITION. Dr. Walter S. Marcantoni

Sensorimotor Functioning. Sensory and Motor Systems. Functional Anatomy of Brain- Behavioral Relationships

Hearing. and other senses

Sound Waves. Sensation and Perception. Sound Waves. Sound Waves. Sound Waves

HEARING. Structure and Function

ID# Exam 2 PS 325, Fall 2009

BCS 221: Auditory Perception BCS 521 & PSY 221

AUDL GS08/GAV1 Signals, systems, acoustics and the ear. Pitch & Binaural listening

Required Slide. Session Objectives

Human cogition. Human Cognition. Optical Illusions. Human cognition. Optical Illusions. Optical Illusions

THE MECHANICS OF HEARING

SUBJECT: Physics TEACHER: Mr. S. Campbell DATE: 15/1/2017 GRADE: DURATION: 1 wk GENERAL TOPIC: The Physics Of Hearing

MECHANISM OF HEARING

COM3502/4502/6502 SPEECH PROCESSING

Auditory Physiology PSY 310 Greg Francis. Lecture 30. Organ of Corti

Categorical Perception

Psychology Chapter 4. Sensation and Perception. Most amazing introduction ever!! Turn to page 77 and prepare to be amazed!

It is important to understand as to how do we hear sounds. There is air all around us. The air carries the sound waves but it is below 20Hz that our

! Can hear whistle? ! Where are we on course map? ! What we did in lab last week. ! Psychoacoustics

NERVOUS SYSTEM & SENSES TEACHER COPY

Speech Spectra and Spectrograms

Ganglion Cells Blind Spot Cornea Pupil Visual Area of the Bipolar Cells Thalamus Rods and Cones Lens Visual cortex of the occipital lobe

Higher Cortical Function

Running head: HEARING-AIDS INDUCE PLASTICITY IN THE AUDITORY SYSTEM 1

Transfer of Sound Energy through Vibrations

Deafness and hearing impairment

Optical Illusions 4/5. Optical Illusions 2/5. Optical Illusions 5/5 Optical Illusions 1/5. Reading. Reading. Fang Chen Spring 2004

SLHS 1301 The Physics and Biology of Spoken Language. Practice Exam 2. b) 2 32

Hearing Lectures. Acoustics of Speech and Hearing. Auditory Lighthouse. Facts about Timbre. Analysis of Complex Sounds

Learning Targets. Module 20. Hearing Explain how the ear transforms sound energy into neural messages.

2/25/2013. Context Effect on Suprasegmental Cues. Supresegmental Cues. Pitch Contour Identification (PCI) Context Effect with Cochlear Implants

Unit VIII Problem 9 Physiology: Hearing

What is sound? Range of Human Hearing. Sound Waveforms. Speech Acoustics 5/14/2016. The Ear. Threshold of Hearing Weighting

ISLAMABAD ACADEMY PHYSICS FOR 10TH CLASS (UNIT # 13)

Receptors / physiology

Perceptual Disorders. Agnosias

SYLLABUS FOR PH.D ENTRANCE TEST IN SPEECH AND HEARING

Lecture Outline. The GIN test and some clinical applications. Introduction. Temporal processing. Gap detection. Temporal resolution and discrimination

Auditory System. Barb Rohrer (SEI )

EVALUATION OF SPEECH PERCEPTION IN PATIENTS WITH SKI SLOPE HEARING LOSS USING ARABIC CONSTANT SPEECH DISCRIMINATION LISTS

Consciousness and Blindsight

to vibrate the fluid. The ossicles amplify the pressure. The surface area of the oval window is

Excellent Network Courses. Department of Neurology Affiliated hospital of Jiangsu University

The Outer and Middle Ear PERIPHERAL AUDITORY SYSTEM HOW WE HEAR. The Ear in Action AUDITORY NEUROPATHY: A CLOSER LOOK. The 3 parts of the ear

The speed at which it travels is a function of the density of the conducting medium.

Learning Objectives.

Music and Hearing in the Older Population: an Audiologist's Perspective

Auditory scene analysis in humans: Implications for computational implementations.

Converting Sound Waves into Neural Signals, Part 1. What happens to initiate neural signals for sound?

HEARING GUIDE PREPARED FOR CLINICAL PROFESSIONALS HEARING.HEALTH.MIL. HCE_ClinicalProvider-Flip_FINAL01.indb 1

Definition Slides. Sensation. Perception. Bottom-up processing. Selective attention. Top-down processing 11/3/2013

Spectrograms (revisited)

= add definition here. Definition Slide

Prof. Greg Francis 7/7/08

Topic 4. Pitch & Frequency

the human 1 of 3 Lecture 6 chapter 1 Remember to start on your paper prototyping

Chapter 17 Sound Sound and Hearing. Properties of Sound Waves 1/20/2017. Pearson Prentice Hall Physical Science: Concepts in Action

The Deaf Brain. Bencie Woll Deafness Cognition and Language Research Centre

Auditory fmri correlates of loudness perception for monaural and diotic stimulation

Systems Neuroscience Oct. 16, Auditory system. http:

Consonant Perception test

Hearing Evaluation: Diagnostic Approach

Sensation and Perception. 8.2 The Senses

The neural basis of sign language processing

Acoustics, signals & systems for audiology. Psychoacoustics of hearing impairment

THE ROLE OF VISUAL SPEECH CUES IN THE AUDITORY PERCEPTION OF SYNTHETIC STIMULI BY CHILDREN USING A COCHLEAR IMPLANT AND CHILDREN WITH NORMAL HEARING

Takwa Adly Gabr Assistant lecturer of Audiology

Audiology Curriculum Foundation Course Linkages

Psy /16 Human Communication. By Joseline

Linguistic Phonetics Fall 2005

Hearing Sound. The Human Auditory System. The Outer Ear. Music 170: The Ear

Music 170: The Ear. Tamara Smyth, Department of Music, University of California, San Diego (UCSD) November 17, 2016

Binaural Hearing. Why two ears? Definitions

ACOUSTIC AND PERCEPTUAL PROPERTIES OF ENGLISH FRICATIVES

Nervous System. Made of two parts. Central Peripheral

Sound. Audition. Physics of Sound. Properties of sound. Perception of sound works the same way as light.

Audition. Sound. Physics of Sound. Perception of sound works the same way as light.

Hearing Screening, Diagnostics and Intervention

THE COCHLEA AND AUDITORY PATHWAY

Fundamentals of Cognitive Psychology, 3e by Ronald T. Kellogg Chapter 2. Multiple Choice

Vision and Audition. This section concerns the anatomy of two important sensory systems, the visual and the auditory systems.

Unit 4: Sensation and Perception

Hearing in the Environment

Transcription:

Outline! Neural representation of speech sounds " Basic intro " Sounds and categories " How do we perceive sounds? " Is speech sounds special?

! What is a phoneme?! It s the basic linguistic unit of speech! Examples of phonemes?! How many phonemes in the word cat?! It comprises of 3 phonemes (k æ t)! kʰ æ t! mk æ t! {m æ t vs. k æ t} vs. {k æ t vs. kʰ æ t}! To be more precise, a phoneme is the smallest contrastive speech unit that brings about a change of meaning.

! Phoneme is the basic contrastive unit of speech! Language = Speech?! How about sign languages?

Language =Speech! Existence of signed languages tells us that that speech is not a necessary modality for language: the meanings and complex structures of language can also be coded as manual gestures.! Nevertheless, no languages spoken by hearing communities are signed.! Thus in the presence of hearing, speech is the preferred modality for language.

How do we classify phonemes! The standard classification of phonemes refers mainly to the motor gestures that produce them! For example, /x/ is Voiceless velar fricative in Arabic!! Distinctive features of sounds! Place of articulation is : velar! Manner of articulation: *fricative! Voicing: voiceless (vocal folds don t vibrate) * Sounds that are produced by bringing the mouth into position to block the passage of the airstream, but not making complete closure so that air moving through the mouth generates audible friction

How our brains perceive categories of sounds! Sounds have distinctive features!! Categorical perception helps us to distinguish between things!! When items that range along a continuum are perceived as being either more or less similar to each other than they really are because of the way the are categorized!

t d p b r l

Do you know where I can find pins! Here is the bin!

But I meant

How do we perceive sounds!! Sound waves are first collected in our outer ear. Then they travel down our ear canal and strike the eardrum, causing it to vibrate. These vibrations are in turn transmitted to our inner ear by the bones of our middle ear (hammer, anvil, and stirrup). Our inner ear plays changes the vibrations created by the sound waves into electrical impulses, or signals, which can be recognized and decoded by our brain. Our brain receives these impulses in its hearing centers and interprets them as a type of sound https://www2.southeastern.edu/academics/faculty/mrossano/perception/powerpoints/chapter%2010.pptx https://www.google.com/url?sa=i&rct=j&q=&esrc=s&source=images&cd=&cad=rja&uact=8&ved=0ahukewjblrxihl7yahubmrqkhzcjdv4qjrwibw&url=http%3a%2f%2fslideplayer.com%2fslide%2f4659288%2f&psig=aovvaw1ddzzcf97s41mgmyjnaobk&ust=1515145836173022

Outer ear Collects sound waves. The configuration of the outer ear serves to amplify sound, particularly at 2000-5000 Hz, a frequency range that is important for speech. Middle ear Transforms the energy of a sound wave into the internal vibrations of the bone structure of the middle ear and transforms these vibrations into a compressional wave in the inner ear. Inner ear Transform the energy of a compressional wave within the inner ear fluid into nerve impulses which can be transmitted to the brain.

what the ear hears!! The healthy ear can perceive frequencies from 20 Hz to 20,000 Hz. The sound frequencies between 500 Hz and 4000 Hz are those that are most important for speech! Amplitude: height of wave; measured in db https://www2.southeastern.edu/academics/faculty/mrossano/perception/powerpoints/chapter%2010.pptx

Auditory input reaches primary auditory cortex about 10 15 msec after stimulus onset (Liegeois- Chauvel, Musolino, & Chauvel, 1991; Celesia, 1976). http://www.sis.pitt.edu/~is1042/html/audtrans.gif

Primary auditory cortex: Brodmann areas 41 and 42. Secondary auditory cortex: Brodmann areas 22 and 37. http://ist-socrates.berkeley.edu/ ~jmp/lo2/6.html http://neurosciencenews.com/auditory-cortex-deaf-hearing-4693/

Task of the listener: Mapping sound to meaning /kæt/ # Requires converting the raw signal into the kinds of sound representations that connect to meaning. $ What are those sound representations like? Sequences of phonemes!

Task of the listener: Mapping sound to meaning! Stored sound representations must abstract away from a huge amount of physical variance in the signal $ Loudness $ Pitch $ Gender $ Speaker identity $ Accent and other variance in pronunciation $ Etc

Categorization inside and outside language # What makes a chair a chair? # What makes a t a t? # Narrowing down the question: What makes a t a t, as opposed to a d?

Distinctive features Major class features t + consonantal - sonorant - syllabic d + consonantal - sonorant - syllabic Manner features - nasal - continuant - lateral - nasal - continuant - lateral Laryngeal features - voiced + voiced Place of articulation features - round + anterior - round + anterior

Voiced # [+ voiced] sounds are produced with the vocal folds close together, in such as way that air passing through them causes them to vibrate. Human vocal tract # Touch your larynx while pronouncing [z] vs. [s] to feel that [z] is voiced and [s] voiceless.

Voicing and stop consonants # Stop consonants: p, t, k, b, d, g $ Produced by causing a complete closure of the vocal tract and then releasing it. # Stops cannot be heard until the closure is released. # All vowels are voiced. # When the stop is released quickly, and the voicing of the vowel starts quickly, the stop is considered [+ voiced]. # The difference between [d] and [t] is that the onset of voicing of the following vowel takes longer for [t] than for [d].

Acoustically, voiced and voiceless stops differ in their Voice Onset Time (VOT) # VOT: The time it takes for the voicing of the vowel to start (voicing = vibration of the vocal folds).

What do we know about the brain bases of phonological categorization/ phoneme representations?

! Evidence: " Lesion data from patients with PWD (pure word deafness) " FMRI studies " MEG studies " Auditory mismtach (ERP) studies

Lesion Data! Early studies of PWD provided the first evidence that speech sounds are processed in the temporal lobe.! lesions causing specific phoneme perception impairment was distinguished from lesions causing other forms of *word deafness, linking the former to damage in the superior temporal gyrus (STG) (Henschen, 1918-1919).! Lesion data from patients with PWD strongly suggest some degree of bilateral phoneme processing, because the vast majority of cases have bilateral STG lesions (Buchman et al., 1986; Poeppel, 2001) *This term was used to designate any speech comprehension impairment. Specific phoneme perception

Lesion Data! *(Wada) testing shows that patients maintain much of their ability to discriminate phonemes even after anesthetization of the language dominant hemisphere (Boatman et al., 1998; Hickok et al., 2008).! https://youtu.be/n5_nx_lz834! https://youtu.be/sbkc_ncpzoo! Data suggest that phoneme perceptual processes are much more bilaterally represented than are later stages of speech comprehension. Nevertheless, PWD after unilateral damage is occasionally reported, and the damage in these instances is nearly always on the left side, suggesting a degree of asymmetry in the bilateral representation (Buchman et al., 1986; Poeppel, 2001). *wada testing is a test used to identify the dominant hemisphere for Language

FUNCTIONAL IMAGING STUDIES! Methodological limitation:! The loud noises produced by rapid gradient switching present a challenge for fmri studies of auditory processing, because these noises not only mask experimental stimuli but also activate many of the brain regions of interest.! Solution:! separating successive image volume acquisitions with a silent period during which the experimental stimuli are presented If the acquisitions are separated by a sufficient amount of time (i.e.,.7 seconds), then the measured BOLD responses primarily reflect activation by the experimental stimuli rather than the noise produced by the previous image acquisition

FMRI Studies! Compared with a silent baseline, speech sounds (whether syllables, words, pseudowords, or reversed speech) activate most of the STG bilaterally (Binder et al., 2000; Wise et al., 1991).! The activated regions include Heschl s gyrus and surrounding dorsal STG areas that contain primary and belt auditory cortex (Rauschecker & Scott, 2009), suggesting that much of the activation is due to low-level, general auditory processes.

FMRI Studies! A variety of nonspeech control stimuli (frequency-modulated tones, amplitudemodulated noise) were introduced to identify areas that might be more speech-specific (e.g., Binder et al., 2000; Zatorre, Evans, Meyer, & Gjedde, 1992).! Findings: primary and belt regions on the dorsal STG respond similarly to speech and simpler nonspeech sounds, whereas areas in the ventral STG and superior temporal sulcus (STS) respond more to speech than to simpler sounds! The pattern suggests a hierarchically organized processing stream running from the primary auditory cortex on the dorsal surface to higher level areas in the lateral and ventral STG Speech sounds Both speech and non speech sounds

FMRI Studies! In contrast to the data from Wada testing and patients with PWD, the fmri studies suggest that, at least in the intact brain, the left auditory system performs most of the computation related specifically to phoneme perception.

MEG Studies An auditory evoked neuromagnetic field M100 field pattern Roberts, Ferrari, Stufflebeam & Poeppel, 2000, J Clin Neurophysiol, Vol 17, No 2, 2000

M100 (N1 in ERP terms)! The auditory N100 is generated in the primary and association auditory cortices in the superior temporal gyrus and in Heschl's gyrus bilaterally (Zouridakis et al., 1998)! It has been found to be sensitive to things such as the predictability of an auditory stimulus, and special features of speech sounds such as voice onset time. For example:! Frequency $ higher frequency shorter latency

ERP Studies! Event-related potentials data were obtained simultaneously with fmri, showing that the difference between phonemic and nonphonemic processing begins at approximately 180 ms after stimulus onset and peaks at approximately 200 ms, manifesting as a larger P2 deflection in the phonemic condition (Liebenthal et al., 2010)! P2 component may reflect the sensationseeking behavior of an individual, and recently it s associated with attention and working memory! Inconsistent findings suggest different effects of intensity on the amplitudes of the auditory N1 and P2 (Paiva et., 2016)

Questions! Is speech special? $ Can brain damage lead to a selective loss of speech perception? $ Are the mechanisms of speech perception different from the mechanisms of auditory perception in general?! Role of audio-visual integration

Disorders of auditory processing! Rare.! Signal from each ear is processed in both hemispheres (contra visual system).

Disorders of auditory processing: CORTICAL DEAFNESS! Inability to hear sounds without apparent damage to the hearing apparatus or brain stem abnormality.! Extensive bilateral damage to auditory cortex (BAs 41 & 42).

Disorders of auditory processing: AUDITORY AGNOSIA! Inability to recognize auditorily presented sounds (e.g., coughing, crying) independent of any deficit in processing spoken language.! Damage in auditory association cortex (BAs 22 & 37)

Disorders of auditory processing: AMUSIA (a subvariaty of AUDITORY AGNOSIA)! The inability to recognize musical tones or to reproduce them (appears mainly as a defect in processing pitch)! Relative sparing of speech and (other) non-speech perception.! Damage generally in right temporal areas.

Disorders of auditory processing: PURE WORD DEAFNESS! Inability to understand spoken words while $ auditory perception is otherwise intact & $ other linguistic skills are intact (reading, speech production)

Disorders of auditory processing: PURE WORD DEAFNESS Either! Bilateral damage to auditory cortex Or! A subcortical lesion in the left hemisphere that severs both ipsilateral and contralateral projections to Wernicke s area (Ziegler,1952; Geschwind, 1965; Coslet et al., 1984; Jones and Dinolt, 1952).

Disorders of auditory processing: PHONAGNOSIA! Impairment in the ability to recognise familiar voices. $ Speech comprehension is intact. $ Intact ability to identify nonverbal sounds. $ Phonagnosia was first described by a study by Van Lancker and Cantor in 1982

Disorders of auditory processing: PHONAGNOSIA! Van Lancker et al., 1988: Double dissociation between voice recognition and voice discrimination. Some patients will perform normally on the discrimination tasks but poorly on the recognition tasks; whereas the other patients will perform normally on the recognition tasks but poorly on the discrimination tasks! Performance on a discrimination task was impaired by a lesion to either temporal lobe.! Performance on a voice recognition task was impaired by lesions to the right parietal lobe.

Questions! Is speech special? $ Can brain damage lead to a selective loss of speech perception? YES, but no one-to-one correspondence between lesion site and type of disorder. E.g., bilateral temporal damage can lead either to cortical deafness or pure word deafness. $ Are the mechanisms of speech perception different from the mechanisms of auditory perception in general?! Role of audio-visual integration

Questions! Is speech special? $ Can brain damage lead to a selective loss of speech perception? YES, but no one-to-one correspondence between lesion site and type of disorder. E.g., bilateral temporal damage can lead either to cortical deafness or pure word deafness. $ Are the mechanisms of speech perception different from the mechanisms of auditory perception in general?! Role of audio-visual integration

Close your eyes, play this movie, and listen. Open your eyes, replay and listen again. Is he saying "ba ba" or "da da"? https://youtu.be/afptc8bvdjk

What is the McGurk effect?! The McGurk effect is a perceptual phenomenon that demonstrates an interaction between hearing and vision in speech perception.! The illusion occurs when the auditory component of one sound is paired with the visual component of another sound, leading to the perception of a third sound.

Role of articulatory gestures on speech perception! When the visual image of a person saying ga is combined with an audio recording of ba, the percept is da (or for some people ga ).! Contrasting articulatory gestures can change the auditory percept even when the signal is clear.

Is audio-visual integration specific to speech?! Saldana and Rosenblum (1993): No. $ Audio visual integration of the plucks and bows of cello playing. % Not only speech, but also other ecologically valid combinations of auditory and visual stimuli can integrate in a complex manner.! Tuomainen et al. (2004, Cognition): Audio-visual integration is used in speech much more than in the perception of other sounds. $ Investigated the effect of incongruent articulatory gestures on speech vs. nonspeech, utilizing the ambiguous status of so-called *sine-wave speech. *sine-wave speech is a technique for synthesizing speech by replacing the formants (main bands of energy) with pure tone whistles. http://www.lifesci.sussex.ac.uk/home/chris_darwin/sws/

Questions! Is speech special? $ Can brain damage lead to a selective loss of speech perception? YES, but no one-to-one correspondence between lesion site and type of disorder. $ Are the mechanisms of speech perception different from the mechanisms of auditory perception in general?! This is a methodologically hard question since it s unclear what could be a control stimulus that is physically highly similar to speech but is not speech.! But the top-down effects in speech perception make it possible to use ambiguous stimuli that may or may not be perceived as speech. % Perceiving a stimulus as speech can enhance activity around auditory cortex. % Role of audio-visual integration may be special for speech.

Further investigations!! Adequate complex controls! % The tone and noise controls used in the studies are far less acoustically complex than speech sounds!!!! The degree to which speech perception requires special mechanisms!