Auditory-Visual Speech Perception Laboratory

Size: px
Start display at page:

Download "Auditory-Visual Speech Perception Laboratory"

Transcription

1 Auditory-Visual Speech Perception Laboratory Research Focus: Identify perceptual processes involved in auditory-visual speech perception Determine the abilities of individual patients to carry out these processes successfully Design intervention strategies using signal processing technologies and training techniques to remedy any deficiencies that may be found. Primary Funding: NIH Grant: DC Al NSF Grant (subcontract): SBR Learning and Intelligent Systems Initiative of the National Science DARPA Grant (subcontract): ONR Award N Staffing: Lab Director: Ken W. Grant Research Associate: Mary T. Cord

2 Auditory-Visual Speech Perception Laboratory Collaborations: Steven Greenberg - The Speech Institute, Oakland, CA David Poeppel - University of Maryland, College Park, MD Virginie van Wassenhove - University of Maryland, College Park, MD

3 Topics Auditory Supplements to Speechreading Bimodal Comodulation Spectro-Temporal Window of Integration

4 Auditory Supplements to Speechreading

5 Selective Needs of Severe-to to-profound Hearing Loss Speech Input Hearing Capacity compressed dynamic range compressed frequency range Speechreading/lipreading as primary channel of spoken language reception

6 Percent Correct Speech Recognition: Sentences B B J B J B J J B B B B B B J J J J J 20 B Auditory-Visual J Auditory J S/N (db) Roughly 6 db improvement in S/N; roughly 30% improvement in intelligibility for NH subjects

7 Speech Recognition: Consonants Percent Correct Recognition Speech-to-Noise Ratio (db) NH - Auditory Consonants HI Auditory Consonants HI-Auditory Sentences ASR Sentences

8 Auditory-Visual vs. Audio Speech Recognition Percent Correct Recognition Speech-to-Noise Ratio (db) NH - Audiovisual Consonants NH - Auditory Consonants HI - Audiovisual Consonants HI Auditory Consonants HI-Auditory Sentences ASR Sentences Roughly 6 db improvement in S/N; roughly 30% improvement in intelligibility for NH subjects.

9 Possible Roles of Speechreading Provide segmental information that is redundant with acoustic information

10 Possible Roles of Speechreading Provide segmental information that is redundant with acoustic information Provide segmental information that is complementary with acoustic information

11 Possible Roles of Speechreading Provide segmental information that is redundant with acoustic information Provide segmental information that is complementary with acoustic information Direct auditory analyses to the target signal who, where, when, what (spectral)

12 Auditory-Visual Speech Recognition: Consonants What information is available through speechreading? Which acoustic signals supplement speechreading? Are there significant audio-visual interactions in speech processing?

13 Visual Feature Recognition 4% 0% 3% Voicing Manner Place Other From Grant, K.W., and Walden, B.E. (1996). J. Acoust. Soc. Am. 100, % %Information Transmitted re: Total Information Received

14 Speech Recognition: Consonants Linguistic feature contributions to visual speech recognition. The top row represents typical feature classifications for speechreading alone (visemes). Each subsequent row represents the effects of adding information about another linguistic feature via an additional input channel (in this case auditory). Note that as additional features are added, consonant confusions associated with speechreading are resolved to a greater and greater extent. Speechreading Voicing Nasality Affrication p,b,m t,d,n g,k f,v T,D s,z S,tS,dZ,Z l r w j p b,m t d,n g k f v T D s z S,tS dz,z l r w j p b m t d n g k f v T D s z S,tS dz,z l r w j p b m t d n g k f v T D s z S ts dz Z l r w j

15 Hypothetical Consonant Recognition - Perfect Feature Transmission 1 Probability Correct A 0 PRE Voicing Manner Place Voicing + Manner Feature(s) Transmitted Auditory consonant recognition based on perfect transmission of indicated feature. Responses within each feature category were uniformly distributed. Predicted AV consonant recognition based on PRE model of integration (Braida, 1991).

16 Designer Acoustic Signals: Minimal Bandwidth, Maximum Benefit Frank and Joe Hardy scanned the wide valley that appeared before them as their yellow sports sedan rounded the crest of a hill. In the distance, a huge cylindrical tower rose from the valley floor. "Looks like a giant barnacle", Joe remarked to his older, dark-haired brother Frank. Biff Hooper, a tall, muscular high school friend of the two amateur detectives, leaned forward from the back seat. "You're looking at the cooling tower. The reactor itself is in the building next to it." Biff's uncle, Jerry Hooper, was a nuclear engineer at the Bayridge Nuclear Power Plant located outside Bayport. He had invited the three boys on a private afternoon tour of the facility. The summer had just begun, and the Hardy brothers were eager for new adventures. WB LPF AM AMFM

17 Designer Acoustic Signals: Minimal Bandwidth, Maximum Benefit Tracking Rate (WPM) SA AM FM AMFM LPF Receiving Condition Tracking Rate (% re: Normal)

18 Designer Acoustic Signals: AM Bands 100 CUNY Sentences Percent Correct / / SA WB 500 Hz 1600 Hz 3150 Hz Prefilter Condition

19 AM Bands - Smoothing-Filter Effects Percent Correct Percent Correct CUNY Sentences 100 IEEE Sentences Smoothing Filter Cutoff (Hz) Prefilter-Carrier SA

20 Problems With Articulation Theory Effective AI With Visual Cues ANSI, Calculated AI

21 Auditory-Visual Spectral Interactions: Consonants A AV Percent Correct Filter Condition Condition (Hz)

22 Feature Distribution re: Center Frequency Percent Information Transmitted (re: Total Received) Voicing + Manner Place From Grant, K.W., and Walden, B.E. (1996). J. Acoust. Soc. Am. 100, Bandwidth Center Frequency (Hz)

23 How It All Works - The Prevailing View Information extracted from both sources independently Integration of extracted information Decision statistic Evaluation Integration Decision From Massaro, 1998

24 A More Expanded View of the Process

25 Auditory Supplements to Speechreading SUMMARY: Speechreading provides information mostly about place-ofarticulation

26 Auditory Supplements to Speechreading SUMMARY: Speechreading provides information mostly about place-ofarticulation Auditory-visual speech recognition is determined primarily by complementary cues between visual and auditory modalities

27 Auditory Supplements to Speechreading SUMMARY: Speechreading provides information mostly about place-ofarticulation Auditory-visual speech recognition is determined primarily by complementary cues between visual and auditory modalities The most intelligible auditory speech signals do not necessarily result in the most intelligible auditory-visual speech signal

28 Auditory Supplements to Speechreading SUMMARY: Speechreading provides information mostly about place-ofarticulation Auditory-visual speech recognition is determined primarily by complementary cues between visual and auditory modalities The most intelligible auditory speech signals do not necessarily result in the most intelligible auditory-visual speech signal Acoustic cues for voicing and manner-or articulation are the best supplements to speechreading

29 Auditory Supplements to Speechreading SUMMARY: Speechreading provides information mostly about place-ofarticulation Auditory-visual speech recognition is determined primarily by complementary cues between visual and auditory modalities The most intelligible auditory speech signals do not necessarily result in the most intelligible auditory-visual speech signal Acoustic cues for voicing and manner-or articulation are the best supplements to speechreading These cues tend to be low frequency

30 Bimodal Coherence: Audio and Visual Comodulation

31 The Independence of Sensory Systems??? Information is extracted independently from A and V modalities Early versus Late Integration Most models are "late integration" models

32 Back to Our Conceptual Framework

33 The Independence of Sensory Systems??? Information is extracted independently from A and V modalities Early versus Late Integration Most models are "late integration" models BUT Speechreading activates primary auditory cortex (cf. Sams et al., 1991)

34 The Independence of Sensory Systems??? Information is extracted independently from A and V modalities Early versus Late Integration Most models are "late integration" models BUT Speechreading activates primary auditory cortex (cf. Sams et al., 1991) Population of neurons in cat Superior Colliculus respond only to bimodal input (cf. Stein and Meredith, 1993)

35 Sensory Integration: Many Questions How is the auditory system modulated by visual speech activity? What is the temporal window governing this interaction?

36 Bimodal Coherence Masking Protection BCMP (Grant and Seitz, 2000, Grant, 2001) Detection of speech in noise is improved by watching a talker (i.e., speechreading) as they produce the target speech signal, provided that the "visible" movement of the lips and acoustic amplitude envelope are highly correlated.

37 Basic Paradigm for BCMP: Exp. 1 Auditory-only speech detection... speech Auditory-visual speech detection... speech noise noise noise noise

38 Methodology for Orthographic BCMP: Exp. 2 Auditory-only speech detection Auditory + orthographic speech detection... speech text... speech noise noise noise noise

39 Methodology for Filtered BCMP: Exp. 3 F1 ( Hz) F2 ( Hz) Auditory-only detection of filtered speech Auditory-visual detection of filtered speech... speech... speech noise noise noise noise

40 Congruent versus Incongruent Speech Masked Threshold (db) A AV M AV UM -21 Sentence 1 Sentence 2 Sentence 3 Average Target Sentence

41 BCMP for Congruent and Incongruent Speech Bimodal Coherence Protection (db) AV M 2.15 AV UM Sentence 1 Sentence 2 Sentence 3 Average Target Sentence

42 Bimodal Coherence Protection (db) BCMP for Orthographic Speech Sentence 1 Sentence 2 Sentence 3 Average Target Sentence

43 Masking Protection (db) BCMP for Filtered Speech AV F1 AV F2 AV WB AV O 0 Sentence 2 Sentence 3 Average Target Sentence

44 Bimodal Masking Release (db) Average BCMP AV WB AV F2 AV F1 AV O Condition Bimodal comodulation masking protection for wideband speech (WB), filtered speech (F2 and F1), and for orthographically cued speech (O).

45 Acoustic Envelope and Lip Area Functions RMS Amplitude (db) WB F1 F2 F3 Lip Area (in 2 ) Watch the log float in the wide river Time (ms)

46 Cross Modality Correlation - Lip Area versus Amplitude Envelope Sentence 3 Sentence 2 WB r=0.52 r=0.35 RMS Envelope r=0.49 r=0.65 F1 F2 r=0.32 r=0.47 r=0.60 F3 r=0.45 Lip Area Lip Area

47 Local Correlations (Lip Area versus F2-Amplitude Envelope Correlation Coefficient Watch the log float in the wide river Time (ms) RMS Amplitude (db)

48 Local Correlations (Lip Area versus F2-Amplitude Envelope Correlation Coefficient Both brothers wear the same size Time (ms) RMS Amplitude (db)

49 Cross Modality Correlations (lip area versus acoustic envelope) Cross Modality Correlations (lip area versus acoustic envelope) S2 S3 Time (ms) S F F S F F2 Time (ms) Correlation Coefficient Peak Amplitude (db) Correlation Coefficient Peak Amplitude (db) Correlation Coefficient Correlation Coefficient Peak Amplitude (db) Peak Amplitude (db)

50 BCMP Summary Speechreading reduces auditory speech detection thresholds by about 1.5 db (range: 1-3 db depending on sentence)

51 BCMP Summary Speechreading reduces auditory speech detection thresholds by about 1.5 db (range: 1-3 db depending on sentence) Amount of BCMP depends on the degree of coherence between acoustic envelope and facial kinematics

52 BCMP Summary Speechreading reduces auditory speech detection thresholds by about 1.5 db (range: 1-3 db depending on sentence) Amount of BCMP depends on the degree of coherence between acoustic envelope and facial kinematics Providing listeners with explicit (orthographic) knowledge of the identity of the target sentence reduces speech detection thresholds by about 0.5 db, independent of the specific target sentence

53 BCMP Summary Speechreading reduces auditory speech detection thresholds by about 1.5 db (range: 1-3 db depending on sentence) Amount of BCMP depends on the degree of coherence between acoustic envelope and facial kinematics Providing listeners with explicit (orthographic) knowledge of the identity of the target sentence reduces speech detection thresholds by about 0.5 db, independent of the specific target sentence Manipulating the degree of coherence between area of mouth opening and acoustic envelope by filtering the target speech has a direct effect on BCMP

54 Temporal Window for Auditory-Visual Integration

55 AUDIO-ALONE ALONE EXPERIMENTS

56 Audio (Alone) Spectral Slit Paradigm The edge of each slit was separated from its nearest neighbor by an octave Can listeners decode spoken sentences using just four narrow (1/3 octave) channels ( slits ) distributed across the spectrum? YES (cf. next slide) What is the intelligibility of each slit alone and in combination with others?

57 Word Intelligibility - Single and Multiple Slits 89% 60% 13% Slit Number CF (Hz) 2% 9% 9% 4% Slit Number CF (Hz)

58 Slit Asynchrony Affects Intelligibility Desynchronizing the slits by more than 25 ms results in a significant decline in intelligibility The effect of asynchrony on intelligibility is relatively symmetrical These data are from a different set of subjects than those participating in the study described earlier - hence slightly different numbers for the baseline conditions

59 Cross-Spectral Temporal Asynchrony Effects TIMIT Word Recognition (%) Bands Bands Slit Asynchrony (ms) From Greenberg, Arai, and Silipo (1998). Proc. ICSLP, Sydney, Dec. 1-4.

60 AUDIO-VISUAL EXPERIMENTS

61 IEEE Sentences Auditory-Visual Tasks Recognition of key words Audio slits Video presented at various temporal asynchronies CV Syllables Recognition of McGurk pairs Audio /pa/, /ba/, /ta/, /da/ Video /ka/, /ga/, /ta/, /da/ Synchrony identification and discrimination Yes/No single interval simultaneity judgments 2IFC adaptive tracking

62 Auditory-Visual Asynchrony - Paradigm

63 Cross-Modality Temporal Asynchrony Effects: Sentences 100 IEEE Word Recognition (%) Audio Alone - Bands Speechreading Alone Audio Delay (ms)

64 Auditory-Visual Integration - by Individual S's These data are complex, but the implications are clear. Video signal leading is better than synchronous for 8 of 9 subjects Audio-visual integration is a complicated, poorly understood process, at least with respect to speech intelligibility Variation across subjects

65 McGurk Synchrony Paradigm

66 Temporal Integration in the McGurk Effect Response Probability /pa/ /ta/ /ka/ Audio Delay (ms)

67 Simultaneity Judgements - Natural vs. McGurk AV Tokens 1 A d V d Probability of Simultaneity A t V t A p V k A b V g Audio Delay (ms)

68 Spectro-Temporal Synchrony Discrimination Filter Band Condition (Hz) WB Audio Lead Audio Lag Audio Delay (ms)

69 Temporal Window of Integration McGurk Synchrony Discrimination McGurk Synchrony Identification Audio Audiovisual McGurk Fusion Identification CV Synchrony Discrimination CV Synchrony Identification IEEE Synchrony Discrimination IEEE Word Identification IEEE Synchrony Discrimination TIMIT Word Identification Audio Delay (ms)

70 Spectro-Temporal Integration: Summary Within Modality (Cross- Spectral Auditory Integration)

71 Spectro-Temporal Integration: Summary Within Modality (Cross- Spectral Auditory Integration) TWI is symmetrical

72 Spectro-Temporal Integration: Summary Within Modality (Cross- Spectral Auditory Integration) TWI is symmetrical TWI roughly 50 ms or less (phoneme?)

73 Spectro-Temporal Integration: Summary Within Modality (Cross- Spectral Auditory Integration) TWI is symmetrical TWI roughly 50 ms or less (phoneme?) Across Modality (Cross-Modal AV Integration)

74 Spectro-Temporal Integration: Summary Within Modality (Cross- Spectral Auditory Integration) TWI is symmetrical TWI roughly 50 ms or less (phoneme?) Across Modality (Cross-Modal AV Integration) TWI is highly asymmetrical favoring visual leads

75 Spectro-Temporal Integration: Summary Within Modality (Cross- Spectral Auditory Integration) TWI is symmetrical TWI roughly 50 ms or less (phoneme?) Across Modality (Cross-Modal AV Integration) TWI is highly asymmetrical favoring visual leads TWI is roughly ms (syllable?)

76 Spectro-Temporal Integration: Summary Within Modality (Cross- Spectral Auditory Integration) TWI is symmetrical TWI roughly 50 ms or less (phoneme?) Across Modality (Cross-Modal AV Integration) TWI is highly asymmetrical favoring visual leads TWI is roughly ms (syllable?) TWI for Incongruent CV's (McGurk Stimuli) is not as wide as TWI for natural congruent CV's

77 Auditory-Visual Speech Perception Laboratory Walter Reed Army Medical Center Army Audiology and Speech Center Washington, DC USA wramc.amedd.army.mil/departments/.army.mil/departments/aasc/avlabavlab

Detection of auditory (cross-spectral) and auditory visual (cross-modal) synchrony

Detection of auditory (cross-spectral) and auditory visual (cross-modal) synchrony Speech Communication 44 (2004) 43 53 www.elsevier.com/locate/specom Detection of auditory (cross-spectral) and auditory visual (cross-modal) synchrony Ken W. Grant a, *, Virginie van Wassenhove b, David

More information

HCS 7367 Speech Perception

HCS 7367 Speech Perception Long-term spectrum of speech HCS 7367 Speech Perception Connected speech Absolute threshold Males Dr. Peter Assmann Fall 212 Females Long-term spectrum of speech Vowels Males Females 2) Absolute threshold

More information

HCS 7367 Speech Perception

HCS 7367 Speech Perception Babies 'cry in mother's tongue' HCS 7367 Speech Perception Dr. Peter Assmann Fall 212 Babies' cries imitate their mother tongue as early as three days old German researchers say babies begin to pick up

More information

MULTI-CHANNEL COMMUNICATION

MULTI-CHANNEL COMMUNICATION INTRODUCTION Research on the Deaf Brain is beginning to provide a new evidence base for policy and practice in relation to intervention with deaf children. This talk outlines the multi-channel nature of

More information

THE ROLE OF VISUAL SPEECH CUES IN THE AUDITORY PERCEPTION OF SYNTHETIC STIMULI BY CHILDREN USING A COCHLEAR IMPLANT AND CHILDREN WITH NORMAL HEARING

THE ROLE OF VISUAL SPEECH CUES IN THE AUDITORY PERCEPTION OF SYNTHETIC STIMULI BY CHILDREN USING A COCHLEAR IMPLANT AND CHILDREN WITH NORMAL HEARING THE ROLE OF VISUAL SPEECH CUES IN THE AUDITORY PERCEPTION OF SYNTHETIC STIMULI BY CHILDREN USING A COCHLEAR IMPLANT AND CHILDREN WITH NORMAL HEARING Vanessa Surowiecki 1, vid Grayden 1, Richard Dowell

More information

Providing Effective Communication Access

Providing Effective Communication Access Providing Effective Communication Access 2 nd International Hearing Loop Conference June 19 th, 2011 Matthew H. Bakke, Ph.D., CCC A Gallaudet University Outline of the Presentation Factors Affecting Communication

More information

Language Speech. Speech is the preferred modality for language.

Language Speech. Speech is the preferred modality for language. Language Speech Speech is the preferred modality for language. Outer ear Collects sound waves. The configuration of the outer ear serves to amplify sound, particularly at 2000-5000 Hz, a frequency range

More information

The Role of Information Redundancy in Audiovisual Speech Integration. A Senior Honors Thesis

The Role of Information Redundancy in Audiovisual Speech Integration. A Senior Honors Thesis The Role of Information Redundancy in Audiovisual Speech Integration A Senior Honors Thesis Printed in Partial Fulfillment of the Requirement for graduation with distinction in Speech and Hearing Science

More information

Speech Reading Training and Audio-Visual Integration in a Child with Autism Spectrum Disorder

Speech Reading Training and Audio-Visual Integration in a Child with Autism Spectrum Disorder University of Arkansas, Fayetteville ScholarWorks@UARK Rehabilitation, Human Resources and Communication Disorders Undergraduate Honors Theses Rehabilitation, Human Resources and Communication Disorders

More information

ELECTROPHYSIOLOGY OF UNIMODAL AND AUDIOVISUAL SPEECH PERCEPTION

ELECTROPHYSIOLOGY OF UNIMODAL AND AUDIOVISUAL SPEECH PERCEPTION AVSP 2 International Conference on Auditory-Visual Speech Processing ELECTROPHYSIOLOGY OF UNIMODAL AND AUDIOVISUAL SPEECH PERCEPTION Lynne E. Bernstein, Curtis W. Ponton 2, Edward T. Auer, Jr. House Ear

More information

Auditory-Visual Integration of Sine-Wave Speech. A Senior Honors Thesis

Auditory-Visual Integration of Sine-Wave Speech. A Senior Honors Thesis Auditory-Visual Integration of Sine-Wave Speech A Senior Honors Thesis Presented in Partial Fulfillment of the Requirements for Graduation with Distinction in Speech and Hearing Science in the Undergraduate

More information

Audio-Visual Integration: Generalization Across Talkers. A Senior Honors Thesis

Audio-Visual Integration: Generalization Across Talkers. A Senior Honors Thesis Audio-Visual Integration: Generalization Across Talkers A Senior Honors Thesis Presented in Partial Fulfillment of the Requirements for graduation with research distinction in Speech and Hearing Science

More information

ILLUSIONS AND ISSUES IN BIMODAL SPEECH PERCEPTION

ILLUSIONS AND ISSUES IN BIMODAL SPEECH PERCEPTION ISCA Archive ILLUSIONS AND ISSUES IN BIMODAL SPEECH PERCEPTION Dominic W. Massaro Perceptual Science Laboratory (http://mambo.ucsc.edu/psl/pslfan.html) University of California Santa Cruz, CA 95064 massaro@fuzzy.ucsc.edu

More information

Tactile Communication of Speech

Tactile Communication of Speech Tactile Communication of Speech RLE Group Sensory Communication Group Sponsor National Institutes of Health/National Institute on Deafness and Other Communication Disorders Grant 2 R01 DC00126, Grant 1

More information

Robust Neural Encoding of Speech in Human Auditory Cortex

Robust Neural Encoding of Speech in Human Auditory Cortex Robust Neural Encoding of Speech in Human Auditory Cortex Nai Ding, Jonathan Z. Simon Electrical Engineering / Biology University of Maryland, College Park Auditory Processing in Natural Scenes How is

More information

Gick et al.: JASA Express Letters DOI: / Published Online 17 March 2008

Gick et al.: JASA Express Letters DOI: / Published Online 17 March 2008 modality when that information is coupled with information via another modality (e.g., McGrath and Summerfield, 1985). It is unknown, however, whether there exist complex relationships across modalities,

More information

FREQUENCY COMPRESSION AND FREQUENCY SHIFTING FOR THE HEARING IMPAIRED

FREQUENCY COMPRESSION AND FREQUENCY SHIFTING FOR THE HEARING IMPAIRED FREQUENCY COMPRESSION AND FREQUENCY SHIFTING FOR THE HEARING IMPAIRED Francisco J. Fraga, Alan M. Marotta National Institute of Telecommunications, Santa Rita do Sapucaí - MG, Brazil Abstract A considerable

More information

Hearing the Universal Language: Music and Cochlear Implants

Hearing the Universal Language: Music and Cochlear Implants Hearing the Universal Language: Music and Cochlear Implants Professor Hugh McDermott Deputy Director (Research) The Bionics Institute of Australia, Professorial Fellow The University of Melbourne Overview?

More information

A Senior Honors Thesis. Brandie Andrews

A Senior Honors Thesis. Brandie Andrews Auditory and Visual Information Facilitating Speech Integration A Senior Honors Thesis Presented in Partial Fulfillment of the Requirements for graduation with distinction in Speech and Hearing Science

More information

Consonant Perception test

Consonant Perception test Consonant Perception test Introduction The Vowel-Consonant-Vowel (VCV) test is used in clinics to evaluate how well a listener can recognize consonants under different conditions (e.g. with and without

More information

Study of perceptual balance for binaural dichotic presentation

Study of perceptual balance for binaural dichotic presentation Paper No. 556 Proceedings of 20 th International Congress on Acoustics, ICA 2010 23-27 August 2010, Sydney, Australia Study of perceptual balance for binaural dichotic presentation Pandurangarao N. Kulkarni

More information

Speech perception in individuals with dementia of the Alzheimer s type (DAT) Mitchell S. Sommers Department of Psychology Washington University

Speech perception in individuals with dementia of the Alzheimer s type (DAT) Mitchell S. Sommers Department of Psychology Washington University Speech perception in individuals with dementia of the Alzheimer s type (DAT) Mitchell S. Sommers Department of Psychology Washington University Overview Goals of studying speech perception in individuals

More information

Optimal Filter Perception of Speech Sounds: Implications to Hearing Aid Fitting through Verbotonal Rehabilitation

Optimal Filter Perception of Speech Sounds: Implications to Hearing Aid Fitting through Verbotonal Rehabilitation Optimal Filter Perception of Speech Sounds: Implications to Hearing Aid Fitting through Verbotonal Rehabilitation Kazunari J. Koike, Ph.D., CCC-A Professor & Director of Audiology Department of Otolaryngology

More information

Best Practice Protocols

Best Practice Protocols Best Practice Protocols SoundRecover for children What is SoundRecover? SoundRecover (non-linear frequency compression) seeks to give greater audibility of high-frequency everyday sounds by compressing

More information

Analysis of Emotion Recognition using Facial Expressions, Speech and Multimodal Information

Analysis of Emotion Recognition using Facial Expressions, Speech and Multimodal Information Analysis of Emotion Recognition using Facial Expressions, Speech and Multimodal Information C. Busso, Z. Deng, S. Yildirim, M. Bulut, C. M. Lee, A. Kazemzadeh, S. Lee, U. Neumann, S. Narayanan Emotion

More information

Time Varying Comb Filters to Reduce Spectral and Temporal Masking in Sensorineural Hearing Impairment

Time Varying Comb Filters to Reduce Spectral and Temporal Masking in Sensorineural Hearing Impairment Bio Vision 2001 Intl Conf. Biomed. Engg., Bangalore, India, 21-24 Dec. 2001, paper PRN6. Time Varying Comb Filters to Reduce pectral and Temporal Masking in ensorineural Hearing Impairment Dakshayani.

More information

Modulation and Top-Down Processing in Audition

Modulation and Top-Down Processing in Audition Modulation and Top-Down Processing in Audition Malcolm Slaney 1,2 and Greg Sell 2 1 Yahoo! Research 2 Stanford CCRMA Outline The Non-Linear Cochlea Correlogram Pitch Modulation and Demodulation Information

More information

Spectrograms (revisited)

Spectrograms (revisited) Spectrograms (revisited) We begin the lecture by reviewing the units of spectrograms, which I had only glossed over when I covered spectrograms at the end of lecture 19. We then relate the blocks of a

More information

Acoustics, signals & systems for audiology. Psychoacoustics of hearing impairment

Acoustics, signals & systems for audiology. Psychoacoustics of hearing impairment Acoustics, signals & systems for audiology Psychoacoustics of hearing impairment Three main types of hearing impairment Conductive Sound is not properly transmitted from the outer to the inner ear Sensorineural

More information

BINAURAL DICHOTIC PRESENTATION FOR MODERATE BILATERAL SENSORINEURAL HEARING-IMPAIRED

BINAURAL DICHOTIC PRESENTATION FOR MODERATE BILATERAL SENSORINEURAL HEARING-IMPAIRED International Conference on Systemics, Cybernetics and Informatics, February 12 15, 2004 BINAURAL DICHOTIC PRESENTATION FOR MODERATE BILATERAL SENSORINEURAL HEARING-IMPAIRED Alice N. Cheeran Biomedical

More information

USING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES

USING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES USING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES Varinthira Duangudom and David V Anderson School of Electrical and Computer Engineering, Georgia Institute of Technology Atlanta, GA 30332

More information

The Deaf Brain. Bencie Woll Deafness Cognition and Language Research Centre

The Deaf Brain. Bencie Woll Deafness Cognition and Language Research Centre The Deaf Brain Bencie Woll Deafness Cognition and Language Research Centre 1 Outline Introduction: the multi-channel nature of language Audio-visual language BSL Speech processing Silent speech Auditory

More information

AUDL GS08/GAV1 Signals, systems, acoustics and the ear. Pitch & Binaural listening

AUDL GS08/GAV1 Signals, systems, acoustics and the ear. Pitch & Binaural listening AUDL GS08/GAV1 Signals, systems, acoustics and the ear Pitch & Binaural listening Review 25 20 15 10 5 0-5 100 1000 10000 25 20 15 10 5 0-5 100 1000 10000 Part I: Auditory frequency selectivity Tuning

More information

Asynchronous glimpsing of speech: Spread of masking and task set-size

Asynchronous glimpsing of speech: Spread of masking and task set-size Asynchronous glimpsing of speech: Spread of masking and task set-size Erol J. Ozmeral, a) Emily Buss, and Joseph W. Hall III Department of Otolaryngology/Head and Neck Surgery, University of North Carolina

More information

ACOUSTIC AND PERCEPTUAL PROPERTIES OF ENGLISH FRICATIVES

ACOUSTIC AND PERCEPTUAL PROPERTIES OF ENGLISH FRICATIVES ISCA Archive ACOUSTIC AND PERCEPTUAL PROPERTIES OF ENGLISH FRICATIVES Allard Jongman 1, Yue Wang 2, and Joan Sereno 1 1 Linguistics Department, University of Kansas, Lawrence, KS 66045 U.S.A. 2 Department

More information

Assessing Hearing and Speech Recognition

Assessing Hearing and Speech Recognition Assessing Hearing and Speech Recognition Audiological Rehabilitation Quick Review Audiogram Types of hearing loss hearing loss hearing loss Testing Air conduction Bone conduction Familiar Sounds Audiogram

More information

2/25/2013. Context Effect on Suprasegmental Cues. Supresegmental Cues. Pitch Contour Identification (PCI) Context Effect with Cochlear Implants

2/25/2013. Context Effect on Suprasegmental Cues. Supresegmental Cues. Pitch Contour Identification (PCI) Context Effect with Cochlear Implants Context Effect on Segmental and Supresegmental Cues Preceding context has been found to affect phoneme recognition Stop consonant recognition (Mann, 1980) A continuum from /da/ to /ga/ was preceded by

More information

Role of F0 differences in source segregation

Role of F0 differences in source segregation Role of F0 differences in source segregation Andrew J. Oxenham Research Laboratory of Electronics, MIT and Harvard-MIT Speech and Hearing Bioscience and Technology Program Rationale Many aspects of segregation

More information

Juan Carlos Tejero-Calado 1, Janet C. Rutledge 2, and Peggy B. Nelson 3

Juan Carlos Tejero-Calado 1, Janet C. Rutledge 2, and Peggy B. Nelson 3 PRESERVING SPECTRAL CONTRAST IN AMPLITUDE COMPRESSION FOR HEARING AIDS Juan Carlos Tejero-Calado 1, Janet C. Rutledge 2, and Peggy B. Nelson 3 1 University of Malaga, Campus de Teatinos-Complejo Tecnol

More information

The effect of wearing conventional and level-dependent hearing protectors on speech production in noise and quiet

The effect of wearing conventional and level-dependent hearing protectors on speech production in noise and quiet The effect of wearing conventional and level-dependent hearing protectors on speech production in noise and quiet Ghazaleh Vaziri Christian Giguère Hilmi R. Dajani Nicolas Ellaham Annual National Hearing

More information

Noise-Robust Speech Recognition in a Car Environment Based on the Acoustic Features of Car Interior Noise

Noise-Robust Speech Recognition in a Car Environment Based on the Acoustic Features of Car Interior Noise 4 Special Issue Speech-Based Interfaces in Vehicles Research Report Noise-Robust Speech Recognition in a Car Environment Based on the Acoustic Features of Car Interior Noise Hiroyuki Hoshino Abstract This

More information

Hearing. and other senses

Hearing. and other senses Hearing and other senses Sound Sound: sensed variations in air pressure Frequency: number of peaks that pass a point per second (Hz) Pitch 2 Some Sound and Hearing Links Useful (and moderately entertaining)

More information

Perceptual congruency of audio-visual speech affects ventriloquism with bilateral visual stimuli

Perceptual congruency of audio-visual speech affects ventriloquism with bilateral visual stimuli Psychon Bull Rev (2011) 18:123 128 DOI 10.3758/s13423-010-0027-z Perceptual congruency of audio-visual speech affects ventriloquism with bilateral visual stimuli Shoko Kanaya & Kazuhiko Yokosawa Published

More information

Lecture 3: Perception

Lecture 3: Perception ELEN E4896 MUSIC SIGNAL PROCESSING Lecture 3: Perception 1. Ear Physiology 2. Auditory Psychophysics 3. Pitch Perception 4. Music Perception Dan Ellis Dept. Electrical Engineering, Columbia University

More information

INTRODUCTION J. Acoust. Soc. Am. 103 (2), February /98/103(2)/1080/5/$ Acoustical Society of America 1080

INTRODUCTION J. Acoust. Soc. Am. 103 (2), February /98/103(2)/1080/5/$ Acoustical Society of America 1080 Perceptual segregation of a harmonic from a vowel by interaural time difference in conjunction with mistuning and onset asynchrony C. J. Darwin and R. W. Hukin Experimental Psychology, University of Sussex,

More information

S everal studies indicate that the identification/recognition. Identification Performance by Right- and Lefthanded Listeners on Dichotic CV Materials

S everal studies indicate that the identification/recognition. Identification Performance by Right- and Lefthanded Listeners on Dichotic CV Materials J Am Acad Audiol 7 : 1-6 (1996) Identification Performance by Right- and Lefthanded Listeners on Dichotic CV Materials Richard H. Wilson* Elizabeth D. Leigh* Abstract Normative data from 24 right-handed

More information

Chapter 40 Effects of Peripheral Tuning on the Auditory Nerve s Representation of Speech Envelope and Temporal Fine Structure Cues

Chapter 40 Effects of Peripheral Tuning on the Auditory Nerve s Representation of Speech Envelope and Temporal Fine Structure Cues Chapter 40 Effects of Peripheral Tuning on the Auditory Nerve s Representation of Speech Envelope and Temporal Fine Structure Cues Rasha A. Ibrahim and Ian C. Bruce Abstract A number of studies have explored

More information

Slow compression for people with severe to profound hearing loss

Slow compression for people with severe to profound hearing loss Phonak Insight February 2018 Slow compression for people with severe to profound hearing loss For people with severe to profound hearing loss, poor auditory resolution abilities can make the spectral and

More information

Prelude Envelope and temporal fine. What's all the fuss? Modulating a wave. Decomposing waveforms. The psychophysics of cochlear

Prelude Envelope and temporal fine. What's all the fuss? Modulating a wave. Decomposing waveforms. The psychophysics of cochlear The psychophysics of cochlear implants Stuart Rosen Professor of Speech and Hearing Science Speech, Hearing and Phonetic Sciences Division of Psychology & Language Sciences Prelude Envelope and temporal

More information

MODALITY, PERCEPTUAL ENCODING SPEED, AND TIME-COURSE OF PHONETIC INFORMATION

MODALITY, PERCEPTUAL ENCODING SPEED, AND TIME-COURSE OF PHONETIC INFORMATION ISCA Archive MODALITY, PERCEPTUAL ENCODING SPEED, AND TIME-COURSE OF PHONETIC INFORMATION Philip Franz Seitz and Ken W. Grant Army Audiology and Speech Center Walter Reed Army Medical Center Washington,

More information

Cross-Frequency Integration for Consonant and Vowel Identification in Bimodal Hearing

Cross-Frequency Integration for Consonant and Vowel Identification in Bimodal Hearing Cross-Frequency Integration for Consonant and Vowel Identification in Bimodal Hearing The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters.

More information

Enrique A. Lopez-Poveda Alan R. Palmer Ray Meddis Editors. The Neurophysiological Bases of Auditory Perception

Enrique A. Lopez-Poveda Alan R. Palmer Ray Meddis Editors. The Neurophysiological Bases of Auditory Perception Enrique A. Lopez-Poveda Alan R. Palmer Ray Meddis Editors The Neurophysiological Bases of Auditory Perception 123 The Neurophysiological Bases of Auditory Perception Enrique A. Lopez-Poveda Alan R. Palmer

More information

Outline.! Neural representation of speech sounds. " Basic intro " Sounds and categories " How do we perceive sounds? " Is speech sounds special?

Outline.! Neural representation of speech sounds.  Basic intro  Sounds and categories  How do we perceive sounds?  Is speech sounds special? Outline! Neural representation of speech sounds " Basic intro " Sounds and categories " How do we perceive sounds? " Is speech sounds special? ! What is a phoneme?! It s the basic linguistic unit of speech!

More information

EEL 6586, Project - Hearing Aids algorithms

EEL 6586, Project - Hearing Aids algorithms EEL 6586, Project - Hearing Aids algorithms 1 Yan Yang, Jiang Lu, and Ming Xue I. PROBLEM STATEMENT We studied hearing loss algorithms in this project. As the conductive hearing loss is due to sound conducting

More information

Hearing Lectures. Acoustics of Speech and Hearing. Auditory Lighthouse. Facts about Timbre. Analysis of Complex Sounds

Hearing Lectures. Acoustics of Speech and Hearing. Auditory Lighthouse. Facts about Timbre. Analysis of Complex Sounds Hearing Lectures Acoustics of Speech and Hearing Week 2-10 Hearing 3: Auditory Filtering 1. Loudness of sinusoids mainly (see Web tutorial for more) 2. Pitch of sinusoids mainly (see Web tutorial for more)

More information

Perceptual Effects of Nasal Cue Modification

Perceptual Effects of Nasal Cue Modification Send Orders for Reprints to reprints@benthamscience.ae The Open Electrical & Electronic Engineering Journal, 2015, 9, 399-407 399 Perceptual Effects of Nasal Cue Modification Open Access Fan Bai 1,2,*

More information

Infant Hearing Development: Translating Research Findings into Clinical Practice. Auditory Development. Overview

Infant Hearing Development: Translating Research Findings into Clinical Practice. Auditory Development. Overview Infant Hearing Development: Translating Research Findings into Clinical Practice Lori J. Leibold Department of Allied Health Sciences The University of North Carolina at Chapel Hill Auditory Development

More information

! Can hear whistle? ! Where are we on course map? ! What we did in lab last week. ! Psychoacoustics

! Can hear whistle? ! Where are we on course map? ! What we did in lab last week. ! Psychoacoustics 2/14/18 Can hear whistle? Lecture 5 Psychoacoustics Based on slides 2009--2018 DeHon, Koditschek Additional Material 2014 Farmer 1 2 There are sounds we cannot hear Depends on frequency Where are we on

More information

Speech (Sound) Processing

Speech (Sound) Processing 7 Speech (Sound) Processing Acoustic Human communication is achieved when thought is transformed through language into speech. The sounds of speech are initiated by activity in the central nervous system,

More information

Speech Spectra and Spectrograms

Speech Spectra and Spectrograms ACOUSTICS TOPICS ACOUSTICS SOFTWARE SPH301 SLP801 RESOURCE INDEX HELP PAGES Back to Main "Speech Spectra and Spectrograms" Page Speech Spectra and Spectrograms Robert Mannell 6. Some consonant spectra

More information

Tactual Display of Consonant Voicing to Supplement Lipreading

Tactual Display of Consonant Voicing to Supplement Lipreading Tactual Display of Consonant Voicing to Supplement Lipreading by Hanfeng Yuan S.B., Shanghai Jiao Tong University (1995) S.M. Massachusetts Institute of Technology (1999) Submitted to the Department of

More information

The role of low frequency components in median plane localization

The role of low frequency components in median plane localization Acoust. Sci. & Tech. 24, 2 (23) PAPER The role of low components in median plane localization Masayuki Morimoto 1;, Motoki Yairi 1, Kazuhiro Iida 2 and Motokuni Itoh 1 1 Environmental Acoustics Laboratory,

More information

CHAPTER 1 INTRODUCTION

CHAPTER 1 INTRODUCTION CHAPTER 1 INTRODUCTION 1.1 BACKGROUND Speech is the most natural form of human communication. Speech has also become an important means of human-machine interaction and the advancement in technology has

More information

Applying the summation model in audiovisual speech perception

Applying the summation model in audiovisual speech perception Applying the summation model in audiovisual speech perception Kaisa Tiippana, Ilmari Kurki, Tarja Peromaa Department of Psychology and Logopedics, Faculty of Medicine, University of Helsinki, Finland kaisa.tiippana@helsinki.fi,

More information

Sound localization psychophysics

Sound localization psychophysics Sound localization psychophysics Eric Young A good reference: B.C.J. Moore An Introduction to the Psychology of Hearing Chapter 7, Space Perception. Elsevier, Amsterdam, pp. 233-267 (2004). Sound localization:

More information

RESEARCH ON SPOKEN LANGUAGE PROCESSING Progress Report No. 27 (2005) Indiana University

RESEARCH ON SPOKEN LANGUAGE PROCESSING Progress Report No. 27 (2005) Indiana University AUDIOVISUAL ASYNCHRONY DETECTION RESEARCH ON SPOKEN LANGUAGE PROCESSING Progress Report No. 27 (2005) Indiana University Audiovisual Asynchrony Detection and Speech Perception in Normal-Hearing Listeners

More information

Topic 4. Pitch & Frequency

Topic 4. Pitch & Frequency Topic 4 Pitch & Frequency A musical interlude KOMBU This solo by Kaigal-ool of Huun-Huur-Tu (accompanying himself on doshpuluur) demonstrates perfectly the characteristic sound of the Xorekteer voice An

More information

Research Article The Acoustic and Peceptual Effects of Series and Parallel Processing

Research Article The Acoustic and Peceptual Effects of Series and Parallel Processing Hindawi Publishing Corporation EURASIP Journal on Advances in Signal Processing Volume 9, Article ID 6195, pages doi:1.1155/9/6195 Research Article The Acoustic and Peceptual Effects of Series and Parallel

More information

RESEARCH ON SPOKEN LANGUAGE PROCESSING Progress Report No. 24 (2000) Indiana University

RESEARCH ON SPOKEN LANGUAGE PROCESSING Progress Report No. 24 (2000) Indiana University COMPARISON OF PARTIAL INFORMATION RESEARCH ON SPOKEN LANGUAGE PROCESSING Progress Report No. 24 (2000) Indiana University Use of Partial Stimulus Information by Cochlear Implant Patients and Normal-Hearing

More information

Computational Perception /785. Auditory Scene Analysis

Computational Perception /785. Auditory Scene Analysis Computational Perception 15-485/785 Auditory Scene Analysis A framework for auditory scene analysis Auditory scene analysis involves low and high level cues Low level acoustic cues are often result in

More information

Perceived Audiovisual Simultaneity in Speech by Musicians and Non-musicians: Preliminary Behavioral and Event-Related Potential (ERP) Findings

Perceived Audiovisual Simultaneity in Speech by Musicians and Non-musicians: Preliminary Behavioral and Event-Related Potential (ERP) Findings The 14th International Conference on Auditory-Visual Speech Processing 25-26 August 2017, Stockholm, Sweden Perceived Audiovisual Simultaneity in Speech by Musicians and Non-musicians: Preliminary Behavioral

More information

What you re in for. Who are cochlear implants for? The bottom line. Speech processing schemes for

What you re in for. Who are cochlear implants for? The bottom line. Speech processing schemes for What you re in for Speech processing schemes for cochlear implants Stuart Rosen Professor of Speech and Hearing Science Speech, Hearing and Phonetic Sciences Division of Psychology & Language Sciences

More information

Evaluating the role of spectral and envelope characteristics in the intelligibility advantage of clear speech

Evaluating the role of spectral and envelope characteristics in the intelligibility advantage of clear speech Evaluating the role of spectral and envelope characteristics in the intelligibility advantage of clear speech Jean C. Krause a and Louis D. Braida Research Laboratory of Electronics, Massachusetts Institute

More information

Visual impairment and speech understanding in older adults with acquired hearing loss

Visual impairment and speech understanding in older adults with acquired hearing loss Visual impairment and speech understanding in older adults with acquired hearing loss Jean-Pierre Gagné École d orthophonie et d audiologie, Université de Montréal, Montréal, Québec, Canada H3C3J7 jean-pierre.gagne@umontreal.ca

More information

Twenty subjects (11 females) participated in this study. None of the subjects had

Twenty subjects (11 females) participated in this study. None of the subjects had SUPPLEMENTARY METHODS Subjects Twenty subjects (11 females) participated in this study. None of the subjects had previous exposure to a tone language. Subjects were divided into two groups based on musical

More information

An Update on Auditory Neuropathy Spectrum Disorder in Children

An Update on Auditory Neuropathy Spectrum Disorder in Children An Update on Auditory Neuropathy Spectrum Disorder in Children Gary Rance PhD The University of Melbourne Sound Foundations Through Early Amplification Meeting, Chicago, Dec 2013 Overview Auditory neuropathy

More information

Simulations of high-frequency vocoder on Mandarin speech recognition for acoustic hearing preserved cochlear implant

Simulations of high-frequency vocoder on Mandarin speech recognition for acoustic hearing preserved cochlear implant INTERSPEECH 2017 August 20 24, 2017, Stockholm, Sweden Simulations of high-frequency vocoder on Mandarin speech recognition for acoustic hearing preserved cochlear implant Tsung-Chen Wu 1, Tai-Shih Chi

More information

WIDEXPRESS. no.30. Background

WIDEXPRESS. no.30. Background WIDEXPRESS no. january 12 By Marie Sonne Kristensen Petri Korhonen Using the WidexLink technology to improve speech perception Background For most hearing aid users, the primary motivation for using hearing

More information

The role of periodicity in the perception of masked speech with simulated and real cochlear implants

The role of periodicity in the perception of masked speech with simulated and real cochlear implants The role of periodicity in the perception of masked speech with simulated and real cochlear implants Kurt Steinmetzger and Stuart Rosen UCL Speech, Hearing and Phonetic Sciences Heidelberg, 09. November

More information

Grozdana Erjavec 1, Denis Legros 1. Department of Cognition, Language and Interaction, University of Paris VIII, France. Abstract. 1.

Grozdana Erjavec 1, Denis Legros 1. Department of Cognition, Language and Interaction, University of Paris VIII, France. Abstract. 1. INTERSPEECH 2013 Effects of Mouth-Only and Whole-Face Displays on Audio-Visual Speech Perception in Noise: Is the Vision of a Talker s Full Face Truly the Most Efficient Solution? Grozdana Erjavec 1, Denis

More information

CONSTRUCTING TELEPHONE ACOUSTIC MODELS FROM A HIGH-QUALITY SPEECH CORPUS

CONSTRUCTING TELEPHONE ACOUSTIC MODELS FROM A HIGH-QUALITY SPEECH CORPUS CONSTRUCTING TELEPHONE ACOUSTIC MODELS FROM A HIGH-QUALITY SPEECH CORPUS Mitchel Weintraub and Leonardo Neumeyer SRI International Speech Research and Technology Program Menlo Park, CA, 94025 USA ABSTRACT

More information

Fitting Frequency Compression Hearing Aids to Kids: The Basics

Fitting Frequency Compression Hearing Aids to Kids: The Basics Fitting Frequency Compression Hearing Aids to Kids: The Basics Presenters: Susan Scollie and Danielle Glista Presented at AudiologyNow! 2011, Chicago Support This work was supported by: Canadian Institutes

More information

Who are cochlear implants for?

Who are cochlear implants for? Who are cochlear implants for? People with little or no hearing and little conductive component to the loss who receive little or no benefit from a hearing aid. Implants seem to work best in adults who

More information

Variability in Word Recognition by Adults with Cochlear Implants: The Role of Language Knowledge

Variability in Word Recognition by Adults with Cochlear Implants: The Role of Language Knowledge Variability in Word Recognition by Adults with Cochlear Implants: The Role of Language Knowledge Aaron C. Moberly, M.D. CI2015 Washington, D.C. Disclosures ASA and ASHFoundation Speech Science Research

More information

Categorical Perception

Categorical Perception Categorical Perception Discrimination for some speech contrasts is poor within phonetic categories and good between categories. Unusual, not found for most perceptual contrasts. Influenced by task, expectations,

More information

Speech conveys not only linguistic content but. Vocal Emotion Recognition by Normal-Hearing Listeners and Cochlear Implant Users

Speech conveys not only linguistic content but. Vocal Emotion Recognition by Normal-Hearing Listeners and Cochlear Implant Users Cochlear Implants Special Issue Article Vocal Emotion Recognition by Normal-Hearing Listeners and Cochlear Implant Users Trends in Amplification Volume 11 Number 4 December 2007 301-315 2007 Sage Publications

More information

The development of a modified spectral ripple test

The development of a modified spectral ripple test The development of a modified spectral ripple test Justin M. Aronoff a) and David M. Landsberger Communication and Neuroscience Division, House Research Institute, 2100 West 3rd Street, Los Angeles, California

More information

Effects of slow- and fast-acting compression on hearing impaired listeners consonantvowel identification in interrupted noise

Effects of slow- and fast-acting compression on hearing impaired listeners consonantvowel identification in interrupted noise Downloaded from orbit.dtu.dk on: Jan 01, 2018 Effects of slow- and fast-acting compression on hearing impaired listeners consonantvowel identification in interrupted noise Kowalewski, Borys; Zaar, Johannes;

More information

Effects of noise and filtering on the intelligibility of speech produced during simultaneous communication

Effects of noise and filtering on the intelligibility of speech produced during simultaneous communication Journal of Communication Disorders 37 (2004) 505 515 Effects of noise and filtering on the intelligibility of speech produced during simultaneous communication Douglas J. MacKenzie a,*, Nicholas Schiavetti

More information

Diagnosis and Management of ANSD: Outcomes of Cochlear Implants versus Hearing Aids

Diagnosis and Management of ANSD: Outcomes of Cochlear Implants versus Hearing Aids Diagnosis and Management of ANSD: Outcomes of Cochlear Implants versus Hearing Aids Gary Rance PhD The University of Melbourne International Paediatric Conference, Shanghai, April 214 Auditory Neuropathy

More information

Although considerable work has been conducted on the speech

Although considerable work has been conducted on the speech Influence of Hearing Loss on the Perceptual Strategies of Children and Adults Andrea L. Pittman Patricia G. Stelmachowicz Dawna E. Lewis Brenda M. Hoover Boys Town National Research Hospital Omaha, NE

More information

Audio-Visual Integration in Multimodal Communication

Audio-Visual Integration in Multimodal Communication Audio-Visual Integration in Multimodal Communication TSUHAN CHEN, MEMBER, IEEE, AND RAM R. RAO Invited Paper In this paper, we review recent research that examines audiovisual integration in multimodal

More information

The functional importance of age-related differences in temporal processing

The functional importance of age-related differences in temporal processing Kathy Pichora-Fuller The functional importance of age-related differences in temporal processing Professor, Psychology, University of Toronto Adjunct Scientist, Toronto Rehabilitation Institute, University

More information

Comment by Delgutte and Anna. A. Dreyer (Eaton-Peabody Laboratory, Massachusetts Eye and Ear Infirmary, Boston, MA)

Comment by Delgutte and Anna. A. Dreyer (Eaton-Peabody Laboratory, Massachusetts Eye and Ear Infirmary, Boston, MA) Comments Comment by Delgutte and Anna. A. Dreyer (Eaton-Peabody Laboratory, Massachusetts Eye and Ear Infirmary, Boston, MA) Is phase locking to transposed stimuli as good as phase locking to low-frequency

More information

Recognition of Multiply Degraded Speech by Young and Elderly Listeners

Recognition of Multiply Degraded Speech by Young and Elderly Listeners Journal of Speech and Hearing Research, Volume 38, 1150-1156, October 1995 Recognition of Multiply Degraded Speech by Young and Elderly Listeners Sandra Gordon-Salant University of Maryland College Park

More information

Binaural Hearing for Robots Introduction to Robot Hearing

Binaural Hearing for Robots Introduction to Robot Hearing Binaural Hearing for Robots Introduction to Robot Hearing 1Radu Horaud Binaural Hearing for Robots 1. Introduction to Robot Hearing 2. Methodological Foundations 3. Sound-Source Localization 4. Machine

More information

Sound Texture Classification Using Statistics from an Auditory Model

Sound Texture Classification Using Statistics from an Auditory Model Sound Texture Classification Using Statistics from an Auditory Model Gabriele Carotti-Sha Evan Penn Daniel Villamizar Electrical Engineering Email: gcarotti@stanford.edu Mangement Science & Engineering

More information

Binaural Hearing. Why two ears? Definitions

Binaural Hearing. Why two ears? Definitions Binaural Hearing Why two ears? Locating sounds in space: acuity is poorer than in vision by up to two orders of magnitude, but extends in all directions. Role in alerting and orienting? Separating sound

More information

Essential feature. Who are cochlear implants for? People with little or no hearing. substitute for faulty or missing inner hair

Essential feature. Who are cochlear implants for? People with little or no hearing. substitute for faulty or missing inner hair Who are cochlear implants for? Essential feature People with little or no hearing and little conductive component to the loss who receive little or no benefit from a hearing aid. Implants seem to work

More information