The Effects of Hearing Loss on Binaural Function

Similar documents
Binaural Hearing. Steve Colburn Boston University

Binaural processing of complex stimuli

Binaural Hearing. Why two ears? Definitions

Sound localization psychophysics

AUDL GS08/GAV1 Signals, systems, acoustics and the ear. Pitch & Binaural listening

Processing in The Superior Olivary Complex

21/01/2013. Binaural Phenomena. Aim. To understand binaural hearing Objectives. Understand the cues used to determine the location of a sound source

Systems Neuroscience Oct. 16, Auditory system. http:

The Auditory Nervous System

Lecture 7 Hearing 2. Raghav Rajan Bio 354 Neurobiology 2 February 04th All lecture material from the following links unless otherwise mentioned:

Auditory System & Hearing

What Is the Difference between db HL and db SPL?

HCS 7367 Speech Perception

Publication VI. c 2007 Audio Engineering Society. Reprinted with permission.

Processing in The Cochlear Nucleus

COM3502/4502/6502 SPEECH PROCESSING

Infant Hearing Development: Translating Research Findings into Clinical Practice. Auditory Development. Overview

Hearing. Juan P Bello

HST.723J, Spring 2005 Theme 3 Report

Spectrograms (revisited)

Acoustics, signals & systems for audiology. Psychoacoustics of hearing impairment

Auditory System. Barb Rohrer (SEI )

Hearing in the Environment

Spatial hearing and sound localization mechanisms in the brain. Henri Pöntynen February 9, 2016

HCS 7367 Speech Perception

The perceptual consequences of binaural hearing

Representation of sound in the auditory nerve

Lauer et al Olivocochlear efferents. Amanda M. Lauer, Ph.D. Dept. of Otolaryngology-HNS

Spatial processing in adults with hearing loss

HEARING AND PSYCHOACOUSTICS

SPHSC 462 HEARING DEVELOPMENT. Overview Review of Hearing Science Introduction

Issues faced by people with a Sensorineural Hearing Loss

Signals, systems, acoustics and the ear. Week 5. The peripheral auditory system: The ear as a signal processor

Hearing Lectures. Acoustics of Speech and Hearing. Auditory Lighthouse. Facts about Timbre. Analysis of Complex Sounds

Deafness and hearing impairment

PERIPHERAL AND CENTRAL AUDITORY ASSESSMENT

Brad May, PhD Johns Hopkins University

Required Slide. Session Objectives

An Auditory System Modeling in Sound Source Localization

Lecture 3: Perception

J Jeffress model, 3, 66ff

Frequency refers to how often something happens. Period refers to the time it takes something to happen.

Auditory Phase Opponency: A Temporal Model for Masked Detection at Low Frequencies

Neural correlates of the perception of sound source separation

The extent to which a position-based explanation accounts for binaural release from informational masking a)

Acoustics Research Institute

Archives LIBRARP.! BINAURAL INTERACTION IN HEARING IMPAIRED LISTENERS. Kaigham Jacob Gabriel 1. B.S. University of Pittsburgh ( 1977 )

Organization. The physics of sound. Measuring sound intensity. Fourier analysis. (c) S-C. Liu, Inst of Neuroinformatics 1

Proceedings of Meetings on Acoustics

Comment by Delgutte and Anna. A. Dreyer (Eaton-Peabody Laboratory, Massachusetts Eye and Ear Infirmary, Boston, MA)

Hearing II Perceptual Aspects

What you re in for. Who are cochlear implants for? The bottom line. Speech processing schemes for

Physiological measures of the precedence effect and spatial release from masking in the cat inferior colliculus.

Localization: Give your patients a listening edge

The role of periodicity in the perception of masked speech with simulated and real cochlear implants

Before we talk about the auditory system we will talk about the sound and waves

Masker-signal relationships and sound level

The Central Auditory System

Prelude Envelope and temporal fine. What's all the fuss? Modulating a wave. Decomposing waveforms. The psychophysics of cochlear

Brian D. Simpson Veridian, 5200 Springfield Pike, Suite 200, Dayton, Ohio 45431

3-D Sound and Spatial Audio. What do these terms mean?

The Structure and Function of the Auditory Nerve

How is the stimulus represented in the nervous system?

Neural Correlates and Mechanisms of Spatial Release From Masking: Single-Unit and Population Responses in the Inferior Colliculus

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 THE DUPLEX-THEORY OF LOCALIZATION INVESTIGATED UNDER NATURAL CONDITIONS

Effect of mismatched place-of-stimulation on the salience of binaural cues in conditions that simulate bilateral cochlear-implant listening

Chapter 11: Sound, The Auditory System, and Pitch Perception

Speech segregation in rooms: Effects of reverberation on both target and interferer

16.400/453J Human Factors Engineering /453. Audition. Prof. D. C. Chandra Lecture 14

Hearing. and other senses

Sound and Hearing. Decibels. Frequency Coding & Localization 1. Everything is vibration. The universe is made of waves.

Essential feature. Who are cochlear implants for? People with little or no hearing. substitute for faulty or missing inner hair

HHS Public Access Author manuscript Ear Hear. Author manuscript; available in PMC 2018 March 01.

Auditory Physiology Richard M. Costanzo, Ph.D.

Linguistic Phonetics. Basic Audition. Diagram of the inner ear removed due to copyright restrictions.

I. INTRODUCTION. NL-5656 AA Eindhoven, The Netherlands. Electronic mail:

Who are cochlear implants for?

Effect of microphone position in hearing instruments on binaural masking level differences

Linguistic Phonetics Fall 2005

Interaural envelope correlation change discrimination in bilateral cochlear implantees: Effects of mismatch, centering, and onset of deafness

THE EAR Dr. Lily V. Hughes, Audiologist

Lateralized speech perception in normal-hearing and hearing-impaired listeners and its relationship to temporal processing

Structure and Function of the Auditory and Vestibular Systems (Fall 2014) Auditory Cortex (3) Prof. Xiaoqin Wang

! Can hear whistle? ! Where are we on course map? ! What we did in lab last week. ! Psychoacoustics

Nature of the Sound Stimulus. Sound is the rhythmic compression and decompression of the air around us caused by a vibrating object.

Binaural interference and auditory grouping

INTRODUCTION J. Acoust. Soc. Am. 103 (2), February /98/103(2)/1080/5/$ Acoustical Society of America 1080

Thresholds for different mammals

The role of spatial cues for processing speech in noise

PHYS 1240 Sound and Music Professor John Price. Cell Phones off Laptops closed Clickers on Transporter energized

Synaptopathy Research Uwe Andreas Hermann

Prescribe hearing aids to:

An Auditory-Model-Based Electrical Stimulation Strategy Incorporating Tonal Information for Cochlear Implant

Models of Plasticity in Spatial Auditory Processing

Hearing. istockphoto/thinkstock

ALD-A26I 659 IMNAINPAGE

Juha Merimaa b) Institut für Kommunikationsakustik, Ruhr-Universität Bochum, Germany

IN EAR TO OUT THERE: A MAGNITUDE BASED PARAMETERIZATION SCHEME FOR SOUND SOURCE EXTERNALIZATION. Griffin D. Romigh, Brian D. Simpson, Nandini Iyer

Hearing the Universal Language: Music and Cochlear Implants

Localization 103: Training BiCROS/CROS Wearers for Left-Right Localization

Transcription:

The Effects of Hearing Loss on Binaural Function Steve Colburn Boston University colburn@bu.edu

Outline What are the major advantages of two ears? How do the left and right signals differ? How do we use this information physiologically? How well can we exploit our two ears for important tasks? Localization of sound sources Attending to source of interest in complex situations Effects of hearing impairment in binaural processing... Interaural parameter sensitivity and speech with speech maskers. (a study by Nathan Spencer in my lab). What limits performance with impairments/aids/implants?

What are the major advantages of two ears? Redundancy Localization of sources Identifying room characteristics (size, shape, wall reflectivity, and???) The Cocktail Party Effect Renoir's Luncheon of the Boating Party

Different signals at the two ears Localization Cues: ITD: far ear delayed ILD: far ear attenuated Spectral Shapes

Interaural Time Delay (ITD) versus azimuth θ Durlach and Colburn, 1978

Interaural Level Difference (ILD) ILD versus azimuth for two frequencies Measurements and sphere calculations

Spectral Shape Information as a localization cue

Outer Ear (beige), Middle Ear (pink), Inner Ear (Blue, cochlea)

Physiological Mechanisms lconsider stimulus coding and binaural processing in physiological structures. lquick review of peripheral physiology lauditory Nerve Responses and Timing Information lsuperior Olivary Nuclei compare left and right responses with sensitivity to ITD or ILD

Middle Ear bones Semicircular Canals Vestibular and Auditory Nerves Cochlear spiral Stapedius muscle attaches to third bone (stapes) and contracts to decrease transmission May respond in anticipation of loud sounds

Recordings from three individual auditory nerve fibers with NO ACOUSTIC STIMULUS

Primary Auditory Nerve Patterns (Kiang, 1965) Note the stochastic nature of the response

Rate-intensity function for auditory nerve fiber

Auditory Nerve Tuning Curves: Level required for detectable response

Synchronization to tone cycles Histograms of responses to tones at various levels Note that histograms here cover two periods of the tonal stimulus

Jeffress model of ITD processing How do process this timing information to compare right and left timings Jeffress postulated Coincidence Detector Network in 1948. Internal delays that are different for each cell in an array Fixed time delay on input patterns would excite compensating delay cell most strongly

Coincidence Network of Jeffress (1948)

ITD-sensitive neuron in MSO

Mechanisms of ITD Processing Coding of time structure to neural pattern lprimary auditory nerve coding Maintenance/sharpening of temporal structure lsharpening in cochlear nucleus (Bushy Cells of CN) Sensitivity to delays of left and right inputs lsingle neuron processing in MSO Preservation to higher neural levels lmaintenance of spatial code for ITD

Binaural Display: (f,τ)-plane Narrowband Cross-correlation function for each f Internal delay limiting function Resulting distribution of activity ITD from ridge From Stern and Trahiotis, 1997

Interaural level (ILD) representation LSO neurons (ILD and onset-itd sensitive) Excitatory input from ipsilateral side Inhibitory input from contralateral side Array of ILD processors versus frequency

LSO Neuron and ILD sensitivity Diranieh and Colburn, 1994 (MODEL)

ILD,f array of LSO neurons LSO neurons are each ILD sensitive LSO neurons are tuned in frequency (like auditory nerve fibers and MSO cells) Provide information about ILD for each frequency band May be particularly important in reverberant environments (Hartmann, 1999)

Interaural Difference (ITD or ILD) Resolution for Human Subjects How small a time delay between the ears is reliably detectable by a human listener? Called a Just-Noticeable Difference in ITD or ILD Best listeners can do better than 10 µs on the order of ten millionths of a second! Best listeners for ILD do about 0.5 decibels

Interaural Difference Sensitivity versus frequency

Envelope carries ITD at HF Temporal synchronization to timing falls off at high frequencies Minimal sensitivity to fine structure ITD above about 1500 Hz in humans But high frequencies carry interaural timing information in synchronization to the envelopes. This is a factor in binaural performance but it is complicated.

Envelope on noise band provides timing cues Neural synchronization to the envelope (black curve with frequencies related to bandwidth) No synchronization to the fine structure (i.e., to the carrier) at high frequencies

How do we handle the cocktail party? Combined stimuli mixed together in each frequency band We want to use the internal distributions over frequency and interaural parameters may be based on interaural time and level or may involve processed, position-based information These displays change over time as different stimuli from different directions dominate. This is a complex challenge!

Sorting it all out The information has to be there. We need to ignore distracting sources Evidence that musicians can do this better Practice makes us better in any case Called informational masking versus energetic We need to exploit the interaural differences as well as the dynamic temporal variations We need to use our knowledge of the signals

Interaural Sensitivity with Hearing Loss (Nathan Spencer Experiments) Primary interest was effects of hearing impairments on binaural processing. Measured interaural discrimination as well as spatial release in speech-on-speech masking. Hypothesized that interaural discrimination ability [just-noticeable differences (JNDs) in interaural time, level, and correlation] would co-vary with spatial separation benefits.

Nathan Spencer Subjects Ten young normal-hearing subjects, aged 20-29 Ten young hearing-impaired subjects. HI1-HI8, HI10, HI11; aged 19-38 One older hearing-impaired subject HI9 aged 70

Hearing Impaired audiograms IM)eC* IM)eC* IM)eC* IM)eC* 31 1 51 71 91 211 31 51 71 91 211 31 51 71 91 211 31 51 71 91 211 231 HI1 HI2 HI3 HI4 HI5 HI6 HI7 HI8 HI9 HI10 25 years old, male 19,m 24,m 37,m 22,f 29,m HI11 18,f 19,f /36/6 2 3 5 9 /36/6 2 3 5 9 /36/6 2 3 5 9 Gsf r vf odz!)l I { * Gsf r vf odz!)l I { * Gsf r vf odz!)l I { * 29,f 70,m 38,f Audiograms symmetric in all cases except for in HI10, an author in this work.

Interaural difference sensitivity tasks Stimuli 1/3 octave, narrow-band noises, centered at either low frequency (500 Hz) or high frequency (2 khz or 4 khz). Measured just-noticeable differences (JNDs), relative to diotic reference (waveforms same in both ears) Interaural time difference (ITD) discrimination Interaural level difference (ILD) discrimination Interaural correlation difference (ICC) discrimination

Measuring interaural difference sensitivity thresholds Acquired adaptive tracks to (Levitt, 1971) 70.7% correct Kept measuring performance in all tasks, over multiple days Calculated thresholds and error bars based on last twelve adaptive tracks

ITD sensitivity measurements Presented four intervals In interval 2 or interval 3, waveform was delayed in one ear, relative to other creating an off-center image in only one interval Subject asked to identify which interval (2 or 3) was different Provided with feedback

Normal Hearing ITD thresholds Mp h! 2 1! ) JUE *! ) µ t* Ijhi!Gsfrvfodz Mp x! Gsf r v f o dz 4 3 2 4 3 2 Opsn bm!i f bsjoh 2 3 4 5 6 7 8 9 : 21 2 3 4 5 6 7 8 9 : 21 µ 2111 211 21 2111 211 21 JUE!) µ t* OI!611!I { OI!611!I {!N f bo OI!5!l I { (Based on envelope timing) Increasing average hearing loss (< 10 db difference cumulative)

Hearing Impaired ITD thresholds (right) Mp h! 2 1! ) JUE *! ) µ t* Ijhi!Gsfrvfodz Mp x! Gsf r v f o dz 4 3 2 4 3 2 Opsn bm!i f bsjoh 2 3 4 5 6 7 8 9 : 21 2 3 4 5 6 7 8 9 : 21 µ! I f bsjoh!jn qbjsf e 2 3 4 5 6 7 8 9 : 2122 2 3 4 5 6 7 8 9 : 2122 µ! 2111 211 21 2111 211 21 JUE!) µ t* Wilcoxon Median Test p>0.05 Increasing average hearing loss

ILD Thresholds (Note log scale) Opsn bm!i f bsjoh I f bsjoh!jn qbjsf e Mp h! 2 1! ) JME *! ) e C* Ijhi!Gsfrvfodz Mp x! Gsf r v f o dz 2 1.2 2 1.2! 2 3 4 5 6 7 8 9 : 2122 2 3 4 5 6 7 8 9 : 21 µ!! 21 2! 23456789:21223 1/2 2 3 4 5 6 7 8 9 : 2122 µ 21 2 1/2 JME!)eC* Wilcoxon Median Test p>0.05 Wilcoxon Median Test p>0.05

Hearing loss as a function of interaural difference sensitivity thresholds for HI 211 IM!bu!611!I{!)eC* Bwf sbhf!i M!)eC* ec!i M 61 : 7 21 22 3 4 2 5 1 21 211 9 86 : 21 22 21 211 2111 JUE!Mpx!Gsf r /!) µ t* 3 7 5 2 4 9 8 6 Lack of correlation found for HI group in ITD at 500 Hz No correlations found for either of the subject groups, for any interaural difference sensitivity task

ILD threshold as a function of ITD threshold JME!)e C*!Mpx!Gsf r /!I jhi!gsf r / 21 4 2 21 4 2 Opsn bm!i f bsjoh r²=.59 r²=.73 Ifbsjoh!Jnqbjsfe 4 6 5 : 42 7 9 8 21 21 22 2 3 3 7 5 r²=.88 : 2 3 4 7 9 8 6 5 r²=.79 21 211 2111 JUE!Mpx!Gsf r vf odz!) µ t* : 3 5 7 21 2 4 6 9 8 6 21 211 2111

ID JND measurements were generally correlated between low and high frequencies, considered across subjects JME!I jhi!g/!)ec* JUE!I jh i!g/!) µ t* 2111 211 21 21 4 2 Opsn bm.i f bsjoh : 5 6 2 3 4 821 7 9 3 : : 21 5 2 48 7 9 3 Ifbsjoh.Jnqbjsfe 5 7 21 2 r²=.73 r²=.88 21 211 21 211 JUE!Mpx!Gsf r vf odz!) µ t* 6 5 r²=.68 : 2 21 37 r²=.94 6 4 JDD!I jhi!g/!)ec* 1.6.21.26.31 2 921 3 7 4 2 4 56 8.36.41.31 2 4 JME!Mpx!Gsf r vf odz!)ec* : r²=.84 3.41.31.21 JDD!Mpx!Gsf r vf odz!)ec* 9 6 52 7 21 : 4 r²=.25

JME!I jhi!)ec*!jme!mpx!)ec* 21 4 2 21 4 2 Opsn bm.i f bsjoh Ifbsjoh.Jnqbjsfe 6 5 : 7 9 22 42 8 21 21 2 3 3 7 5 : 2 3 4 7 89 6 5 : 3 5 7 21 2 4 4 6 9 8 6!JDD!Mpx!)eC*.21.31.41 8 : 4 8 56 21 2 3 47 921 : 7 22 5 2 96 3 21 211 21 211 2111 JUE!Mpx!Gsf r vf odz!) µ t*

Four, highly trained Normal-Hearing listeners at 500 Hz (Koehnke et al. 1986) Previous studies: individual differences in inter-aural difference sensitivity, not in monaural level sensitivity MONAURAL discrimination INTERAURAL discrimination

Interaural Discrimination Conclusions Large inter-subject variability Interaural difference sensitivity thresholds did not correlate with hearing loss ITD and ILD thresholds highly correlated among both NH and HI subject groups Consistent with my seeing remarkable subjects who were excellent in binaural tasks with severe hearing loss

Now turn to the Cocktail Party Problem Understanding speech in competition with other speech can be challenging. The famous psychologist George A. Miller once wrote: "It is said that the best place to hide a leaf is in the forest, and presumably the best place to hide a voice is among other voices" (p. 118). Miller, G.A. (1947) "The masking of speech," Psychol. Bulletin, 44, 105-129

E.Colin Cherry s 1953 summary: How do we recognize what one person is saying when others are speaking at the same time (the cocktail party problem')? Factors: lvoices from different directions llip-reading, gestures, and the like ldifferent speaking voices, pitches, speeds, laccents differing ltransition-probabilities (subject matter, voice dynamics, syntax, )

Obtaining SRTs Used procedures developed by Carr (2010) (now Carr Levy) Adaptive algorithm used to determine the target/masker (T/M) ratio (db) for 50% correct Masker levels fixed Normal hearing Each masker set to ~45 db above SRT in quiet (65 db SPL) Hearing impaired Each masker set ~25 db above SRT in quiet in HI, when possible Target level varied For HI 11, discomfort thresholds limited masker level, so target could be limited by audibility (less than 10 db above quiet threshold, Duquesnoy, 1983)

Sources ~ 1.5m away from head Colocated task 10 m ~In front Difficult case, in general All talkers in the same position No separation-related cues 6m

Symmetric maskers condition 10 m ~In front ~1.5m Moderately difficult ~60⁰ ~60⁰ Interaural differences No long-term better ear 6m

Anti-symmetric maskers condition Sources ~ 1.5m from head Expected to be easier than colocated ~60⁰ ~In front Long-term better ear, contralateral to ear with dominant masker Interauraltime differences Interaural level differences Measured performance both right-ear better and left-ear better

Sources ~ 1.5m away from head Colocated task 10 m ~In front Difficult case, in general All talkers in the same position No separation-related cues 6m

Opsn bm!i f bsjoh Colocated SRTs! Ifbsjoh!Jnqbjsfe! TSU!U0N!)eC* 1.21.31! µ 2 3 4 5 6 7 8 9 : 21! 2 3 4 5 6 7 8 9 : 2122 µ Limited by audibility Wilcoxon Median Test p>0.05

Anti-symmetric condition Sources ~ 1.5m from head Expected to be easier than colocated ~60⁰ ~In front Long-term better ear, contralateral to ear with dominant masker Interauraltime differences Interaural level differences Measured performance both right-ear better and left-ear better

Anti-symmetric SRTs Opsn bm!i f bsjoh! I f bsjoh!jn qbjsf e! TSU!U0N!)eC* 1.21.31! 2 3 4 5 6 7 8 9 : 21 µ! 2 3 4 5 6 7 8 9 : 2122 µ Spatial release from masking Individual differences are evident Wilcoxon Median Test p<0.05

Determining benefits of spatial separation Total benefit = colocated - spatially separated Colocated Monaural benefit = Colocated separated monaural Separated monaural is measured at better ear Monaural Binaural benefit =Total benefit monaural benefit Binaural

Anti-symmetric benefits of spatial separation Upubm)eC* Npobvsbm!)eC* 31 21 1 31 21 1 31 OI!Cf of gju IJ!Cfofgju!! Large monaural benefit Wilcoxon Median Test p<0.05 p<0.05 Cjobvsbm!)eC* 21 1 µ µ 2 3 4 5 6 7 8 9 : 21 2 3 4 5 6 7 8 9 : 2122 p<0.05

Anti-symmetric binaural benefit Cjobvsbm)eC* 21 6 1.6 OI!Cf of gju 2 3 4 5 6 7 8 9 : 21 IJ!Cfofgju µ µ 2 3 4 5 6 7 8 9 : 2122 High inter-subject variability in both groups Wilcoxon Median Test p<0.05

Anti-symmetric binaural benefit as a function of hearing loss Cjobvsbm)eC* 21 6 1.6 3 2 4 5 6 1 31 51 71 91 211 Nf bo!i M!)eC* 7 8 9 21 : 22 r 2 =.55

Anti-symmetric binaural benefit as a function of interaural difference sensitivity r 2 shown in table Red means p<0.05 NH HI 1-10 HI 1-8 (<70 db HL) ILD 500 ILD 4k Binaural Benefit r 2 <0.2 r 2 =0.67 Binaural Benefit r 2 <.2 r 2 <.2 Binaural Benefit N=8 r 2 =.53 N=6 r 2 = 0.64 ITD 500 ITD 4k r 2 =0.51 r 2 =0.39 r 2 <.2 r 2 N/A N=8 0.54 N=4 r 2 N/A Correlation for most interaural difference sensitivity thresholds Correlation for no interaural difference sensitivity thresholds Correlation for some interaural difference sensitivity thresholds

Symmetric task 10 m ~In front ~1.5m Moderately difficult ~60⁰ ~60⁰ Interaural differences No long-term better ear 6m

Symmetric binaural benefit 21 OI!Cf of gju IJ!Cfofgju Cjobvsbm)eC* 6 1 p<0.05 2 3 4 5 6 7 8 9 : 21 µ µ 2 3 4 5 6 7 8 9 : 2122 High inter-subject variability among each subject group Wilcoxon Median Test

Symmetric binaural benefit Binaural benefit did not correlate with hearing loss in either group Binaural benefit did not correlate with interaural difference sensitivity thresholds in either groups

Data Conclusions Hearing loss a key factor in the binaural benefit for the antisymmetric condition but not for the symmetric condition In most cases, hearing-impaired listeners benefited from binaural listening Binaural benefits mostly positive for anti-symmetric condition Binaural benefits were always positive for HI in the symmetric condition In the anti-symmetric condition, binaural benefit was correlated with some interaural difference sensitivity thresholds for both normal-hearing and subset of hearing-impaired listeners In the symmetric condition, binaural benefit not correlated with interaural difference sensitivity thresholds

Time window processing factor? In the symmetric case, with speech maskers, listeners benefit from selectively listening in optimal direction (and/or to the better ear) differently in different time intervals This factor is distinct from interaural resolution factors and from attenuation factors

Speech Spectrogram lfrequency (khz) ltime (about 3.5 seconds)

Effect of EC window length on SRT (1x Jitter) 6 Tzn n f usjd Inconsistent better-ear Bouj.t zn n Consistent better-ear Qsf ejduf e!tsu)ec* 1.6.21.26.31 31 61 311 31 61 311 Xjoepx)nt*

Overall Conclusions for this study Binaural benefit in anti-symmetric condition for HI group decreases with hearing loss Interaural difference sensitivity a key factor for binaural benefit for anti-symmetric speech condition Correlations were observed for normal hearing group and for hearing-impaired subgroup with less-than-severe average hearing loss Model predictions show large effect of jitter standard deviation Interaural difference sensitivity and hearing loss each unrelated to binaural benefit for symmetric condition Neither interaural difference sensitivity nor hearing loss correlated with binaural benefit in the symmetric condition Modeling results show influence of processing window duration Suggest independent estimation of processing window length

Take home message? The processing in multiple speech interference situations is very complex Depends on locations, nature of sources, positions of maskers, individual ability to use the multiple cues, and possibly on temporal window factors. Neither the audiogram nor abilities to do simple discrimination tasks are predictors of abilities in complex environments Tests to be done with hearing aids in complex environments seem important to develop.

Ongoing work on this topic (with Theo Goverts at VUMC) Attempt to characterize difficult environments Recording (binaurally) difficult environments as reported by listeners with hearing impairments Analysis of recordings and comparisons with subjective impressions Work to develop a speech-in-complexenvironments test that could be used to evaluate hearing aids in the clinic.

Stable World Model Internal Representations Real Sources in the real Environment Physical Acoustics L(t) R(t) Peripheral Physiology ITD,f array ILD,f array freq spectra Compare Actual to Current Model of Sources and Environment Listener s Learned Transformations ITD,f array ILD,f array freq spectra Hypothesis Non-auditory inputs Fast (Add/Delete/Move Source) Slow (Pinna) Update Mechanism Colburn and Kulkarni (2005)

Acknowledgements IRC hearing-aid company consortium National Institutes of Health Grant support: NIDCD R01 DC00100 Many Colleagues, especially Gerald Kidd, Theo Goverts, Suzanne Carr Levy, Nathan Spencer

The End!!