Neural correlates of the perception of sound source separation

Similar documents
Comment by Delgutte and Anna. A. Dreyer (Eaton-Peabody Laboratory, Massachusetts Eye and Ear Infirmary, Boston, MA)

Physiological measures of the precedence effect and spatial release from masking in the cat inferior colliculus.

Behavioral/Systems/Cognitive. Sasha Devore 1,2,3 and Bertrand Delgutte 1,2,3 1

Binaural Hearing. Why two ears? Definitions

Sound localization psychophysics

AUDL GS08/GAV1 Signals, systems, acoustics and the ear. Pitch & Binaural listening

HST.723J, Spring 2005 Theme 3 Report

21/01/2013. Binaural Phenomena. Aim. To understand binaural hearing Objectives. Understand the cues used to determine the location of a sound source

Accurate Sound Localization in Reverberant Environments Is Mediated by Robust Encoding of Spatial Cues in the Auditory Midbrain

Effect of spectral content and learning on auditory distance perception

Binaural Hearing. Steve Colburn Boston University

Systems Neuroscience Oct. 16, Auditory system. http:

Binaural interference and auditory grouping

Representation of sound in the auditory nerve

Binaurally-coherent jitter improves neural and perceptual ITD sensitivity in normal and electric hearing

Effect of source spectrum on sound localization in an everyday reverberant room

The neural code for interaural time difference in human auditory cortex

Isolating mechanisms that influence measures of the precedence effect: Theoretical predictions and behavioral tests

Trading Directional Accuracy for Realism in a Virtual Auditory Display

Auditory System & Hearing

Publication VI. c 2007 Audio Engineering Society. Reprinted with permission.

Localization in the Presence of a Distracter and Reverberation in the Frontal Horizontal Plane. I. Psychoacoustical Data

Models of Plasticity in Spatial Auditory Processing

Neural Correlates and Mechanisms of Spatial Release From Masking: Single-Unit and Population Responses in the Inferior Colliculus

Neural Recording Methods

Combating the Reverberation Problem

Modeling Physiological and Psychophysical Responses to Precedence Effect Stimuli

Hearing in the Environment

Hearing II Perceptual Aspects

Juha Merimaa b) Institut für Kommunikationsakustik, Ruhr-Universität Bochum, Germany

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 THE DUPLEX-THEORY OF LOCALIZATION INVESTIGATED UNDER NATURAL CONDITIONS

J Jeffress model, 3, 66ff

IN EAR TO OUT THERE: A MAGNITUDE BASED PARAMETERIZATION SCHEME FOR SOUND SOURCE EXTERNALIZATION. Griffin D. Romigh, Brian D. Simpson, Nandini Iyer

Rhythm and Rate: Perception and Physiology HST November Jennifer Melcher

Neural System Model of Human Sound Localization

3-D Sound and Spatial Audio. What do these terms mean?

The Structure and Function of the Auditory Nerve

3-D SOUND IMAGE LOCALIZATION BY INTERAURAL DIFFERENCES AND THE MEDIAN PLANE HRTF. Masayuki Morimoto Motokuni Itoh Kazuhiro Iida

How is the stimulus represented in the nervous system?

A Model of Visually Guided Plasticity of the Auditory Spatial Map in the Barn Owl

Information conveyed by inferior colliculus neurons about stimuli with aligned and misaligned sound localization cues

Spatial hearing and sound localization mechanisms in the brain. Henri Pöntynen February 9, 2016

Temporal weighting functions for interaural time and level differences. IV. Effects of carrier frequency

Response Properties of Neighboring Neurons in the Auditory Midbrain for Pure-Tone Stimulation: A Tetrode Study

Auditory Phase Opponency: A Temporal Model for Masked Detection at Low Frequencies

Congenital and Prolonged Adult-Onset Deafness Cause Distinct Degradations in Neural ITD Coding with Bilateral Cochlear Implants

P. M. Zurek Sensimetrics Corporation, Somerville, Massachusetts and Massachusetts Institute of Technology, Cambridge, Massachusetts 02139

The role of low frequency components in median plane localization

The use of interaural time and level difference cues by bilateral cochlear implant users

I. INTRODUCTION. J. Acoust. Soc. Am. 113 (3), March /2003/113(3)/1631/15/$ Acoustical Society of America

B. G. Shinn-Cunningham Hearing Research Center, Departments of Biomedical Engineering and Cognitive and Neural Systems, Boston, Massachusetts 02215

EFFECTS OF TEMPORAL FINE STRUCTURE ON THE LOCALIZATION OF BROADBAND SOUNDS: POTENTIAL IMPLICATIONS FOR THE DESIGN OF SPATIAL AUDIO DISPLAYS

The role of spatial cues for processing speech in noise

Phase Locking of Auditory-Nerve Fibers to the Envelopes of High-Frequency Sounds: Implications for Sound Localization

Localization in the Presence of a Distracter and Reverberation in the Frontal Horizontal Plane: II. Model Algorithms

HEARING AND PSYCHOACOUSTICS

Study of sound localization by owls and its relevance to humans

On the influence of interaural differences on onset detection in auditory object formation. 1 Introduction

I. INTRODUCTION. J. Acoust. Soc. Am. 112 (3), Pt. 1, Sep /2002/112(3)/1037/9/$ Acoustical Society of America

Chapter 40 Effects of Peripheral Tuning on the Auditory Nerve s Representation of Speech Envelope and Temporal Fine Structure Cues

An Auditory System Modeling in Sound Source Localization

Discrimination of temporal fine structure by birds and mammals

Binaural detection with narrowband and wideband reproducible noise maskers. IV. Models using interaural time, level, and envelope differences

J. Acoust. Soc. Am. 114 (2), August /2003/114(2)/1009/14/$ Acoustical Society of America

Effect of mismatched place-of-stimulation on the salience of binaural cues in conditions that simulate bilateral cochlear-implant listening

Role of F0 differences in source segregation

Speech segregation in rooms: Effects of reverberation on both target and interferer

Spectro-temporal response fields in the inferior colliculus of awake monkey

Structure and Function of the Auditory and Vestibular Systems (Fall 2014) Auditory Cortex (3) Prof. Xiaoqin Wang

A NOVEL HEAD-RELATED TRANSFER FUNCTION MODEL BASED ON SPECTRAL AND INTERAURAL DIFFERENCE CUES

Chapter 11: Sound, The Auditory System, and Pitch Perception

Sound Localization PSY 310 Greg Francis. Lecture 31. Audition

Spike-Timing Codes Enhance the Representation of Multiple Simultaneous Sound-Localization Cues in the Inferior Colliculus

The masking of interaural delays

Topic 4. Pitch & Frequency

Interaural Time Sensitivity Dominated by Cochlea-Induced Envelope Patterns

The Auditory Nervous System

Coincidence Detection in Pitch Perception. S. Shamma, D. Klein and D. Depireux

Neural coding of ITD with bilateral cochlear implants: Effects of auditory experience

Processing in The Superior Olivary Complex

HCS 7367 Speech Perception

Cortical codes for sound localization

Neural coding of time-varying interaural time differences and its relation to perception. Nathaniel J. Zuk

Robust Neural Encoding of Speech in Human Auditory Cortex

Discrimination and identification of azimuth using spectral shape a)

Dynamic-range compression affects the lateral position of sounds

The development of a modified spectral ripple test

Proceedings of Meetings on Acoustics

Neurobiology of Hearing (Salamanca, 2012) Auditory Cortex (2) Prof. Xiaoqin Wang

Signals, systems, acoustics and the ear. Week 5. The peripheral auditory system: The ear as a signal processor

INTRODUCTION J. Acoust. Soc. Am. 100 (4), Pt. 1, October /96/100(4)/2352/13/$ Acoustical Society of America 2352

INTRODUCTION J. Acoust. Soc. Am. 103 (2), February /98/103(2)/1080/5/$ Acoustical Society of America 1080

The perceptual consequences of binaural hearing

Perceptual Plasticity in Spatial Auditory Displays

Review Auditory Processing of Interaural Timing Information: New Insights

Temporal coding in the sub-millisecond range: Model of barn owl auditory pathway

Brian D. Simpson Veridian, 5200 Springfield Pike, Suite 200, Dayton, Ohio 45431

Prelude Envelope and temporal fine. What's all the fuss? Modulating a wave. Decomposing waveforms. The psychophysics of cochlear

9/29/14. Amanda M. Lauer, Dept. of Otolaryngology- HNS. From Signal Detection Theory and Psychophysics, Green & Swets (1966)

The Effects of Hearing Loss on Binaural Function

Transcription:

Neural correlates of the perception of sound source separation Mitchell L. Day 1,2 * and Bertrand Delgutte 1,2,3 1 Department of Otology and Laryngology, Harvard Medical School, Boston, MA 02115, USA. 2 Eaton-Peabody Laboratories, Massachusetts Eye and Ear Infirmary, Boston, MA 02114, USA. 3 Research Laboratory of Electronics, Massachusetts Institute of Technology, Cambridge, MA 02139, USA. *e-mail: day@meei.harvard.edu. As two sound sources become spatially separated in the horizontal plane, the binaural cues used for sound localization become distorted from their values for each sound in isolation. Because firing rates of most neurons in the inferior colliculus (IC) are sensitive to these binaural cues, we hypothesized that these neurons would be sensitive to source separation. We examined changes in the target azimuth tuning functions of IC neurons in unanesthetized rabbits caused by the concurrent presentation of an interferer at a fixed spatial location. Both target and interferer were broadband noise bursts, uncorrelated with each other. Signal detection analysis of firing rates of individual IC neurons shows that responses are correlated with psychophysical performance on segregation of spatially-separated sources. The analysis also highlights the role of neural sensitivity to interaural time differences of cochlea-induced envelopes in performing this task. Psychophysical performance on source segregation was also compared to the performance of two contrasting maximum-likelihood classifiers operating on the firing rates of the population of IC neurons. The population-pattern classifier had access to the firing rates of every neuron in the population, while the twochannel classifier operated on the summed firing rates from each side of the brain. Unlike the two-channel classifier, the population-pattern classifier could segregate the sources accurately, suggesting that some of the information contained in the variability of azimuth tuning functions across IC neurons is used to segregate sources. Key words: interaural time difference; interaural level difference; interaural correlation; sound localization; neural decoding; inferior colliculus; rabbit. 1. Introduction Sound reaching the two ears exhibits both an interaural time difference (ITD) and an interaural level difference (ILD) which are used to determine the lateral position of the sound source. Both binaural cues become distorted in the presence of a spatially-separated interferer, and these changes are likely used by humans and other species to perceive spatially-separated sources. Best et al. (2004) investigated the ability of human listeners to perceptually segregate two broadband noise sources, uncorrelated with each other, at various spatial separation. The two sources could be distinguished from a single source with nearly perfect accuracy when separated by as little as 15 when one of the sources was fixed at 0 (straight-ahead). A greater spatial separation was required to achieve perceptual segregation when one of the sources was located more laterally. We recorded from single neurons in the IC of unanesthetized rabbits to search for neural correlates of human perception of source separation. A target broadband noise source was presented at various azimuths in the frontal horizontal plane in the presence of a concurrent, spatially-separated noise source, uncorrelated with the target. We analyzed IC responses in two ways to compare to psychophysical results. First, we characterized neural 1

signal detection based on changes in firing rate with spatial separation for each individual IC neuron. Next, we performed a classification of single sources versus two spatially-separated sources based on the population neural activity using either of two neural decoding strategies that represent opposite extremes of how information may be combined across IC neurons. 2. Methods Methods for surgery, acoustic stimulus delivery, and single-unit electrophysiology for the unanesthetized, head-fixed rabbit preparation were essentially the same as described previously (Devore and Delgutte, 2010). Sound stimuli in virtual acoustic space were presented binaurally through ear inserts and were first filtered with rabbit directional transfer functions (DTFs) associated with each azimuthal location in the horizontal plane. Two additional sets of DTFs with altered binaural cues were created for comparison to the standard condition in which ITD and ILD naturally co-vary with azimuth. For the ITD-only condition, all magnitude spectra were fixed to the spectrum at 0, allowing ITD to vary naturally with azimuth while ILD and monaural levels remained fixed. Similarly, for the fixed-itd condition, all phase spectra were fixed to the spectrum at 0, allowing ILD and monaural levels to vary naturally with azimuth while ITD remained fixed. Target and interferer stimuli were both frozen, broadband (0.1 18 khz), 300-ms noise bursts, uncorrelated with each other, presented concurrently every 500 ms. Target and interferer were each presented at a sound level 20 db above the neural threshold at 0. For each IC neuron, target azimuth tuning functions (average firing rate over the stimulus duration vs. target azimuth) were collected in 15 steps with the interferer either co-located with the target (equivalent to a single-source condition) or fixed at 0 ( central interferer ). Azimuth tuning functions were measured under standard, ITD-only, and fixed-itd conditions. An additional set of responses was collected under the standard condition with both target and interferer azimuths independently varied in 30 steps covering every possible spatial combination of target and interferer in the frontal hemifield. Eight stimulus trials were collected for each spatial combination. We investigated the ability of the rate responses of individual neurons to signal the separation of the target from a central interferer (i.e., neural signal detection). For each spatial separation between target and interferer, we calculated a neural sensitivity index d to characterize the change in neural firing rate between co-located and separated source conditions. The neural detection threshold was defined as the (interpolated) target azimuth closest to 0 at which d =1 (e.g., Fig. 2B). Neural detection thresholds were calculated from responses collected under standard, ITD-only, and fixed-itd conditions. We also investigated the ability of two contrasting neural decoding strategies based on population IC activity to detect the separation of target and interferer anywhere in the frontal hemifield. For both classifiers, a conditional probability density of the population activity was estimated for each target/interferer combination, assuming multivariate Gaussian densities. Left and right populations of IC neurons were created by making a mirror twin of every neuron. The population-pattern classifier operated on a 2N-dimensional vector corresponding to the firing rates of the N neurons in our sample and their mirror twins. The two-channel classifier operated on the 2-dimensional vector formed by the summed firing rates across the left and right IC populations. For 500 iterations of a Monte Carlo procedure, a random stimulus trial was removed from each neuron before the mean and variance of each dimension were calculated. The removed trial was then used to test the classifier performance in that iteration by classifying it to the spatial combination having the maximum likelihood. Classifier performance was assessed as the fraction of test trials in which the classifier 2

correctly distinguished between a single source and spatially-separated sources, regardless of localization accuracy. 3. Results To investigate the physiological correlates of perception of source separation, we measured the responses to concurrent target and interferer stimuli from 99 neurons in the right IC of two female, Dutch-belted rabbits. Best frequencies (BFs) ranged from 0.19 to 17 khz. A characteristic feature observed in many neurons was a pronounced peak or notch at 0 in the target azimuth tuning function with a central interferer (Fig. 1B, solid line, notch ), while there was no such feature in the single-source tuning function (Fig. 1A, solid line). In other words, the firing rate was either sharply suppressed or enhanced as the target was separated by just 15 in either direction from the central interferer. Rate suppression or enhancement was only found in neurons with high BFs (>1.5 khz), and occurred in a majority (72%) of these neurons. Figure 1. Target azimuth tuning functions of one IC neuron under altered-cue and standard conditions for a single source (A) and with a central interferer (B). (C) Target ITD tuning functions from a cross-correlation model using a cochlear filter with center frequency, f center =3.19 khz. Interferer location (solid triangles); ITD equal to one period of f center (open triangles). The restriction of rate suppression and enhancement to high-bf neurons suggests the involvement of ILDs or envelope ITDs, as these are the binaural cues to which high-bf neurons are sensitive (Delgutte et al., 1995; Joris, 2003). To identify the relevant cues, we compared azimuth tuning functions between altered-cue and standard conditions. For the high-bf neuron in Fig. 1A, the single-source tuning function in the fixed-itd condition resembles that in the standard condition, while the tuning function in the ITD-only condition is nearly flat, suggesting that tuning is primarily determined by ILD. In contrast, in the presence of a central interferer (Fig. 1B), the fixed-itd tuning function no longer resembles the standard tuning function, while the ITD-only tuning function shows rate enhancement similar to that observed in the standard condition. Therefore, the rate enhancement with target separation from a central interferer was dependent on sensitivity to ITD, although the single-source tuning function was primarily dependent on ILD. This neuron was representative in that rate suppression or enhancement were consistently observed in the ITDonly condition but not the fixed-itd condition. Rate suppression or enhancement can be explained by a cross-correlation model operating on cochlea-induced envelopes of the ear input signals. In this model, left and right sound stimuli were bandpass filtered (0.25-oct bandwidth) to mimic cochlear processing, and the normalized correlation (Trahiotis et al., 2005) was computed between the Hilbert envelopes of the left and right bandpass-filtered waveforms. As the ITD of a broadband source is varied, the interaural cross-correlation coefficient (IACC) decreases slowly (Fig. 1C, dashed line) because the envelopes being cross-correlated only contain low frequencies. 3

However, when a central interferer is introduced (Fig. 1C, solid line), the IACC changes more rapidly with target ITD, oscillating at the period of the filter center frequency (3.19 khz in Fig. 1C). While the envelopes in this case are still limited to low frequencies, the rapid oscillation occurs because the waveform interactions between the target and interferer that determine the mixture envelope are dependent on the temporal fine structure after cochlear filtering. The behavior of IC neurons that are sensitive to envelope ITD is consistent with model predictions. While the firing rates of IC neurons typically change little when a single source moves by 15 (Fig. 1A, dashed line), the azimuth tuning functions with a central interferer show sharp changes in firing rate as the cochlea-induced envelopes become strongly decorrelated when the target is separated by 15 from the interferer. Figure 2. Neural signal detection. (A) Target azimuth tuning function in the presence of a central interferer for a single IC neuron (bars: 1 SD). Triangle indicates interferer location. (B) d calculated from the data in (A) showing the detection thresholds (dotted lines) where d = 1 (dashed line). (C) Distribution of detection thresholds across high-bf neurons (N = 58) showing medians (dashed lines) and interquartile ranges (boxes). Best et al. (2004) showed that the perceptual segregation performance for two broadband noise sources (with one fixed at 0 ) remains good when the stimuli are highpass filtered, but is degraded in the fixed-itd condition, suggesting the use of envelope ITD for segregation. The sharp rate suppression or enhancement seen in high-bf IC neurons is a possible neural correlate of this ability. To test this idea, we calculated a neural detection threshold for the spatial separation of the target from the central interferer for each high-bf neuron in our sample (Fig. 2A and B). The median detection threshold across the population of high-bf neurons was small (< 8 ) and nearly the same for ITD-only and standard conditions, but was 4 times larger for the fixed-itd condition (Fig. 2C). Thus, the sensitivity of IC neurons to envelope decorrelation plays a key role in the neural detection of source separation for high-pass sounds. 4

Figure 3. Neural decoding. (A) Listener performance on the perception of separate sources as a target was separated from an interferer [modified from Best et al. (2004), Fig. 3e]. Each line indicates performance for an interferer fixed at the location marked by a triangle with matching grayscale. Performance of the population-pattern (B) and two-channel (C) classifiers on the same task. Mean rates of the two-channel (D) and population-pattern (E) classifiers for two spatial combinations (single source at 30 and two sources at 30 and 60 ). Best et al. (2004) further showed that the amount of spatial separation between target and interferer necessary for perceptual segregation increases when the interferer is placed more laterally (Fig. 3A). We tested the ability of two neural decoding strategies to segregate two sources anywhere in the frontal hemifield based on the firing rates of the population of IC neurons. The population-pattern classifier takes as input the vector of firing rates for every neuron in the population and classifies the population firing pattern into either single source or two sources. This strategy represents the best performance that can be achieved from the firing rates of IC neurons under the classifier assumptions. In contrast the two-channel classifier operates only on the summed rates across IC neurons for each side of the brain. Such a two-channel model of sound localization has been shown to be consistent with some psychophysical results for single sources (Stecker et al., 2005; Devore et al., 2009; Lesica et al., 2010). The population-pattern classifier was able to perfectly segregate sources with 30 separation when the interferer was fixed at 0 or 30, but required a wider separation at 60 and 90 to attain perfect segregation (Fig. 3B), similar to human perception. The two-channel classifier, however, failed to correctly identify single sources at 0, 30 and 60 (Fig. 3C). For example, the two-channel classifier often mistook a single source at 30 for two sources at 30 and 60, respectively. The classifier made this error because the left and right summed rates for a single source at 30 are very similar to those for 30 and 60 (Fig. 3D). The population-pattern classifier, on the other hand, has much more information available in its inputs. In Fig. 3E, the firing rates of each neuron in our sample in response to a single source at 30 are shown in decreasing order (gray bars). Maintaining the same order, the pattern of firing rates in response to sources at 30 and 60 (black bars) differed greatly from the pattern for 30. The population-pattern classifier can use information contained in the variability of azimuth tuning functions across neurons to accurately segregate sources, while this information is averaged away with the two-channel classifier. 5

4. Discussion The majority of high-bf neurons in the IC show a sharp rate suppression or enhancement when one broadband source is separated by just 15 from another one located at 0. These sharp rate changes are due to neural sensitivity to the decorrelation of cochleainduced envelopes that occurs with source separation. We showed that this rate suppression or enhancement is a neural correlate of the perceptual segregation of concurrent, spatiallyseparated sources. Further, we showed that a maximum-likelihood classifier operating on the pattern of firing rates across the population of IC neurons can account for psychophysical data on perceptual segregation anywhere in the frontal hemifield by making use of the information contained in the variability of azimuth tuning functions across neurons. The success of the population-pattern classifier on source segregation does not prove that this decoding strategy is used in the brain. This classifier assumes that some higher auditory center keeps track of the response of every IC neuron. At minimum, our result is a proof-of-principle that the information contained in rabbit IC firing rates is sufficient to approximate human performance on source segregation. However, the failure of the twochannel classifier does suggest that this particular neural decoding strategy cannot be used to segregate sources, even though it works adequately for the localization of a single source. Our results show that neurons that respond similarly to a single source at a given location can respond in dramatically different ways in the presence of a spatially-separated interferer (Fig. 3E). It is likely that the neural decoding strategy used at higher auditory centers takes into account at least some of this variability across neurons. The failure of the two-channel classifier on source separation does not necessarily rule out its possible use for localization and lateralization. It is possible that different decoding strategies are used for different spatial hearing tasks. The segregation study of Best et al. (2004) required subjects to respond either one source or two sources. However, a similar study of the localization of a target broadband noise in the presence of an interferer found that human listeners usually perceived a single auditory event and less often two spatially diffuse events when there were no onset or offset differences between the two stimuli (Braasch, 2002). As such, the subjects responses in the Best et al. study may be better interpreted as single source and diffuse source, and similarly the population-pattern classifier may be detecting the diffuseness of the sound percept rather than two distinct source locations. The neural mechanisms involved in source separation are relevant to sound localization in reverberation, which can be interpreted as a superposition of many attenuated, separated (but coherent) sources. The effects of envelope decorrelation have also been observed in the responses of IC neurons under reverberant conditions (Devore and Delgutte, 2010). Therefore it is likely that the neural mechanisms for the perception of source separation are more generally involved in the coding of spatial properties of the surrounding environment. Acknowledgments This work was supported by NIDCD (US) grants R01 DC002258 and P30 DC005209. References Best, V., van Schaik, A., Carlile, S., 2004. Separation of concurrent broadband sound sources by human listeners. J. Acoust. Soc. Am. 115, 324-36. Braasch, J., 2002. Localization in the presence of a distracter and reverberation in the frontal horizontal plane: II. Model algorithms. Acta Acustica United with Acustica 88, 956-969. Delgutte, B., Joris, P.X., Litovsky, R.Y., Yin, T.C., 1995. Relative importance of different acoustic cues to the directional sensitivity of inferior-colliculus neurons. In: Manley, 6

G.A., Klump, G.M., Koeppl, C., Fastl, H., Oeckinghaus, H., (Eds.), Advances in Hearing Research, World Scientific Publishing, Singapore, pp. 288-99. Devore, S., Delgutte, B., 2010. Effects of reverberation on the directional sensitivity of auditory neurons across the tonotopic axis: influences of interaural time and level differences. J. Neurosci. 30, 7826-37. Devore, S., Ihlefeld, A., Hancock, K., Shinn-Cunningham, B., Delgutte, B., 2009. Accurate sound localization in reverberant environments is mediated by robust encoding of spatial cues in the auditory midbrain. Neuron 62, 123-34. Joris, P.X., 2003. Interaural time sensitivity dominated by cochlea-induced envelope patterns. J. Neurosci. 23, 6345-50. Lesica, N.A., Lingner, A., Grothe, B., 2010. Population coding of interaural time differences in gerbils and barn owls. J. Neurosci. 30, 11696-702. Stecker, G.C., Harrington, I.A., Middlebrooks, J.C., 2005. Location coding by opponent neural populations in the auditory cortex. PLoS Biol. 3, e78. Trahiotis, C., Bernstein, L.R., Stern, R.M., Buell, T.N., 2005. Interaural correlation as the basis of a working model of binaural processing: an introduction. In: Popper, A.N., Fay, R.R., (Eds.), Sound Source Localization, Springer, New York. 7