Lecture 8: Spatial sound

Size: px
Start display at page:

Download "Lecture 8: Spatial sound"

Transcription

1 EE E6820: Speech & Audio Processing & Recognition Lecture 8: Spatial sound Spatial acoustics Binaural perception Synthesizing spatial audio Extracting spatial sounds Dan Ellis <dpwe@ee.columbia.edu> E6820 SAPR - Dan Ellis L08 - Spatilal sound

2 1 Spatial acoustics Received sound = source + channel - so far, only considered ideal source waveform Sound carries information on its spatial origin - e.g. ripples in the lake - great evolutionary significance The basis of scene analysis? - yes and no - try blocking an ear E6820 SAPR - Dan Ellis L08 - Spatilal sound

3 Ripples in the lake Listener Source Wavefront c m/s) Energy 1/r 2 Effect of relative position on sound - delay = r/c - energy decay ~ 1/r 2 - absorption ~ G(f) r - direct energy plus reflections Give cues for recovering source position Describe wavefront by its normal E6820 SAPR - Dan Ellis L08 - Spatilal sound

4 Recovering spatial information Source direction as wavefront normal - moving plane found from timing at 3 points B pressure t/c = s = AB cosθ θ A C wavefront time - need to solve correspondence elevation φ range r Space: need 3 parameters - e.g. 2 angles and range azimuth θ E6820 SAPR - Dan Ellis L08 - Spatilal sound

5 The effect of the environment Reflection causes additional wavefronts reflection diffraction & shadowing freq / Hz scattering, absorption - many paths many echoes Reverberant effect - causal smearing of signal energy Dry speech airvib16 freq / Hz reverb from hlwy time / sec E6820 SAPR - Dan Ellis L08 - Spatilal sound time / sec

6 Reverberation impulse response Exponential decay of reflections: h room (t) ~e -t /T freq / Hz hlwy16-128pt window t time / s Frequency-dependent - greater absorption at high frequencies faster decay Size-dependent - larger rooms longer delays slower decay Sabine s equation: 0.049V RT 60 = Sα Time constant as size, absorption E6820 SAPR - Dan Ellis L08 - Spatilal sound

7 Outline Spatial acoustics Binaural perception - The sound at the two ears - Available cues - Perceptual phenomena Synthesizing spatial audio Extracting spatial sounds E6820 SAPR - Dan Ellis L08 - Spatilal sound

8 2 Binaural perception R L head shadow (high freq) source path length difference What is the information in the 2 ear signals? - the sound of the source(s) (L+R) - the position of the source(s) (L-R) Example waveforms (ShATR database) shatr78m3 waveform Left Right E6820 SAPR - Dan Ellis L08 - Spatilal sound time / s

9 Main cues to spatial hearing Interaural time difference (ITD) - from different path lengths around head - dominates in low frequency (< 1.5 khz) - max ~ 750 µs ambiguous for freqs > 600 Hz Interaural intensity difference (IID) - from head shadowing of far ear - negligable for LF; increases with frequency Spectral detail (from pinna relfections) useful for elevation & range Direct-to-reverberant useful for range E6820 SAPR - Dan Ellis L08 - Spatilal sound

10 Head-Related Transfer Fns (HRTFs) Capture source coupling as impulse responses { ()} l θφr,, ()r t, θφr,, t Collection: ( RIGHT HRIR_021 0 el HRIR_021 0 el 1 HRIR_021 0 el 0 az 45 0 Azimuth / deg 0 1 HRIR_021 0 el 0 az LEFT time / ms time / ms Highly individual! E6820 SAPR - Dan Ellis L08 - Spatilal sound

11 Cone of confusion azimuth θ Cone of confusion (approx equal ITD) Interaural timing cue dominates (below 1kHz) - from differing path lengths to two ears But: only resolves to a cone - Up/down? Front/back? E6820 SAPR - Dan Ellis L08 - Spatilal sound

12 Further cues Pinna causes elevation-dependent coloration Monaural perception - separate coloration from source spectrum? Head motion - synchronized spectral changes - also for ITD (front/back) etc. E6820 SAPR - Dan Ellis L08 - Spatilal sound

13 Combining multiple cues Both ITD and ILD influence azimuth; What happens when they disagree? Identical signals to both ears image is centered l(t) r(t) t 1 ms t l(t) Delaying right channel moves image to left r(t) t t Attenuating left channel returns image to center l(t) r(t) t t - around 0.1 ms / db E6820 SAPR - Dan Ellis L08 - Spatilal sound

14 Binaural position estimation Imperfect results: (Arruda, Kistler & Wightman 1992) 180 Judged Azimuth (Deg) Target Azimuth (Deg) - listening to wrong hrtfs errors - front/back reversals stay on cone of confusion E6820 SAPR - Dan Ellis L08 - Spatilal sound

15 The Precedence Effect Reflections give misleading spatial cues R l(t) r(t) direct reflected t R /c t But: Spatial impression based on 1st wavefront then switches off for ~50 ms -.. even if reflections are louder -.. leads to impression of room E6820 SAPR - Dan Ellis L08 - Spatilal sound

16 Binaural Masking Release Adding noise to reveal target Tone + noise to one ear: tone is masked + t t Identical noise to other ear: tone is audible + t t t - why does this make sense? Binaural Masking Level Difference up to 12dB - greatest for noise in phase, tone anti-phase E6820 SAPR - Dan Ellis L08 - Spatilal sound

17 Outline Spatial acoustics Binaural perception Synthesizing spatial audio - Position - Environment Extracting spatial sounds E6820 SAPR - Dan Ellis L08 - Spatilal sound

18 3 Synthesizing spatial audio Goal: recreate realistic soundfield - hi-fi experience - synthetic environments (VR) Constraints - resources - information (individual HRTFs) - delivery mechanism (headphones) Source material types - live recordings (actual soundfields) - synthetic (studio mixing, virtual environments) E6820 SAPR - Dan Ellis L08 - Spatilal sound

19 Classic stereo L R Intensity panning : no timing modifications, just vary level ±20 db - works as long as listener is equidistant Surround sound: extra channels in center, sides,... - same basic effect - pan between pairs E6820 SAPR - Dan Ellis L08 - Spatilal sound

20 Simulating reverberation Can characterize reverb by impulse response - spatial cues are important - record in stereo - IRs of ~ 1 sec very long convolution Image model: reflections as duplicate sources virtual (image) sources reflected path source listener Early echos in room impulse response: direct path early echos h room (t) Actual reflection may be h ref (t), not δ(t) E6820 SAPR - Dan Ellis L08 - Spatilal sound t

21 Artificial reverberation Reproduce perceptually salient aspects - early echo pattern ( room size impression) - overall decay tail ( wall materials...) - interaural coherence ( spaciousness) Nested allpass filters (Gardner 92) Allpass x[n] -g + z -k + y[n] H(z) = h[n] z -k - g 1 - g z -k 1-g 2 g(1-g 2 ) g 2 (1-g 2 ) g,k g -g k 2k 3k n Nested+Cascade Allpass 50,0.5 30,0.7 20,0.3 Synthetic Reverb + AP 0 AP 1 AP 2 g LPF + + a 0 a 1 a 2 E6820 SAPR - Dan Ellis L08 - Spatilal sound

22 Synthetic binaural audio Source convolved with {L,R} HRTFs gives precise positioning -...for headphone presentation - can combine multiple sources (by adding) Where to get HRTFs? - measured set, but: specific to individual, discrete - interpolate by linear crossfade, PCA basis set - or: parametric model - delay, shadow, pinna Delay Shadow Pinna z -t DL (θ) 1 - b L (θ)z az t Σ p kl (θ,φ) z -t PkL(θ,φ) + Source Room echo K E z -t E z -t DR (θ) 1 - b R (θ)z az t Σ p kr (θ,φ) z -t PkR(θ,φ) Head motion cues? - head tracking + fast updates + (after Brown & Duda '97) E6820 SAPR - Dan Ellis L08 - Spatilal sound

23 Transaural sound Binaural signals without headphones? Can cross-cancel wrap-around signals - speakers S L,R, ears E L,R, binaural signals B L,R. 1 S L = H LL ( B L H RL S R ) B L M B R 1 S R = H RR ( B R H LR S L ) S L S R H LR H RL H LL H RR E L E R Narrow sweet spot - head motion? E6820 SAPR - Dan Ellis L08 - Spatilal sound

24 Soundfield reconstruction Stop thinking about ears just reconstruct pressure + spatial derivatives p(t) / z p(t) / y p(t) / x p(x,y,z,t) - ears in reconstructed field receive same sounds Complex reconstruction setup (ambisonics) - able to preserve head motion cues? E6820 SAPR - Dan Ellis L08 - Spatilal sound

25 Outline Spatial acoustics Binaural perception Synthesizing spatial audio Extracting spatial sounds - Microphone arrays - Modeling binaural processing E6820 SAPR - Dan Ellis L08 - Spatilal sound

26 4 Extracting spatial sounds Given access to soundfield, can we recover separate components? - degrees of freedom: >N signals from N sensors is hard - but: people can do it (somewhat) Information-theoretic approach - use only very general constraints - rely on precision measurements Anthropic approach - examine human perception - attempt to use same information E6820 SAPR - Dan Ellis L08 - Spatilal sound

27 Microphone arrays Signals from multiple microphones can be combined to enhance/cancel certain sources Coincident mics with diff. directional gains m 1 s 1 a a a 22 a 21 m 2 s 2 m 1 a 11 a 12 s = 1 m 2 a 21 a 22 s 2 s 1ˆ = A 1 s 2ˆ m Microphone arrays (endfire) 0 λ = 4D -20 λ = 2D -40 λ = D + D + D + D E6820 SAPR - Dan Ellis L08 - Spatilal sound

28 Adaptive Beamforming & Independent Component Analysis (ICA) Formulate mathematical criteria to optimize Beamforming: Drive interference to zero - cancel energy during nontarget intervals ICA: maximize mutual independence of outputs - from higher-order moments during overlap m 1 a 11 a s x 12 1 m 2 a 21 a 22 s 2 δ MutInfo δa Limited by separation model parameter space - only NxN? E6820 SAPR - Dan Ellis L08 - Spatilal sound

29 Binaural models Human listeners do better? - certainly given only 2 channels Extract ITD and IID cues? Target azimuth Center freq / Hz Interaural cross-correlation lag / ms - cross-correlation finds timing differences - consume counter-moving pulses - how to achieve IID, trading - vertical cues... E6820 SAPR - Dan Ellis L08 - Spatilal sound

30 Nonlinear filtering How to separate sounds based on direction? - estimate direction locally - choose target direction - remove energy from other directions E.g. Kollmeier, Peissig & Hohman 93 frequency FFT Modulus l analysis L w L w 2 Smooth (1-a) (1-az -1 ) S LL OLA- FFT synthesis l' Crosscorrelation L w R * w Smooth (1-a) (1-az -1 ) S LR Gain factor calc g time FFT Modulus r analysis R w R w 2 Smooth (1-a) (1-az -1 ) S RR OLA- FFT synthesis r' X w (mh,2 πk/nt) - IID from L w / R w ; ITD (IPD) from arg{l w R * w } - match to IID/IPD template for desired direction - also reverberation? E6820 SAPR - Dan Ellis L08 - Spatilal sound

31 Summary Spatial sound - sampling at more than one point gives information on origin direction Binaural perception - time & intensity cues used between/within ears Sound rendering - conventional stereo - HRTF-based Spatial analysis - optimal linear techniques - elusive auditory models E6820 SAPR - Dan Ellis L08 - Spatilal sound

32 References B.C.J. Moore, An introduction to the psychology of hearing (4th ed.) Academic, J. Blauert, Spatial Hearing (revised ed.), MIT Press, R.O. Duda, Sound Localization Research, E6820 SAPR - Dan Ellis L08 - Spatilal sound

Binaural Hearing. Why two ears? Definitions

Binaural Hearing. Why two ears? Definitions Binaural Hearing Why two ears? Locating sounds in space: acuity is poorer than in vision by up to two orders of magnitude, but extends in all directions. Role in alerting and orienting? Separating sound

More information

Sound localization psychophysics

Sound localization psychophysics Sound localization psychophysics Eric Young A good reference: B.C.J. Moore An Introduction to the Psychology of Hearing Chapter 7, Space Perception. Elsevier, Amsterdam, pp. 233-267 (2004). Sound localization:

More information

3-D Sound and Spatial Audio. What do these terms mean?

3-D Sound and Spatial Audio. What do these terms mean? 3-D Sound and Spatial Audio What do these terms mean? Both terms are very general. 3-D sound usually implies the perception of point sources in 3-D space (could also be 2-D plane) whether the audio reproduction

More information

Using Source Models in Speech Separation

Using Source Models in Speech Separation Using Source Models in Speech Separation Dan Ellis Laboratory for Recognition and Organization of Speech and Audio Dept. Electrical Eng., Columbia Univ., NY USA dpwe@ee.columbia.edu http://labrosa.ee.columbia.edu/

More information

IN EAR TO OUT THERE: A MAGNITUDE BASED PARAMETERIZATION SCHEME FOR SOUND SOURCE EXTERNALIZATION. Griffin D. Romigh, Brian D. Simpson, Nandini Iyer

IN EAR TO OUT THERE: A MAGNITUDE BASED PARAMETERIZATION SCHEME FOR SOUND SOURCE EXTERNALIZATION. Griffin D. Romigh, Brian D. Simpson, Nandini Iyer IN EAR TO OUT THERE: A MAGNITUDE BASED PARAMETERIZATION SCHEME FOR SOUND SOURCE EXTERNALIZATION Griffin D. Romigh, Brian D. Simpson, Nandini Iyer 711th Human Performance Wing Air Force Research Laboratory

More information

21/01/2013. Binaural Phenomena. Aim. To understand binaural hearing Objectives. Understand the cues used to determine the location of a sound source

21/01/2013. Binaural Phenomena. Aim. To understand binaural hearing Objectives. Understand the cues used to determine the location of a sound source Binaural Phenomena Aim To understand binaural hearing Objectives Understand the cues used to determine the location of a sound source Understand sensitivity to binaural spatial cues, including interaural

More information

Binaural Hearing. Steve Colburn Boston University

Binaural Hearing. Steve Colburn Boston University Binaural Hearing Steve Colburn Boston University Outline Why do we (and many other animals) have two ears? What are the major advantages? What is the observed behavior? How do we accomplish this physiologically?

More information

AUDL GS08/GAV1 Signals, systems, acoustics and the ear. Pitch & Binaural listening

AUDL GS08/GAV1 Signals, systems, acoustics and the ear. Pitch & Binaural listening AUDL GS08/GAV1 Signals, systems, acoustics and the ear Pitch & Binaural listening Review 25 20 15 10 5 0-5 100 1000 10000 25 20 15 10 5 0-5 100 1000 10000 Part I: Auditory frequency selectivity Tuning

More information

3-D SOUND IMAGE LOCALIZATION BY INTERAURAL DIFFERENCES AND THE MEDIAN PLANE HRTF. Masayuki Morimoto Motokuni Itoh Kazuhiro Iida

3-D SOUND IMAGE LOCALIZATION BY INTERAURAL DIFFERENCES AND THE MEDIAN PLANE HRTF. Masayuki Morimoto Motokuni Itoh Kazuhiro Iida 3-D SOUND IMAGE LOCALIZATION BY INTERAURAL DIFFERENCES AND THE MEDIAN PLANE HRTF Masayuki Morimoto Motokuni Itoh Kazuhiro Iida Kobe University Environmental Acoustics Laboratory Rokko, Nada, Kobe, 657-8501,

More information

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 THE DUPLEX-THEORY OF LOCALIZATION INVESTIGATED UNDER NATURAL CONDITIONS

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 THE DUPLEX-THEORY OF LOCALIZATION INVESTIGATED UNDER NATURAL CONDITIONS 19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 27 THE DUPLEX-THEORY OF LOCALIZATION INVESTIGATED UNDER NATURAL CONDITIONS PACS: 43.66.Pn Seeber, Bernhard U. Auditory Perception Lab, Dept.

More information

Computational Perception /785. Auditory Scene Analysis

Computational Perception /785. Auditory Scene Analysis Computational Perception 15-485/785 Auditory Scene Analysis A framework for auditory scene analysis Auditory scene analysis involves low and high level cues Low level acoustic cues are often result in

More information

Binaural processing of complex stimuli

Binaural processing of complex stimuli Binaural processing of complex stimuli Outline for today Binaural detection experiments and models Speech as an important waveform Experiments on understanding speech in complex environments (Cocktail

More information

Literature Overview - Digital Hearing Aids and Group Delay - HADF, June 2017, P. Derleth

Literature Overview - Digital Hearing Aids and Group Delay - HADF, June 2017, P. Derleth Literature Overview - Digital Hearing Aids and Group Delay - HADF, June 2017, P. Derleth Historic Context Delay in HI and the perceptual effects became a topic with the widespread market introduction of

More information

Lecture 8: Spatial sound. Spatial acoustics

Lecture 8: Spatial sound. Spatial acoustics EE E682: Speech & Audio Processing & Recogniion Lecure 8: Spaial sound 1 2 3 4 Spaial acousics Binaural percepion Synhesizing spaial audio Exracing spaial sounds Dan Ellis hp://www.ee.columbia.edu/~dpwe/e682/

More information

Abstract. 1. Introduction. David Spargo 1, William L. Martens 2, and Densil Cabrera 3

Abstract. 1. Introduction. David Spargo 1, William L. Martens 2, and Densil Cabrera 3 THE INFLUENCE OF ROOM REFLECTIONS ON SUBWOOFER REPRODUCTION IN A SMALL ROOM: BINAURAL INTERACTIONS PREDICT PERCEIVED LATERAL ANGLE OF PERCUSSIVE LOW- FREQUENCY MUSICAL TONES Abstract David Spargo 1, William

More information

Effect of spectral content and learning on auditory distance perception

Effect of spectral content and learning on auditory distance perception Effect of spectral content and learning on auditory distance perception Norbert Kopčo 1,2, Dávid Čeljuska 1, Miroslav Puszta 1, Michal Raček 1 a Martin Sarnovský 1 1 Department of Cybernetics and AI, Technical

More information

Hearing. Juan P Bello

Hearing. Juan P Bello Hearing Juan P Bello The human ear The human ear Outer Ear The human ear Middle Ear The human ear Inner Ear The cochlea (1) It separates sound into its various components If uncoiled it becomes a tapering

More information

Hearing II Perceptual Aspects

Hearing II Perceptual Aspects Hearing II Perceptual Aspects Overview of Topics Chapter 6 in Chaudhuri Intensity & Loudness Frequency & Pitch Auditory Space Perception 1 2 Intensity & Loudness Loudness is the subjective perceptual quality

More information

An Auditory System Modeling in Sound Source Localization

An Auditory System Modeling in Sound Source Localization An Auditory System Modeling in Sound Source Localization Yul Young Park The University of Texas at Austin EE381K Multidimensional Signal Processing May 18, 2005 Abstract Sound localization of the auditory

More information

This will be accomplished using maximum likelihood estimation based on interaural level

This will be accomplished using maximum likelihood estimation based on interaural level Chapter 1 Problem background 1.1 Overview of the proposed work The proposed research consists of the construction and demonstration of a computational model of human spatial hearing, including long term

More information

HEARING AND PSYCHOACOUSTICS

HEARING AND PSYCHOACOUSTICS CHAPTER 2 HEARING AND PSYCHOACOUSTICS WITH LIDIA LEE I would like to lead off the specific audio discussions with a description of the audio receptor the ear. I believe it is always a good idea to understand

More information

Challenges in microphone array processing for hearing aids. Volkmar Hamacher Siemens Audiological Engineering Group Erlangen, Germany

Challenges in microphone array processing for hearing aids. Volkmar Hamacher Siemens Audiological Engineering Group Erlangen, Germany Challenges in microphone array processing for hearing aids Volkmar Hamacher Siemens Audiological Engineering Group Erlangen, Germany SIEMENS Audiological Engineering Group R&D Signal Processing and Audiology

More information

Lecture 3: Perception

Lecture 3: Perception ELEN E4896 MUSIC SIGNAL PROCESSING Lecture 3: Perception 1. Ear Physiology 2. Auditory Psychophysics 3. Pitch Perception 4. Music Perception Dan Ellis Dept. Electrical Engineering, Columbia University

More information

Ear Beamer. Aaron Lucia, Niket Gupta, Matteo Puzella, Nathan Dunn. Department of Electrical and Computer Engineering

Ear Beamer. Aaron Lucia, Niket Gupta, Matteo Puzella, Nathan Dunn. Department of Electrical and Computer Engineering Ear Beamer Aaron Lucia, Niket Gupta, Matteo Puzella, Nathan Dunn The Team Aaron Lucia CSE Niket Gupta CSE Matteo Puzella EE Nathan Dunn CSE Advisor: Prof. Mario Parente 2 A Familiar Scenario 3 Outlining

More information

A NOVEL HEAD-RELATED TRANSFER FUNCTION MODEL BASED ON SPECTRAL AND INTERAURAL DIFFERENCE CUES

A NOVEL HEAD-RELATED TRANSFER FUNCTION MODEL BASED ON SPECTRAL AND INTERAURAL DIFFERENCE CUES A NOVEL HEAD-RELATED TRANSFER FUNCTION MODEL BASED ON SPECTRAL AND INTERAURAL DIFFERENCE CUES Kazuhiro IIDA, Motokuni ITOH AV Core Technology Development Center, Matsushita Electric Industrial Co., Ltd.

More information

Using Speech Models for Separation

Using Speech Models for Separation Using Speech Models for Separation Dan Ellis Comprising the work of Michael Mandel and Ron Weiss Laboratory for Recognition and Organization of Speech and Audio Dept. Electrical Eng., Columbia Univ., NY

More information

Neural System Model of Human Sound Localization

Neural System Model of Human Sound Localization in Advances in Neural Information Processing Systems 13 S.A. Solla, T.K. Leen, K.-R. Müller (eds.), 761 767 MIT Press (2000) Neural System Model of Human Sound Localization Craig T. Jin Department of Physiology

More information

Effect of microphone position in hearing instruments on binaural masking level differences

Effect of microphone position in hearing instruments on binaural masking level differences Effect of microphone position in hearing instruments on binaural masking level differences Fredrik Gran, Jesper Udesen and Andrew B. Dittberner GN ReSound A/S, Research R&D, Lautrupbjerg 7, 2750 Ballerup,

More information

HST.723J, Spring 2005 Theme 3 Report

HST.723J, Spring 2005 Theme 3 Report HST.723J, Spring 2005 Theme 3 Report Madhu Shashanka shashanka@cns.bu.edu Introduction The theme of this report is binaural interactions. Binaural interactions of sound stimuli enable humans (and other

More information

Hearing in the Environment

Hearing in the Environment 10 Hearing in the Environment Click Chapter to edit 10 Master Hearing title in the style Environment Sound Localization Complex Sounds Auditory Scene Analysis Continuity and Restoration Effects Auditory

More information

Spectral and Spatial Parameter Resolution Requirements for Parametric, Filter-Bank-Based HRTF Processing*

Spectral and Spatial Parameter Resolution Requirements for Parametric, Filter-Bank-Based HRTF Processing* Spectral and Spatial Parameter Resolution Requirements for Parametric, Filter-Bank-Based HRTF Processing* JEROEN BREEBAART, 1 AES Member, FABIAN NATER, 2 (jeroen.breebaart@philips.com) (fnater@vision.ee.ethz.ch)

More information

Influence of multi-microphone signal enhancement algorithms on auditory movement detection in acoustically complex situations

Influence of multi-microphone signal enhancement algorithms on auditory movement detection in acoustically complex situations Syddansk Universitet Influence of multi-microphone signal enhancement algorithms on auditory movement detection in acoustically complex situations Lundbeck, Micha; Hartog, Laura; Grimm, Giso; Hohmann,

More information

Sound Localization PSY 310 Greg Francis. Lecture 31. Audition

Sound Localization PSY 310 Greg Francis. Lecture 31. Audition Sound Localization PSY 310 Greg Francis Lecture 31 Physics and psychology. Audition We now have some idea of how sound properties are recorded by the auditory system So, we know what kind of information

More information

Chapter 3. Sounds, Signals, and Studio Acoustics

Chapter 3. Sounds, Signals, and Studio Acoustics Chapter 3 Sounds, Signals, and Studio Acoustics Sound Waves Compression/Rarefaction: speaker cone Sound travels 1130 feet per second Sound waves hit receiver Sound waves tend to spread out as they travel

More information

Brian D. Simpson Veridian, 5200 Springfield Pike, Suite 200, Dayton, Ohio 45431

Brian D. Simpson Veridian, 5200 Springfield Pike, Suite 200, Dayton, Ohio 45431 The effects of spatial separation in distance on the informational and energetic masking of a nearby speech signal Douglas S. Brungart a) Air Force Research Laboratory, 2610 Seventh Street, Wright-Patterson

More information

Localization in the Presence of a Distracter and Reverberation in the Frontal Horizontal Plane: II. Model Algorithms

Localization in the Presence of a Distracter and Reverberation in the Frontal Horizontal Plane: II. Model Algorithms 956 969 Localization in the Presence of a Distracter and Reverberation in the Frontal Horizontal Plane: II. Model Algorithms Jonas Braasch Institut für Kommunikationsakustik, Ruhr-Universität Bochum, Germany

More information

BINAURAL DICHOTIC PRESENTATION FOR MODERATE BILATERAL SENSORINEURAL HEARING-IMPAIRED

BINAURAL DICHOTIC PRESENTATION FOR MODERATE BILATERAL SENSORINEURAL HEARING-IMPAIRED International Conference on Systemics, Cybernetics and Informatics, February 12 15, 2004 BINAURAL DICHOTIC PRESENTATION FOR MODERATE BILATERAL SENSORINEURAL HEARING-IMPAIRED Alice N. Cheeran Biomedical

More information

Audibility of time differences in adjacent head-related transfer functions (HRTFs) Hoffmann, Pablo Francisco F.; Møller, Henrik

Audibility of time differences in adjacent head-related transfer functions (HRTFs) Hoffmann, Pablo Francisco F.; Møller, Henrik Aalborg Universitet Audibility of time differences in adjacent head-related transfer functions (HRTFs) Hoffmann, Pablo Francisco F.; Møller, Henrik Published in: Audio Engineering Society Convention Papers

More information

Recognition & Organization of Speech & Audio

Recognition & Organization of Speech & Audio Recognition & Organization of Speech & Audio Dan Ellis http://labrosa.ee.columbia.edu/ Outline 1 2 3 Introducing Projects in speech, music & audio Summary overview - Dan Ellis 21-9-28-1 1 Sound organization

More information

William A. Yost and Sandra J. Guzman Parmly Hearing Institute, Loyola University Chicago, Chicago, Illinois 60201

William A. Yost and Sandra J. Guzman Parmly Hearing Institute, Loyola University Chicago, Chicago, Illinois 60201 The precedence effect Ruth Y. Litovsky a) and H. Steven Colburn Hearing Research Center and Department of Biomedical Engineering, Boston University, Boston, Massachusetts 02215 William A. Yost and Sandra

More information

Angular Resolution of Human Sound Localization

Angular Resolution of Human Sound Localization Angular Resolution of Human Sound Localization By Simon Skluzacek A senior thesis submitted to the Carthage College Physics & Astronomy Department in partial fulfillment of the requirements for the Bachelor

More information

Binaural Hearing for Robots Introduction to Robot Hearing

Binaural Hearing for Robots Introduction to Robot Hearing Binaural Hearing for Robots Introduction to Robot Hearing 1Radu Horaud Binaural Hearing for Robots 1. Introduction to Robot Hearing 2. Methodological Foundations 3. Sound-Source Localization 4. Machine

More information

Discrimination and identification of azimuth using spectral shape a)

Discrimination and identification of azimuth using spectral shape a) Discrimination and identification of azimuth using spectral shape a) Daniel E. Shub b Speech and Hearing Bioscience and Technology Program, Division of Health Sciences and Technology, Massachusetts Institute

More information

Speech segregation in rooms: Effects of reverberation on both target and interferer

Speech segregation in rooms: Effects of reverberation on both target and interferer Speech segregation in rooms: Effects of reverberation on both target and interferer Mathieu Lavandier a and John F. Culling School of Psychology, Cardiff University, Tower Building, Park Place, Cardiff,

More information

HearIntelligence by HANSATON. Intelligent hearing means natural hearing.

HearIntelligence by HANSATON. Intelligent hearing means natural hearing. HearIntelligence by HANSATON. HearIntelligence by HANSATON. Intelligent hearing means natural hearing. Acoustic environments are complex. We are surrounded by a variety of different acoustic signals, speech

More information

Systems Neuroscience Oct. 16, Auditory system. http:

Systems Neuroscience Oct. 16, Auditory system. http: Systems Neuroscience Oct. 16, 2018 Auditory system http: www.ini.unizh.ch/~kiper/system_neurosci.html The physics of sound Measuring sound intensity We are sensitive to an enormous range of intensities,

More information

HOW TO USE THE SHURE MXA910 CEILING ARRAY MICROPHONE FOR VOICE LIFT

HOW TO USE THE SHURE MXA910 CEILING ARRAY MICROPHONE FOR VOICE LIFT HOW TO USE THE SHURE MXA910 CEILING ARRAY MICROPHONE FOR VOICE LIFT Created: Sept 2016 Updated: June 2017 By: Luis Guerra Troy Jensen The Shure MXA910 Ceiling Array Microphone offers the unique advantage

More information

HCS 7367 Speech Perception

HCS 7367 Speech Perception Long-term spectrum of speech HCS 7367 Speech Perception Connected speech Absolute threshold Males Dr. Peter Assmann Fall 212 Females Long-term spectrum of speech Vowels Males Females 2) Absolute threshold

More information

J. Acoust. Soc. Am. 116 (2), August /2004/116(2)/1057/9/$ Acoustical Society of America

J. Acoust. Soc. Am. 116 (2), August /2004/116(2)/1057/9/$ Acoustical Society of America The role of head-induced interaural time and level differences in the speech reception threshold for multiple interfering sound sources John F. Culling a) School of Psychology, Cardiff University, P.O.

More information

Lecture 9: Speech Recognition: Front Ends

Lecture 9: Speech Recognition: Front Ends EE E682: Speech & Audio Processing & Recognition Lecture 9: Speech Recognition: Front Ends 1 2 Recognizing Speech Feature Calculation Dan Ellis http://www.ee.columbia.edu/~dpwe/e682/

More information

J Jeffress model, 3, 66ff

J Jeffress model, 3, 66ff Index A Absolute pitch, 102 Afferent projections, inferior colliculus, 131 132 Amplitude modulation, coincidence detector, 152ff inferior colliculus, 152ff inhibition models, 156ff models, 152ff Anatomy,

More information

Systems for Improvement of the Communication in Passenger Compartments

Systems for Improvement of the Communication in Passenger Compartments Systems for Improvement of the Communication in Passenger Compartments Tim Haulick/ Gerhard Schmidt thaulick@harmanbecker.com ETSI Workshop on Speech and Noise in Wideband Communication 22nd and 23rd May

More information

Spatial hearing and sound localization mechanisms in the brain. Henri Pöntynen February 9, 2016

Spatial hearing and sound localization mechanisms in the brain. Henri Pöntynen February 9, 2016 Spatial hearing and sound localization mechanisms in the brain Henri Pöntynen February 9, 2016 Outline Auditory periphery: from acoustics to neural signals - Basilar membrane - Organ of Corti Spatial

More information

Frequency Tracking: LMS and RLS Applied to Speech Formant Estimation

Frequency Tracking: LMS and RLS Applied to Speech Formant Estimation Aldebaro Klautau - http://speech.ucsd.edu/aldebaro - 2/3/. Page. Frequency Tracking: LMS and RLS Applied to Speech Formant Estimation ) Introduction Several speech processing algorithms assume the signal

More information

Synthesis of Spatially Extended Virtual Sources with Time-Frequency Decomposition of Mono Signals

Synthesis of Spatially Extended Virtual Sources with Time-Frequency Decomposition of Mono Signals PAPERS Synthesis of Spatially Extended Virtual Sources with Time-Frequency Decomposition of Mono Signals TAPANI PIHLAJAMÄKI, AES Student Member, OLLI SANTALA, AES Student Member, AND (tapani.pihlajamaki@aalto.fi)

More information

THE PHYSICAL AND PSYCHOPHYSICAL BASIS OF SOUND LOCALIZATION

THE PHYSICAL AND PSYCHOPHYSICAL BASIS OF SOUND LOCALIZATION CHAPTER 2 THE PHYSICAL AND PSYCHOPHYSICAL BASIS OF SOUND LOCALIZATION Simon Carlile 1. PHYSICAL CUES TO A SOUND S LOCATION 1.1. THE DUPLEX THEORY OF AUDITORY LOCALIZATION Traditionally, the principal cues

More information

Role of F0 differences in source segregation

Role of F0 differences in source segregation Role of F0 differences in source segregation Andrew J. Oxenham Research Laboratory of Electronics, MIT and Harvard-MIT Speech and Hearing Bioscience and Technology Program Rationale Many aspects of segregation

More information

Lateralized speech perception in normal-hearing and hearing-impaired listeners and its relationship to temporal processing

Lateralized speech perception in normal-hearing and hearing-impaired listeners and its relationship to temporal processing Lateralized speech perception in normal-hearing and hearing-impaired listeners and its relationship to temporal processing GUSZTÁV LŐCSEI,*, JULIE HEFTING PEDERSEN, SØREN LAUGESEN, SÉBASTIEN SANTURETTE,

More information

The basic mechanisms of directional sound localization are well documented. In the horizontal plane, interaural difference INTRODUCTION I.

The basic mechanisms of directional sound localization are well documented. In the horizontal plane, interaural difference INTRODUCTION I. Auditory localization of nearby sources. II. Localization of a broadband source Douglas S. Brungart, a) Nathaniel I. Durlach, and William M. Rabinowitz b) Research Laboratory of Electronics, Massachusetts

More information

Trading Directional Accuracy for Realism in a Virtual Auditory Display

Trading Directional Accuracy for Realism in a Virtual Auditory Display Trading Directional Accuracy for Realism in a Virtual Auditory Display Barbara G. Shinn-Cunningham, I-Fan Lin, and Tim Streeter Hearing Research Center, Boston University 677 Beacon St., Boston, MA 02215

More information

A Microphone-Array-Based System for Restoring Sound Localization with Occluded Ears

A Microphone-Array-Based System for Restoring Sound Localization with Occluded Ears Restoring Sound Localization with Occluded Ears Adelbert W. Bronkhorst TNO Human Factors P.O. Box 23, 3769 ZG Soesterberg The Netherlands adelbert.bronkhorst@tno.nl Jan A. Verhave TNO Human Factors P.O.

More information

Discrete Signal Processing

Discrete Signal Processing 1 Discrete Signal Processing C.M. Liu Perceptual Lab, College of Computer Science National Chiao-Tung University http://www.cs.nctu.edu.tw/~cmliu/courses/dsp/ ( Office: EC538 (03)5731877 cmliu@cs.nctu.edu.tw

More information

Auditory System & Hearing

Auditory System & Hearing Auditory System & Hearing Chapters 9 and 10 Lecture 17 Jonathan Pillow Sensation & Perception (PSY 345 / NEU 325) Spring 2015 1 Cochlea: physical device tuned to frequency! place code: tuning of different

More information

Evidence base for hearing aid features:

Evidence base for hearing aid features: Evidence base for hearing aid features: { the ʹwhat, how and whyʹ of technology selection, fitting and assessment. Drew Dundas, PhD Director of Audiology, Clinical Assistant Professor of Otolaryngology

More information

Perceptual Plasticity in Spatial Auditory Displays

Perceptual Plasticity in Spatial Auditory Displays Perceptual Plasticity in Spatial Auditory Displays BARBARA G. SHINN-CUNNINGHAM, TIMOTHY STREETER, and JEAN-FRANÇOIS GYSS Hearing Research Center, Boston University Often, virtual acoustic environments

More information

The role of low frequency components in median plane localization

The role of low frequency components in median plane localization Acoust. Sci. & Tech. 24, 2 (23) PAPER The role of low components in median plane localization Masayuki Morimoto 1;, Motoki Yairi 1, Kazuhiro Iida 2 and Motokuni Itoh 1 1 Environmental Acoustics Laboratory,

More information

Auditory Scene Analysis: phenomena, theories and computational models

Auditory Scene Analysis: phenomena, theories and computational models Auditory Scene Analysis: phenomena, theories and computational models July 1998 Dan Ellis International Computer Science Institute, Berkeley CA Outline 1 2 3 4 The computational

More information

HAT Process: Determining HAT for a Student

HAT Process: Determining HAT for a Student HAT Process: Determining HAT for a Student (Approved DSR or request Considerations for IEP Team) Audio & HI / TC determine who will be on team to determine HAT Team completes Needs Identification section

More information

Auditory scene analysis in humans: Implications for computational implementations.

Auditory scene analysis in humans: Implications for computational implementations. Auditory scene analysis in humans: Implications for computational implementations. Albert S. Bregman McGill University Introduction. The scene analysis problem. Two dimensions of grouping. Recognition

More information

Juha Merimaa b) Institut für Kommunikationsakustik, Ruhr-Universität Bochum, Germany

Juha Merimaa b) Institut für Kommunikationsakustik, Ruhr-Universität Bochum, Germany Source localization in complex listening situations: Selection of binaural cues based on interaural coherence Christof Faller a) Mobile Terminals Division, Agere Systems, Allentown, Pennsylvania Juha Merimaa

More information

Computational Auditory Scene Analysis: An overview and some observations. CASA survey. Other approaches

Computational Auditory Scene Analysis: An overview and some observations. CASA survey. Other approaches CASA talk - Haskins/NUWC - Dan Ellis 1997oct24/5-1 The importance of auditory illusions for artificial listeners 1 Dan Ellis International Computer Science Institute, Berkeley CA

More information

Improved method for accurate sound localization

Improved method for accurate sound localization Acoust. Sci. & Tech. 7, 3 () PAPER Improved method for accurate sound localization Akihiro Kudo, Hiroshi Higuchi, Haruhide Hokari and Shoji Shimada Nagaoka University of Technology, 13 1 Kamitomioka-machi,

More information

Localization: Give your patients a listening edge

Localization: Give your patients a listening edge Localization: Give your patients a listening edge For those of us with healthy auditory systems, localization skills are often taken for granted. We don t even notice them, until they are no longer working.

More information

Binaural synthesis Møller, Henrik; Jensen, Clemen Boje; Hammershøi, Dorte; Sørensen, Michael Friis

Binaural synthesis Møller, Henrik; Jensen, Clemen Boje; Hammershøi, Dorte; Sørensen, Michael Friis Aalborg Universitet Binaural synthesis Møller, Henrik; Jensen, Clemen Boje; Hammershøi, Dorte; Sørensen, Michael Friis Published in: Proceedings of 15th International Congress on Acoustics, ICA'95, Trondheim,

More information

INTRODUCTION J. Acoust. Soc. Am. 103 (2), February /98/103(2)/1080/5/$ Acoustical Society of America 1080

INTRODUCTION J. Acoust. Soc. Am. 103 (2), February /98/103(2)/1080/5/$ Acoustical Society of America 1080 Perceptual segregation of a harmonic from a vowel by interaural time difference in conjunction with mistuning and onset asynchrony C. J. Darwin and R. W. Hukin Experimental Psychology, University of Sussex,

More information

Localization in the Presence of a Distracter and Reverberation in the Frontal Horizontal Plane. I. Psychoacoustical Data

Localization in the Presence of a Distracter and Reverberation in the Frontal Horizontal Plane. I. Psychoacoustical Data 942 955 Localization in the Presence of a Distracter and Reverberation in the Frontal Horizontal Plane. I. Psychoacoustical Data Jonas Braasch, Klaus Hartung Institut für Kommunikationsakustik, Ruhr-Universität

More information

THE RELATION BETWEEN SPATIAL IMPRESSION AND THE PRECEDENCE EFFECT. Masayuki Morimoto

THE RELATION BETWEEN SPATIAL IMPRESSION AND THE PRECEDENCE EFFECT. Masayuki Morimoto THE RELATION BETWEEN SPATIAL IMPRESSION AND THE PRECEDENCE EFFECT Masayuki Morimoto Environmental Acoustics Laboratory, Faculty of Engineering, Kobe University Rokko Nada Kobe 657-85 Japan mrmt@kobe-u.ac.jp

More information

BASIC NOTIONS OF HEARING AND

BASIC NOTIONS OF HEARING AND BASIC NOTIONS OF HEARING AND PSYCHOACOUSICS Educational guide for the subject Communication Acoustics VIHIAV 035 Fülöp Augusztinovicz Dept. of Networked Systems and Services fulop@hit.bme.hu 2018. október

More information

Sound, Mixtures, and Learning

Sound, Mixtures, and Learning Sound, Mixtures, and Learning Dan Ellis Laboratory for Recognition and Organization of Speech and Audio (LabROSA) Electrical Engineering, Columbia University http://labrosa.ee.columbia.edu/

More information

Signals, systems, acoustics and the ear. Week 5. The peripheral auditory system: The ear as a signal processor

Signals, systems, acoustics and the ear. Week 5. The peripheral auditory system: The ear as a signal processor Signals, systems, acoustics and the ear Week 5 The peripheral auditory system: The ear as a signal processor Think of this set of organs 2 as a collection of systems, transforming sounds to be sent to

More information

The effect of wearing conventional and level-dependent hearing protectors on speech production in noise and quiet

The effect of wearing conventional and level-dependent hearing protectors on speech production in noise and quiet The effect of wearing conventional and level-dependent hearing protectors on speech production in noise and quiet Ghazaleh Vaziri Christian Giguère Hilmi R. Dajani Nicolas Ellaham Annual National Hearing

More information

HCS 7367 Speech Perception

HCS 7367 Speech Perception Babies 'cry in mother's tongue' HCS 7367 Speech Perception Dr. Peter Assmann Fall 212 Babies' cries imitate their mother tongue as early as three days old German researchers say babies begin to pick up

More information

Speech intelligibility in simulated acoustic conditions for normal hearing and hearing-impaired listeners

Speech intelligibility in simulated acoustic conditions for normal hearing and hearing-impaired listeners Speech intelligibility in simulated acoustic conditions for normal hearing and hearing-impaired listeners Ir i s Arw e i l e r 1, To r b e n Po u l s e n 2, a n d To r s t e n Da u 1 1 Centre for Applied

More information

Publication VI. c 2007 Audio Engineering Society. Reprinted with permission.

Publication VI. c 2007 Audio Engineering Society. Reprinted with permission. VI Publication VI Hirvonen, T. and Pulkki, V., Predicting Binaural Masking Level Difference and Dichotic Pitch Using Instantaneous ILD Model, AES 30th Int. Conference, 2007. c 2007 Audio Engineering Society.

More information

16.400/453J Human Factors Engineering /453. Audition. Prof. D. C. Chandra Lecture 14

16.400/453J Human Factors Engineering /453. Audition. Prof. D. C. Chandra Lecture 14 J Human Factors Engineering Audition Prof. D. C. Chandra Lecture 14 1 Overview Human ear anatomy and hearing Auditory perception Brainstorming about sounds Auditory vs. visual displays Considerations for

More information

Auditory principles in speech processing do computers need silicon ears?

Auditory principles in speech processing do computers need silicon ears? * with contributions by V. Hohmann, M. Kleinschmidt, T. Brand, J. Nix, R. Beutelmann, and more members of our medical physics group Prof. Dr. rer.. nat. Dr. med. Birger Kollmeier* Auditory principles in

More information

Acoustic Sensing With Artificial Intelligence

Acoustic Sensing With Artificial Intelligence Acoustic Sensing With Artificial Intelligence Bowon Lee Department of Electronic Engineering Inha University Incheon, South Korea bowon.lee@inha.ac.kr bowon.lee@ieee.org NVIDIA Deep Learning Day Seoul,

More information

Acoustics Research Institute

Acoustics Research Institute Austrian Academy of Sciences Acoustics Research Institute Modeling Modelingof ofauditory AuditoryPerception Perception Bernhard BernhardLaback Labackand andpiotr PiotrMajdak Majdak http://www.kfs.oeaw.ac.at

More information

Representation of sound in the auditory nerve

Representation of sound in the auditory nerve Representation of sound in the auditory nerve Eric D. Young Department of Biomedical Engineering Johns Hopkins University Young, ED. Neural representation of spectral and temporal information in speech.

More information

Advanced Audio Interface for Phonetic Speech. Recognition in a High Noise Environment

Advanced Audio Interface for Phonetic Speech. Recognition in a High Noise Environment DISTRIBUTION STATEMENT A Approved for Public Release Distribution Unlimited Advanced Audio Interface for Phonetic Speech Recognition in a High Noise Environment SBIR 99.1 TOPIC AF99-1Q3 PHASE I SUMMARY

More information

Digital. hearing instruments have burst on the

Digital. hearing instruments have burst on the Testing Digital and Analog Hearing Instruments: Processing Time Delays and Phase Measurements A look at potential side effects and ways of measuring them by George J. Frye Digital. hearing instruments

More information

On the influence of interaural differences on onset detection in auditory object formation. 1 Introduction

On the influence of interaural differences on onset detection in auditory object formation. 1 Introduction On the influence of interaural differences on onset detection in auditory object formation Othmar Schimmel Eindhoven University of Technology, P.O. Box 513 / Building IPO 1.26, 56 MD Eindhoven, The Netherlands,

More information

LabROSA Research Overview

LabROSA Research Overview LabROSA Research Overview Dan Ellis Laboratory for Recognition and Organization of Speech and Audio Dept. Electrical Eng., Columbia Univ., NY USA dpwe@ee.columbia.edu http://labrosa.ee.columbia.edu/ 1.

More information

The development of a modified spectral ripple test

The development of a modified spectral ripple test The development of a modified spectral ripple test Justin M. Aronoff a) and David M. Landsberger Communication and Neuroscience Division, House Research Institute, 2100 West 3rd Street, Los Angeles, California

More information

Two Modified IEC Ear Simulators for Extended Dynamic Range

Two Modified IEC Ear Simulators for Extended Dynamic Range Two Modified IEC 60318-4 Ear Simulators for Extended Dynamic Range Peter Wulf-Andersen & Morten Wille The international standard IEC 60318-4 specifies an occluded ear simulator, often referred to as a

More information

The use of interaural time and level difference cues by bilateral cochlear implant users

The use of interaural time and level difference cues by bilateral cochlear implant users The use of interaural time and level difference cues by bilateral cochlear implant users Justin M. Aronoff, a) Yang-soo Yoon, and Daniel J. Freed b) Communication and Neuroscience Division, House Ear Institute,

More information

White Paper: micon Directivity and Directional Speech Enhancement

White Paper: micon Directivity and Directional Speech Enhancement White Paper: micon Directivity and Directional Speech Enhancement www.siemens.com Eghart Fischer, Henning Puder, Ph. D., Jens Hain Abstract: This paper describes a new, optimized directional processing

More information

ICaD 2013 ADJUSTING THE PERCEIVED DISTANCE OF VIRTUAL SPEECH SOURCES BY MODIFYING BINAURAL ROOM IMPULSE RESPONSES

ICaD 2013 ADJUSTING THE PERCEIVED DISTANCE OF VIRTUAL SPEECH SOURCES BY MODIFYING BINAURAL ROOM IMPULSE RESPONSES ICaD 213 6 1 july, 213, Łódź, Poland international Conference on auditory Display ADJUSTING THE PERCEIVED DISTANCE OF VIRTUAL SPEECH SOURCES BY MODIFYING BINAURAL ROOM IMPULSE RESPONSES Robert Albrecht

More information

Sound Analysis Research at LabROSA

Sound Analysis Research at LabROSA Sound Analysis Research at LabROSA Dan Ellis Laboratory for Recognition and Organization of Speech and Audio Dept. Electrical Eng., Columbia Univ., NY USA dpwe@ee.columbia.edu http://labrosa.ee.columbia.edu/

More information

Sound source localization in real sound fields based on empirical statistics of interaural parameters a)

Sound source localization in real sound fields based on empirical statistics of interaural parameters a) Sound source localization in real sound fields based on empirical statistics of interaural parameters a) Johannes Nix b and Volker Hohmann Medizinische Physik, Carl von Ossietzky Universität Oldenburg,

More information