Robust Neural Encoding of Speech in Human Auditory Cortex
|
|
- Chester Lloyd
- 5 years ago
- Views:
Transcription
1 Robust Neural Encoding of Speech in Human Auditory Cortex Nai Ding, Jonathan Z. Simon Electrical Engineering / Biology University of Maryland, College Park
2 Auditory Processing in Natural Scenes How is the stable perception of sound generated from degraded acoustics?
3 Auditory Processing in Natural Scenes How is the stable perception of sound generated from degraded acoustics? Magnetoencephalography (MEG) MEG measures spatially synchronized dendritic current.
4 Outline Cortical Encoding of Speech in MEG Representation of Spectro-temporal Features Cortical Code despite Energetic Masking Speech in Stationary Noise Cortical Code despite Informational Masking Segregation of Simultaneous Speakers
5 MEG Response to Speech Speech Stimulus MEG Response frequency time time
6 MEG Response to Speech Speech Stimulus STRF MEG Response frequency time time correlation Predictive Power Large-scale synchronized cortical activity is phase locked to slow temporal modulations of speech frequency (Hz) Ding & Simon (in press) J. Neurophysiol.
7 Neural Reconstruction The temporal envelope of speech can be reconstructed from the MEG response. stimulus speech envelope speech envelope reconstructed from MEG response 2 seconds Subject: R1747
8 Outline Cortical Encoding of Speech in MEG Representation of Spectro-temporal Features Neural Coding under Energetic Masking Speech in Stationary Noise Neural Coding under Informational Masking Segregation of Simultaneous Speakers
9 Speech Embedded in Noise Clean Speech SNR: 6 db SNR: -2 db SNR: -9 db Spectrogram Intelligibility: 100 % 70 % 5 % Envelope 6 db 1 second 10 participants; 2 minutes of stimulus in each condition
10 Neural Reconstruction of Speech The temporal envelope of the underlying speech is reconstructed neurally from cortical response. +6 db Reconstruction Accuracy 1 s -6 db Neural Reconstruction Speech Envelope Correlation C SNR
11 Contrast Gain Control Neural compensation for noise-induced loss of stimulus contrast Amplitude-Intensity Function Amplitude Growth Rate response 30 db stimulus 12 db C SNR
12 Adaptive Encoding of Modulations power Modulation Spectrum of Stimulus 0 18 db frequency (Hz) Noise noisier Speech coherence Response Spectrum Noise contains more energy at higher.2 modulation rate, and therefore interfere with speech more at.1 high modulation rates. frequency (Hz)
13 Adaptive Encoding of Modulations power Neural sensitivity profile shifts away from the modulation rates heavily corrupted by noise. Modulation Spectrum of Stimulus 18 db noisier coherence.2.1 Response Spectrum Cutoff Frequency Hz C SNR frequency (Hz) frequency (Hz)
14 Outline Cortical Encoding of Speech in MEG Representation of Spectro-temporal Features Neural Coding under Energetic Masking Speech in Stationary Noise Neural Coding under Informational Masking Segregation of Simultaneous Speakers
15 Diotic Speech Segregation Two speakers, one male and one female, were mixed and presented diotically. The subjects were instructed to focus on one or the other speaker. The MEG response is modeled using two STRFs, one for each speaker. speech mixture Stream 1 Stream 2 MEG signal
16 Neural Unmixing of Concurrent Speakers frequency (khz) frequency (khz) Attended Unattended time (ms) Neurally decoded envelope is more correlated with the attended speaker in >90% of single trials. Correlation Attended P << Unattended
17 Summary 1. Neural processing adapts to noise. 2. Simultaneous speakers can be neurally segregated and processed differently. 3. Cortical encoding is precise yet dynamic: modulated by both stimulus acoustics (bottom-up) and attention (top-down), and leading to a robust encoding of speech in natural scenes.
18 Acknowledgement We thank Stephen David, David Poeppel, Mary Howard, Shihab Shamma, and Monita Chatterjee for discussions! SfN poster: /KK6 (Sunday, 10-11) Contact: Nai Ding Jonathan Z. Simon
19 Thank you!
20 Adaptive Encoding of Modulations Neural sensitivity profile shifts away from the modulation rates heavily corrupted by noise. Modulation Spectrum of Stimulus Response Spectrum 18 db noisier coherence frequency (Hz) frequency (Hz)
21 STRF from MEG and LFP Frequency (khz) MEG STRF Time (s) f (khz) LFP from ferret AI LFP time (s) (in collaboration with Stephen David and Shihab Shamma)
Cortical Encoding of Auditory Objects at the Cocktail Party. Jonathan Z. Simon University of Maryland
Cortical Encoding of Auditory Objects at the Cocktail Party Jonathan Z. Simon University of Maryland ARO Presidential Symposium, February 2013 Introduction Auditory Objects Magnetoencephalography (MEG)
More informationNeural Representations of the Cocktail Party in Human Auditory Cortex
Neural Representations of the Cocktail Party in Human Auditory Cortex Jonathan Z. Simon Department of Electrical & Computer Engineering Department of Biology Institute for Systems Research University of
More informationCortical Encoding of Auditory Objects in the Cocktail Party Problem. Jonathan Z. Simon University of Maryland
Cortical Encoding of Auditory Objects in the Cocktail Party Problem Jonathan Z. Simon University of Maryland Introduction Auditory Objects Magnetoencephalography (MEG) Decoding Neural Signals/Encoding
More informationNeural Representations of the Cocktail Party in Human Auditory Cortex
Neural Representations of the Cocktail Party in Human Auditory Cortex Jonathan Z. Simon Department of Biology Department of Electrical & Computer Engineering Institute for Systems Research University of
More informationNeural Representations of the Cocktail Party in Human Auditory Cortex
Neural Representations of the Cocktail Party in Human Auditory Cortex Jonathan Z. Simon Department of Biology Department of Electrical & Computer Engineering Institute for Systems Research University of
More informationNeural Representations of Speech, and Speech in Noise, in Human Auditory Cortex
Neural Representations of Speech, and Speech in Noise, in Human Auditory Cortex Jonathan Z. Simon Department of Biology Department of Electrical & Computer Engineering Institute for Systems Research University
More informationNeural Representations of Speech at the Cocktail Party in Human Auditory Cortex
Neural Representations of Speech at the Cocktail Party in Human Auditory Cortex Jonathan Z. Simon Department of Electrical & Computer Engineering Department of Biology Institute for Systems Research University
More informationCompeting Streams at the Cocktail Party
Competing Streams at the Cocktail Party A Neural and Behavioral Study of Auditory Attention Jonathan Z. Simon Neuroscience and Cognitive Sciences / Biology / Electrical & Computer Engineering University
More informationNeural Representations of Speech in Human Auditory Cortex
Neural Representations of Speech in Human Auditory Cortex Jonathan Z. Simon Department of Electrical & Computer Engineering Department of Biology Institute for Systems Research University of Maryland http://www.isr.umd.edu/labs/cssl/simonlab
More informationOver-representation of speech in older adults originates from early response in higher order auditory cortex
Over-representation of speech in older adults originates from early response in higher order auditory cortex Christian Brodbeck, Alessandro Presacco, Samira Anderson & Jonathan Z. Simon Overview 2 Puzzle
More informationEffects of aging on temporal synchronization of speech in noise investigated in the cortex by using MEG and in the midbrain by using EEG techniques
Hearing Brain Lab Computational Sensorimotor Systems Lab Effects of aging on temporal synchronization of speech in noise investigated in the cortex by using MEG and in the midbrain by using EEG techniques
More informationUSING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES
USING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES Varinthira Duangudom and David V Anderson School of Electrical and Computer Engineering, Georgia Institute of Technology Atlanta, GA 30332
More informationComputational Perception /785. Auditory Scene Analysis
Computational Perception 15-485/785 Auditory Scene Analysis A framework for auditory scene analysis Auditory scene analysis involves low and high level cues Low level acoustic cues are often result in
More informationChapter 40 Effects of Peripheral Tuning on the Auditory Nerve s Representation of Speech Envelope and Temporal Fine Structure Cues
Chapter 40 Effects of Peripheral Tuning on the Auditory Nerve s Representation of Speech Envelope and Temporal Fine Structure Cues Rasha A. Ibrahim and Ian C. Bruce Abstract A number of studies have explored
More informationThe role of periodicity in the perception of masked speech with simulated and real cochlear implants
The role of periodicity in the perception of masked speech with simulated and real cochlear implants Kurt Steinmetzger and Stuart Rosen UCL Speech, Hearing and Phonetic Sciences Heidelberg, 09. November
More informationModulation and Top-Down Processing in Audition
Modulation and Top-Down Processing in Audition Malcolm Slaney 1,2 and Greg Sell 2 1 Yahoo! Research 2 Stanford CCRMA Outline The Non-Linear Cochlea Correlogram Pitch Modulation and Demodulation Information
More informationNeurobiology of Hearing (Salamanca, 2012) Auditory Cortex (2) Prof. Xiaoqin Wang
Neurobiology of Hearing (Salamanca, 2012) Auditory Cortex (2) Prof. Xiaoqin Wang Laboratory of Auditory Neurophysiology Department of Biomedical Engineering Johns Hopkins University web1.johnshopkins.edu/xwang
More informationSpectro-temporal response fields in the inferior colliculus of awake monkey
3.6.QH Spectro-temporal response fields in the inferior colliculus of awake monkey Versnel, Huib; Zwiers, Marcel; Van Opstal, John Department of Biophysics University of Nijmegen Geert Grooteplein 655
More informationHCS 7367 Speech Perception
Long-term spectrum of speech HCS 7367 Speech Perception Connected speech Absolute threshold Males Dr. Peter Assmann Fall 212 Females Long-term spectrum of speech Vowels Males Females 2) Absolute threshold
More informationHCS 7367 Speech Perception
Babies 'cry in mother's tongue' HCS 7367 Speech Perception Dr. Peter Assmann Fall 212 Babies' cries imitate their mother tongue as early as three days old German researchers say babies begin to pick up
More informationABSTRACT. in humans is not yet fully understood. In this dissertation, we develop a computational
ABSTRACT Title of dissertation: NEUROMORPHIC MODEL FOR SOUND SOURCE SEGREGATION Lakshmi Krishnan, Doctor of Philosophy, 2015 Dissertation directed by: Professor Shihab Shamma Department of Electrical and
More informationAuditory scene analysis in humans: Implications for computational implementations.
Auditory scene analysis in humans: Implications for computational implementations. Albert S. Bregman McGill University Introduction. The scene analysis problem. Two dimensions of grouping. Recognition
More informationInfant Hearing Development: Translating Research Findings into Clinical Practice. Auditory Development. Overview
Infant Hearing Development: Translating Research Findings into Clinical Practice Lori J. Leibold Department of Allied Health Sciences The University of North Carolina at Chapel Hill Auditory Development
More informationFINE-TUNING THE AUDITORY SUBCORTEX Measuring processing dynamics along the auditory hierarchy. Christopher Slugocki (Widex ORCA) WAS 5.3.
FINE-TUNING THE AUDITORY SUBCORTEX Measuring processing dynamics along the auditory hierarchy. Christopher Slugocki (Widex ORCA) WAS 5.3.2017 AUDITORY DISCRIMINATION AUDITORY DISCRIMINATION /pi//k/ /pi//t/
More informationAUDL GS08/GAV1 Signals, systems, acoustics and the ear. Pitch & Binaural listening
AUDL GS08/GAV1 Signals, systems, acoustics and the ear Pitch & Binaural listening Review 25 20 15 10 5 0-5 100 1000 10000 25 20 15 10 5 0-5 100 1000 10000 Part I: Auditory frequency selectivity Tuning
More informationChapter 5. Summary and Conclusions! 131
! Chapter 5 Summary and Conclusions! 131 Chapter 5!!!! Summary of the main findings The present thesis investigated the sensory representation of natural sounds in the human auditory cortex. Specifically,
More informationPrecise Spike Timing and Reliability in Neural Encoding of Low-Level Sensory Stimuli and Sequences
Precise Spike Timing and Reliability in Neural Encoding of Low-Level Sensory Stimuli and Sequences Temporal Structure In the World Representation in the Brain Project 1.1.2 Feldman and Harris Labs Temporal
More informationABSTRACT. Professor, Jonathan Z. Simon, Department of Electrical and Computer Engineering
ABSTRACT Title of Document: TEMPORAL CODING OF SPEECH IN HUMAN AUDITORY CORTEX Nai Ding, Doctor of Philosophy, 2012 Directed By: Professor, Jonathan Z. Simon, Department of Electrical and Computer Engineering
More informationEffect of informational content of noise on speech representation in the aging midbrain and cortex
J Neurophysiol 116: 2356 2367, 2016. First published September 7, 2016; doi:10.1152/jn.00373.2016. Effect of informational content of noise on speech representation in the aging midbrain and cortex Alessandro
More informationStructure and Function of the Auditory and Vestibular Systems (Fall 2014) Auditory Cortex (3) Prof. Xiaoqin Wang
580.626 Structure and Function of the Auditory and Vestibular Systems (Fall 2014) Auditory Cortex (3) Prof. Xiaoqin Wang Laboratory of Auditory Neurophysiology Department of Biomedical Engineering Johns
More informationInteraction between Attention and Bottom-Up Saliency Mediates the Representation of Foreground and Background in an Auditory Scene
Interaction between Attention and Bottom-Up Saliency Mediates the Representation of Foreground and Background in an Auditory Scene Mounya Elhilali 1., Juanjuan Xiang 2., Shihab A. Shamma 3,4, Jonathan
More informationEffects of Cochlear Hearing Loss on the Benefits of Ideal Binary Masking
INTERSPEECH 2016 September 8 12, 2016, San Francisco, USA Effects of Cochlear Hearing Loss on the Benefits of Ideal Binary Masking Vahid Montazeri, Shaikat Hossain, Peter F. Assmann University of Texas
More informationCombination of Bone-Conducted Speech with Air-Conducted Speech Changing Cut-Off Frequency
Combination of Bone-Conducted Speech with Air-Conducted Speech Changing Cut-Off Frequency Tetsuya Shimamura and Fumiya Kato Graduate School of Science and Engineering Saitama University 255 Shimo-Okubo,
More informationRole of F0 differences in source segregation
Role of F0 differences in source segregation Andrew J. Oxenham Research Laboratory of Electronics, MIT and Harvard-MIT Speech and Hearing Bioscience and Technology Program Rationale Many aspects of segregation
More informationAdaptation of Classification Model for Improving Speech Intelligibility in Noise
1: (Junyoung Jung et al.: Adaptation of Classification Model for Improving Speech Intelligibility in Noise) (Regular Paper) 23 4, 2018 7 (JBE Vol. 23, No. 4, July 2018) https://doi.org/10.5909/jbe.2018.23.4.511
More information10/15/2016. Hearing loss. Aging. Cognition. Aging, Cognition, and Hearing Loss: Clinical Implications
Aging, Cognition, and Loss: Clinical Implications Samira Anderson, Au.D., Ph.D. MAC Conference 2016 1 Aging loss Cognition Frank Lin, M.D., Ph.D., Baltimore Longitudinal Study of Aging 2 Cognitive ability
More informationChallenges in microphone array processing for hearing aids. Volkmar Hamacher Siemens Audiological Engineering Group Erlangen, Germany
Challenges in microphone array processing for hearing aids Volkmar Hamacher Siemens Audiological Engineering Group Erlangen, Germany SIEMENS Audiological Engineering Group R&D Signal Processing and Audiology
More informationBinaural Hearing. Why two ears? Definitions
Binaural Hearing Why two ears? Locating sounds in space: acuity is poorer than in vision by up to two orders of magnitude, but extends in all directions. Role in alerting and orienting? Separating sound
More informationAsynchronous glimpsing of speech: Spread of masking and task set-size
Asynchronous glimpsing of speech: Spread of masking and task set-size Erol J. Ozmeral, a) Emily Buss, and Joseph W. Hall III Department of Otolaryngology/Head and Neck Surgery, University of North Carolina
More informationAnalysis of in-vivo extracellular recordings. Ryan Morrill Bootcamp 9/10/2014
Analysis of in-vivo extracellular recordings Ryan Morrill Bootcamp 9/10/2014 Goals for the lecture Be able to: Conceptually understand some of the analysis and jargon encountered in a typical (sensory)
More informationAuditory principles in speech processing do computers need silicon ears?
* with contributions by V. Hohmann, M. Kleinschmidt, T. Brand, J. Nix, R. Beutelmann, and more members of our medical physics group Prof. Dr. rer.. nat. Dr. med. Birger Kollmeier* Auditory principles in
More informationNoise-Robust Speech Recognition in a Car Environment Based on the Acoustic Features of Car Interior Noise
4 Special Issue Speech-Based Interfaces in Vehicles Research Report Noise-Robust Speech Recognition in a Car Environment Based on the Acoustic Features of Car Interior Noise Hiroyuki Hoshino Abstract This
More informationHearing II Perceptual Aspects
Hearing II Perceptual Aspects Overview of Topics Chapter 6 in Chaudhuri Intensity & Loudness Frequency & Pitch Auditory Space Perception 1 2 Intensity & Loudness Loudness is the subjective perceptual quality
More informationAuditory Scene Analysis. Dr. Maria Chait, UCL Ear Institute
Auditory Scene Analysis Dr. Maria Chait, UCL Ear Institute Expected learning outcomes: Understand the tasks faced by the auditory system during everyday listening. Know the major Gestalt principles. Understand
More informationLATERAL INHIBITION MECHANISM IN COMPUTATIONAL AUDITORY MODEL AND IT'S APPLICATION IN ROBUST SPEECH RECOGNITION
LATERAL INHIBITION MECHANISM IN COMPUTATIONAL AUDITORY MODEL AND IT'S APPLICATION IN ROBUST SPEECH RECOGNITION Lu Xugang Li Gang Wang Lip0 Nanyang Technological University, School of EEE, Workstation Resource
More information21/01/2013. Binaural Phenomena. Aim. To understand binaural hearing Objectives. Understand the cues used to determine the location of a sound source
Binaural Phenomena Aim To understand binaural hearing Objectives Understand the cues used to determine the location of a sound source Understand sensitivity to binaural spatial cues, including interaural
More informationSound Localization PSY 310 Greg Francis. Lecture 31. Audition
Sound Localization PSY 310 Greg Francis Lecture 31 Physics and psychology. Audition We now have some idea of how sound properties are recorded by the auditory system So, we know what kind of information
More informationPower and phase properties of oscillatory neural responses in the presence of background activity
J Comput Neurosci (2013) 34:337 343 DOI 10.1007/s10827-012-0424-6 Power and phase properties of oscillatory neural responses in the presence of background activity Nai Ding & Jonathan Z. Simon Received:
More informationAn Auditory-Model-Based Electrical Stimulation Strategy Incorporating Tonal Information for Cochlear Implant
Annual Progress Report An Auditory-Model-Based Electrical Stimulation Strategy Incorporating Tonal Information for Cochlear Implant Joint Research Centre for Biomedical Engineering Mar.7, 26 Types of Hearing
More informationIntroduction to Computational Neuroscience
Introduction to Computational Neuroscience Lecture 5: Data analysis II Lesson Title 1 Introduction 2 Structure and Function of the NS 3 Windows to the Brain 4 Data analysis 5 Data analysis II 6 Single
More informationIntegration of Visual Information in Auditory Cortex Promotes Auditory Scene Analysis through Multisensory Binding
Article Integration of Visual Information in Auditory Cortex Promotes Auditory Scene Analysis through Multisensory Binding Graphical Abstract Authors Huriye Atilgan, Stephen M. Town, Katherine C. Wood,
More informationSound localization psychophysics
Sound localization psychophysics Eric Young A good reference: B.C.J. Moore An Introduction to the Psychology of Hearing Chapter 7, Space Perception. Elsevier, Amsterdam, pp. 233-267 (2004). Sound localization:
More informationChapter 59 Temporal Coherence and the Streaming of Complex Sounds
Chapter 59 Temporal Coherence and the Streaming of Complex Sounds Shihab Shamma, Mounya Elhilali, Ling Ma, Christophe Micheyl, ndrew J. Oxenham, Daniel Pressnitzer, Pingbo Yin, and Yanbo Xu bstract Humans
More informationAuditory Physiology PSY 310 Greg Francis. Lecture 30. Organ of Corti
Auditory Physiology PSY 310 Greg Francis Lecture 30 Waves, waves, waves. Organ of Corti Tectorial membrane Sits on top Inner hair cells Outer hair cells The microphone for the brain 1 Hearing Perceptually,
More informationTask Difficulty and Performance Induce Diverse Adaptive Patterns in Gain and Shape of Primary Auditory Cortical Receptive Fields
Article Task Difficulty and Performance Induce Diverse Adaptive Patterns in Gain and Shape of Primary Auditory Cortical Receptive Fields Serin Atiani, 1 Mounya Elhilali, 3 Stephen V. David, 2 Jonathan
More informationComment by Delgutte and Anna. A. Dreyer (Eaton-Peabody Laboratory, Massachusetts Eye and Ear Infirmary, Boston, MA)
Comments Comment by Delgutte and Anna. A. Dreyer (Eaton-Peabody Laboratory, Massachusetts Eye and Ear Infirmary, Boston, MA) Is phase locking to transposed stimuli as good as phase locking to low-frequency
More informationCompeting Streams at the Cocktail Party: Exploring the Mechanisms of Attention and Temporal Integration
12084 The Journal of Neuroscience, September 8, 2010 30(36):12084 12093 Behavioral/Systems/Cognitive Competing Streams at the Cocktail Party: Exploring the Mechanisms of Attention and Temporal Integration
More informationOn the influence of interaural differences on onset detection in auditory object formation. 1 Introduction
On the influence of interaural differences on onset detection in auditory object formation Othmar Schimmel Eindhoven University of Technology, P.O. Box 513 / Building IPO 1.26, 56 MD Eindhoven, The Netherlands,
More informationNeural Correlates of Auditory Perceptual Awareness under Informational Masking
Neural Correlates of Auditory Perceptual Awareness under Informational Masking Alexander Gutschalk 1*, Christophe Micheyl 2, Andrew J. Oxenham 2 PLoS BIOLOGY 1 Department of Neurology, Ruprecht-Karls-Universität
More informationSound, Mixtures, and Learning
Sound, Mixtures, and Learning Dan Ellis Laboratory for Recognition and Organization of Speech and Audio (LabROSA) Electrical Engineering, Columbia University http://labrosa.ee.columbia.edu/
More informationNeural correlates of the perception of sound source separation
Neural correlates of the perception of sound source separation Mitchell L. Day 1,2 * and Bertrand Delgutte 1,2,3 1 Department of Otology and Laryngology, Harvard Medical School, Boston, MA 02115, USA.
More informationAcoustics, signals & systems for audiology. Psychoacoustics of hearing impairment
Acoustics, signals & systems for audiology Psychoacoustics of hearing impairment Three main types of hearing impairment Conductive Sound is not properly transmitted from the outer to the inner ear Sensorineural
More informationTOWARD A BRAIN INTERFACE FOR TRACKING ATTENDED AUDITORY SOURCES
216 IEEE INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING, SEPT. 13 16, 216, SALERNO, ITALY TOWARD A BRAIN INTERFACE FOR TRACKING ATTENDED AUDITORY SOURCES Marzieh Haghighi 1, Mohammad
More informationRepresentation of sound in the auditory nerve
Representation of sound in the auditory nerve Eric D. Young Department of Biomedical Engineering Johns Hopkins University Young, ED. Neural representation of spectral and temporal information in speech.
More informationProcessing Interaural Cues in Sound Segregation by Young and Middle-Aged Brains DOI: /jaaa
J Am Acad Audiol 20:453 458 (2009) Processing Interaural Cues in Sound Segregation by Young and Middle-Aged Brains DOI: 10.3766/jaaa.20.7.6 Ilse J.A. Wambacq * Janet Koehnke * Joan Besing * Laurie L. Romei
More informationAtypical processing of prosodic changes in natural speech stimuli in school-age children with Asperger syndrome
Atypical processing of prosodic changes in natural speech stimuli in school-age children with Asperger syndrome Riikka Lindström, PhD student Cognitive Brain Research Unit University of Helsinki 31.8.2012
More informationWhat you re in for. Who are cochlear implants for? The bottom line. Speech processing schemes for
What you re in for Speech processing schemes for cochlear implants Stuart Rosen Professor of Speech and Hearing Science Speech, Hearing and Phonetic Sciences Division of Psychology & Language Sciences
More informationHow is the stimulus represented in the nervous system?
How is the stimulus represented in the nervous system? Eric Young F Rieke et al Spikes MIT Press (1997) Especially chapter 2 I Nelken et al Encoding stimulus information by spike numbers and mean response
More informationLinguistic Phonetics Fall 2005
MIT OpenCourseWare http://ocw.mit.edu 24.963 Linguistic Phonetics Fall 2005 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. 24.963 Linguistic Phonetics
More informationEncoding of natural sounds by variance of the cortical local field potential
J Neurophysiol 115: 2389 2398, 2016. First published February 24, 2016; doi:10.1152/jn.00652.2015. Encoding of natural sounds by variance of the cortical local field potential Nai Ding, 1 Jonathan Z. Simon,
More informationHearing Lectures. Acoustics of Speech and Hearing. Auditory Lighthouse. Facts about Timbre. Analysis of Complex Sounds
Hearing Lectures Acoustics of Speech and Hearing Week 2-10 Hearing 3: Auditory Filtering 1. Loudness of sinusoids mainly (see Web tutorial for more) 2. Pitch of sinusoids mainly (see Web tutorial for more)
More informationInfluence of acoustic complexity on spatial release from masking and lateralization
Influence of acoustic complexity on spatial release from masking and lateralization Gusztáv Lőcsei, Sébastien Santurette, Torsten Dau, Ewen N. MacDonald Hearing Systems Group, Department of Electrical
More informationAuditory Scene Analysis: phenomena, theories and computational models
Auditory Scene Analysis: phenomena, theories and computational models July 1998 Dan Ellis International Computer Science Institute, Berkeley CA Outline 1 2 3 4 The computational
More informationINTRODUCTION J. Acoust. Soc. Am. 103 (2), February /98/103(2)/1080/5/$ Acoustical Society of America 1080
Perceptual segregation of a harmonic from a vowel by interaural time difference in conjunction with mistuning and onset asynchrony C. J. Darwin and R. W. Hukin Experimental Psychology, University of Sussex,
More information= + Auditory Scene Analysis. Week 9. The End. The auditory scene. The auditory scene. Otherwise known as
Auditory Scene Analysis Week 9 Otherwise known as Auditory Grouping Auditory Streaming Sound source segregation The auditory scene The auditory system needs to make sense of the superposition of component
More informationReport. Direct Recordings of Pitch Responses from Human Auditory Cortex
Current Biology 0,, June, 00 ª00 Elsevier Ltd. Open access under CC BY license. DOI 0.0/j.cub.00.0.0 Direct Recordings of Pitch Responses from Human Auditory Cortex Report Timothy D. Griffiths,, * Sukhbinder
More informationAuditory-Visual Speech Perception Laboratory
Auditory-Visual Speech Perception Laboratory Research Focus: Identify perceptual processes involved in auditory-visual speech perception Determine the abilities of individual patients to carry out these
More informationSystems Neuroscience Oct. 16, Auditory system. http:
Systems Neuroscience Oct. 16, 2018 Auditory system http: www.ini.unizh.ch/~kiper/system_neurosci.html The physics of sound Measuring sound intensity We are sensitive to an enormous range of intensities,
More informationRevisiting the right-ear advantage for speech: Implications for speech displays
INTERSPEECH 2014 Revisiting the right-ear advantage for speech: Implications for speech displays Nandini Iyer 1, Eric Thompson 2, Brian Simpson 1, Griffin Romigh 1 1 Air Force Research Laboratory; 2 Ball
More informationIsolating the energetic component of speech-on-speech masking with ideal time-frequency segregation
Isolating the energetic component of speech-on-speech masking with ideal time-frequency segregation Douglas S. Brungart a Air Force Research Laboratory, Human Effectiveness Directorate, 2610 Seventh Street,
More informationBinaural processing of complex stimuli
Binaural processing of complex stimuli Outline for today Binaural detection experiments and models Speech as an important waveform Experiments on understanding speech in complex environments (Cocktail
More informationABSTRACT. Auditory streaming: behavior, physiology, and modeling
ABSTRACT Title of Document: Auditory streaming: behavior, physiology, and modeling Ling Ma, Doctor of Philosophy, 2011 Directed By: Professor Shihab A. Shamma, Department of Electrical and computer Engineering,
More informationLateralized speech perception in normal-hearing and hearing-impaired listeners and its relationship to temporal processing
Lateralized speech perception in normal-hearing and hearing-impaired listeners and its relationship to temporal processing GUSZTÁV LŐCSEI,*, JULIE HEFTING PEDERSEN, SØREN LAUGESEN, SÉBASTIEN SANTURETTE,
More informationChapter 11: Sound, The Auditory System, and Pitch Perception
Chapter 11: Sound, The Auditory System, and Pitch Perception Overview of Questions What is it that makes sounds high pitched or low pitched? How do sound vibrations inside the ear lead to the perception
More informationSpectral fingerprints of large-scale neuronal interactions
Nature Reviews Neuroscience AOP, published online 11 January 212; doi:1.138/nrn3137 REVIEWS Spectral fingerprints of large-scale neuronal interactions Markus Siegel 1 *, Tobias H. Donner 2 * and Andreas
More informationA. SEK, E. SKRODZKA, E. OZIMEK and A. WICHER
ARCHIVES OF ACOUSTICS 29, 1, 25 34 (2004) INTELLIGIBILITY OF SPEECH PROCESSED BY A SPECTRAL CONTRAST ENHANCEMENT PROCEDURE AND A BINAURAL PROCEDURE A. SEK, E. SKRODZKA, E. OZIMEK and A. WICHER Institute
More informationSpeech recognition in noisy environments: A survey
T-61.182 Robustness in Language and Speech Processing Speech recognition in noisy environments: A survey Yifan Gong presented by Tapani Raiko Feb 20, 2003 About the Paper Article published in Speech Communication
More informationEffects of speaker's and listener's environments on speech intelligibili annoyance. Author(s)Kubo, Rieko; Morikawa, Daisuke; Akag
JAIST Reposi https://dspace.j Title Effects of speaker's and listener's environments on speech intelligibili annoyance Author(s)Kubo, Rieko; Morikawa, Daisuke; Akag Citation Inter-noise 2016: 171-176 Issue
More informationAmbiguity in the recognition of phonetic vowels when using a bone conduction microphone
Acoustics 8 Paris Ambiguity in the recognition of phonetic vowels when using a bone conduction microphone V. Zimpfer a and K. Buck b a ISL, 5 rue du Général Cassagnou BP 734, 6831 Saint Louis, France b
More informationCarl Wernicke s Contribution to Theories of Conceptual Representation in the Cerebral Cortex. Nicole Gage and Gregory Hickok Irvine, California
Carl Wernicke s Contribution to Theories of Conceptual Representation in the Cerebral Cortex Nicole Gage and Gregory Hickok Irvine, California Acknowledgments Christian Sekirnjak, Ph.D. San Diego, CA Heidi
More informationAUTOCORRELATION AND CROSS-CORRELARION ANALYSES OF ALPHA WAVES IN RELATION TO SUBJECTIVE PREFERENCE OF A FLICKERING LIGHT
AUTOCORRELATION AND CROSS-CORRELARION ANALYSES OF ALPHA WAVES IN RELATION TO SUBJECTIVE PREFERENCE OF A FLICKERING LIGHT Y. Soeta, S. Uetani, and Y. Ando Graduate School of Science and Technology, Kobe
More informationEEG reveals divergent paths for speech envelopes during selective attention
EEG reveals divergent paths for speech envelopes during selective attention Cort Horton a, Michael D Zmura a, and Ramesh Srinivasan a,b a Dept. of Cognitive Sciences, University of California, Irvine,
More informationWho are cochlear implants for?
Who are cochlear implants for? People with little or no hearing and little conductive component to the loss who receive little or no benefit from a hearing aid. Implants seem to work best in adults who
More informationTowards a biological theory of phone5c percep5on
Towards a biological theory of phone5c percep5on Q Can we build a theory of phone5c percep5on from the ground up using neurological primi5ves? A so far, so good Q Why would this be desirable? A Many theories
More informationLinguistic Phonetics. Basic Audition. Diagram of the inner ear removed due to copyright restrictions.
24.963 Linguistic Phonetics Basic Audition Diagram of the inner ear removed due to copyright restrictions. 1 Reading: Keating 1985 24.963 also read Flemming 2001 Assignment 1 - basic acoustics. Due 9/22.
More informationCombating the Reverberation Problem
Combating the Reverberation Problem Barbara Shinn-Cunningham (Boston University) Martin Cooke (Sheffield University, U.K.) How speech is corrupted by reverberation DeLiang Wang (Ohio State University)
More informationAuditory Scene Analysis
1 Auditory Scene Analysis Albert S. Bregman Department of Psychology McGill University 1205 Docteur Penfield Avenue Montreal, QC Canada H3A 1B1 E-mail: bregman@hebb.psych.mcgill.ca To appear in N.J. Smelzer
More informationTelephone Based Automatic Voice Pathology Assessment.
Telephone Based Automatic Voice Pathology Assessment. Rosalyn Moran 1, R. B. Reilly 1, P.D. Lacy 2 1 Department of Electronic and Electrical Engineering, University College Dublin, Ireland 2 Royal Victoria
More informationOscillations: From Neuron to MEG
Oscillations: From Neuron to MEG Educational Symposium, MEG UK 2014, Nottingham, Jan 8th 2014 Krish Singh CUBRIC, School of Psychology Cardiff University What are we trying to achieve? Bridge the gap from
More informationHearing. and other senses
Hearing and other senses Sound Sound: sensed variations in air pressure Frequency: number of peaks that pass a point per second (Hz) Pitch 2 Some Sound and Hearing Links Useful (and moderately entertaining)
More information