Using Speech Models for Separation
|
|
- Derick Beasley
- 5 years ago
- Views:
Transcription
1 Using Speech Models for Separation Dan Ellis Comprising the work of Michael Mandel and Ron Weiss Laboratory for Recognition and Organization of Speech and Audio Dept. Electrical Eng., Columbia Univ., NY USA 1. Eigenvoice Speaker Models. Spatial Parameter Models in Reverb 3. Combining Source + Spatial Speech Models for Separation - Dan Ellis /19
2 1. Speech Separation Using Models Cooke & Lee s Monaural Speech Separation Task pairs of short, grammatically-constrained utterances: <command:><color:><preposition:><letter:5><number:1><adverb:> e.g. "bin white by R again" task: report letter + number for white Separation depends on source constraints the more the better - ASR model mix separation identify target energy t-f masking + resynthesis ASR speech models words vs. mix identify speech models find best words model words source knowledge Speech Models for Separation - Dan Ellis /19
3 Speech Mixture Recognition Kristjansson, Hershey et al. Speech recognizers contain speech models ASR is just argmax P(W X) Recognize mixtures with Factorial HMM one model+state sequence for each voice exploit sequence constraints, speaker differences model model 1 observations / time separation relies on detailed speaker model Speech Models for Separation - Dan Ellis /19
4 Eigenvoices Idea: Find speaker model parameter space Kuhn et al. 9, Weiss & Ellis 7,, 9 generalize without losing detail? Eigenvoice model: µ = µ + U w + B h Speaker models Speaker subspace bases adapted mean eigenvoice weights channel channel model voice bases bases weights Speech Models for Separation - Dan Ellis /19
5 Eigenvoice Bases Mean Voice 1 Mean model states x 3 bins = 9, dimensions Frequency (khz) b d g p t k jh ch s z f th v dh m n l r w y iy ih eh ey ae aa aw ay ah ao owuw ax Eigenvoice dimension Eigencomponents shift formants/ coloration additional components for acoustic channel Frequency (khz) Frequency (khz) Frequency (khz) b d g p t k jh ch s z f th v dh m n l r w y iy ih eh ey ae aa aw ay ah ao owuw ax Eigenvoice dimension b d g p t k jh ch s z f th v dh m n l r w y iy ih eh ey ae aa aw ay ah ao owuw ax Eigenvoice dimension 3 b d g p t k jh ch s z f th v dh m n l r w y iy ih eh ey ae aa aw ay ah ao owuw ax Speech Models for Separation - Dan Ellis /19
6 Speaker-Adapted Separation Factorial HMM analysis with tuning of source model parameters = eigenvoice speaker adaptation Weiss & Ellis Speech Models for Separation - Dan Ellis /19
7 Speaker-Adapted Separation Speech Models for Separation - Dan Ellis /19
8 Speaker-Adapted Separation Eigenvoices for Speech Separation task speaker adapted (SA) performs midway between speaker-dependent (SD) & speaker-indep (SI) SI SA SD Speech Models for Separation - Dan Ellis /19
9 . Spatial Models & Reverb or 3 sources in reverberation assume just ears Mandel & Ellis 7 Model interaural spectrum of each source as stationary level and time differences: Speech Models for Separation - Dan Ellis /19
10 IPD, ILD Distributions Source at 75 in reverberation IPD IPD residual ILD IPD residual offsets phase by constant ωτ IPD can be fit by single Gaussian ILD needs frequency-dependence Speech Models for Separation - Dan Ellis /19
11 Model-Based EM Source Separation and Localization (MESSL) Re-estimate source parameters Mandel & Ellis 9 Assign spectrogram points to sources can model more sources than sensors flexible initialization Speech Models for Separation - Dan Ellis /19
12 MESSL Results Modeling uncertainty improves results tradeoff between constraints & noisiness.5 db. db.77 db 1.35 db 9.1 db -.7 db Speech Models for Separation - Dan Ellis /19
13 MESSL Results Speech recognizer (Digits) 1 Ground truth masks 1 Algorithmic masks Percent correct Human DP Oracle Oracle OracleAllRev Mixes Target to masker ratio (db) Human Sawada Mouba MESSL G MESSL ΩΩ DUET Mixes Target to masker ratio (db) Speech Models for Separation - Dan Ellis /19
14 3. Combining Spatial + Speech Models Weiss, Mandel & Ellis Interaural parameters give ILDi(ω), ITDi, Pr(X(t, ω) = Si(t, ω)) Speech source model can give Pr(S i (t, ω) is speech signal) Can combine into one big EM framework... E-step M-step u is: Pr(cell from source i) phoneme sequence Θ is: interaural params speaker params Speech Models for Separation - Dan Ellis /19
15 Parameter estimation and source separation MESSL-SP (Source Prior) Ron Weiss Underdetermined Source Separation Using Speaker Subspace Models Speech Models for Separation - Dan Ellis May, 9 7 / /19
16 MESSL-SP Results Source models function as priors Interaural parameter spatial separation source model prior improves spatial estimate Frequency (khz) Frequency (khz) Ground truth (1. db) MESSL (5. db) Time (sec) DUET (3. db) MESSL SP (1.1 db) Time (sec) D FD BSS (5.1 db) Speech Models for Separation - Dan Ellis /19 MESSL EV (1.37 db) Time (sec) 1....
17 MESSL-SP Results SNR improvement vs. source angle separation sources (anechoic) sources (reverb) SNR improvement (db) SNR improvement (db) sources (anechoic) sources (reverb) Ground truth MESSL!EV MESSL!SP MESSL S!FD!BSS DUET Separation (degrees) Separation (degrees) Speech Models for Separation - Dan Ellis /19
18 Future Work Better parametric speaker models limitations of eigenvoices varying style Understanding reverb & ASR early echoes what spoils ASR? Models of other sources eigeninstruments? Speech Models for Separation - Dan Ellis /19
19 Summary & Conclusions Source models provide the constraints to make scene analysis possible Eigenvoices (model subspace) can be used to provide detailed models that generalize Spatial parameters can identify more sources than models in reverb (MESSL) Can combine source + spatial models Speech Models for Separation - Dan Ellis /19
Using Source Models in Speech Separation
Using Source Models in Speech Separation Dan Ellis Laboratory for Recognition and Organization of Speech and Audio Dept. Electrical Eng., Columbia Univ., NY USA dpwe@ee.columbia.edu http://labrosa.ee.columbia.edu/
More informationLabROSA Research Overview
LabROSA Research Overview Dan Ellis Laboratory for Recognition and Organization of Speech and Audio Dept. Electrical Eng., Columbia Univ., NY USA dpwe@ee.columbia.edu http://labrosa.ee.columbia.edu/ 1.
More informationLecture 9: Speech Recognition: Front Ends
EE E682: Speech & Audio Processing & Recognition Lecture 9: Speech Recognition: Front Ends 1 2 Recognizing Speech Feature Calculation Dan Ellis http://www.ee.columbia.edu/~dpwe/e682/
More informationComputational Perception /785. Auditory Scene Analysis
Computational Perception 15-485/785 Auditory Scene Analysis A framework for auditory scene analysis Auditory scene analysis involves low and high level cues Low level acoustic cues are often result in
More informationUsing the Soundtrack to Classify Videos
Using the Soundtrack to Classify Videos Dan Ellis Laboratory for Recognition and Organization of Speech and Audio Dept. Electrical Eng., Columbia Univ., NY USA dpwe@ee.columbia.edu http://labrosa.ee.columbia.edu/
More informationAdaptation of Classification Model for Improving Speech Intelligibility in Noise
1: (Junyoung Jung et al.: Adaptation of Classification Model for Improving Speech Intelligibility in Noise) (Regular Paper) 23 4, 2018 7 (JBE Vol. 23, No. 4, July 2018) https://doi.org/10.5909/jbe.2018.23.4.511
More informationSingle-Channel Sound Source Localization Based on Discrimination of Acoustic Transfer Functions
3 Single-Channel Sound Source Localization Based on Discrimination of Acoustic Transfer Functions Ryoichi Takashima, Tetsuya Takiguchi and Yasuo Ariki Graduate School of System Informatics, Kobe University,
More informationSound, Mixtures, and Learning
Sound, Mixtures, and Learning Dan Ellis Laboratory for Recognition and Organization of Speech and Audio (LabROSA) Electrical Engineering, Columbia University http://labrosa.ee.columbia.edu/
More informationRecognition & Organization of Speech and Audio
Recognition & Organization of Speech and Audio Dan Ellis Electrical Engineering, Columbia University http://www.ee.columbia.edu/~dpwe/ Outline 1 2 3 4 5 Introducing Tandem modeling
More informationAutomatic audio analysis for content description & indexing
Automatic audio analysis for content description & indexing Dan Ellis International Computer Science Institute, Berkeley CA Outline 1 2 3 4 5 Auditory Scene Analysis (ASA) Computational
More informationComputational Auditory Scene Analysis: An overview and some observations. CASA survey. Other approaches
CASA talk - Haskins/NUWC - Dan Ellis 1997oct24/5-1 The importance of auditory illusions for artificial listeners 1 Dan Ellis International Computer Science Institute, Berkeley CA
More informationLecture 4: Auditory Perception. Why study perception?
EE E682: Speech & Audio Processing & Recognition Lecture 4: Auditory Perception 1 2 3 4 5 6 Motivation: Why & how Auditory physiology Psychophysics: Detection & discrimination Pitch perception Speech perception
More informationRecognition & Organization of Speech & Audio
Recognition & Organization of Speech & Audio Dan Ellis http://labrosa.ee.columbia.edu/ Outline 1 2 3 Introducing Projects in speech, music & audio Summary overview - Dan Ellis 21-9-28-1 1 Sound organization
More informationRobustness, Separation & Pitch
Robustness, Separation & Pitch or Morgan, Me & Pitch Dan Ellis Columbia / ICSI dpwe@ee.columbia.edu http://labrosa.ee.columbia.edu/ 1. Robustness and Separation 2. An Academic Journey 3. Future COLUMBIA
More informationHCS 7367 Speech Perception
Babies 'cry in mother's tongue' HCS 7367 Speech Perception Dr. Peter Assmann Fall 212 Babies' cries imitate their mother tongue as early as three days old German researchers say babies begin to pick up
More informationRecognition & Organization of Speech and Audio
Recognition & Organization of Speech and Audio Dan Ellis Electrical Engineering, Columbia University http://www.ee.columbia.edu/~dpwe/ Outline 1 2 3 4 Introducing Robust speech recognition
More informationBinaural processing of complex stimuli
Binaural processing of complex stimuli Outline for today Binaural detection experiments and models Speech as an important waveform Experiments on understanding speech in complex environments (Cocktail
More informationComputational Auditory Scene Analysis: Principles, Practice and Applications
Computational Auditory Scene Analysis: Principles, Practice and Applications Dan Ellis International Computer Science Institute, Berkeley CA Outline 1 2 3 4 5 Auditory Scene Analysis
More informationSound, Mixtures, and Learning
Sound, Mixtures, and Learning Dan Ellis oratory for Recognition and Organization of Speech and Audio () Electrical Engineering, Columbia University http://labrosa.ee.columbia.edu/
More informationLecture 8: Spatial sound
EE E6820: Speech & Audio Processing & Recognition Lecture 8: Spatial sound 1 2 3 4 Spatial acoustics Binaural perception Synthesizing spatial audio Extracting spatial sounds Dan Ellis
More informationHCS 7367 Speech Perception
Long-term spectrum of speech HCS 7367 Speech Perception Connected speech Absolute threshold Males Dr. Peter Assmann Fall 212 Females Long-term spectrum of speech Vowels Males Females 2) Absolute threshold
More informationTandem acoustic modeling: Neural nets for mainstream ASR?
Tandem acoutic modeling: for maintream ASR? Dan Elli International Computer Science Intitute Berkeley CA dpwe@ici.berkeley.edu Outline 2 3 Tandem acoutic modeling Inide Tandem ytem: What going on? Future
More informationGeneral Soundtrack Analysis
General Soundtrack Analysis Dan Ellis oratory for Recognition and Organization of Speech and Audio () Electrical Engineering, Columbia University http://labrosa.ee.columbia.edu/
More informationLecture 3: Perception
ELEN E4896 MUSIC SIGNAL PROCESSING Lecture 3: Perception 1. Ear Physiology 2. Auditory Psychophysics 3. Pitch Perception 4. Music Perception Dan Ellis Dept. Electrical Engineering, Columbia University
More informationRobust Neural Encoding of Speech in Human Auditory Cortex
Robust Neural Encoding of Speech in Human Auditory Cortex Nai Ding, Jonathan Z. Simon Electrical Engineering / Biology University of Maryland, College Park Auditory Processing in Natural Scenes How is
More informationSound Analysis Research at LabROSA
Sound Analysis Research at LabROSA Dan Ellis Laboratory for Recognition and Organization of Speech and Audio Dept. Electrical Eng., Columbia Univ., NY USA dpwe@ee.columbia.edu http://labrosa.ee.columbia.edu/
More informationCHAPTER 1 INTRODUCTION
CHAPTER 1 INTRODUCTION 1.1 BACKGROUND Speech is the most natural form of human communication. Speech has also become an important means of human-machine interaction and the advancement in technology has
More informationEffects of Cochlear Hearing Loss on the Benefits of Ideal Binary Masking
INTERSPEECH 2016 September 8 12, 2016, San Francisco, USA Effects of Cochlear Hearing Loss on the Benefits of Ideal Binary Masking Vahid Montazeri, Shaikat Hossain, Peter F. Assmann University of Texas
More informationHearing in the Environment
10 Hearing in the Environment Click Chapter to edit 10 Master Hearing title in the style Environment Sound Localization Complex Sounds Auditory Scene Analysis Continuity and Restoration Effects Auditory
More informationNoise-Robust Speech Recognition Technologies in Mobile Environments
Noise-Robust Speech Recognition echnologies in Mobile Environments Mobile environments are highly influenced by ambient noise, which may cause a significant deterioration of speech recognition performance.
More informationSpeech Separation in Humans and Machines
Speech Separation in Humans and Machines Dan Ellis Laboratory for Recognition and Organization of Speech and Audio Dept. Electrical Eng., Columbia Univ., NY USA dpwe@ee.columbia.edu http://labrosa.ee.columbia.edu/
More informationAuditory Scene Analysis: phenomena, theories and computational models
Auditory Scene Analysis: phenomena, theories and computational models July 1998 Dan Ellis International Computer Science Institute, Berkeley CA Outline 1 2 3 4 The computational
More informationEffect of microphone position in hearing instruments on binaural masking level differences
Effect of microphone position in hearing instruments on binaural masking level differences Fredrik Gran, Jesper Udesen and Andrew B. Dittberner GN ReSound A/S, Research R&D, Lautrupbjerg 7, 2750 Ballerup,
More informationRole of F0 differences in source segregation
Role of F0 differences in source segregation Andrew J. Oxenham Research Laboratory of Electronics, MIT and Harvard-MIT Speech and Hearing Bioscience and Technology Program Rationale Many aspects of segregation
More informationNOISE REDUCTION ALGORITHMS FOR BETTER SPEECH UNDERSTANDING IN COCHLEAR IMPLANTATION- A SURVEY
INTERNATIONAL JOURNAL OF RESEARCH IN COMPUTER APPLICATIONS AND ROBOTICS ISSN 2320-7345 NOISE REDUCTION ALGORITHMS FOR BETTER SPEECH UNDERSTANDING IN COCHLEAR IMPLANTATION- A SURVEY Preethi N.M 1, P.F Khaleelur
More information21/01/2013. Binaural Phenomena. Aim. To understand binaural hearing Objectives. Understand the cues used to determine the location of a sound source
Binaural Phenomena Aim To understand binaural hearing Objectives Understand the cues used to determine the location of a sound source Understand sensitivity to binaural spatial cues, including interaural
More informationSound localization psychophysics
Sound localization psychophysics Eric Young A good reference: B.C.J. Moore An Introduction to the Psychology of Hearing Chapter 7, Space Perception. Elsevier, Amsterdam, pp. 233-267 (2004). Sound localization:
More informationAcoustic feedback suppression in binaural hearing aids
Aud Vest Res (2016);25(4):207-214. RESEARCH ARTICLE Acoustic feedback suppression in binaural hearing aids Fatmah Tarrab 1*, Zouheir Marmar 1, Isam Alamine 2 1 - Department of Biomedical Engineering, Faculty
More informationAcoustic Signal Processing Based on Deep Neural Networks
Acoustic Signal Processing Based on Deep Neural Networks Chin-Hui Lee School of ECE, Georgia Tech chl@ece.gatech.edu Joint work with Yong Xu, Yanhui Tu, Qing Wang, Tian Gao, Jun Du, LiRong Dai Outline
More informationSUPPRESSION OF MUSICAL NOISE IN ENHANCED SPEECH USING PRE-IMAGE ITERATIONS. Christina Leitner and Franz Pernkopf
2th European Signal Processing Conference (EUSIPCO 212) Bucharest, Romania, August 27-31, 212 SUPPRESSION OF MUSICAL NOISE IN ENHANCED SPEECH USING PRE-IMAGE ITERATIONS Christina Leitner and Franz Pernkopf
More informationAuditory Scene Analysis in Humans and Machines
Auditory Scene Analysis in Humans and Machines Dan Ellis Laboratory for Recognition and Organization of Speech and Audio Dept. Electrical Eng., Columbia Univ., NY USA dpwe@ee.columbia.edu http://labrosa.ee.columbia.edu/
More informationGENERALIZATION OF SUPERVISED LEARNING FOR BINARY MASK ESTIMATION
GENERALIZATION OF SUPERVISED LEARNING FOR BINARY MASK ESTIMATION Tobias May Technical University of Denmark Centre for Applied Hearing Research DK - 2800 Kgs. Lyngby, Denmark tobmay@elektro.dtu.dk Timo
More informationVoice Detection using Speech Energy Maximization and Silence Feature Normalization
, pp.25-29 http://dx.doi.org/10.14257/astl.2014.49.06 Voice Detection using Speech Energy Maximization and Silence Feature Normalization In-Sung Han 1 and Chan-Shik Ahn 2 1 Dept. of The 2nd R&D Institute,
More informationSpeech enhancement based on harmonic estimation combined with MMSE to improve speech intelligibility for cochlear implant recipients
INERSPEECH 1 August 4, 1, Stockholm, Sweden Speech enhancement based on harmonic estimation combined with MMSE to improve speech intelligibility for cochlear implant recipients Dongmei Wang, John H. L.
More informationEnhancement of Reverberant Speech Using LP Residual Signal
IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, VOL. 8, NO. 3, MAY 2000 267 Enhancement of Reverberant Speech Using LP Residual Signal B. Yegnanarayana, Senior Member, IEEE, and P. Satyanarayana Murthy
More informationI. INTRODUCTION. OMBARD EFFECT (LE), named after the French otorhino-laryngologist
IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 18, NO. 6, AUGUST 2010 1379 Unsupervised Equalization of Lombard Effect for Speech Recognition in Noisy Adverse Environments Hynek Bořil,
More informationEffects of speaker's and listener's environments on speech intelligibili annoyance. Author(s)Kubo, Rieko; Morikawa, Daisuke; Akag
JAIST Reposi https://dspace.j Title Effects of speaker's and listener's environments on speech intelligibili annoyance Author(s)Kubo, Rieko; Morikawa, Daisuke; Akag Citation Inter-noise 2016: 171-176 Issue
More informationCodebook driven short-term predictor parameter estimation for speech enhancement
Codebook driven short-term predictor parameter estimation for speech enhancement Citation for published version (APA): Srinivasan, S., Samuelsson, J., & Kleijn, W. B. (2006). Codebook driven short-term
More informationInfluence of acoustic complexity on spatial release from masking and lateralization
Influence of acoustic complexity on spatial release from masking and lateralization Gusztáv Lőcsei, Sébastien Santurette, Torsten Dau, Ewen N. MacDonald Hearing Systems Group, Department of Electrical
More informationIMPROVING CHANNEL SELECTION OF SOUND CODING ALGORITHMS IN COCHLEAR IMPLANTS. Hussnain Ali, Feng Hong, John H. L. Hansen, and Emily Tobey
2014 IEEE International Conference on Acoustic, Speech and Signal Processing (ICASSP) IMPROVING CHANNEL SELECTION OF SOUND CODING ALGORITHMS IN COCHLEAR IMPLANTS Hussnain Ali, Feng Hong, John H. L. Hansen,
More informationLATERAL INHIBITION MECHANISM IN COMPUTATIONAL AUDITORY MODEL AND IT'S APPLICATION IN ROBUST SPEECH RECOGNITION
LATERAL INHIBITION MECHANISM IN COMPUTATIONAL AUDITORY MODEL AND IT'S APPLICATION IN ROBUST SPEECH RECOGNITION Lu Xugang Li Gang Wang Lip0 Nanyang Technological University, School of EEE, Workstation Resource
More informationLateralized speech perception in normal-hearing and hearing-impaired listeners and its relationship to temporal processing
Lateralized speech perception in normal-hearing and hearing-impaired listeners and its relationship to temporal processing GUSZTÁV LŐCSEI,*, JULIE HEFTING PEDERSEN, SØREN LAUGESEN, SÉBASTIEN SANTURETTE,
More informationAmbiguity in the recognition of phonetic vowels when using a bone conduction microphone
Acoustics 8 Paris Ambiguity in the recognition of phonetic vowels when using a bone conduction microphone V. Zimpfer a and K. Buck b a ISL, 5 rue du Général Cassagnou BP 734, 6831 Saint Louis, France b
More informationAUDL GS08/GAV1 Signals, systems, acoustics and the ear. Pitch & Binaural listening
AUDL GS08/GAV1 Signals, systems, acoustics and the ear Pitch & Binaural listening Review 25 20 15 10 5 0-5 100 1000 10000 25 20 15 10 5 0-5 100 1000 10000 Part I: Auditory frequency selectivity Tuning
More informationAuditory principles in speech processing do computers need silicon ears?
* with contributions by V. Hohmann, M. Kleinschmidt, T. Brand, J. Nix, R. Beutelmann, and more members of our medical physics group Prof. Dr. rer.. nat. Dr. med. Birger Kollmeier* Auditory principles in
More informationAcoustic Sensing With Artificial Intelligence
Acoustic Sensing With Artificial Intelligence Bowon Lee Department of Electronic Engineering Inha University Incheon, South Korea bowon.lee@inha.ac.kr bowon.lee@ieee.org NVIDIA Deep Learning Day Seoul,
More informationBinaural Hearing. Why two ears? Definitions
Binaural Hearing Why two ears? Locating sounds in space: acuity is poorer than in vision by up to two orders of magnitude, but extends in all directions. Role in alerting and orienting? Separating sound
More informationBinaural Hearing for Robots Introduction to Robot Hearing
Binaural Hearing for Robots Introduction to Robot Hearing 1Radu Horaud Binaural Hearing for Robots 1. Introduction to Robot Hearing 2. Methodological Foundations 3. Sound-Source Localization 4. Machine
More informationThe importance of phase in speech enhancement
Available online at www.sciencedirect.com Speech Communication 53 (2011) 465 494 www.elsevier.com/locate/specom The importance of phase in speech enhancement Kuldip Paliwal, Kamil Wójcicki, Benjamin Shannon
More informationBINAURAL DICHOTIC PRESENTATION FOR MODERATE BILATERAL SENSORINEURAL HEARING-IMPAIRED
International Conference on Systemics, Cybernetics and Informatics, February 12 15, 2004 BINAURAL DICHOTIC PRESENTATION FOR MODERATE BILATERAL SENSORINEURAL HEARING-IMPAIRED Alice N. Cheeran Biomedical
More informationAuditory scene analysis in humans: Implications for computational implementations.
Auditory scene analysis in humans: Implications for computational implementations. Albert S. Bregman McGill University Introduction. The scene analysis problem. Two dimensions of grouping. Recognition
More informationModulation and Top-Down Processing in Audition
Modulation and Top-Down Processing in Audition Malcolm Slaney 1,2 and Greg Sell 2 1 Yahoo! Research 2 Stanford CCRMA Outline The Non-Linear Cochlea Correlogram Pitch Modulation and Demodulation Information
More informationTHE ROLE OF VISUAL SPEECH CUES IN THE AUDITORY PERCEPTION OF SYNTHETIC STIMULI BY CHILDREN USING A COCHLEAR IMPLANT AND CHILDREN WITH NORMAL HEARING
THE ROLE OF VISUAL SPEECH CUES IN THE AUDITORY PERCEPTION OF SYNTHETIC STIMULI BY CHILDREN USING A COCHLEAR IMPLANT AND CHILDREN WITH NORMAL HEARING Vanessa Surowiecki 1, vid Grayden 1, Richard Dowell
More informationRecognition & Organization of Speech and Audio
Recognition & Organization of Speech and Audio Dan Ellis Electrical Engineering, Columbia University Outline 1 2 3 4 5 Sound organization Background & related work Existing projects
More informationPerceptual Evaluation of Signal Enhancement Strategies
Perceptual Evaluation of Signal Enhancement Strategies Koen Eneman, Heleen Luts, Jan Wouters, Michael Büchler, Norbert Dillier, Wouter Dreschler, Matthias Froehlich, Giso Grimm, Niklas Harlander, Volker
More informationTopics in Linguistic Theory: Laboratory Phonology Spring 2007
MIT OpenCourseWare http://ocw.mit.edu 24.91 Topics in Linguistic Theory: Laboratory Phonology Spring 27 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationThe Institution of Engineers Australia Communications Conference Sydney, October
The Institution of Engineers Australia Communications Conference Sydney, October 20-22 1992 A Comparison of Noise Reduction Techniques for Speech Recognition in Telecommunications Environments A.G. MAHER,
More informationThe effect of wearing conventional and level-dependent hearing protectors on speech production in noise and quiet
The effect of wearing conventional and level-dependent hearing protectors on speech production in noise and quiet Ghazaleh Vaziri Christian Giguère Hilmi R. Dajani Nicolas Ellaham Annual National Hearing
More informationIMPROVING INTELLIGIBILITY IN NOISE OF HMM-GENERATED SPEECH VIA NOISE-DEPENDENT AND -INDEPENDENT METHODS
IMPROVING INTELLIGIBILITY IN NOISE OF HMM-GENERATED SPEECH VIA NOISE-DEPENDENT AND -INDEPENDENT METHODS Cassia Valentini-Botinhao 1, Elizabeth Godoy 2, Yannis Stylianou 2, Bastian Sauert 3 Simon King 1
More informationFrequency refers to how often something happens. Period refers to the time it takes something to happen.
Lecture 2 Properties of Waves Frequency and period are distinctly different, yet related, quantities. Frequency refers to how often something happens. Period refers to the time it takes something to happen.
More informationAdvanced Audio Interface for Phonetic Speech. Recognition in a High Noise Environment
DISTRIBUTION STATEMENT A Approved for Public Release Distribution Unlimited Advanced Audio Interface for Phonetic Speech Recognition in a High Noise Environment SBIR 99.1 TOPIC AF99-1Q3 PHASE I SUMMARY
More informationCommunication quality for students with a hearing impairment: An experiment evaluating speech intelligibility and annoyance
Communication quality for students with a hearing impairment: An experiment evaluating speech intelligibility and annoyance Johan Odelius, Örjan Johansson, Anders Ågren Division of Sound and Vibration,
More informationDifference Evaluation Method for Differential Diagnosis of Frequency Specific Hearing Loss
Dietl, H. and Weiss, S. (2004) Difference evaluation method for differential diagnosis of frequency-specific hearing loss. In: IPEM Meeting on Signal Processing Applications in Clinical Neurophysiology,
More informationFrequency Tracking: LMS and RLS Applied to Speech Formant Estimation
Aldebaro Klautau - http://speech.ucsd.edu/aldebaro - 2/3/. Page. Frequency Tracking: LMS and RLS Applied to Speech Formant Estimation ) Introduction Several speech processing algorithms assume the signal
More informationUsing blind source separation techniques to improve speech recognition in bilateral cochlear implant patients
Using blind source separation techniques to improve speech recognition in bilateral cochlear implant patients Kostas Kokkinakis and Philipos C. Loizou a Department of Electrical Engineering, The University
More informationInternational Journal of Scientific & Engineering Research, Volume 5, Issue 3, March ISSN
International Journal of Scientific & Engineering Research, Volume 5, Issue 3, March-214 1214 An Improved Spectral Subtraction Algorithm for Noise Reduction in Cochlear Implants Saeed Kermani 1, Marjan
More informationBrian D. Simpson Veridian, 5200 Springfield Pike, Suite 200, Dayton, Ohio 45431
The effects of spatial separation in distance on the informational and energetic masking of a nearby speech signal Douglas S. Brungart a) Air Force Research Laboratory, 2610 Seventh Street, Wright-Patterson
More informationRobust Speech Detection for Noisy Environments
Robust Speech Detection for Noisy Environments Óscar Varela, Rubén San-Segundo and Luis A. Hernández ABSTRACT This paper presents a robust voice activity detector (VAD) based on hidden Markov models (HMM)
More informationIN EAR TO OUT THERE: A MAGNITUDE BASED PARAMETERIZATION SCHEME FOR SOUND SOURCE EXTERNALIZATION. Griffin D. Romigh, Brian D. Simpson, Nandini Iyer
IN EAR TO OUT THERE: A MAGNITUDE BASED PARAMETERIZATION SCHEME FOR SOUND SOURCE EXTERNALIZATION Griffin D. Romigh, Brian D. Simpson, Nandini Iyer 711th Human Performance Wing Air Force Research Laboratory
More informationOn the influence of interaural differences on onset detection in auditory object formation. 1 Introduction
On the influence of interaural differences on onset detection in auditory object formation Othmar Schimmel Eindhoven University of Technology, P.O. Box 513 / Building IPO 1.26, 56 MD Eindhoven, The Netherlands,
More informationPerception Optimized Deep Denoising AutoEncoders for Speech Enhancement
Perception Optimized Deep Denoising AutoEncoders for Speech Enhancement Prashanth Gurunath Shivakumar, Panayiotis Georgiou University of Southern California, Los Angeles, CA, USA pgurunat@uscedu, georgiou@sipiuscedu
More informationChallenges in microphone array processing for hearing aids. Volkmar Hamacher Siemens Audiological Engineering Group Erlangen, Germany
Challenges in microphone array processing for hearing aids Volkmar Hamacher Siemens Audiological Engineering Group Erlangen, Germany SIEMENS Audiological Engineering Group R&D Signal Processing and Audiology
More informationSonic Spotlight. Binaural Coordination: Making the Connection
Binaural Coordination: Making the Connection 1 Sonic Spotlight Binaural Coordination: Making the Connection Binaural Coordination is the global term that refers to the management of wireless technology
More informationLiterature Overview - Digital Hearing Aids and Group Delay - HADF, June 2017, P. Derleth
Literature Overview - Digital Hearing Aids and Group Delay - HADF, June 2017, P. Derleth Historic Context Delay in HI and the perceptual effects became a topic with the widespread market introduction of
More informationSpeech intelligibility in background noise with ideal binary time-frequency masking
Speech intelligibility in background noise with ideal binary time-frequency masking DeLiang Wang a Department of Computer Science & Engineering and Center for Cognitive Science, The Ohio State University,
More informationLinguistic Phonetics Fall 2005
MIT OpenCourseWare http://ocw.mit.edu 24.963 Linguistic Phonetics Fall 2005 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. 24.963 Linguistic Phonetics
More informationProceedings of Meetings on Acoustics
Proceedings of Meetings on Acoustics Volume 19, 2013 http://acousticalsociety.org/ ICA 2013 Montreal Montreal, Canada 2-7 June 2013 Psychological and Physiological Acoustics Session 4aPPb: Binaural Hearing
More informationUSING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES
USING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES Varinthira Duangudom and David V Anderson School of Electrical and Computer Engineering, Georgia Institute of Technology Atlanta, GA 30332
More informationIssues faced by people with a Sensorineural Hearing Loss
Issues faced by people with a Sensorineural Hearing Loss Issues faced by people with a Sensorineural Hearing Loss 1. Decreased Audibility 2. Decreased Dynamic Range 3. Decreased Frequency Resolution 4.
More information16.400/453J Human Factors Engineering /453. Audition. Prof. D. C. Chandra Lecture 14
J Human Factors Engineering Audition Prof. D. C. Chandra Lecture 14 1 Overview Human ear anatomy and hearing Auditory perception Brainstorming about sounds Auditory vs. visual displays Considerations for
More informationAutomatic Live Monitoring of Communication Quality for Normal-Hearing and Hearing-Impaired Listeners
Automatic Live Monitoring of Communication Quality for Normal-Hearing and Hearing-Impaired Listeners Jan Rennies, Eugen Albertin, Stefan Goetze, and Jens-E. Appell Fraunhofer IDMT, Hearing, Speech and
More informationSpeech Enhancement Based on Deep Neural Networks
Speech Enhancement Based on Deep Neural Networks Chin-Hui Lee School of ECE, Georgia Tech chl@ece.gatech.edu Joint work with Yong Xu and Jun Du at USTC 1 Outline and Talk Agenda In Signal Processing Letter,
More informationBroadband Wireless Access and Applications Center (BWAC) CUA Site Planning Workshop
Broadband Wireless Access and Applications Center (BWAC) CUA Site Planning Workshop Lin-Ching Chang Department of Electrical Engineering and Computer Science School of Engineering Work Experience 09/12-present,
More informationSpeech intelligibility in simulated acoustic conditions for normal hearing and hearing-impaired listeners
Speech intelligibility in simulated acoustic conditions for normal hearing and hearing-impaired listeners Ir i s Arw e i l e r 1, To r b e n Po u l s e n 2, a n d To r s t e n Da u 1 1 Centre for Applied
More informationThe Effects of Hearing Loss on Binaural Function
The Effects of Hearing Loss on Binaural Function Steve Colburn Boston University colburn@bu.edu Outline What are the major advantages of two ears? How do the left and right signals differ? How do we use
More informationSpectrograms (revisited)
Spectrograms (revisited) We begin the lecture by reviewing the units of spectrograms, which I had only glossed over when I covered spectrograms at the end of lecture 19. We then relate the blocks of a
More informationNatural statistics of binaural sounds
Natural statistics of binaural sounds Wiktor M lynarski 1 and Jürgen Jost 1,2 1 Max-Planck Institute for Mathematics in the Sciences, Leipzig, Germany 2 Santa Fe Institute, Santa Fe, New Mexico, USA August
More informationSpeech Enhancement Using Deep Neural Network
Speech Enhancement Using Deep Neural Network Pallavi D. Bhamre 1, Hemangi H. Kulkarni 2 1 Post-graduate Student, Department of Electronics and Telecommunication, R. H. Sapat College of Engineering, Management
More informationSpeech and Sound Use in a Remote Monitoring System for Health Care
Speech and Sound Use in a Remote System for Health Care M. Vacher J.-F. Serignat S. Chaillol D. Istrate V. Popescu CLIPS-IMAG, Team GEOD Joseph Fourier University of Grenoble - CNRS (France) Text, Speech
More informationHST.723J, Spring 2005 Theme 3 Report
HST.723J, Spring 2005 Theme 3 Report Madhu Shashanka shashanka@cns.bu.edu Introduction The theme of this report is binaural interactions. Binaural interactions of sound stimuli enable humans (and other
More information