Tandem modeling investigations

Size: px
Start display at page:

Download "Tandem modeling investigations"

Transcription

1 Tandem modeling investigations Dan Ellis International Computer Science Institute, Berkeley CA Outline What makes Tandem successful? Can we make Tandem better? Does Tandem work with LVCSR tricks? Tandem investigations - Dan Ellis

2 +w +w 1 Combo over msg: +20% What makes Tandem work? plp (with Manuel Reyes) Input sound Pre-nonlinearity over posteriors: +12% + PCA orthog'n Gauss mix models HTK decoder s ah t Combo over plp: +20% msg KLT over direct: +8% Combo-into-HTK over Combo-into-noway: +15% Words Combo over mfcc: +25% NN over HTK: +15% Tandem over HTK: +35% Tandem over hybrid: +25% Tandem combo over HTK mfcc baseline: +53% Model diversity? - try a phone-based GMM model - try training the NN model to HTK state labels Discriminative network training? - (try posteriors from GMM & Bayes) Tandem investigations - Dan Ellis

3 Phone vs. word models plp KLT orthog'n Gauss mix models HTK decoder Input sound +w s ah t Words Trained on phoneme targets Trained to subword states Try a phone-based HTK model (instead of whole-word models) Try training NN model to subword-state labels net outputs; reduce to 40 in KLT Results (Aurora2k, HTK-baseline WER ratio): System test A: matched test B: var noise test C: var chan Tandem PLP baseline 63.5% 70l.3% 59.5% Phone-based HTK sys 63.6% 72.5% 61.5% Subword-based NN sys 63.1% 62.8% 55.1% Diversity doesn t help - subword units may be good for NN Tandem investigations - Dan Ellis

4 2 Enhancements to Tandem-Aurora More tandem-feature-domain processing: norm / deltas? KLT orthog'n norm / deltas? Gauss mix models +w Results (HTK baseline WER ratio): System test A: matched test B: var noise test C: var chan PLP: Tandem baseline 63.5% 70l.3% 59.5% PLP: norm - KLT 72.6% 71.2% 63.6% PLP: KLT - norm 57.8% 58.8% 51.3% PLP: KLT - delta 59.0% 60.2% 52.9% PLP: KLT - delta - norm 58.1% 59.9% 48.9% PLP: delta - KLT - norm 54.7% 53.6% 46.9% - delta-klt-norm: 80% Tdm baseline WER Tandem investigations - Dan Ellis

5 Best effort Tandem system Deltas & norms help PLP: try on combo (PLP+MSG) system: System test A: matched test B: var noise test C: var chan PLP+MSG: baseline 51.1% 52.0% 45.6% PLP+MSG: dlt-klt-nrm 50.9% 50.5% 43.6% PLP+MSG: KLT-nrm 48.3% 49.5% 39.4% - deltas hurt for MSG: features too sluggish? Deltas help clean, norms help noisy: baseline K-D K-N WER / % 40 4 WER / % clean SNR / db Tandem investigations - Dan Ellis

6 +w +w 3 Tandem for LVCSR: the SPINE task (with Rita Singh/CMU & Sunil Sivadas/OGI) Noisy spontaneous speech, ~5000 word vocab Recognition: PLP feature calculation 1 Tandem feature calculation SPHINX recognizer Input sound Pre-nonlinearity outputs + PCA decorrelation GMM Subword likelihoods HMM decoder s ah t Words MSG feature calculation 2 MLLR adaptation - same tandem features - NN training from Broadcast News boot + iterate - GMM-HMM has context-dependence, MLLR Tandem investigations - Dan Ellis

7 SPINE-Tandem results Evaluation WER results: Features (dimensions) CI system CD system CD + MLLR MFCC + d + dd (39) 69.5% 35.1% 33.5% Tandem features (56) 47.6% 35.7% 32.8% - much better for CI systems - differences evaporate with CD, MLLR Not quite fair: - CD senones optimized for MFCC - worth 2-3% absolute? Not unexpected: - NN confounds CD variants - Tandem space very nonlinear - bad for MLLR Any hope? - more training data / train CD classes /... Tandem investigations - Dan Ellis

Recognition & Organization of Speech and Audio

Recognition & Organization of Speech and Audio Recognition & Organization of Speech and Audio Dan Ellis Electrical Engineering, Columbia University http://www.ee.columbia.edu/~dpwe/ Outline 1 2 3 4 Introducing Robust speech recognition

More information

Tandem acoustic modeling: Neural nets for mainstream ASR?

Tandem acoustic modeling: Neural nets for mainstream ASR? Tandem acoutic modeling: for maintream ASR? Dan Elli International Computer Science Intitute Berkeley CA dpwe@ici.berkeley.edu Outline 2 3 Tandem acoutic modeling Inide Tandem ytem: What going on? Future

More information

Recognition & Organization of Speech and Audio

Recognition & Organization of Speech and Audio Recognition & Organization of Speech and Audio Dan Ellis Electrical Engineering, Columbia University http://www.ee.columbia.edu/~dpwe/ Outline 1 2 3 4 5 Introducing Tandem modeling

More information

Recognition & Organization of Speech & Audio

Recognition & Organization of Speech & Audio Recognition & Organization of Speech & Audio Dan Ellis http://labrosa.ee.columbia.edu/ Outline 1 2 3 Introducing Projects in speech, music & audio Summary overview - Dan Ellis 21-9-28-1 1 Sound organization

More information

Recognition & Organization of Speech and Audio

Recognition & Organization of Speech and Audio Recognition & Organization of Speech and Audio Dan Ellis Electrical Engineering, Columbia University Outline 1 2 3 4 5 Sound organization Background & related work Existing projects

More information

Using Source Models in Speech Separation

Using Source Models in Speech Separation Using Source Models in Speech Separation Dan Ellis Laboratory for Recognition and Organization of Speech and Audio Dept. Electrical Eng., Columbia Univ., NY USA dpwe@ee.columbia.edu http://labrosa.ee.columbia.edu/

More information

Lecture 9: Speech Recognition: Front Ends

Lecture 9: Speech Recognition: Front Ends EE E682: Speech & Audio Processing & Recognition Lecture 9: Speech Recognition: Front Ends 1 2 Recognizing Speech Feature Calculation Dan Ellis http://www.ee.columbia.edu/~dpwe/e682/

More information

Sound, Mixtures, and Learning

Sound, Mixtures, and Learning Sound, Mixtures, and Learning Dan Ellis Laboratory for Recognition and Organization of Speech and Audio (LabROSA) Electrical Engineering, Columbia University http://labrosa.ee.columbia.edu/

More information

Sound, Mixtures, and Learning

Sound, Mixtures, and Learning Sound, Mixtures, and Learning Dan Ellis oratory for Recognition and Organization of Speech and Audio () Electrical Engineering, Columbia University http://labrosa.ee.columbia.edu/

More information

General Soundtrack Analysis

General Soundtrack Analysis General Soundtrack Analysis Dan Ellis oratory for Recognition and Organization of Speech and Audio () Electrical Engineering, Columbia University http://labrosa.ee.columbia.edu/

More information

Speech recognition in noisy environments: A survey

Speech recognition in noisy environments: A survey T-61.182 Robustness in Language and Speech Processing Speech recognition in noisy environments: A survey Yifan Gong presented by Tapani Raiko Feb 20, 2003 About the Paper Article published in Speech Communication

More information

Noise-Robust Speech Recognition Technologies in Mobile Environments

Noise-Robust Speech Recognition Technologies in Mobile Environments Noise-Robust Speech Recognition echnologies in Mobile Environments Mobile environments are highly influenced by ambient noise, which may cause a significant deterioration of speech recognition performance.

More information

EECS 433 Statistical Pattern Recognition

EECS 433 Statistical Pattern Recognition EECS 433 Statistical Pattern Recognition Ying Wu Electrical Engineering and Computer Science Northwestern University Evanston, IL 60208 http://www.eecs.northwestern.edu/~yingwu 1 / 19 Outline What is Pattern

More information

A Comparison of Deep Neural Network Training Methods for Large Vocabulary Speech Recognition

A Comparison of Deep Neural Network Training Methods for Large Vocabulary Speech Recognition A Comparison of Deep Neural Network Training Methods for Large Vocabulary Speech Recognition LászlóTóth and Tamás Grósz MTA-SZTE Research Group on Artificial Intelligence Hungarian Academy of Sciences

More information

Visualization tools & demos and the ICSI Realization group

Visualization tools & demos and the ICSI Realization group Visualization tools & demos and the ICSI Realization group Dan Ellis International Computer Science Institute, Berkeley CA Outline 1 2 3 4 Visualization: Goals and philosophy Tour

More information

Review of SPRACH/Thisl meetings Cambridge UK, 1998sep03/04

Review of SPRACH/Thisl meetings Cambridge UK, 1998sep03/04 Review of SPRACH/Thisl meetings Cambridge UK, 1998sep03/04 Dan Ellis International Computer Science Institute, Berkeley CA Outline 1 2 3 4 5 SPRACH overview The SPRACH Broadcast

More information

Broadband Wireless Access and Applications Center (BWAC) CUA Site Planning Workshop

Broadband Wireless Access and Applications Center (BWAC) CUA Site Planning Workshop Broadband Wireless Access and Applications Center (BWAC) CUA Site Planning Workshop Lin-Ching Chang Department of Electrical Engineering and Computer Science School of Engineering Work Experience 09/12-present,

More information

Automatic audio analysis for content description & indexing

Automatic audio analysis for content description & indexing Automatic audio analysis for content description & indexing Dan Ellis International Computer Science Institute, Berkeley CA Outline 1 2 3 4 5 Auditory Scene Analysis (ASA) Computational

More information

Sonic Spotlight. SmartCompress. Advancing compression technology into the future

Sonic Spotlight. SmartCompress. Advancing compression technology into the future Sonic Spotlight SmartCompress Advancing compression technology into the future Speech Variable Processing (SVP) is the unique digital signal processing strategy that gives Sonic hearing aids their signature

More information

Using the Soundtrack to Classify Videos

Using the Soundtrack to Classify Videos Using the Soundtrack to Classify Videos Dan Ellis Laboratory for Recognition and Organization of Speech and Audio Dept. Electrical Eng., Columbia Univ., NY USA dpwe@ee.columbia.edu http://labrosa.ee.columbia.edu/

More information

Robustness, Separation & Pitch

Robustness, Separation & Pitch Robustness, Separation & Pitch or Morgan, Me & Pitch Dan Ellis Columbia / ICSI dpwe@ee.columbia.edu http://labrosa.ee.columbia.edu/ 1. Robustness and Separation 2. An Academic Journey 3. Future COLUMBIA

More information

Adaptation of Classification Model for Improving Speech Intelligibility in Noise

Adaptation of Classification Model for Improving Speech Intelligibility in Noise 1: (Junyoung Jung et al.: Adaptation of Classification Model for Improving Speech Intelligibility in Noise) (Regular Paper) 23 4, 2018 7 (JBE Vol. 23, No. 4, July 2018) https://doi.org/10.5909/jbe.2018.23.4.511

More information

Acoustic Signal Processing Based on Deep Neural Networks

Acoustic Signal Processing Based on Deep Neural Networks Acoustic Signal Processing Based on Deep Neural Networks Chin-Hui Lee School of ECE, Georgia Tech chl@ece.gatech.edu Joint work with Yong Xu, Yanhui Tu, Qing Wang, Tian Gao, Jun Du, LiRong Dai Outline

More information

VIRTUAL ASSISTANT FOR DEAF AND DUMB

VIRTUAL ASSISTANT FOR DEAF AND DUMB VIRTUAL ASSISTANT FOR DEAF AND DUMB Sumanti Saini, Shivani Chaudhry, Sapna, 1,2,3 Final year student, CSE, IMSEC Ghaziabad, U.P., India ABSTRACT For the Deaf and Dumb community, the use of Information

More information

Auditory-based Automatic Speech Recognition

Auditory-based Automatic Speech Recognition Auditory-based Automatic Speech Recognition Werner Hemmert, Marcus Holmberg Infineon Technologies, Corporate Research Munich, Germany {Werner.Hemmert,Marcus.Holmberg}@infineon.com David Gelbart International

More information

SPEECH recordings taken from realistic environments typically

SPEECH recordings taken from realistic environments typically IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING 1 Coupled Dictionaries for Exemplar-based Speech Enhancement and Automatic Speech Recognition Deepak Baby, Student Member, IEEE, Tuomas Virtanen,

More information

Modulation and Top-Down Processing in Audition

Modulation and Top-Down Processing in Audition Modulation and Top-Down Processing in Audition Malcolm Slaney 1,2 and Greg Sell 2 1 Yahoo! Research 2 Stanford CCRMA Outline The Non-Linear Cochlea Correlogram Pitch Modulation and Demodulation Information

More information

I. INTRODUCTION. OMBARD EFFECT (LE), named after the French otorhino-laryngologist

I. INTRODUCTION. OMBARD EFFECT (LE), named after the French otorhino-laryngologist IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 18, NO. 6, AUGUST 2010 1379 Unsupervised Equalization of Lombard Effect for Speech Recognition in Noisy Adverse Environments Hynek Bořil,

More information

Amplitude Compression: Timing is Everything

Amplitude Compression: Timing is Everything Amplitude Compression: Timing is Everything Andrea Pittman Arizona State University Funded by a grant from Oticon A/S Output (db SPL) Wide Dynamic Range Compression 100 80 60 40 20 40 60 80 100 Input (db

More information

Speech and Sound Use in a Remote Monitoring System for Health Care

Speech and Sound Use in a Remote Monitoring System for Health Care Speech and Sound Use in a Remote System for Health Care M. Vacher J.-F. Serignat S. Chaillol D. Istrate V. Popescu CLIPS-IMAG, Team GEOD Joseph Fourier University of Grenoble - CNRS (France) Text, Speech

More information

Computational Auditory Scene Analysis: An overview and some observations. CASA survey. Other approaches

Computational Auditory Scene Analysis: An overview and some observations. CASA survey. Other approaches CASA talk - Haskins/NUWC - Dan Ellis 1997oct24/5-1 The importance of auditory illusions for artificial listeners 1 Dan Ellis International Computer Science Institute, Berkeley CA

More information

Robust Speech Detection for Noisy Environments

Robust Speech Detection for Noisy Environments Robust Speech Detection for Noisy Environments Óscar Varela, Rubén San-Segundo and Luis A. Hernández ABSTRACT This paper presents a robust voice activity detector (VAD) based on hidden Markov models (HMM)

More information

Single-Channel Sound Source Localization Based on Discrimination of Acoustic Transfer Functions

Single-Channel Sound Source Localization Based on Discrimination of Acoustic Transfer Functions 3 Single-Channel Sound Source Localization Based on Discrimination of Acoustic Transfer Functions Ryoichi Takashima, Tetsuya Takiguchi and Yasuo Ariki Graduate School of System Informatics, Kobe University,

More information

Mitigating the Effects of Non-Stationary Unseen Noises on Language Recognition Performance

Mitigating the Effects of Non-Stationary Unseen Noises on Language Recognition Performance INTERSPEECH 2015 Mitigating the Effects of Non-Stationary Unseen Noises on Language Recognition Performance Luciana Ferrer 1, Mitchell McLaren 2, Aaron Lawson 2, Martin Graciarena 2 1 Departamento de Computación,

More information

LATERAL INHIBITION MECHANISM IN COMPUTATIONAL AUDITORY MODEL AND IT'S APPLICATION IN ROBUST SPEECH RECOGNITION

LATERAL INHIBITION MECHANISM IN COMPUTATIONAL AUDITORY MODEL AND IT'S APPLICATION IN ROBUST SPEECH RECOGNITION LATERAL INHIBITION MECHANISM IN COMPUTATIONAL AUDITORY MODEL AND IT'S APPLICATION IN ROBUST SPEECH RECOGNITION Lu Xugang Li Gang Wang Lip0 Nanyang Technological University, School of EEE, Workstation Resource

More information

Exploiting visual information for NAM recognition

Exploiting visual information for NAM recognition Exploiting visual information for NAM recognition Panikos Heracleous, Denis Beautemps, Viet-Anh Tran, Hélène Loevenbruck, Gérard Bailly To cite this version: Panikos Heracleous, Denis Beautemps, Viet-Anh

More information

Advanced Audio Interface for Phonetic Speech. Recognition in a High Noise Environment

Advanced Audio Interface for Phonetic Speech. Recognition in a High Noise Environment DISTRIBUTION STATEMENT A Approved for Public Release Distribution Unlimited Advanced Audio Interface for Phonetic Speech Recognition in a High Noise Environment SBIR 99.1 TOPIC AF99-1Q3 PHASE I SUMMARY

More information

Speech Enhancement Based on Deep Neural Networks

Speech Enhancement Based on Deep Neural Networks Speech Enhancement Based on Deep Neural Networks Chin-Hui Lee School of ECE, Georgia Tech chl@ece.gatech.edu Joint work with Yong Xu and Jun Du at USTC 1 Outline and Talk Agenda In Signal Processing Letter,

More information

Online Speaker Adaptation of an Acoustic Model using Face Recognition

Online Speaker Adaptation of an Acoustic Model using Face Recognition Online Speaker Adaptation of an Acoustic Model using Face Recognition Pavel Campr 1, Aleš Pražák 2, Josef V. Psutka 2, and Josef Psutka 2 1 Center for Machine Perception, Department of Cybernetics, Faculty

More information

CONSTRUCTING TELEPHONE ACOUSTIC MODELS FROM A HIGH-QUALITY SPEECH CORPUS

CONSTRUCTING TELEPHONE ACOUSTIC MODELS FROM A HIGH-QUALITY SPEECH CORPUS CONSTRUCTING TELEPHONE ACOUSTIC MODELS FROM A HIGH-QUALITY SPEECH CORPUS Mitchel Weintraub and Leonardo Neumeyer SRI International Speech Research and Technology Program Menlo Park, CA, 94025 USA ABSTRACT

More information

SPEECH TO TEXT CONVERTER USING GAUSSIAN MIXTURE MODEL(GMM)

SPEECH TO TEXT CONVERTER USING GAUSSIAN MIXTURE MODEL(GMM) SPEECH TO TEXT CONVERTER USING GAUSSIAN MIXTURE MODEL(GMM) Virendra Chauhan 1, Shobhana Dwivedi 2, Pooja Karale 3, Prof. S.M. Potdar 4 1,2,3B.E. Student 4 Assitant Professor 1,2,3,4Department of Electronics

More information

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2014

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2014 UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2014 Exam policy: This exam allows two one-page, two-sided cheat sheets (i.e. 4 sides); No other materials. Time: 2 hours. Be sure to write

More information

Oregon Graduate Institute of Science and Technology,

Oregon Graduate Institute of Science and Technology, SPEAKER RECOGNITION AT OREGON GRADUATE INSTITUTE June & 6, 997 Sarel van Vuuren and Narendranath Malayath Hynek Hermansky and Pieter Vermeulen, Oregon Graduate Institute, Portland, Oregon Oregon Graduate

More information

Using Speech Models for Separation

Using Speech Models for Separation Using Speech Models for Separation Dan Ellis Comprising the work of Michael Mandel and Ron Weiss Laboratory for Recognition and Organization of Speech and Audio Dept. Electrical Eng., Columbia Univ., NY

More information

EEL 6586, Project - Hearing Aids algorithms

EEL 6586, Project - Hearing Aids algorithms EEL 6586, Project - Hearing Aids algorithms 1 Yan Yang, Jiang Lu, and Ming Xue I. PROBLEM STATEMENT We studied hearing loss algorithms in this project. As the conductive hearing loss is due to sound conducting

More information

arxiv: v1 [cs.cv] 13 Mar 2018

arxiv: v1 [cs.cv] 13 Mar 2018 RESOURCE AWARE DESIGN OF A DEEP CONVOLUTIONAL-RECURRENT NEURAL NETWORK FOR SPEECH RECOGNITION THROUGH AUDIO-VISUAL SENSOR FUSION Matthijs Van keirsbilck Bert Moons Marian Verhelst MICAS, Department of

More information

The Unifi-EV Protocol for Evalita 2009

The Unifi-EV Protocol for Evalita 2009 AI*IA 2009: International Conference on Artificial Intelligence EVALITA Workshop Reggio Emilia, 12 Dicembre 2009 The Unifi-EV2009-1 Protocol for Evalita 2009 Prof. Ing. Monica Carfagni, Ing. Matteo Nunziati

More information

Audio-Visual Speech Recognition Using Bimodal-Trained Bottleneck Features for a Person with Severe Hearing Loss

Audio-Visual Speech Recognition Using Bimodal-Trained Bottleneck Features for a Person with Severe Hearing Loss INTERSPEECH 2016 September 8 12, 2016, San Francisco, USA Audio-Visual Speech Recognition Using Bimodal-Trained Features for a Person with Severe Hearing Loss Yuki Takashima 1, Ryo Aihara 1, Tetsuya Takiguchi

More information

A Vision-based Affective Computing System. Jieyu Zhao Ningbo University, China

A Vision-based Affective Computing System. Jieyu Zhao Ningbo University, China A Vision-based Affective Computing System Jieyu Zhao Ningbo University, China Outline Affective Computing A Dynamic 3D Morphable Model Facial Expression Recognition Probabilistic Graphical Models Some

More information

Improving the Noise-Robustness of Mel-Frequency Cepstral Coefficients for Speech Processing

Improving the Noise-Robustness of Mel-Frequency Cepstral Coefficients for Speech Processing ISCA Archive http://www.isca-speech.org/archive ITRW on Statistical and Perceptual Audio Processing Pittsburgh, PA, USA September 16, 2006 Improving the Noise-Robustness of Mel-Frequency Cepstral Coefficients

More information

Noise-Robust Speech Recognition in a Car Environment Based on the Acoustic Features of Car Interior Noise

Noise-Robust Speech Recognition in a Car Environment Based on the Acoustic Features of Car Interior Noise 4 Special Issue Speech-Based Interfaces in Vehicles Research Report Noise-Robust Speech Recognition in a Car Environment Based on the Acoustic Features of Car Interior Noise Hiroyuki Hoshino Abstract This

More information

Voice Detection using Speech Energy Maximization and Silence Feature Normalization

Voice Detection using Speech Energy Maximization and Silence Feature Normalization , pp.25-29 http://dx.doi.org/10.14257/astl.2014.49.06 Voice Detection using Speech Energy Maximization and Silence Feature Normalization In-Sung Han 1 and Chan-Shik Ahn 2 1 Dept. of The 2nd R&D Institute,

More information

Re/Habilitation of the Hearing Impaired. Better Hearing Philippines Inc.

Re/Habilitation of the Hearing Impaired. Better Hearing Philippines Inc. Re/Habilitation of the Hearing Impaired Better Hearing Philippines Inc. Nature of Hearing Loss Decreased Audibility Decreased Dynamic Range Decreased Frequency Resolution Decreased Temporal Resolution

More information

Evaluating Auditory Contexts and Their Impacts on Hearing Aid Outcomes with Mobile Phones

Evaluating Auditory Contexts and Their Impacts on Hearing Aid Outcomes with Mobile Phones Evaluating Auditory Contexts and Their Impacts on Hearing Aid Outcomes with Mobile Phones Syed Shabih Hasan, Octav Chipara Department of Computer Science/Aging Mind and Brain Initiative (AMBI) Yu-Hsiang

More information

Research Article Automatic Speaker Recognition for Mobile Forensic Applications

Research Article Automatic Speaker Recognition for Mobile Forensic Applications Hindawi Mobile Information Systems Volume 07, Article ID 698639, 6 pages https://doi.org//07/698639 Research Article Automatic Speaker Recognition for Mobile Forensic Applications Mohammed Algabri, Hassan

More information

COMPARISON BETWEEN GMM-SVM SEQUENCE KERNEL AND GMM: APPLICATION TO SPEECH EMOTION RECOGNITION

COMPARISON BETWEEN GMM-SVM SEQUENCE KERNEL AND GMM: APPLICATION TO SPEECH EMOTION RECOGNITION Journal of Engineering Science and Technology Vol. 11, No. 9 (2016) 1221-1233 School of Engineering, Taylor s University COMPARISON BETWEEN GMM-SVM SEQUENCE KERNEL AND GMM: APPLICATION TO SPEECH EMOTION

More information

HCS 7367 Speech Perception

HCS 7367 Speech Perception Long-term spectrum of speech HCS 7367 Speech Perception Connected speech Absolute threshold Males Dr. Peter Assmann Fall 212 Females Long-term spectrum of speech Vowels Males Females 2) Absolute threshold

More information

Review: Logistic regression, Gaussian naïve Bayes, linear regression, and their connections

Review: Logistic regression, Gaussian naïve Bayes, linear regression, and their connections Review: Logistic regression, Gaussian naïve Bayes, linear regression, and their connections New: Bias-variance decomposition, biasvariance tradeoff, overfitting, regularization, and feature selection Yi

More information

The SRI System for the NIST OpenSAD 2015 Speech Activity Detection Evaluation

The SRI System for the NIST OpenSAD 2015 Speech Activity Detection Evaluation INTERSPEECH 2016 September 8 12, 2016, San Francisco, USA The SRI System for the NIST OpenSAD 2015 Speech Activity Detection Evaluation Martin Graciarena 1, Luciana Ferrer 2, Vikramjit Mitra 1 1 SRI International,

More information

Noisy training for deep neural networks in speech recognition

Noisy training for deep neural networks in speech recognition Yin et al. EURASIP Journal on Audio, Speech, and Music Processing (15) 15:2 DOI.1186/s13636-14-47- RESEARCH Open Access Noisy training for deep neural networks in speech recognition Shi Yin 1,4,ChaoLiu

More information

Introduction to Machine Learning. Katherine Heller Deep Learning Summer School 2018

Introduction to Machine Learning. Katherine Heller Deep Learning Summer School 2018 Introduction to Machine Learning Katherine Heller Deep Learning Summer School 2018 Outline Kinds of machine learning Linear regression Regularization Bayesian methods Logistic Regression Why we do this

More information

Gender Based Emotion Recognition using Speech Signals: A Review

Gender Based Emotion Recognition using Speech Signals: A Review 50 Gender Based Emotion Recognition using Speech Signals: A Review Parvinder Kaur 1, Mandeep Kaur 2 1 Department of Electronics and Communication Engineering, Punjabi University, Patiala, India 2 Department

More information

Ambiguity in the recognition of phonetic vowels when using a bone conduction microphone

Ambiguity in the recognition of phonetic vowels when using a bone conduction microphone Acoustics 8 Paris Ambiguity in the recognition of phonetic vowels when using a bone conduction microphone V. Zimpfer a and K. Buck b a ISL, 5 rue du Général Cassagnou BP 734, 6831 Saint Louis, France b

More information

Interpreting Speech Results to Optimise Hearing aid Fittings

Interpreting Speech Results to Optimise Hearing aid Fittings Interpreting Speech Results to Optimise Hearing aid Fittings Josephine Marriage PhD josephine@chears.co.uk BAA Monday 18 th November 2013 How has the role of families changed with introduction of NHSP?

More information

Practical Bayesian Optimization of Machine Learning Algorithms. Jasper Snoek, Ryan Adams, Hugo LaRochelle NIPS 2012

Practical Bayesian Optimization of Machine Learning Algorithms. Jasper Snoek, Ryan Adams, Hugo LaRochelle NIPS 2012 Practical Bayesian Optimization of Machine Learning Algorithms Jasper Snoek, Ryan Adams, Hugo LaRochelle NIPS 2012 ... (Gaussian Processes) are inadequate for doing speech and vision. I still think they're

More information

Error Detection based on neural signals

Error Detection based on neural signals Error Detection based on neural signals Nir Even- Chen and Igor Berman, Electrical Engineering, Stanford Introduction Brain computer interface (BCI) is a direct communication pathway between the brain

More information

The WIDEX DREAM family has just got even better.

The WIDEX DREAM family has just got even better. THE DREAM CONTINUES The WIDEX DREAM family has just got even better. The new DREAM FASHION model gives you the chance to offer users the absolute latest in Behind-the-ear technology and design. Small,

More information

Optimising Outcomes for Speech Mapping

Optimising Outcomes for Speech Mapping Optimising Outcomes for Speech Mapping Jan Pollard Chief Audiologist Sonic Craig Lett Audiologist ECS Back to Basics What is the point of all of this? Optimising the fitting for best patient outcomes and

More information

LIE DETECTION SYSTEM USING INPUT VOICE SIGNAL K.Meena 1, K.Veena 2 (Corresponding Author: K.Veena) 1 Associate Professor, 2 Research Scholar,

LIE DETECTION SYSTEM USING INPUT VOICE SIGNAL K.Meena 1, K.Veena 2 (Corresponding Author: K.Veena) 1 Associate Professor, 2 Research Scholar, International Journal of Pure and Applied Mathematics Volume 117 No. 8 2017, 121-125 ISSN: 1311-8080 (printed version); ISSN: 1314-3395 (on-line version) url: http://www.ijpam.eu doi: 10.12732/ijpam.v117i8.25

More information

HISTOGRAM EQUALIZATION BASED FRONT-END PROCESSING FOR NOISY SPEECH RECOGNITION

HISTOGRAM EQUALIZATION BASED FRONT-END PROCESSING FOR NOISY SPEECH RECOGNITION 2 th May 216. Vol.87. No.2 25-216 JATIT & LLS. All rights reserved. HISTOGRAM EQUALIZATION BASED FRONT-END PROCESSING FOR NOISY SPEECH RECOGNITION 1 IBRAHIM MISSAOUI, 2 ZIED LACHIRI National Engineering

More information

Computational Auditory Scene Analysis: Principles, Practice and Applications

Computational Auditory Scene Analysis: Principles, Practice and Applications Computational Auditory Scene Analysis: Principles, Practice and Applications Dan Ellis International Computer Science Institute, Berkeley CA Outline 1 2 3 4 5 Auditory Scene Analysis

More information

SoundRecover2 the first adaptive frequency compression algorithm More audibility of high frequency sounds

SoundRecover2 the first adaptive frequency compression algorithm More audibility of high frequency sounds Phonak Insight April 2016 SoundRecover2 the first adaptive frequency compression algorithm More audibility of high frequency sounds Phonak led the way in modern frequency lowering technology with the introduction

More information

Speech Enhancement Based on Spectral Subtraction Involving Magnitude and Phase Components

Speech Enhancement Based on Spectral Subtraction Involving Magnitude and Phase Components Speech Enhancement Based on Spectral Subtraction Involving Magnitude and Phase Components Miss Bhagat Nikita 1, Miss Chavan Prajakta 2, Miss Dhaigude Priyanka 3, Miss Ingole Nisha 4, Mr Ranaware Amarsinh

More information

Analysis of Emotion Recognition using Facial Expressions, Speech and Multimodal Information

Analysis of Emotion Recognition using Facial Expressions, Speech and Multimodal Information Analysis of Emotion Recognition using Facial Expressions, Speech and Multimodal Information C. Busso, Z. Deng, S. Yildirim, M. Bulut, C. M. Lee, A. Kazemzadeh, S. Lee, U. Neumann, S. Narayanan Emotion

More information

An Examination of Speech In Noise and its Effect on Understandability for Natural and Synthetic Speech. Brian Langner and Alan W Black CMU-LTI

An Examination of Speech In Noise and its Effect on Understandability for Natural and Synthetic Speech. Brian Langner and Alan W Black CMU-LTI An Examination of Speech In Noise and its Effect on Understandability for Natural and Synthetic Speech Brian Langner and Alan W Black CMU-LTI-4-187 FINAL Language Technologies Institute School of Computer

More information

What is motor planning?

What is motor planning? Motor Planning What is motor planning? Children with ASD have been found to have deficits and delays in: Gross and Fine motor skills Adaptive or daily living skills (eating, dressing bathing) Balance and

More information

Reviewing the connection between speech and obstructive sleep apnea

Reviewing the connection between speech and obstructive sleep apnea DOI 10.1186/s12938-016-0138-5 BioMedical Engineering OnLine RESEARCH Open Access Reviewing the connection between speech and obstructive sleep apnea Fernando Espinoza Cuadros 1*, Rubén Fernández Pozo 1,

More information

Speaker Independent Isolated Word Speech to Text Conversion Using Auto Spectral Subtraction for Punjabi Language

Speaker Independent Isolated Word Speech to Text Conversion Using Auto Spectral Subtraction for Punjabi Language International Journal of Scientific and Research Publications, Volume 7, Issue 7, July 2017 469 Speaker Independent Isolated Word Speech to Text Conversion Using Auto Spectral Subtraction for Punjabi Language

More information

We are IntechOpen, the world s leading publisher of Open Access books Built by scientists, for scientists. International authors and editors

We are IntechOpen, the world s leading publisher of Open Access books Built by scientists, for scientists. International authors and editors We are IntechOpen, the world s leading publisher of Open Access books Built by scientists, for scientists 3,900 116,000 120M Open access books available International authors and editors Downloads Our

More information

A New Paradigm for the Evaluation of Forensic Evidence

A New Paradigm for the Evaluation of Forensic Evidence A New Paradigm for the Evaluation of Forensic Evidence and its implementation in forensic voice comparison Geoffrey Stewart Morrison Ewald Enzinger p(e H ) p p(e H ) d Abstract & Biographies In Europe

More information

AN-ftlf 644 FINAL REPORT FOR CONTRACT IWIM NNI4-0U-C-NSI (OUR) IA1 (U) SIGNITION INC LOS ALAMOS NN HEARING RESEARCH LRD G ZWEIG 30 SEP 67

AN-ftlf 644 FINAL REPORT FOR CONTRACT IWIM NNI4-0U-C-NSI (OUR) IA1 (U) SIGNITION INC LOS ALAMOS NN HEARING RESEARCH LRD G ZWEIG 30 SEP 67 AN-ftlf 644 FINAL REPORT FOR CONTRACT IWIM NNI4-0U-C-NSI (OUR) IA1 (U) SIGNITION INC LOS ALAMOS NN HEARING RESEARCH LRD G ZWEIG 30 SEP 67 MM14-86-C-MSI WICLMSSIFIED F/O 26/1 ML IIII 1.0 t 2.0 *~ W3 MICROCOPY

More information

Analysis of Speech Recognition Techniques for use in a Non-Speech Sound Recognition System

Analysis of Speech Recognition Techniques for use in a Non-Speech Sound Recognition System Analysis of Recognition Techniques for use in a Sound Recognition System Michael Cowling, Member, IEEE and Renate Sitte, Member, IEEE Griffith University Faculty of Engineering & Information Technology

More information

Lyric3 Programming Features

Lyric3 Programming Features Lyric3 Programming Features Introduction To help your patients experience the most optimal Lyric3 fitting, our clinical training staff has been working with providers around the country to understand their

More information

VITHEA: On-line word naming therapy in Portuguese for aphasic patients exploiting automatic speech recognition

VITHEA: On-line word naming therapy in Portuguese for aphasic patients exploiting automatic speech recognition VITHEA: On-line word naming therapy in Portuguese for aphasic patients exploiting automatic speech recognition Anna Pompili, Pedro Fialho and Alberto Abad L 2 F - Spoken Language Systems Lab, INESC-ID

More information

Auditory Scene Analysis: phenomena, theories and computational models

Auditory Scene Analysis: phenomena, theories and computational models Auditory Scene Analysis: phenomena, theories and computational models July 1998 Dan Ellis International Computer Science Institute, Berkeley CA Outline 1 2 3 4 The computational

More information

Enhanced Feature Extraction for Speech Detection in Media Audio

Enhanced Feature Extraction for Speech Detection in Media Audio INTERSPEECH 2017 August 20 24, 2017, Stockholm, Sweden Enhanced Feature Extraction for Speech Detection in Media Audio Inseon Jang 1, ChungHyun Ahn 1, Jeongil Seo 1, Younseon Jang 2 1 Media Research Division,

More information

Auditory principles in speech processing do computers need silicon ears?

Auditory principles in speech processing do computers need silicon ears? * with contributions by V. Hohmann, M. Kleinschmidt, T. Brand, J. Nix, R. Beutelmann, and more members of our medical physics group Prof. Dr. rer.. nat. Dr. med. Birger Kollmeier* Auditory principles in

More information

The role of periodicity in the perception of masked speech with simulated and real cochlear implants

The role of periodicity in the perception of masked speech with simulated and real cochlear implants The role of periodicity in the perception of masked speech with simulated and real cochlear implants Kurt Steinmetzger and Stuart Rosen UCL Speech, Hearing and Phonetic Sciences Heidelberg, 09. November

More information

EXEMPLAR SELECTION TECHNIQUES FOR SPARSE REPRESENTATIONS OF SPEECH USING MULTIPLE DICTIONARIES. Emre Yılmaz, Jort F. Gemmeke, and Hugo Van hamme

EXEMPLAR SELECTION TECHNIQUES FOR SPARSE REPRESENTATIONS OF SPEECH USING MULTIPLE DICTIONARIES. Emre Yılmaz, Jort F. Gemmeke, and Hugo Van hamme EXEMPLAR SELECTION TECHNIQUES FOR SPARSE REPRESENTATIONS OF SPEECH USING MULTIPLE DICTIONARIES Emre Yılmaz, Jort F. Gemmeke, and Hugo Van hamme Dept. ESAT, KU Leuven, Leuven, Belgium ABSTRACT This paper

More information

Robust Neural Encoding of Speech in Human Auditory Cortex

Robust Neural Encoding of Speech in Human Auditory Cortex Robust Neural Encoding of Speech in Human Auditory Cortex Nai Ding, Jonathan Z. Simon Electrical Engineering / Biology University of Maryland, College Park Auditory Processing in Natural Scenes How is

More information

Issues faced by people with a Sensorineural Hearing Loss

Issues faced by people with a Sensorineural Hearing Loss Issues faced by people with a Sensorineural Hearing Loss Issues faced by people with a Sensorineural Hearing Loss 1. Decreased Audibility 2. Decreased Dynamic Range 3. Decreased Frequency Resolution 4.

More information

Cost-Utility Analysis (CUA) Explained

Cost-Utility Analysis (CUA) Explained Pharmaceutical Management Agency Cost-Utility Analysis (CUA) Explained Cost-Utility Analysis (CUA) at PHARMAC Questions and Answers go to page 9 >> This document explains the process that PHARMAC generally

More information

SPEECH PERCEPTION IN A 3-D WORLD

SPEECH PERCEPTION IN A 3-D WORLD SPEECH PERCEPTION IN A 3-D WORLD A line on an audiogram is far from answering the question How well can this child hear speech? In this section a variety of ways will be presented to further the teacher/therapist

More information

Having your cake and eating it too: multiple dimensions and a composite

Having your cake and eating it too: multiple dimensions and a composite Having your cake and eating it too: multiple dimensions and a composite Perman Gochyyev and Mark Wilson UC Berkeley BEAR Seminar October, 2018 outline Motivating example Different modeling approaches Composite

More information

Lecture 3: Perception

Lecture 3: Perception ELEN E4896 MUSIC SIGNAL PROCESSING Lecture 3: Perception 1. Ear Physiology 2. Auditory Psychophysics 3. Pitch Perception 4. Music Perception Dan Ellis Dept. Electrical Engineering, Columbia University

More information

Sonic Spotlight. Binaural Coordination: Making the Connection

Sonic Spotlight. Binaural Coordination: Making the Connection Binaural Coordination: Making the Connection 1 Sonic Spotlight Binaural Coordination: Making the Connection Binaural Coordination is the global term that refers to the management of wireless technology

More information

Noise Spectrum Estimation using Gaussian Mixture Model-based Speech Presence Probability for Robust Speech Recognition

Noise Spectrum Estimation using Gaussian Mixture Model-based Speech Presence Probability for Robust Speech Recognition Noise Spectru Estiation using Gaussian Mixture Model-bed Speech Presence Probability for Robust Speech Recognition M. J. Ala 2 P. Kenny P. Duouchel 2 D. O'Shaughnessy 3 CRIM Montreal Canada 2 ETS Montreal

More information

A New Paradigm for Forensic Science. Geoffrey Stewart Morrison Ewald Enzinger. p p(e H )

A New Paradigm for Forensic Science. Geoffrey Stewart Morrison Ewald Enzinger. p p(e H ) A New Paradigm for Forensic Science Geoffrey Stewart Morrison Ewald Enzinger p(e H ) p p(e H ) d Abstract In Europe there has been a great deal of concern about the logically correct way to evaluate the

More information

Validation Studies. How well does this work??? Speech perception (e.g., Erber & Witt 1977) Early Development... History of the DSL Method

Validation Studies. How well does this work??? Speech perception (e.g., Erber & Witt 1977) Early Development... History of the DSL Method DSL v5.: A Presentation for the Ontario Infant Hearing Program Associates The Desired Sensation Level (DSL) Method Early development.... 198 Goal: To develop a computer-assisted electroacoustic-based procedure

More information

THE INDIRECT EFFECT IN MULTIPLE MEDIATORS MODEL BY STRUCTURAL EQUATION MODELING ABSTRACT

THE INDIRECT EFFECT IN MULTIPLE MEDIATORS MODEL BY STRUCTURAL EQUATION MODELING ABSTRACT European Journal of Business, Economics and Accountancy Vol. 4, No. 3, 016 ISSN 056-6018 THE INDIRECT EFFECT IN MULTIPLE MEDIATORS MODEL BY STRUCTURAL EQUATION MODELING Li-Ju Chen Department of Business

More information