SRIRAM GANAPATHY. Indian Institute of Science, Phone: +91-(80) Bangalore, India, Fax: +91-(80)

Similar documents
SRIRAM GANAPATHY. Indian Institute of Science, Phone: +91-(80) Bangalore, India, Fax: +91-(80)

The SRI System for the NIST OpenSAD 2015 Speech Activity Detection Evaluation

Panying Rong, Ph.D. MGH Institute of Health Professions

Speech recognition in noisy environments: A survey

CONSTRUCTING TELEPHONE ACOUSTIC MODELS FROM A HIGH-QUALITY SPEECH CORPUS

USING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES

Acoustic Signal Processing Based on Deep Neural Networks

Speech Enhancement Based on Deep Neural Networks

PROFESSIONAL EXPERIENCE September, 2003 Present University of Drytown Drytown, CA

Speech Enhancement Based on Spectral Subtraction Involving Magnitude and Phase Components

Speech Enhancement Using Deep Neural Network

Noise-Robust Speech Recognition in a Car Environment Based on the Acoustic Features of Car Interior Noise

Advanced Audio Interface for Phonetic Speech. Recognition in a High Noise Environment

Yonghee Oh, Ph.D. Department of Speech, Language, and Hearing Sciences University of Florida 1225 Center Drive, Room 2128 Phone: (352)

Modulation and Top-Down Processing in Audition

Using Source Models in Speech Separation

Effects of Cochlear Hearing Loss on the Benefits of Ideal Binary Masking

Jing Shen CURRICULUM VITAE Contact Information Education Academic and Research Experience

Performance Comparison of Speech Enhancement Algorithms Using Different Parameters

Adaptation of Classification Model for Improving Speech Intelligibility in Noise

Enhanced Feature Extraction for Speech Detection in Media Audio

HISTOGRAM EQUALIZATION BASED FRONT-END PROCESSING FOR NOISY SPEECH RECOGNITION

Robust system for patient specific classification of ECG signal using PCA and Neural Network

EMOTION DETECTION FROM VOICE BASED CLASSIFIED FRAME-ENERGY SIGNAL USING K-MEANS CLUSTERING

Robust Neural Encoding of Speech in Human Auditory Cortex

Analysis of Speech Recognition Techniques for use in a Non-Speech Sound Recognition System

EXEMPLAR SELECTION TECHNIQUES FOR SPARSE REPRESENTATIONS OF SPEECH USING MULTIPLE DICTIONARIES. Emre Yılmaz, Jort F. Gemmeke, and Hugo Van hamme

Bio-Data. Siddheshwar nagar, Trikoni Banglow Rd., Amrutdham, Panchavati, Nashik Mob: University Institute Degree Year

Broadband Wireless Access and Applications Center (BWAC) CUA Site Planning Workshop

General Soundtrack Analysis

CURRICULUM VITAE. Postdoctoral Fellow, Auditory & Speech Sciences Laboratory. University of South Florida

Mark John Francis Gales

Fig. 1 High level block diagram of the binary mask algorithm.[1]

Gender Based Emotion Recognition using Speech Signals: A Review

FIR filter bank design for Audiogram Matching

SUPPRESSION OF MUSICAL NOISE IN ENHANCED SPEECH USING PRE-IMAGE ITERATIONS. Christina Leitner and Franz Pernkopf

Hearing Lectures. Acoustics of Speech and Hearing. Auditory Lighthouse. Facts about Timbre. Analysis of Complex Sounds

An Auditory-Model-Based Electrical Stimulation Strategy Incorporating Tonal Information for Cochlear Implant

arxiv: v1 [cs.cv] 13 Mar 2018

Auditory Perception: Sense of Sound /785 Spring 2017

Lecturer: T. J. Hazen. Handling variability in acoustic conditions. Computing and applying confidence scores

An active unpleasantness control system for indoor noise based on auditory masking

MULTI-MODAL FETAL ECG EXTRACTION USING MULTI-KERNEL GAUSSIAN PROCESSES. Bharathi Surisetti and Richard M. Dansereau

Recognition of Sleep Dependent Memory Consolidation with Multi-modal Sensor Data

Recognition & Organization of Speech & Audio

Speech Enhancement, Human Auditory system, Digital hearing aid, Noise maskers, Tone maskers, Signal to Noise ratio, Mean Square Error

Sound Texture Classification Using Statistics from an Auditory Model

DESIGN OF AN EFFECTIVE DIGITAL RECONFIGURABLE FILTER BANK

Interpreting Deep Neural Networks and their Predictions

University of Southern California, Los Angeles, California USA M.S., Electrical Engineering (2012)

Elias B Issa, PhD. Curriculum Vitae 10/2015

HCS 7367 Speech Perception

Parametric Optimization and Analysis of Adaptive Equalization Algorithms for Noisy Speech Signals

Elias B Issa, PhD. Curriculum Vitae 08/2016. RESEARCH INTEREST The neural mechanisms and computational principles underlying high-level vision

Research Article Automatic Speaker Recognition for Mobile Forensic Applications

TESTS OF ROBUSTNESS OF GMM SPEAKER VERIFICATION IN VoIP TELEPHONY

An Artificial Neural Network Architecture Based on Context Transformations in Cortical Minicolumns

July 2014-present Postdoctoral Fellowship, in the Department of Experimental Psychology,

Speech, Hearing and Language: work in progress. Volume 11

Research Assistant at Martinos Center for Biomedical Imaging Harvard Medical School Advisor: Moshe Bar, Ph.D.

Mitigating the Effects of Non-Stationary Unseen Noises on Language Recognition Performance

CHAPTER 1 INTRODUCTION

CURRICULUM VITAE. Date prepared: 09/01/2017

NOISE REDUCTION ALGORITHMS FOR BETTER SPEECH UNDERSTANDING IN COCHLEAR IMPLANTATION- A SURVEY

Education. Professional Experience and Employment History

Mammogram Analysis: Tumor Classification

Lecture 9: Speech Recognition: Front Ends

Empirical Mode Decomposition based Feature Extraction Method for the Classification of EEG Signal

Prof. Yousef Bassyouni Mahdy الا ستاذ الدكتور يوسف بسيونى مهدى. Fax: ٢٠-٨٨-٣٤٧٦٧٨

Chapter 5. Summary and Conclusions! 131

Mammogram Analysis: Tumor Classification

An Auditory System Modeling in Sound Source Localization

A Consumer-friendly Recap of the HLAA 2018 Research Symposium: Listening in Noise Webinar

IMPROVING INTELLIGIBILITY IN NOISE OF HMM-GENERATED SPEECH VIA NOISE-DEPENDENT AND -INDEPENDENT METHODS

Analysis of Emotion Recognition using Facial Expressions, Speech and Multimodal Information

A Review on Dysarthria speech disorder

Simulations of high-frequency vocoder on Mandarin speech recognition for acoustic hearing preserved cochlear implant

Domain Adversarial Training for Accented Speech Recognition

Curriculum Vitae Outstanding Graduate Student Center for Hearing Research, University of California, Irvine

Improving the Noise-Robustness of Mel-Frequency Cepstral Coefficients for Speech Processing

Combination of Bone-Conducted Speech with Air-Conducted Speech Changing Cut-Off Frequency

Detection of Parkinson S Disease by Speech Analysis

SPEECH TO TEXT CONVERTER USING GAUSSIAN MIXTURE MODEL(GMM)

Frequency Tracking: LMS and RLS Applied to Speech Formant Estimation

Single-Channel Sound Source Localization Based on Discrimination of Acoustic Transfer Functions

Implementation of Spectral Maxima Sound processing for cochlear. implants by using Bark scale Frequency band partition

Automatic Judgment System for Chinese Retroflex and Dental Affricates Pronounced by Japanese Students

Genesis of wearable DSP structures for selective speech enhancement and replacement to compensate severe hearing deficits

Discrimination between ictal and seizure free EEG signals using empirical mode decomposition

CURRICULUM VITAE Michelle A. Aldridge, Ph.D., C.C.C Inwood Rd. 214/ (home)

BINAURAL DICHOTIC PRESENTATION FOR MODERATE BILATERAL SENSORINEURAL HEARING-IMPAIRED

Perception Optimized Deep Denoising AutoEncoders for Speech Enhancement

Using the Soundtrack to Classify Videos

CURRICULUM VITAE. Dr. Abdulnasir Yahya Hossen

Enhancement of Reverberant Speech Using LP Residual Signal

SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING

Analysis of Face Mask Effect on Speaker Recognition

VITHEA: On-line word naming therapy in Portuguese for aphasic patients exploiting automatic speech recognition

Novel Speech Signal Enhancement Techniques for Tamil Speech Recognition using RLS Adaptive Filtering and Dual Tree Complex Wavelet Transform

LATERAL INHIBITION MECHANISM IN COMPUTATIONAL AUDITORY MODEL AND IT'S APPLICATION IN ROBUST SPEECH RECOGNITION

Transcription:

SRIRAM GANAPATHY Assistant Professor, Email: sriramg@iisc.ac.in Electrical Engineering, Web: http://www.leap.ee.iisc.ac.in/sriram Indian Institute of Science, Phone: +91-(80)-2293-2433 Bangalore, India, 560012. Fax: +91-(80)-2360-0444 Interests My research interests include signal processing, machine learning, deep learning, auditory neuroscience, robust speech recognition and audio applications like information extraction, enhancement and coding. Current Position I am an Assistant Professor at the Electrical Engineering, Indian Institute of Science, Bangalore. I manage the Learning and Extraction of Acoustic Patterns (LEAP) laboratory where the activities are focussed on information extraction and analysis of acoustic signals like speech and audio. Education PhD, Center of Language and Speech Processing (4.0/4.0) Jan. 2009-Dec. 2011 Johns Hopkins University, Baltimore, USA. Master of Engineering, Signal Processing (7.4/8.0) Aug. 2004-July. 2006 Indian Institute of Science, Bangalore, India. Bachelor of Technology, Electronics and Communications (82%) Oct. 2000- June. 2004 College of Engineering, Trivandrum, India Skills Programming: MATLAB, C, Python, HTML. Tools: HTK, QuickNet, Kaldi, Latex, Theano, Tensorflow. Operating Systems: Unix/Linux, Windows, Mac OS X Honors and Awards Young Investigator Award, Prathiksha Trust, Indian Institute of Science Bangalore. Microsoft Award - Best Summer Workshop Award in Artificial Social Intelligence. Early Career Research Award, Department of Science and Technology, India, May 2017 IEEE Senior Member, April, 2017. Master Inventor, IBM T.J. Watson Research Center, New York, USA, 2014.!1

Best tutorial presentation, Interspeech 2014 on The Art and Science of Speech Feature Engineering, Singapore. Best outgoing student award for M.E. Signal Processing, Dept. of ECE, Indian Institute of Science, Bangalore, 2006. Experience Asst. Prof, Electrical Engg. IISc Bangalore. Jan 2016 - Leading the research activities in learning and extraction of acoustic patterns (LEAP) labs - working on understanding the human encoding of speech and langauge, classifying natural and synthetic speech (spoof detection), speaker verification in multi-lingual conditions, unsupervised filter learning and noise robust speech recognition, low complexity wake-up word detection, language recognition. Research Staff Member, IBM T.J Watson Center, NY, USA. Dec.2011 - Dec. 2015 Research on signal analysis and processing of noisy and degraded radio channel speech for biometric applications like speaker and language recognition as well as speech activity detection. These technologies are developed for the U.S. Government under the Defense Advanced Research Project Agency. Full-time employment (40 hours per week). Research Intern, IBM T.J Watson Center, NY, USA. June 2010 - Aug 2010 The main focus of this internship was to develop feature normalization techniques for speaker verification in far-field reverberant environments. Full-time employment (40 hours per week). Research Assistant, Idiap Research Institute, Switzerland. Oct. 2006 - Jan 2009 The goal of this work was to investigate the use of long-term energy summarization of speech signals for speech recognition and coding applications. Full-time employment (40 hours per week). Teaching Course Instructor E9 205 - Machine Learning for Signal Processing, Indian Institute of Science, Aug- Dec 2017. Course Instructor for E9-261 - Speech Information Processing, Indian Institute of Science, Jan-April, 2017. (Jointly given with Dr. Prasanta Ghosh). Course Instructor - Machine Learning for Sensory Signals, Center for Continuing Education, Indian Institute of Science, Jan-Apr 2017. Course Instructor E9 205 - Machine Learning for Signal Processing, Indian Institute of Science, Aug- Dec 2016. Course Instructor for E9-261 - Speech Information Processing, Indian Institute of Science, Jan-April, 2016. (Jointly given with Dr. Prasanta Ghosh). Teaching Assistant for Information processing of sensory signals, for Spring 2009, Spring 2010, Johns Hopkins University.!2

External Talks and Seminars Deep Learning - Theory and Practice, Rajiv Gandhi Institute of Technology, Kottayam, Kerala, April, 2017. Introduction to Deep Learning, DRDO CAIR labs, Bangalore, Dec. 2016. Machine Learning for Speech Processing, IIIT- Bangalore, Sept. 2016. Modulation Spectrum Processing of Speech, Summer School of Advances in Speech and Audio Processing, DA-IICT, July 2016. Introduction to Deep Learning, Naval Physical Oceanographic Laboratory, DRDO, July 2016. Deep Learning for Speech Processing, IBM India Research Labs, February 2016. The Art and Science of Speech Feature Engineering, Tutorial at Interspeech, Singapore, Sept. 2014. Algorithms in Speech Signal Processing, College of Engineering, Trivandrum, India, Nov., 2013. Robust Processing of Noisy and Degraded Channel Speech, Computational Science an Artificial Intelligence Laboratory, MIT, Cambridge, USA, Oct., 2013. Dealing with Noisy Speech Using Autoregressive Models, Idiap Research Institute, Martigny, Switzerland, Aug, 2013. Signal analysis using autoregressive models of amplitude modulation, Electrical and Computer Engineering, University of Texas, Dallas, Feb. 2013. Frequency Domain Linear Predictive Analysis of Speech, Indian Institute of Science, Bangalore, India, Nov. 2011. Signal modeling with long term feature processing, Raytheon BBN Technologies, Cambridge, MA, USA, July 2011. Professional Memberships IEEE Signal Processing Society - Senior Member. International Speech Communication Association (ISCA). Patents Method for System Combination in Audio Analytics Application, with IBM Watson Center, USA. [Filed June 2015]. Spectral Noise Shaping in Audio Coding Based on Spectral Dynamics in Frequency Sub-bands, with Qualcomm Inc, [Approved Nov. 2011].!3

Temporal Masking in Audio Coding Based on Spectral Dynamics in Frequency Subbands, with Qualcomm Inc, [Approved May 2009]. Publications I. Thesis S. Ganapathy, "Signal Analysis using Autoregressive Models of Amplitude Modulation ", Johns Hopkins University, Jan. 2012 II. Peer Reviewed Journals S. Gutta, V.S. Kadimesetty, S. K. Kalva, M. Pramanik, S. Ganapathy and P. K. Yalavarthy, "Deep Neural Network Based Bandwidth Enhancement of Photoacoustic Data", Journal of Biomedical Optics, October 2017. G. Kocavs, L. Toth, D. V. Compernolle and S. Ganapathy, "Increasing the Robustness of CNN Acoustic Models using ARMA Spectrogram Features and Channel Dropout", Elsevier Pattern Recognition Letters, September 2017. P. Agrawal and S. Ganapathy, "Unsupervised Modulation Filter Learning for Noise-Robust Speech Recognition", Journal of Acoustical Society of America, September 2017. S. Ganapathy, "Multi-variate Autoregressive Spectrogram Modeling for Noisy Speech Recognition", IEEE Signal Processing Letters, July 2017. S. Ganapathy, M. Omar, Auditory Motivated Front-end for Noisy Speech Using Spectro-temporal Modulation Filtering, Journal of Acoustical Society of America, EL343-349, Vol. 136(5), Nov. 2014. S. Ganapathy, S. H. Mallidi and H. Hermansky, "Robust Feature Extraction Using Modulation Filtering of Autoregressive Models", IEEE Transactions on Audio, Speech and Language Processing, Vol. 22(8), pp. 1285-1295, Aug. 2014. S. Ganapathy and J. Pelecanos, "Enhancing Frequency Shifted Speech Signals in Single Side Band Communication", IEEE Signal Processing Letters, Vol. 20(12), pp. 1231-1234, Oct. 2013. S. Ganapathy and H. Hermansky, "Temporal Resolution Analysis in Frequency Domain Linear Prediction", Journal of Acoustical Society of America, EL436-442, Vol. 132(5), Oct. 2012. S. Ganapathy, S. Thomas and H. Hermansky, "Temporal envelope compensation for robust phoneme recognition using modulation spectrum", Journal of Acoustical Society of America, Vol. 128(6), pp. 3769-3780, Dec. 2010.!4

S. Ganapathy, P. Motlicek and H. Hermansky, "Autoregressive Models Of Amplitude Modulations In Audio Compression", IEEE Transactions on Audio, Speech and Language Processing, Vol. 18(6), pp. 1624-1631, Aug. 2010. P. Motlicek, S. Ganapathy, H. Hermansky and H. Garudadri, "Wide-Band Audio Coding based on Frequency Domain Linear Prediction", EURASIP Journal on Audio, Speech, and Music Processing, Vol. 2010(3), pp. 1-14, Jan. 2010. S. Ganapathy, S. Thomas and H. Hermansky, "Modulation Frequency Features For Phoneme Recognition In Noisy Speech", Journal of Acoustical Society of America, EL8-12, Vol. 125(1), Jan. 2009. S. Thomas, S. Ganapathy and H. Hermansky, "Recognition of Reverberant Speech Using Frequency Domain Linear Prediction", IEEE Signal Processing Letters, Vol. 15, pp. 681-684 Nov. 2008. III. Conferences Ansari T, R. Kumar, S. Singh, S. Ganapathy "Unsupervised HMM Posteriograms for Language Independent Acoustic Modeling in Zero Resource Conditions", IEEE ASRU, Dec. 2017. Ansari T, R. Kumar, S. Singh, S. Ganapathy "Deep Learning Methods For Unsupervised Acoustic Modeling - LEAP Submission to ZeroSpeech Challenge 2017", IEEE ASRU, Dec. 2017. A. Siddhant, P. Jyothi and S. Ganapathy "Leveraging Native Language Speech For Accent Identfication Using Deep Siamese Networks", IEEE ASRU, Dec. 2017. P. Agarwal, S. Ganapathy, Speech Representation Learning Using Unsupervised Data-Driven Modulation Filtering for Robust ASR, Interspeech 2017, Stockholm, Sweden. N. Kumar, R. K. Das, S. Jelil, Dhanush B K, H. Kashyap,K. S. Murty, S. Ganapathy, R. Sinha and S. R. M. Prasanna, IITG-Indigo System for NIST 2016 SRE Challenge, Interspeech 2017, Stockholm, Sweden. Dhanush B. K., Suparna S., Aarthy R., Likhita C., Shashank D., Harish H., S. Ganapathy, Factor Analysis Methods for Joint Speaker Verification and Spoof Detection, ICASSP, New Orleans, March, 2017. S. Sadjadi, J. Pelecanos and S. Ganapathy, "The IBM Speaker Recognition System: Recent Advances and Error Analysis", Interspeech, San Francisco, September, 2016. D. Dimitriadis, S. Thomas and S. Ganapathy, "An investigation on the use of ivectors for improved ASR robustness", Interspeech, San Francisco, Sept. 2016. S. Sadjadi, S. Ganapathy and J. Pelecanos, "The IBM 2016 Speaker Recognition System", Odyssey, Spain, June, 2016.!5

S. Sadjadi, S. Ganapathy and J. Pelecanos, Speaker Age Estimation On Conversational Telephone Speech Using Senone Posterior Based I-vectors, ICASSP, Shanghai, 2016. S. Ganapathy, S. Thomas, D. Dimitriadis, S. Rennie "Investigating Factor Analysis Features for Deep Neural Networks In Noisy Speech Recognition", Interspeech, Dresden, Germany, Sept. 2015. S. Ganapathy, "Robust Speech Processing Using ARMA Spectrograms", ICASSP, Brisbane, April, 2015. S. Sadjadi, J. Pelecanos and S. Ganapathy, "Nearest Neighbor Discriminant Analysis for Language Recognition", ICASSP, Brisbane, April, 2015. S. Ganapathy, K. J. Han, M. Omar, M. V. Segbroeck and S. Narayan, "Robust Language Identification Using Convolutional Neural Networks", Interspeech, Singapore, 2014. S. Thomas, S. Ganapathy, G. Saon and H. Soltau, Analyzing Convolutional Neural Networks For Speech Activity Detection In Mismatched Acoustic Conditions, ICASSP, Florence, 2014. M. Omar and S. Ganapathy, Shift-Invariant Features for Speech Activity Detection in Adverse Radio- Frequency Channel Conditions, ICASSP, Florence, 2014. G. Saon, S. Thomas, H. Soltau, S. Ganapathy and B. Kingsbury, "The IBM Speech Activity Detection System for the DARPA RATS Program", Interspeech, Lyon, Aug. 2013. K. J. Han, S. Ganapathy, M Li, M. Omar and S. Narayan, "TRAP Language Identification System for RATS Phase II Evaluation", Interspeech, Lyon, Aug. 2013. H. Mallidi, S. Ganapathy and H. Hermansky, "Robust Speaker Recognition Using Spectro-Temporal Autoregressive Models", Interspeech, Lyon, Aug. 2013. S. Ganapathy, M. Omar and J. Pelecanos, "Unsupervised Channel Adaptation For Language Identification Using Co-training", ICASSP, Vancouver, May, 2013. S. Ganapathy, M. Omar and J. Pelecanos, "Noisy Channel Adaptation in Language Identification", IEEE SLT, Miami, Dec, 2012. S. Ganapathy and H. Hermansky, "Robust Phoneme Recognition Using High Resolution Temporal Envelopes", Interspeech, Portland, Sept. 2012. S. Thomas, S. Ganapathy, A. Jansen and H. Hermansky, "Data-driven Posterior Features for Low Resource Speech Recognition Applications", Interspeech, Portland, Sept. 2012. S. Ganapathy, S. Thomas and H. Hermansky, "Feature Extraction Using 2-D Autoregressive Models For Speaker Recognition", ISCA Speaker Odyssey, June 2012.!6

S. Thomas, H. Mallidi, S. Ganapathy and H. Hermansky, "Adaptation Transforms of Auto-Associative Neural Networks as Features for Speaker Verification", ISCA Speaker Odyssey, June 2012. D. Gomero et al. "The UMD-JHU 2011 Speaker Recognition System", ICASSP, Japan, Mar. 2012. S. Thomas, S. Ganapathy and H. Hermansky, "Multilingual MLP Features For Low-resource LVCSR Systems", ICASSP, Japan, Mar. 2012. S. Ganapathy, P. Rajan and H. Hermansky, "Multi-layer Perceptron Based Speech Activity Detection for Speaker Verification", IEEE WASPAA, Oct. 2011. H. Mallidi, S. Ganapathy and H. Hermansky, "Modulation spectrum analysis for recognition of reverberant speech", Interspeech, Italy, Aug. 2011. S. Ganapathy, J. Pelecanos and M. Omar, "Feature Normalization for Speaker Verification in Room Reverberation", ICASSP, Prague, May 2011. S. Garimella, S. Ganapathy and H. Hermansky, "Sparse Auto-associative Neural Networks: Theory and Application to Speech Recognition", Interspeech, Japan, Sept. 2010. S. Thomas, S. Ganapathy and H. Hermansky, "Cross-lingual and Multi-stream Posterior Features for Low-resource LVCSR Systems", Proc. of Interspeech, Japan, Sept. 2010. S. Thomas, K. Patil, S. Ganapathy, N. Mesgarani, H. Hermansky, "A Phoneme Recognition Framework based on Auditory Spectro-Temporal Receptive Fields", Proc. of Interspeech, Japan, Sept. 2010. S. Ganapathy, S. Thomas and H. Hermansky, "Robust Spectro-Temporal Features Based on Autoregressive Models of Hilbert Envelopes", ICASSP, Dallas, USA, March 2010. S. Ganapathy, S. Thomas and H. Hermansky, "Comparison of Modulation Features For Phoneme Recognition", ICASSP, Dallas, USA, March 2010. S. Ganapathy, S. Thomas, and H. Hermansky, "Temporal Envelope Subtraction for Robust Speech Recognition Using Modulation Spectrum", IEEE ASRU, 2009. S. Ganapathy, S. Thomas, P. Motlicek and H. Hermansky, "Applications of Signal Analysis Using Autoregressive Models for Amplitude Modulation", IEEE WASPAA 2009. S. Ganapathy, S. Thomas and H. Hermansky, "Static and Dynamic Modulation Spectrum for Speech Recognition", Proc. of Interspeech, Brighton, UK, Sept. 2009. S. Thomas, S. Ganapathy and H. Hermansky, "Tandem Representations of Spectral Envelope and Modulation Frequency Features for ASR", Proc. of Interspeech, Brighton, UK, Sept. 2009.!7

S. Thomas, S. Ganapathy and H. Hermansky, "Phoneme Recognition Using Spectral Envelope and Modulation Frequency Features", ICASSP, Taiwan, April 2009. S. Ganapathy, S. Thomas and H. Hermansky, "Front-end for Far-field Speech Recognition based on Frequency Domain Linear Prediction", Proc. of INTERSPEECH, Brisbane, Australia, Sep 2008. S. Thomas, S. Ganapathy and H. Hermansky, "Hilbert Envelope Based Specto-Temporal Features for Phoneme Recognition in Telephone Speech", Proc. of INTERSPEECH, Brisbane, Australia, Sep 2008. S. Ganapathy, P. Motlicek, H. Hermansky and H. Garudadri, "Spectral Noise Shaping: Improvements in Speech/Audio Codec Based on Linear Prediction in Spectral Domain", Proc. of INTERSPEECH, Brisbane, Australia, Sep 2008. P. Motlicek, S. Ganapathy, H. Hermansky, H. Garudadri and Marios Athineos, "Perceptually motivated Sub-band Decomposition for FDLP Audio Coding", in Lecture Notes In Artificial Intelligence, Springer- Verlag Berlin, Heidelberg, 2008. S. Thomas, S. Ganapathy and H. Hermansky, "Spectro-Temporal Features for Automatic Speech Recognition using Linear Prediction in Spectral Domain", Proc. of EUSIPCO, Lausanne, Switzerland, Aug 2008. S. Ganapathy, P. Motlicek, H. Hermansky and H. Garudadri, "Autoregressive Modelling of Hilbert Envelopes for Wide-band Audio Coding", AES 124th Convention, Audio Engineering Society, May 2008. S. Ganapathy, P. Motlicek, H. Hermansky and H. Garudadri, ""Temporal Masking for Bit-rate Reduction in Audio Codec Based on Frequency Domain Linear Prediction", Proc. of ICASSP, April 2008. S. Thomas, S. Ganapathy and H. Hermansky, "Hilbert Envelope Based Features for Far-Field Speech Recognition", Lecture Notes in Computer Science, Springer Berlin, Heidelberg 2008. P. Motlicek, H. Hermansky, S. Ganapathy and H. Garudadri, "Frequency Domain Linear Prediction for QMF Sub-bands and Applications to Audio Coding", Lecture Notes in Computer Science, Springer Berlin, Heidelberg 2007. P. Motlicek, H. Hermansky, S. Ganapathy and H. Garudadri, "Non- Uniform Speech/Audio Coding Exploiting Predictability of Temporal Evolution of Spectral Envelopes", Lecture Notes in Computer Science, Springer Berlin, Heidelberg 2007.!8