Image Understanding and Machine Vision, Optical Society of America, June [8] H. Ney. The Use of a One-Stage Dynamic Programming Algorithm for

Similar documents
original image 1 image 2 not adaptive grayvalue modification Figure 6. adaptive gray-value modication, example Xl Xr exponential hamming

CS-E Deep Learning Session 4: Convolutional Networks

Recognition of sign language gestures using neural networks

Analysis of Speech Recognition Techniques for use in a Non-Speech Sound Recognition System

Noise-Robust Speech Recognition Technologies in Mobile Environments

1. INTRODUCTION. Vision based Multi-feature HGR Algorithms for HCI using ISL Page 1

Advanced Audio Interface for Phonetic Speech. Recognition in a High Noise Environment

Comparison of Lip Image Feature Extraction Methods for Improvement of Isolated Word Recognition Rate

Research Proposal on Emotion Recognition

Facial expression recognition with spatiotemporal local descriptors

Gesture Recognition using Marathi/Hindi Alphabet

Analysis of Emotion Recognition using Facial Expressions, Speech and Multimodal Information

Effect of Sensor Fusion for Recognition of Emotional States Using Voice, Face Image and Thermal Image of Face

Audio-Visual Integration in Multimodal Communication

Modulation and Top-Down Processing in Audition

Edge Detection Techniques Using Fuzzy Logic

M.Sc. in Cognitive Systems. Model Curriculum

Frequency Tracking: LMS and RLS Applied to Speech Formant Estimation

EEL 6586, Project - Hearing Aids algorithms

Noise-Robust Speech Recognition in a Car Environment Based on the Acoustic Features of Car Interior Noise

Computational Auditory Scene Analysis: An overview and some observations. CASA survey. Other approaches

Consonant Perception test

1 Pattern Recognition 2 1

Cognitive Processes PSY 334. Chapter 2 Perception

Katsunari Shibata and Tomohiko Kawano

FREQUENCY COMPRESSION AND FREQUENCY SHIFTING FOR THE HEARING IMPAIRED

SPEECH TO TEXT CONVERTER USING GAUSSIAN MIXTURE MODEL(GMM)

TWO HANDED SIGN LANGUAGE RECOGNITION SYSTEM USING IMAGE PROCESSING

Improved Intelligent Classification Technique Based On Support Vector Machines

A Model of Visually Guided Plasticity of the Auditory Spatial Map in the Barn Owl

Sparse Coding in Sparse Winner Networks

CHAPTER 1 INTRODUCTION

Requirements for Maintaining Web Access for Hearing-Impaired Individuals

QUANTIFICATION OF PROGRESSION OF RETINAL NERVE FIBER LAYER ATROPHY IN FUNDUS PHOTOGRAPH

A Smart Texting System For Android Mobile Users

CONSTRUCTING TELEPHONE ACOUSTIC MODELS FROM A HIGH-QUALITY SPEECH CORPUS

Chapter 17. Speech Recognition and Understanding Systems

Juan Carlos Tejero-Calado 1, Janet C. Rutledge 2, and Peggy B. Nelson 3

A Perceptron Reveals the Face of Sex

Deep Learning for Lip Reading using Audio-Visual Information for Urdu Language

Perceptual Effects of Nasal Cue Modification

Contributions of the piriform fossa of female speakers to vowel spectra

INTELLIGENT LIP READING SYSTEM FOR HEARING AND VOCAL IMPAIRMENT

HCS 7367 Speech Perception

Mammogram Analysis: Tumor Classification

R Jagdeesh Kanan* et al. International Journal of Pharmacy & Technology

REVIEW ON ARRHYTHMIA DETECTION USING SIGNAL PROCESSING

Understanding Convolutional Neural

PROFESSIONAL EXPERIENCE September, 2003 Present University of Drytown Drytown, CA

Mammography is a most effective imaging modality in early breast cancer detection. The radiographs are searched for signs of abnormality by expert

N RISCE 2K18 ISSN International Journal of Advance Research and Innovation

Sound Interfaces Engineering Interaction Technologies. Prof. Stefanie Mueller HCI Engineering Group

A Hierarchical Artificial Neural Network Model for Giemsa-Stained Human Chromosome Classification

Biometric Authentication through Advanced Voice Recognition. Conference on Fraud in CyberSpace Washington, DC April 17, 1997

Image Processing of Eye for Iris Using. Canny Edge Detection Technique

Face Analysis : Identity vs. Expressions

ILLUSIONS AND ISSUES IN BIMODAL SPEECH PERCEPTION

ECG Beat Recognition using Principal Components Analysis and Artificial Neural Network

An Artificial Neural Network Architecture Based on Context Transformations in Cortical Minicolumns

Ambiguity in the recognition of phonetic vowels when using a bone conduction microphone

Automatic audio analysis for content description & indexing

A Lip Reading Application on MS Kinect Camera

Real Time Sign Language Processing System

Design of Palm Acupuncture Points Indicator

B657: Final Project Report Holistically-Nested Edge Detection

Computational Perception /785. Auditory Scene Analysis

Clustering of MRI Images of Brain for the Detection of Brain Tumor Using Pixel Density Self Organizing Map (SOM)

Proceedings of Meetings on Acoustics

AVR Based Gesture Vocalizer Using Speech Synthesizer IC

Acoustic-Labial Speaker Verication. (luettin, genoud,

Using.a Saliency Map for Active Spatial Selective Attention: Implementation & Initial Results

Morton-Style Factorial Coding of Color in Primary Visual Cortex

CHAPTER 5 WAVELET BASED DETECTION OF VENTRICULAR ARRHYTHMIAS WITH NEURAL NETWORK CLASSIFIER

Identification of Sickle Cells using Digital Image Processing. Academic Year Annexure I

Spectrograms (revisited)

Classification of electrocardiographic ST-T segments human expert vs artificial neural network

An Edge-Device for Accurate Seizure Detection in the IoT

CPSC81 Final Paper: Facial Expression Recognition Using CNNs

Development of an Electronic Glove with Voice Output for Finger Posture Recognition

A Brain Computer Interface System For Auto Piloting Wheelchair

ANN predicts locoregional control using molecular marker profiles of. Head and Neck squamous cell carcinoma

Mammogram Analysis: Tumor Classification

Lecture 9: Speech Recognition: Front Ends

Intelligent Agents. Philipp Koehn. 16 February 2017

Convolutional Neural Networks (CNN)

ACOUSTIC AND PERCEPTUAL PROPERTIES OF ENGLISH FRICATIVES

A Survey on Localizing Optic Disk

Audiovisual to Sign Language Translator

Flexible, High Performance Convolutional Neural Networks for Image Classification

Price list Tools for Research & Development

arxiv: v2 [cs.cv] 19 Dec 2017

What you re in for. Who are cochlear implants for? The bottom line. Speech processing schemes for

Gender Based Emotion Recognition using Speech Signals: A Review

Using Deep Convolutional Networks for Gesture Recognition in American Sign Language

Detection of microcalcifications in digital mammogram using wavelet analysis

LIE DETECTION SYSTEM USING INPUT VOICE SIGNAL K.Meena 1, K.Veena 2 (Corresponding Author: K.Veena) 1 Associate Professor, 2 Research Scholar,

PMR5406 Redes Neurais e Lógica Fuzzy. Aula 5 Alguns Exemplos

COMPARATIVE STUDY ON FEATURE EXTRACTION METHOD FOR BREAST CANCER CLASSIFICATION

A Consumer-friendly Recap of the HLAA 2018 Research Symposium: Listening in Noise Webinar

group by pitch: similar frequencies tend to be grouped together - attributed to a common source.

Transcription:

Image Understanding and Machine Vision, Optical Society of America, June 1989. [8] H. Ney. The Use of a One-Stage Dynamic Programming Algorithm for Connected Word Recognition. IEEE International Conference on Acoustics, Speech, and Signal Processing, April 1984. [9] E. Petajan, B. Bischoff, D. Bodoff, and N.M. Brooke. An Improved Automatic Lipreading System to enhance Speech Recognition. In ACM SIGCHI, 1988. [10] D.A. Pomerleau. Neural Network Perception for Mobile Robot Guidance. PhD Thesis, CMU. CMU- CS-92-115, February 1992. [11] P.W. Rander. Facetracking Using a Template Based Approach. Personal Communication. [12] D.E. Rumelhart, G.E. Hinton, and R.J. Williams. Learning Internal Representations by Error Propagation. Parallel Distributed Processing Vol. 1. MIT Press, 1986. [13] David G. Stork, Greg Wolff, and Earl Levine. Neural Network Lipreading System for Improved Speech Recognition. In IJCNN, June 1992. [14] A. Waibel, T. Hanazawa, G. Hinton, K. Shikano, and K. Lang. Phoneme Recognition Using Time-Delay Neural Networks. IEEE Transactions on Acoustics, Speech, and Signal Processing, 37(3):328-339, March 1989. [15] B.P. Yuhas, M.H. Goldstein, and T.J. Sejnowski. Integration of Acoustic and Visual Speech Signals using Neural Networks. IEEE Communications Magazine, [16] John B. Hampshire II and Alexander H. Waibel. A Novel Objective Function for Improved Phoneme Recognition Using Time-Delay Neural Networks. IEEE Transactions on Neural Networks, 1(2), June 1990. [17] P. Haffner, M. Frannzini, and A. Waibel. Integrating Time Alignment and Connectionist Networks for High Performance Continuous Speech Recognition. Proc. International Conference on Acoustics, Speech, and Signal Processing, 1991.

VII. CONCLUSION AND FUTURE WORK We have shown how a state-of-the-art speech recognition system can be improved by considering additional visual information for the recognition process. This is true for optimal recording conditions but even more for nonoptimal recording conditions as they usually exist in real world applications. Experiments were performed on the connected letter recognition task, but similar results can be expected for continuous speech recognition as well. Work is in progress to integrate not only the time independent weight sharing but also position independent weight sharing for the visual TDNN, in order to locate and track the lips. We are also on the way to largely increase our database in order to achieve better recognition rates and to train speaker independently. Investigations of different approaches are still in progress in order to combine visual and acoustic features and to apply different preprocessing to the visual data. A V A V /ae/ /ie/ /m/ /#/ /#/ /eh/ ACKNOWLEDGEMENTS We appreciate the help from the DEC on campus research center (CEC) for the initial data acquisition. This research is sponsored in part by the Land Baden Württemberg (Landesschwerpunktprogramm Neuroinformatik), and the National Science Foundation. Figure 3: Entropy-Weights malized grey-value images. Experiments with 2D-FFT images are still in progress. In our initial 2D-FFT simulations we come to visual recognition errors, which are on average about 8% higher than the grey-level coding recognition errors. We also took a closer look to the dynamic behavior of the entropy-weights. Figure 3 shows the weights from the acoustic and visual TDNN to the combined layer over time during the letter sequence M-I-E was spoken. The upper dots represent the acoustic weight A and the lower dots the visual weight V, where A=0.5+ (entropy(visual-tdnn)-entropy(acoustic-tdnn))/2k and V=1.0-A. Big white dots represent weights close to 1.0 and big black dots weights close to 0.0. K is the maximum entropy difference in the training set. At the end of the /m/-phoneme when the lips are closed, V is higher than A. Obviously there the visual hypotheses are more certain than the acoustic ones. During the /ie/-phoneme the acoustic hypotheses are more certain than the visual ones, which also makes sense. VI. OTHER APPROACHES The interest in automated speechreading (or lipreading) is growing recently. As a non-connectionistic approach the work of Petajan et al. [9] should be mentioned. Yuhas et al. [15] did use a neural network for vowel recognition, working on static images. Stork et al. [13] used a conventional TDNN (without DTW) for speechreading. They limited the task to recognize 10 isolated letters and used artificial markers on the lips. No visual feature extraction was integrated into their model. Also of interest are some psychological studies about human speechreading and their approach to describe the human performance. This measurements could also be applied to the performance analysis of automated speechreading systems. Dodd and Campbell [3], and Demorest and Bernstein [2] did some valuable work in this area. REFERENCES [1] Christian Benoit, Tahar Lallouache, Tayeb Mohamadi, and Christian Abry. A Set of French Visemes for Visual Speech Synthesis. Talking Machines: Theories, Models, and Designs, 1992. [2] M.E. Demorest and L.E. Bernstein. Computational Explorations of Speechreading. In Submission. [3] B. Dodd and R. Campbell. Hearing by Eye: The Psychology fo Lipreading. Lawrence Erlbaum Press, 1987. [4] C.G. Fischer. Confusion among visually perceived consonants. J. Speech Hearing Res., 11, 1968. [5] P. Haffner and A. Waibel. Multi-State Time Delay Neural Networks for Continuous Speech Recognition. In Neural Information Processing Systems (NIPS 4). Morgan Kaufmann, April 1992. [6] H. Hild and A. Waibel. Connected Letter Recognition with a Multi-State Time Delay Neural Network. To appear in Neural Information Processing Systems (NIPS 5). [7] K. Mase and A. Pentland. LIP READING: Automatic Visual Recognition of Spoken Words. Proc.

In a MS-TDNN the hierarchy continues above the phone-state layer with the Multi-State (MS) units [17] or better the DTW layer and word layer. In the forward pass of the network the DTW layer performs the Dynamic Time Warping algorithm [8] with the phoneme hyptheses as input to find the optimal path for the word models (German alphabet). The activations of the phone-state units along the optimal paths are accumulated in the word layer. The word unit with the highest score represents the recognized letter. In a second learning phase (after the TDNN bootstrapping) the networks are trained to fit word targets. The error derivatives are backpropagated from the word units through the best path in the DTW layer down to the front-end TDNNs, ensuring that the network is optimized for the actual evaluation task, which is letter and not phoneme recognition. As before, the acoustic and visual subnets are trained individually. In the final combined mode of the recognizer, a combined phone-state layer is included between the front-end TDNNs and the DTW layer. The activation of each combined phone-state unit is the weighted sum of the regarding acoustic phone-state unit and visual phone-state unit. We call these weights entropy-weights, because their values are proportional to the relative entropy between all acoustic phone-state activations and all visual phone-state activations. Hypotheses with higher uncertainty (higher entropy) are weighted lower than hypotheses with lower uncertainty. IV. PHONEME AND VISEME CODING For the acoustic classification we use a set of 65 phoneme-states (phoneme-to-phoneme transition states included). They represent a reasonable choice of smallest acoustic distinguishable units in German speech, and the TDNN architecture is very well suited to be trained as a classifier for them. For visual features this will be different. Distinct sounds are generated by distinct vocal tract positions, and voiced/ unvoiced excitations. External features of the vocal tract like the lips, part of the tongue and teeth, contribute only in part to the sound generation. I.e. /b/ and /p/ are generated by similar lip-movements, and cannot be distinguished with pure visual information. Training a TDNN to classify /b/ and /p/ based only on visual information would lead to recognition rates not better than guessing, or the net perhaps would get sensitive for features which are uncorelated to the produced speech. This leads to the design of a smaller set of visual distinguishable units in speech, so called visemes. We investigate a new set of 42 visemes and a 1-to-n mapping from the viseme set to the phoneme set. The mapping is necessary for the combined layer, in order to calculate the combined acoustic and visual hyphotheses for the DTW layer. For example the hypotheses for /b/ and /p/ are built out of the same viseme /b_or_p/ but the different phonemes /b/ and /p/ respectly. V. EXPERIMENTS Our database consists of 114 and 350 letter sequences spelled by two male speakers. They consist of names and random sequences. The first data set was split into 75 training and 39 test sequences (speaker msm). The second data set was split into 200 training and 150 test sequences (speaker mcb). Best results were achieved with 15 hidden units in the acoustic subnet and 7 hidden units in the visual subnet. Obviously visual speech data contains less information than acoustic data. Therefore better generalization was achieved with as little as 7 hidden units. Backpropagation was applied with a learning rate of 0.05 and momentum of 0.5. We applied different error functions to compute the error derivatives. For bootstrapping the McClelland error measure was applied, and for the global training on letter targets the Classification Figure of Merit [16] was applied. Acoustic Visual Combined msm/clean 88.8% 31.6% 93.2% msm/noisy 47.2% 31.6% 75.6% mcb/clean 97.0% 46.9% 97.2% mcb/noisy 59.0% 46.9% 69.6% Table 1: Results in word accuracy (words correct minus insertion and deletion errors) Table 1 summarizes the recognition performance results on the sentence level. Errors are misclassified words, insertion, and deletion errors. For speaker msm, we get an error reduction on clean data from 11.2% (acoustic only) down to 6.8% with additional visual data. With noise added to the acoustic data, the error rate was 52.8%, and could be reduced down to 24.4% with lipreading, which means an error reduction to less than half of the pure acoustic recognition. For speaker mcb, we could not get the same error reduction. Obviously the pronunciation of speaker mcb was better, but doing that, he was not moving his lips so much. It also should be noted that in the pure visual recognition a lot of the errors are caused by insertion and deletion errors. When we presented the letters with known boundaries, we came to visual recognition rates of up to 50.2%. The results of table 1 were achieved with histogram-nor-

reported that necessary techniques like edge detection are automatically learned by multi-layer perceptrons [10]. We apply two alternative preprocessing techniques: Histogram normalized grey-value coding, or 2 dimensional Fourier transformation. In both cases we just consider an area of interest (AOI) centered around the lips, and low pass filter these AOIs. The AOIs were initially segmented by hand, but an automatic procedure is now also available 1 [11]. Grey-Value coding: We found that a 24x16 pixel resolution is enough to recognize lip shapes and movements (Figure 1). Each of these AOI pixels is the average greyvalue of a small square in the original image (low pass filter). The grey-levels are rescaled in such a way that the darkest/brightest 5% in the histogram are coded with -1.0/ 1.0. The remaining 90% are scaled linear between -1.0 and 1.0. 2D-FFT: The AOI is rescaled to a 64x64 pixel image so that the 2 dimensional FFT results also with 64x64 coefficients. We just consider the log magnitudes of the first 13x13 FFT coefficients and rescale them to [-1.0, 1.0]. (After multiplying the complex FFT space with a 13x13 window and applying the inverse FFT, we could still recognize in the resulting low passed original image the distinct lip shapes and movements.) The motivation for considering the FFT is, that this coding is spatial shift invariant. It makes the recognition more stable against inaccurate AOI positioning. Phoneme Hypotheses DTW Layer Combined Hyptheses features of the lips change over time in a similar manner like the acoustic speech signal, so we use for the visual classification a similar TDNN architecture. Figure 2 shows the architecture. The preprocessed acoustic and visual data are fed into two front-end TDNNs [14], respectively. Each TDNN consists of an input layer, one hidden layer and the phone-state layer. Backpropagation was applied to train the networks in a bootstrapping phase, to fit phoneme targets. Entropy Word Hypotheses Viseme Hypotheses Figure 1: Typical AOIs III. THE MODULAR MS-TDNN ARCHITECTURE Time Delay Neural Networks are very well suited for low-level phoneme classification [14]. Because they act as general time-shift invariant feature extractors they are not limited to acoustic speech input only. In our case we want to process image sequences as well. The extracted visual Preprocessing TDNN Hidden Units Acoustic Input FFT Acoustic Preproc. Hidden Units Visual Input Histogr. Normaliz. OR Find Lip-AOI 2D-FFT 1. We don t want to hide the fact, that the complex task of finding and tracking the lips is very critical to the success of our recognizer. The Real Time Face Tracking System we are using right now is a template matching based approach and also used in a wider context of multi-modal human computer interface research, were tracking other parts of the face like for example the eyes is involved as well. Work is in progess to solve this problem with the connectionist architecture as well. Figure 2: Neural Network Architecture

BIMODAL SENSOR INTEGRATION ON THE EXAMPLE OF SPEECH-READING Christoph Bregler*, Stefan Manke University of Karlsruhe Department of Computer Science Am Fasanengarten 5 7500 Karlsruhe 1 Germany bregler@icsi.berkeley.edu, manke@ira.uka.de Hermann Hild, Alex Waibel Carnegie Mellon University School of Computer Science Pittsburgh Pennsylvania 15213 U.S.A. hhild@cs.cmu.edu, ahw@cs.cmu.edu Abstract -- In this paper we show how recognition performance in automated speech perception can be significantly improved by additional Lipreading, so called speech-reading. We show this on an extension of an existing state-of-theart speech recognition system, a modular MS-TDNN. The acoustic and visual speech data is preclassified in two separate front-end phoneme TDNNs and combined to acousticvisual hypotheses for the Dynamic Time Warping algorithm. This is shown on a connected word recognition problem, the notoriously difficult letter spelling task. With speech-reading we could reduce the error rate up to half of the error rate of the pure acoustic recognition. I. INTRODUCTION Recent development in the design of human computer interfaces (HCI) requests a new field of research: Multi- Modal Recognition. There are strong research efforts in all kind of alternative computer input methods on the way, like speech input, or visual input, but usually all different recognition methods are treated separately. When it comes to real-world-applications, single recognition methods have poor recognition performance or are limited to very small domains. Besides considering a lot of background knowledge, humans come to much better recognition performance in combining or substituting the different input sources itself in a flexible way. In speech recognition the main source is the acoustic signal, but if the signal is of low quality or ambiguous, visual information, i.e. lipmovements can contribute to the recognition process as well. In fact this is a well studied phenomena in human psychology often called speech-reading. Research with hearing impaired people and also normal hearing people has shown, that lipreading is subconsciously involved into human perception in a significant manner [3]. We investigate this phenomena on a state-of-the-art speech recognition architecture, a so called Multi State Time-Delay Neural Network (MS-TDNN [6]). We extended the architecture in a way, that besides the acoustic speech signals also visual signals (i.e. lipmovements) * The author is now with International Computer Science Institute, 1947 Center Street, Berkeley, CA 94704 are integrated in the learning and recognition process. A speaker is set in front of a microphone and a camera, spelling German names and random letter sequences. The spelling task is seen as a connected word recognition problem. As words we take the highly ambiguous 26 German letters. We do not care about high quality recordings, we even degrade the acoustic signal with artificial noise to simulate real-world conditions. The camera covers the full face, but as network input we just consider an area of interest (AOI) centered around the lips. We show how recognition performance with additional visual input further improves over acoustic recognition only. II. ACOUSTIC AND VISUAL INPUT REPRESENTATION Our bimodal input devices are a microphone and a conventional NTSC camera. The acoustic signal is sampled with a 16KHz rate and 12 bit accuracy. Every 10ms we compute 16 Melscale Fourier coefficients on the Hamming Windowed speech samples as input for the later classifier. Due to the NTSC standard, the video images are sampled with a much lower frequency. We digitize roughly every 33 ms a 256x256 pixel area covering the face of the speaker. We use 8bit gray-value coding for each pixel. Also timestamps were saved, because the correct synchronization between audio and video signals is critical for the later classification. In vision there does not exist such a well established body of preprocessing techniques like in speech recognition. Basically preprocessing for connectionist vision is split in two different schools of believes : 1. Apply as much as possible preprocessing in a manner, that just some few characteristic features are fed into the classifier. 2. Apply only transformations with very little information reduction and let the learning scheme find out the characteristic features. Edge detection would be an example for the first case and 2D-FFT for the second case. We follow the second approach, because we believe that conventional image processing techniques may cut out in an early stage important features, which could be useful for the later global learning scheme. In fact it has been