AUDIOVISUAL SPEECH RECOGNITION USING MULTISCALE NONLINEAR IMAGE DECOMPOSITION

Similar documents
SCALE BASED FEATURES FOR AUDIOVISUAL SPEECH RECOGNITION

Lip reading from scale-space measurements

2-2-2, Hikaridai, Seika-cho, Soraku-gun, Kyoto , Japan 2 Graduate School of Information Science, Nara Institute of Science and Technology

A New Manifold Representation for Visual Speech Recognition

LOW-DIMENSIONAL MOTION FEATURES FOR AUDIO-VISUAL SPEECH RECOGNITION

A Comparison of Visual Features for Audio-Visual Automatic Speech Recognition

Data Mining Chapter 3: Visualizing and Exploring Data Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University

SPEECH FEATURE EXTRACTION USING WEIGHTED HIGHER-ORDER LOCAL AUTO-CORRELATION

Sparse Coding Based Lip Texture Representation For Visual Speaker Identification

Nonlinear Manifold Learning for Visual Speech Recognition

Automatic Segmentation of Color Lip Images Based on Morphological Filter

Audio-visual interaction in sparse representation features for noise robust audio-visual speech recognition

3D LIP TRACKING AND CO-INERTIA ANALYSIS FOR IMPROVED ROBUSTNESS OF AUDIO-VIDEO AUTOMATIC SPEECH RECOGNITION

Multi-pose lipreading and audio-visual speech recognition

Mouth Center Detection under Active Near Infrared Illumination

Critique: Efficient Iris Recognition by Characterizing Key Local Variations

Face Recognition for Mobile Devices

Chapter 3. Speech segmentation. 3.1 Preprocessing

Production of Video Images by Computer Controlled Cameras and Its Application to TV Conference System

of Manchester The University COMP14112 Markov Chains, HMMs and Speech Revision

Speech Driven Synthesis of Talking Head Sequences

A Multimodal Sensor Fusion Architecture for Audio-Visual Speech Recognition

Multifactor Fusion for Audio-Visual Speaker Recognition

Automatic speech reading by oral motion tracking for user authentication system

Hybrid Biometric Person Authentication Using Face and Voice Features

FACE RECOGNITION USING INDEPENDENT COMPONENT

Particle Filtering. CS6240 Multimedia Analysis. Leow Wee Kheng. Department of Computer Science School of Computing National University of Singapore

Filtering Images. Contents

Research Article Audio-Visual Speech Recognition Using Lip Information Extracted from Side-Face Images

Probabilistic Facial Feature Extraction Using Joint Distribution of Location and Texture Information

Audio-Visual Speech Recognition Using MPEG-4 Compliant Visual Features

Features for Audio-Visual Speech Recognition

Bengt J. Borgström, Student Member, IEEE, and Abeer Alwan, Senior Member, IEEE

2. Basic Task of Pattern Classification

Articulatory Features for Robust Visual Speech Recognition

Dietrich Paulus Joachim Hornegger. Pattern Recognition of Images and Speech in C++

Short Survey on Static Hand Gesture Recognition

Local Filter Selection Boosts Performance of Automatic Speechreading

Applications Video Surveillance (On-line or off-line)

GYROPHONE RECOGNIZING SPEECH FROM GYROSCOPE SIGNALS. Yan Michalevsky (1), Gabi Nakibly (2) and Dan Boneh (1)

Locating 1-D Bar Codes in DCT-Domain

A GENERIC FACE REPRESENTATION APPROACH FOR LOCAL APPEARANCE BASED FACE VERIFICATION

Hand Gesture Extraction by Active Shape Models

Illumination-Robust Face Recognition based on Gabor Feature Face Intrinsic Identity PCA Model

ECG782: Multidimensional Digital Signal Processing

WATERMARKING FOR LIGHT FIELD RENDERING 1

Biometrics Technology: Multi-modal (Part 2)

SOUND EVENT DETECTION AND CONTEXT RECOGNITION 1 INTRODUCTION. Toni Heittola 1, Annamaria Mesaros 1, Tuomas Virtanen 1, Antti Eronen 2

Audio-visual speech recognition using deep bottleneck features and high-performance lipreading

Confusability of Phonemes Grouped According to their Viseme Classes in Noisy Environments

The Analysis of Animate Object Motion using Neural Networks and Snakes

AUTOMATIC VIDEO INDEXING

Broad field that includes low-level operations as well as complex high-level algorithms

ADVANCED IMAGE PROCESSING METHODS FOR ULTRASONIC NDE RESEARCH C. H. Chen, University of Massachusetts Dartmouth, N.

M I RA Lab. Speech Animation. Where do we stand today? Speech Animation : Hierarchy. What are the technologies?

Image Enhancement Techniques for Fingerprint Identification

Maximum Likelihood Beamforming for Robust Automatic Speech Recognition

Robust Lip Tracking by Combining Shape, Color and Motion

Robust speech recognition using features based on zero crossings with peak amplitudes

Invariant Recognition of Hand-Drawn Pictograms Using HMMs with a Rotating Feature Extraction

Mono-font Cursive Arabic Text Recognition Using Speech Recognition System

Recognition, SVD, and PCA

Lecture 8 Object Descriptors

Wavelet Transform in Face Recognition

Audio Visual Isolated Oriya Digit Recognition Using HMM and DWT

UNIVERSITY OF OSLO. Faculty of Mathematics and Natural Sciences

Facial Expression Detection Using Implemented (PCA) Algorithm

Convolution Neural Networks for Chinese Handwriting Recognition

Combining Audio and Video for Detection of Spontaneous Emotions

Locating Salient Object Features

The Analysis of Animate Object Motion using Neural Networks and Snakes

(and/or region-based image segmentation). These methods work indeed quite well in prole images and even in frontal images if the speaker wears lipstic

Haresh D. Chande #, Zankhana H. Shah *

Adaptive Quantization for Video Compression in Frequency Domain

Multi-modal Translation and Evaluation of Lip-synchronization using Noise Added Voice

The segmentation of images via scale-space trees

Modeling the Spectral Envelope of Musical Instruments

Further Studies of a FFT-Based Auditory Spectrum with Application in Audio Classification

Face Detection and Recognition in an Image Sequence using Eigenedginess

An Adaptive Eigenshape Model

An Automatic Face Identification System Using Flexible Appearance Models

A Novel Template Matching Approach To Speaker-Independent Arabic Spoken Digit Recognition

Shape Descriptor using Polar Plot for Shape Recognition.

The Novel Approach for 3D Face Recognition Using Simple Preprocessing Method

Lecture 7: Segmentation. Thursday, Sept 20

LOCAL APPEARANCE BASED FACE RECOGNITION USING DISCRETE COSINE TRANSFORM

Digital Image Processing

Image and speech recognition Exercises

Joint Processing of Audio and Visual Information for Speech Recognition

Matching and Recognition in 3D. Based on slides by Tom Funkhouser and Misha Kazhdan

Dynamic Time Warping

A Dynamic Human Model using Hybrid 2D-3D Representations in Hierarchical PCA Space

HMM-Based Handwritten Amharic Word Recognition with Feature Concatenation

Image Coding with Active Appearance Models

Face recognition using Singular Value Decomposition and Hidden Markov Models

Emotion Classification

Joint design of data analysis algorithms and user interface for video applications

Video shot segmentation using late fusion technique

Audio-Visual Speech Processing System for Polish with Dynamic Bayesian Network Models

Task analysis based on observing hands and objects by vision

Transcription:

AUDIOVISUAL SPEECH RECOGNITION USING MULTISCALE NONLINEAR IMAGE DECOMPOSITION Iain Matthews, J. Andrew Bangham and Stephen Cox School of Information Systems, University of East Anglia, Norwich, NR4 7TJ, UK Email: {iam,ab,sjc}@sys.uea.ac.uk ABSTRACT There has recently been increasing interest in the idea of enhancing speech recognition by the use of visual information derived from the face of the talker. This paper demonstrates the use of a nonlinear image decomposition, in the form of a sieve, applied to the task of visual speech recognition of a database of the letters A Z for ten talkers. A scale histogram is formed directly from the grayscale pixels of a window containing the talkers mouth on a per frame basis. Results are presented for visual-only, audio-only and for early and late-integrated audiovisual cases. 1. INTRODUCTION It has already been shown [1,6,8,1,12,14,15,17] that the incorporation of visual information with acoustic speech recognition leads to a more robust recogniser. While the visual cues of speech alone are unable to discriminate between all phonemes (e.g. [b] [p]) they do represent a useful separate channel that can be used to derive speech information. Degradation of one modality, e.g. interfering noise or cross-talk for audio, or occlusion for video, may be compensated to some extent by information from the other modality. In some cases information in each modality is complementary, e.g. the phonemes [m] [n] vary only by place of articulation and are acoustically similar but visually dissimilar. A major problem in audiovisual speech recognition is that of visual feature extraction. There are two general methods for extracting these features: model based or data driven. The model based approach immediately reduces the dimensionality of the problem to that of the model and allows the direct incorporation of any a priori knowledge of visual speech features. The problem arises that we do not know exactly which visual speech features to use, e.g. lip position/rounding/protrusion, presence and position of teeth, tongue etc. and these features are difficult to extract because they require the use of complex tracking algorithms. Using data driven methods we do not have to explicitly define the visual features as they are automatically learnt by the classifier. The main problem with images lies in how to reduce the dimensionality while retaining as much essential information as possible. This problem has been addressed for example by [6, 7, 15]. In this paper we use a nonlinear image decomposition method, a sieve [2 5], to decompose an image into the granularity domain. This completely describes the image in terms of granules that have the attributes position, amplitude and scale. A visual feature vector is formed using only the scale information in an attempt to define a feature that is relatively intensity and position invariant and yet entirely derived from the image itself. This scale-based visual feature vector is used for Hidden Markov Model (HMM) recognition for visual only, early integrated and late integrated audiovisual cases. 2. SIEVE DECOMPOSITION A sieve or datasieve is an algorithm that uses rank or morphological filters to simplify signals over multiple scales, that preserves scalespace causality, and can reversibly transform a signal to a granularity domain. Sieves represent the development of mathematical morphology to form an alternative to wavelet decomposition. A sieve is defined by, Φ m(x) = φ(φ m 1(X)) where Φ (X) = X The operator φ m may be an open/close (M sieve) or close/open (N sieve) or a recursive equivalent. In the one-dimensional case used in this paper, Φ m : Z Z is based on a series of increasing scale operations. Defining φ m,m = 1,2,...,m as recursive median, ρ mf(x) = med(ρ mf(x m+1),..., ρ mf(x 1),f(x),...,f(x+m 1)) gives the recursive median or R sieve, R m(x) = ρ m(r m 1(X)), R (X) = X the granularity of which is obtained from, Gran R(X)(m) = (R m(x)) (R m+1(x)) The granularity consists of the set of granules G that represent the non-zero intervals in the granule functions and are characterised by the triplet {position, amplitude, scale}. The sieve transform maps the signal into a set of granules, S : Z G The inverse, S 1, may be obtained by summing the re-expanded granules. 1

3. VISUAL FEATURE EXTRACTION Utterance "D G M" The goal of visual feature extraction is to obtain information about the current image frame that is robust across varying intensity, scale, translation, rotation, viewing angle and talkers. This is an even more difficult task for pixel based systems, but the use of e.g. active shape models [11], deformable templates [9] or dynamic contours [1] to track the lip contours removes the possibility of learning any other visual cues that may be significant. The one dimensional recursive median sieve defined above can be used to decompose two dimensional image by scanning in a given direction, e.g. over each column of the image for the vertical case as this is the direction in which most motion occurs during speech. In practice the sieve transform is applied to an entire image in a single pass. The resulting granularity contains the position, amplitude and scale information of the set of granules that describe that image. As a transform of the original image it contains all the information. We must discard the amplitude attribute of the granules as this largely codes the intensity of the image and we require the visual feature to be independent of intensity variation. The position variation of a granule is dependent on inter-frame differences of the image feature to which it belongs. To use position information would require the identification and tracking of interesting granules (image features), which is counter to the data-driven paradigm. The scale parameter is relatively insensitive to to intensity variations (until quantisation effects become large) and translation. As the image scene varies between frames (e.g. mouth opens, teeth become visible) the number of granules of a given scale in the image will change. We can now collapse the granularity spectrum, by ignoring position and amplitude information, and form a scale histogram. This is the number of granules of each scale summed across the entire image. This provides a simple method of substantially reducing the dimensionality of the raw image data to that of the maximum scale used in the sieve transform, pixels in this work. Example scale histograms are show in Figure 1. The top panel shows typical frames from the image sequence a neutral frame and the centre frame from each utterance. The mouth region was roughly located and the scale histogram of the region obtained. This is plotted as intensity, white represents a large number of granules and scale increases down the y-axis. The number of granules clearly change whenever the mouth moves. The bottom panel shows the corresponding audio signal. Figure 2 shows a single utterance in more detail. The utterances are isolated letters and, as expected, the visual cues can be seen to begin before and end after the acoustic signal. 4. DATABASE An audiovisual database was recorded for ten talkers, five male (two with moustaches) and five female. An example frame from each talker is shown in Figure 7. Each talker repeated each of the letters A to Z three times, a total of 78 utterances. Recording took place in the University TV studio under normal studio lighting con- Sound Scale Histogram 1 2 3 1.5 1.5.5 1 x 1 4 2 8 1 12 Frame 1 2 3 4 5 6 7 8 9 1 11 Sample Figure 1: Example scale histogram and audio waveform for the utterance sequence D-G-M. Sound Scale Histogram 1 2 3 1 Utterance "J" 5 1 15 2 25 Frame.2.4.6.8 1 1.2 1.4 1.6 1.8 2 2.2 Sample x 1 4 Figure 2: Example scale histogram and audio waveform for the utterance J. ditions. Three cameras simultaneously recorded different views of the talker: full face, mouth only and a side view. All recording was done to tape, the full face to SVHS quality. The output of a high quality tie clip microphone was adjusted for each talker through a sound mixing desk and fed to all video recorders. An autocue presented the letters A to Z three times in a nonsequential order. Each talker was asked to return their mouth to the neutral position after each utterance and allowed to simply watch the autocue. No attempt at restraining was made but talkers were asked not to move their mouth out of frame of the mouth close up camera. For this work only the full face data has been used. All 78 utterances have been digitised at quarter frame PAL (376 288) resolution and full PAL frame rate (25Hz) using the standard frame grab- x 1 4 2

ber hardware of a Macintosh Quadra 6AV. All video was digitised to 8-bit grayscale. was simultaneously digitised using 16-bit resolution at 22.5kHz to ensure audiovisual time alignment. Each utterance movie was hand segmented using the video channel so that each image sequence began and ended with the mouth in the neutral position. The audio data within this window was then hand labelled as silence letter silence. 5. UNIMODAL RECOGNITION The oral region was manually extracted from each of the utterance movies of the database. This was done by positioning a window of 8 pixels centrally on the mouth image of the middle frame of each image sequence. Although there is some head motion of the talkers the mouth always stays within this region. Scale histograms were generated for all utterance movies by applying a vertical one dimensional sieve to each frame. The vertical dimension of all images is pixels so this is the dimensionality of each scale histogram. All recognition experiments were performed using the first two utterances from each of the ten talkers as a training set (2 training examples per utterance) and the third utterance from each talker as a test set (1 test examples per utterance). Classification was done using 1 state, left to right HMMs with each state associated with a single Gaussian density with a diagonal covariance matrix. All HMMs were implemented using the HTK Hidden Markov Model Toolkit V1.4. to 2ms frame rates. Figure 4 shows the results for the best visual Average PCA1 PCA2 Method PCA1 (DC) PCA2 (DC) Normal +Delta 15 23.85 3. 3 25. 28.8 Corr 23.85 25. Cov 28.46 3.77 Corr 21.92 27.31 Cov 34.23 33.46 Corr 28.46 29.62 Cov 24.62 28.46 Corr 27.31 29.62 Cov 23.8 3.77 Table 1: Summary of visual-only results. case (2 PCA coefficients calculated using the covariance matrix of non-dc preserved scale histograms) on a per-talker basis with audio only results for comparison. The audio features consisted of 12 MFCC coefficients plus an energy term and delta coefficients, i.e. 26 coefficients, calculated at a frame rate of 2ms. The same HMM topology and training/test data was used for audio only as visual only tasks. The variation in visual only performance is clearly 1 Unimodal Recognition To further reduce the size of the visual feature vector principle component analysis was applied to the entire data set. Figure 3 shows the variance accounted for as a function of the number of PCA coefficients used. Experiments have used 1 (95%) and 2 (99%) coefficients. Further experiments have compared simple averag- 1 98.57 Eigenvalue Contribution 9 8 7 86.15 9 94.84 3 34.23 2 % Variance Accounted 8 7 1 1 2 3 4 5 6 7 8 9 1 Average Figure 4: Unimodal results for each talker. Overall accuracy: visual only 34.23%, audio only 86.15%. 1 2 3 No. coeffs used Figure 3: Principle Component Analysis. ing of adjacent values and using scale histograms generated from zero padded images (this has the effect of preserving the dc, i.e. largest scale, values) and performing PCA using both covariance and correlation matrices (to ignore or imply equal importance of all variables). All results are summarised in Table 1 which also shows results for classification with appended delta coefficients. All visual results were obtained using visual features interpolated from ms much greater than that for audio only (2 65% correct), and there is little correlation between poor audio and poor visual performance. 6. BIMODAL RECOGNITION The two extremes for bimodal integration are the early and late methods. For early integration a single classifier is used with composite audiovisual features. Late integration combines the result of two independent classifiers, one audio, one visual. Further methods have been proposed by [13] and implemented by [1] which extend these methods by adding a recoding step. Both early and late integration methods have been implemented. 3

6.1. Early Integration For early integration a composite audiovisual feature vector was formed by concatenating audio and visual features to form 46 coefficient audiovisual features. The HMM topology described for unimodal recognition was used. Figure 5 shows the recognition accuracy results for audio-only, visual-only and early integrated audiovisual tasks over varying signal to noise ratios (SNR). The appropriate amount to Gaussian white noise was added to take the audio utterances from clean (25 3dB) to the desired SNR. 9 8 7 3 Updated version of ICSLP 96, volume 1, pages 38 41 Results Summary Late Early A problem with early integration is how to segment the utterances our database is labeled for audio and visual data independently to allow the inclusion of the visual articulation that occurs before and after the acoustic utterance. Both methods are shown in Figure 5 and the longer visual segmentation gives improved results over audio segmentation. Both methods improve over audio only results but below 2dB accuracy fall below that of visual alone. 2 1 Clean 2dB 1dB 6dB db 6dB SNR Figure 6: Late integration results. 9 Early Integration 7. CONCLUSIONS 8 7 3 2 1 Seg. Seg. Clean 2dB 1dB 6dB db 6dB SNR 6.2. Late Integration Figure 5: Early integration results. For late integration a method of combining the output of two independent audio and visual classifiers must be found. One method, when there is disagreement, is to assign the output to the classifier which is most confident. A confidence measure may be formed by examining the normalised log-likelihoods of each classifier. A simple method is to form the ratio of the maximum normalised loglikelihood and the night highest, if this ratio is high the next best candidate for classification was much lower than the best candidate. The output is then assigned to the classifier with the highest ratio. { A if A1/A 2 > V 1/V 2 output = V if A 1/A 2 < V 1/V 2 The results in Figure 6 show that for low SNR late integration gives poor results compared to audio alone, this is due to the visual classifier, by this confidence measure, being confidently wrong on several occasions. At higher SNR performance is improved over early integrated audiovisual, but falls below visual only at 1dB. These results indicate that the scale histogram visual speech feature vector can be successfully used in visual speech recognition. A recogniser using only visual information attained an average performance of 34% across ten talkers in a multi-talker, isolated letter recognition task. We have also demonstrated two methods of combining this information with standard audio features to improve recognition under noisy conditions using both early and late integration methods. Future work will focus on finding more effective ways of integrating the audio and visual information with the aim of ensuring that the combined performance is always at least as good as the performance using either modality [1, 13, 14, 17] and in deriving more discriminative features from a scale histogram to increase robustness across talkers. 8. REFERENCES 1. A. Adjoudani and C. Benoît. On the integration of auditory and visual parameters in an HMM-based ASR. In Stork and Hennecke [16], pages 461 471. 2. J. A. Bangham, P. Chardaire, C. J. Pye, and P. D. Ling. Mulitscale nonlinear decomposition: The sieve decomposition theorem. IEEE Trans. Pattern Analysis and Machine Intelligence, 18(5):529 539, 1996. 3. J. A. Bangham, R. Harvey, P. Ling, and R. V. Aldridge. Nonlinear scale-space from n-dimensional sieves. Proc. European Conference on Computer Vision, 1:189 198, 1996. 4. J. A. Bangham, P. Ling, and R. Harvey. Scale-space from nonlinear filters. IEEE Trans. Pattern Analysis and Machine Intelligence, 18(5):52 528, 1996. 5. J. A. Bangham, P. Ling, and R. Young. Mulitscale recursive medians, scale-space and transforms with applications to image processing. IEEE Trans. Image Processing, 5(6):143 148, 1996. 6. C. Bregler, S. M. Omohundro, and Y. Konig. A hybrid approach to bimodal speech recognition. In 28th Annual Asilomar 4

Conference on Signals, Systems, and Computers, volume 1, pages 556 5, Pacific Grove, CA, Nov. 1994. 7. N. M. Brooke and S. D. Scott. PCA image coding schemes and visual speech intelligibility. Proc. Institute of Acoustics, 16(5):123 129, 1994. 8. N. M. Brooke, M. J. Tomlinson, and R. K. Moore. Automatic speech recognition that includes visual speech cues. Proc. Institute of Acoustics, 16(5):15 22, 1994. 9. M. E. Hennecke, D. G. Stork, and K. V. Prasad. Visionary speech: Looking ahead to practical speechreading systems. In Stork and Hennecke [16], pages 331 349. 1. R. Kaucic, B. Dalton, and A. Blake. Real-time lip tracking for audio-visual speech recognition applications. In B. Buxton and R. Cipolla, editors, Proc. European Conference on Computer Vision, volume II of Lecture Notes in Computer Science, pages 376 387, Cambridge, Apr. 1996. Springer-Verlag. 11. J. Luettin, N. A. Thacker, and S. W. Beet. speech recognition using active shape models and hidden markov models. In Proc. International Conference on Acoustics, Speech and Signal Processing, volume 2, pages 817 82, Atlanta, GA, May 1996. IEEE. 12. E. D. Petajan, B. J. Bischoff, D. A. Bodoff, and N. M. Brooke. An improved automatic lipreading system to enhance speech recognition. Technical Report TM 11251-87112-11, AT&T Bell Labs, Oct. 1987. 13. J. Robert-Ribes, M. Piquemal, J.-L. Schwartz, and P. Escudier. Exploiting sensor fusion architectures and stimuli complementarity in AV speech recognition. In Stork and Hennecke [16], pages 193 21. 14. P. L. Silsbee. visual sensory integration using hidden markov models. In Stork and Hennecke [16], pages 489 496. 15. P. L. Silsbee and A. C. Bovik. Medium vocabulary audiovisual speech recognition. In New Advances and Trends in Speech Recogntion and Coding, pages 13 16. NATO ASI, 1993. 16. D. G. Stork and M. E. Hennecke, editors. Speechreading by Humans and Machines: Models, Systems and Applications, volume 1 of NATO ASI Series F: Computer and Systems Sciences. Springer-Verlag, Berlin, 1996. 17. M. J. Tomlinson, M. J. Russell, and N. M. Brooke. Integrating audio and visual information to provide highly robust speech recognition. In Proc. International Conference on Acoustics, Speech and Signal Processing, volume 2, pages 821 824, Atlanta, GA, May 1996. IEEE. 1 2 3 4 5 6 7 8 9 1 Figure 7: Example frame from each talker in the database. 5