Acoustic to Articulatory Mapping using Memory Based Regression and Trajectory Smoothing

Similar documents
Trajectory Mixture Density Networks with Multiple Mixtures for Acoustic-articulatory Inversion

Chapter 3. Speech segmentation. 3.1 Preprocessing

Pitch Prediction from Mel-frequency Cepstral Coefficients Using Sparse Spectrum Recovery

Locally Weighted Learning

Applying Supervised Learning

M I RA Lab. Speech Animation. Where do we stand today? Speech Animation : Hierarchy. What are the technologies?

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS

Artificial Neural Network-Based Prediction of Human Posture

Perceptron as a graph

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition

Clustering K-means. Machine Learning CSEP546 Carlos Guestrin University of Washington February 18, Carlos Guestrin

Text-Independent Speaker Identification

Extending reservoir computing with random static projections: a hybrid between extreme learning and RC

Density estimation. In density estimation problems, we are given a random from an unknown density. Our objective is to estimate

EM Algorithm with Split and Merge in Trajectory Clustering for Automatic Speech Recognition

Linear Separability. Linear Separability. Capabilities of Threshold Neurons. Capabilities of Threshold Neurons. Capabilities of Threshold Neurons

Homework. Gaussian, Bishop 2.3 Non-parametric, Bishop 2.5 Linear regression Pod-cast lecture on-line. Next lectures:

Instance-based Learning

Gas Distribution Modeling Using Sparse Gaussian Process Mixture Models

Neetha Das Prof. Andy Khong

Clustering K-means. Machine Learning CSEP546 Carlos Guestrin University of Washington February 18, Carlos Guestrin

Particle Swarm Optimization applied to Pattern Recognition

HIDDEN Markov model (HMM)-based statistical parametric

The Un-normalized Graph p-laplacian based Semi-supervised Learning Method and Speech Recognition Problem

Neural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

Establishing Virtual Private Network Bandwidth Requirement at the University of Wisconsin Foundation

Introduction to Trajectory Clustering. By YONGLI ZHANG

Interferogram Analysis using Active Instance-Based Learning

Locally Weighted Learning for Control. Alexander Skoglund Machine Learning Course AASS, June 2005

Locally Weighted Least Squares Regression for Image Denoising, Reconstruction and Up-sampling

10-701/15-781, Fall 2006, Final

Research on the New Image De-Noising Methodology Based on Neural Network and HMM-Hidden Markov Models

Tying well information and seismic data

Homework 4: Clustering, Recommenders, Dim. Reduction, ML and Graph Mining (due November 19 th, 2014, 2:30pm, in class hard-copy please)

Performance Measures

Large Scale Data Analysis Using Deep Learning

Chap.12 Kernel methods [Book, Chap.7]

COMPUTATIONAL INTELLIGENCE SEW (INTRODUCTION TO MACHINE LEARNING) SS18. Lecture 6: k-nn Cross-validation Regularization

Voronoi Region. K-means method for Signal Compression: Vector Quantization. Compression Formula 11/20/2013

International Journal of Scientific Research & Engineering Trends Volume 4, Issue 6, Nov-Dec-2018, ISSN (Online): X

Box-Cox Transformation for Simple Linear Regression

Simulation of Zhang Suen Algorithm using Feed- Forward Neural Networks

Audio-visual interaction in sparse representation features for noise robust audio-visual speech recognition

Machine Learning and Pervasive Computing

2014, IJARCSSE All Rights Reserved Page 461

Biometrics Technology: Image Processing & Pattern Recognition (by Dr. Dickson Tong)

Facial Expression Classification with Random Filters Feature Extraction

Voice Conversion for Non-Parallel Datasets Using Dynamic Kernel Partial Least Squares Regression

Invariant Recognition of Hand-Drawn Pictograms Using HMMs with a Rotating Feature Extraction

Introduction to Objective Analysis

CS 543: Final Project Report Texture Classification using 2-D Noncausal HMMs

SPEECH FEATURE EXTRACTION USING WEIGHTED HIGHER-ORDER LOCAL AUTO-CORRELATION

Spatial Analysis and Modeling (GIST 4302/5302) Guofeng Cao Department of Geosciences Texas Tech University

Four equations are necessary to evaluate these coefficients. Eqn

EE368 Project Report CD Cover Recognition Using Modified SIFT Algorithm

An Algorithm For Training Multilayer Perceptron (MLP) For Image Reconstruction Using Neural Network Without Overfitting.

The Approach of Mean Shift based Cosine Dissimilarity for Multi-Recording Speaker Clustering

A Novel Template Matching Approach To Speaker-Independent Arabic Spoken Digit Recognition

3 Nonlinear Regression

Modeling time series with hidden Markov models

Gender-dependent acoustic models fusion developed for automatic subtitling of Parliament meetings broadcasted by the Czech TV

Louis Fourrier Fabien Gaie Thomas Rolf

Toward a robust 2D spatio-temporal self-organization

MACHINE LEARNING: CLUSTERING, AND CLASSIFICATION. Steve Tjoa June 25, 2014

1 Case study of SVM (Rob)

Nearest Neighbor Predictors

A Lazy Approach for Machine Learning Algorithms

Model Assessment and Selection. Reference: The Elements of Statistical Learning, by T. Hastie, R. Tibshirani, J. Friedman, Springer

2-2-2, Hikaridai, Seika-cho, Soraku-gun, Kyoto , Japan 2 Graduate School of Information Science, Nara Institute of Science and Technology

Classifying Depositional Environments in Satellite Images

Classification of Subject Motion for Improved Reconstruction of Dynamic Magnetic Resonance Imaging

Overview Citation. ML Introduction. Overview Schedule. ML Intro Dataset. Introduction to Semi-Supervised Learning Review 10/4/2010

Big Data Methods. Chapter 5: Machine learning. Big Data Methods, Chapter 5, Slide 1

Introduction to ANSYS DesignXplorer

Speech User Interface for Information Retrieval

An Introduction to PDF Estimation and Clustering

A Comparison of Visual Features for Audio-Visual Automatic Speech Recognition

A *69>H>N6 #DJGC6A DG C<>C::G>C<,8>:C8:H /DA 'D 2:6G, ()-"&"3 -"(' ( +-" " " % '.+ % ' -0(+$,

The Fly & Anti-Fly Missile

5 Learning hypothesis classes (16 points)

Instance-based Learning

ENSEMBLE RANDOM-SUBSET SVM

Accelerometer Gesture Recognition

Spatial Enhancement Definition

An Empirical Study of Hoeffding Racing for Model Selection in k-nearest Neighbor Classification

Available online Journal of Scientific and Engineering Research, 2016, 3(4): Research Article

CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, :59pm, PDF to Canvas [100 points]

Directionally Selective Fractional Wavelet Transform Using a 2-D Non-Separable Unbalanced Lifting Structure

Recent advances in Metamodel of Optimal Prognosis. Lectures. Thomas Most & Johannes Will

application of learning vector quantization algorithms. In Proceedings of the International Joint Conference on

Global Journal of Engineering Science and Research Management

RESAMPLING METHODS. Chapter 05

Machine Learning: k-nearest Neighbors. Lecture 08. Razvan C. Bunescu School of Electrical Engineering and Computer Science

Comparative Evaluation of Feature Normalization Techniques for Speaker Verification

Random projection for non-gaussian mixture models

K-Nearest Neighbors. Jia-Bin Huang. Virginia Tech Spring 2019 ECE-5424G / CS-5824

>> x2=-5.01:0.25:5.01; >> [x, y]=meshgrid(x1,x2); >> f=mexhat(x,y); >> mesh(x1,x2,f); The training data and test data sets with 100 instances each are

Human Motion Synthesis by Motion Manifold Learning and Motion Primitive Segmentation

6.034 Quiz 2, Spring 2005

Transcription:

Acoustic to Articulatory Mapping using Memory Based Regression and Trajectory Smoothing Samer Al Moubayed Center for Speech Technology, Department of Speech, Music, and Hearing, KTH, Sweden. sameram@kth.se Abstract This paper investigates the memory based regression for the task of acoustic-to-articulatory inversion. In memory based regression, a local model is built for each test sample using the k nearest neighbors, and an articulatory vector is estimated. In this paper we investigates different regression models, from the basic Average estimate, to a polynomial estimate. The paper also investigates the importance of optimizing the number of nearest neighbors (k) on the error function. For this purpose, we use a neural network to estimate the number of neighbors used for regression. This method is used to estimate the articulatory values, and their first two derivatives. The values and their derivatives are then used on an utterance level to temporally smooth and estimated articulatory parameters using the Trajectory Likelihood Maximization method. We apply this method on the MOCHA database, which consists of 167000 acoustic-articulatory frames. These steps combined, and using 15 fold cross validation on MOCHA, show that memory based regression with optimizing the number of neighbors, followed by trajectory smoothing can decrease the Mean Square Error beyond the state-of-the-art methods. Index Terms: Acoustic-Articulatory Inversion, Visual Speech, articulation, memory based learning, local regression 1 Introduction Lazy learning methods defer processing of training data until a query needs to be answered. This usually involves storing the training data in memory, and finding relevant data in the database to answer a particular query. This type of learning is also referred to as memory-based learning. Relevance is often measured using a distance function, with nearby points having high relevance. One form of lazy learning finds a set of nearest neighbors and selects or votes on the predictions made by each of the stored points. In most learning methods a single global model is used to fit all of the training data. Since the query to be answered is known during processing of training data, training query-specific local models is possible in lazy learning. Local models attempt to fit the training data only in a region around the location of the query (the query point). Examples of types of local models include nearest neighbor, weighted average, and locally weighted regression. Since memory based regression, aims at solving a golbal complicated function estimation task by building linear global models for every test sample, this method is usualy advantages in cases of noise, unbalanced class representation in the training data and variable data density in different regions in the input space. In memory based learning, parameter optimization plays a major role in controling the accuracy of the system, such parameters are k, the number of the nearest neighbors to be used in the regression, and the local model specific parameters and weights. In this paper, we study the application of memory based regression to map acoustic input to articulatory parameters. This application has been, and is, crucial to many tasks, such as articulatory speech synthesis, speech driven facial animation, and the modeling of systems for speech recognition systems. Speech acoustic parameters are highly temporal, where any acoustic sample in time is highly affected by its context of what speech has been pronounced and what is going to be pronunced. In this study, we propose to use a global smoothing method called Maximum Likelihood Trajectory (MLT). Where not only the place of the articulatory parameters are estimated by the local models, but their speed and acceleration. These estimations are then optimized to minimize the global function of the absolute place of the parameters and their derivatives. The paper is organized as follows, Section 2 provides an overview of the MOCHA database, which is used in this study. Section 3 presents the four different local regression methods to be tested. Section 4 presents studies on the effects of the number of k-nearest neighbors and how to optimize it. In Section 5, we explain the MLT smoothing method. Section 6 presents the experiments and evaluation results on the MOCHA dataset. In Section 6 we discuss the different aspects of the paper, and in Section 7 we draw some conclusions and future work. 2 The MOCHA database This study uses the simultanuously recorded Acoustic-EMA data from the MOCHA database [1] consisting of 460 TIMIT sentences spoken by two speakers (one male and one female). The phonetic labels were converted to SAMPA codes from the British English SpeechDat database, giving a total number of 44 phonemes including silence and breath. The acoustic features were the first 14 MFCCs (including the 0th component) computed at 10 ms frame rate. Delta Mel Frequency Cepstral Coefficients (MFCC) were computed using a Hamming window over 5 frames and added to the acoustic features resulting in a total of 28 dimensions. The 14 articulatory channels consisted of the X- and Y-axis trajectories of 7 EMA coils. These were low-pass filtered and downsampled to 100 Hz, in order to correspond to the acoustic frame shift rate. The delta features for the articulatory measurements were computed just like it was done for the MFCC. The articulatory features vectors were normalized to zero mean with a standard deviation of one. The data constituted to 167000 frames of acoustic-articulatory

Figure 1: An example of the function estimated using the different local regression methods. a) Average. b) Weighted Average. c) Second order polynomial fit. data, labeled with the 44 phonetic lables, including Silence and Breath. 3 Local Regression Methods The specific goal of this study is to test methods on estimating the absolute value of the articulatory parameter (a dependent variable) from the acoustic parameters (independent variable). Usualy this is done by estimating the mapping function by building a model from training data. In the case of local regression, this model is built -online- for each new sample. This model is usualy simple, in relation to a more complicated model needed to model all the data at the same time. To estimate this model, only a set of the training data, which is the most similar to this test sample, are used. In [2], a detailed review is presented on the different models, and methods, used for local regression, and their features. In this study, we will build local models using three regression methods, varaying in complexity and requirements. This models are presented in details in the following sections. 3.1 Average Averaging is the simplest possible method to estimate a value using a set of examples. In averaging, the estimated value is the average of the values of the k-nearest neighbors. Figure 1a presents an example of the estimated contour of sample data with k=3. The advantage of this method is its simplicity, while it treats all the neighbors with the same importance. 3.2 Weighted Average In the weighted average method, the nearer neighbors are given more importance than the farther neighbors using a weighting. This weighting is a function of the distance of each neighbor from the new sample. A very common weighting scheme usualy used is what is called a Normal Regression Kernel or a Gaussian Regression Kernel, where it is assumed that the neighbors around the new sample are normally distributed, and the sample constituts and mean of its distribution. Hence, the Gaussian function is used to define a weighting for each of the neighbors values. The weighting is calculated according to the following equation: Where d is the Ecludian distance for the neighbor i and sigma is the width of the gaussian (the standard deviation) of the distribution. sigma is usualy optimized using cross validation tests on the available training data. Figure 1b presents the estimated contour of sample data with k=3 using weighted regression with a Gaussian kernel. 3.3 Polynomial Regression Local models (often polynomials) have been used for over a century to smooth regularly sampled time series and interpolate and extrapolate from data arranged on rectangular grids. [3], [4] and [5] suggested using a weighted regression on irregularly spaced data to fit a local polynomial model at each point a function evaluation was desired. All of the available data points were used. Each data point was weighted by a function of its distance to the desired point in the regression. Many authors have suggested fitting a polynomial surface only to nearby points also using distance weighted regression [6]. Polynomials require a large amount of neighbors (training data), compared to the previous two methods, to adequately fit a polynomial curve. Figure 1c presents the estimated contour of sample data with k=3 using using polynomial fit of the second order. 4 Optimizing the k-nearest Neighbors One of the major parameters which need optimization is the number of nearest neighbors used for regression. 30000 datapoints were randomly taken from the MOCHA dataset to test the effect of k (the number of nearest neighbors) using the Weighted Average regression method described earlier. Figure 2 shows the effect of incrasing k on the Mean Square Error (MSE) of the 30000 articulatory points. It is known commonly, that increasing k usualy yields to a lower estimation error, but suffers from more expensive computational poiwer to deal with a big number of k. The number k, though, still highly depends on the specific sample we need to estimate, since the training data might not always provide the same

Figure 2: Left:A plot of the sorted mean square error for the 30000 points with differing number of k (1..20), with lower mean square error for increasing k. Right: The average mean square error over k. Figure 3: Left: The MSE error over k when the test data is part of the training data. Right: The MSE error over k when the test data is not part of the training data. information to all data points in the articulatory space. To illustrate this in an example, 2000 test points were estimated using the weighted averaging method, with 10000 training points, for a k ranging from 1 to 20. This 2000 test points, where once taken from the same training data (the 10000 points) for one test, and from an external dataset for the other test. Figure 3 shows the MSE error over k for the test points which had a matching point in the training data, and for those which did not. We can see that, actually, increasing k is sometimes disadvantages, since building local models, still estimates a general model in a local space, and hence this model is over-generalized for some test samples (those which have very similar training samples). To test this, we calculated the error from the first test with the 30000 samples, with an optimized k (from these 1 to 20) where the error was minimized, we saw that the mean square error is significantly decreased when k value is optimized compared to this of k=20, as shown in Figure 4. Since optimizing k can increase the MSE error significantly, we trained a neural network, consisting of 3 layers, with 50 hidden neurons, to estimate k using acoustic input vectors. The neural network was trained using back-propagation for 100 iterations, with 20000 training samples, 5000 for validation, and 5000 for testing. After training, the network was used to estimate k for the test sample of 5000. Figure 5 shows the MSE error for the test set using k=20, and k estimated using the trained neural network, and k is optimized for the first 20 k-s. The figure shows that the Figure 4: The difference in MSE between k=20, and when k is optimized from the set of k=[1..20] neural network, has resulted in a decreased MSE regression error by estimating the number of nearest neighbors. 5 Maximum Likelihood Trajectory Smoothing In order to use a constraint on articulatory movements, we apply a parameter generation algorithm based on ML using dynamic features [7] to the GMM-based mapping. Hiroya et al. also use this technique in the inversion mapping with the HMM-based speech production model [8] where the smoothing idea is explained in

Figure 5: The difference in MSE between 1- k=20, 2- k is estimated using a neural network NN, and 3- when k is optimized from the set of k=[1..20] Figure 8: A plot of one of the articulatory parameters, showing the correct trajectory in blue, the estimated one without smoothing in red, and the final estimated trajectory in black. dataset. Figure 6: MSE error for the three regression methods. details. The basic idea is to estimate a new parameter vector where the dynamics of this vector are as similar as possible to the estimated dynamics of the model. This is practically done by finding an inverse of matrix representing the vector and its dynamics. 6 Evaluation and Results 15 cross validation studies have been carried out using 150000 frames from the MOCHA dataset. The first test is to evaluate the three proposed regression models. Figure6 presents the MSE error over the three methods using k=20 nearest neighbors, showing that the weighted average using a Gaussian kernel gives the lowest error compared to the other two methods. Using these results, the final cross validation test conducted is the trajectory smoothing, using the neural network to estimate k, and then using the weighted average regression technique. These tests are done on a sentence level, where a full utterance articulatory parameters are estimated, along with their speed and acceleration using the same neighbors. Figure 7 presents the processing flow chart applied to estimate the articulatory parameters. Figure 8 shows a sample of test data, with the correct articulatory trajectory, and the estimated one without smoothing, and the smoothed one using Maximum Likelihood Trajectory smoothing. Concerning technical details, the full algorithms were implemented in Matlab, with speed optimized code. Nearest neighbors regression is typically very ineffecient, if not optimized, due to the need for sorting all the training data by their distance from each new test sample. In these tests, no clustering or decision trees were used which might compensate accuracy over speed, hence the processing time, after optimizing the code, consumed about 3 days of processing for the 15 fold cross validation on the 150000 Table 1: The MSE on 15 fold cross valiadtion Method MSE k=2 no smoothing 0.73 k is optimized using NN, no smoothing 0.67 k is optimized and trajectories smoothed 0.63 the same but only on data without silence 0.61 GMM and trajectory smoothing, without silence 0.65 Table 1 presents the final cross validation results when using k=20 (without smoothing), and when k is optimizedusing NN (without smoothing), and when k is optimized using NN, and with trajectory smoothing, and we compare this to the reported state of the art on the same dataset, using Gaussian Mixture Models with trajectory smoothing. We can see that our results, using memory based regression, with optimizing the number of the neighbors and the regression methods. 7 Discussion The experiments in this paper investigate the possibility of using memory based regression and maximum likelihood trajectory smoothing to estimate articulatory parameters directly from the acoustics. The experiments show the importance of optimizing the number of nearest neighbors to be used in the regression, the local regression model to be used, and the need for a global utterance based trajectory smoothing to take context into account. Many other experiments will be usefull to be tried out, as in building an input vector from multiple frames, and hence taking into account more context, this on one hand might enhance the regression accuracy since context plays a big role in defining the articulatory parameters, on the other hand, this step might make the training data more sparse, and the computational processing more demanding. To compensate for the big computational consumption of memory based learning, some parameter search and clustering might be established to speed up the search for the neighbors us-

Figure 7: The overall flow chart of the system presenting the different steps. ing discretization and quantization on the acoustic vector, nevertheless, this might hinder the estimation accuracy. Regarding the input feature vector, in addition to taking more context into account to represent the feature space, one can try to add more acoustic parameters as the fundamental frequency and the symbolic feature representing the phone represented in that specific vector (if phonetic boundaries are obtained). 8 Conclusions In this study, we tested the use of memory based regression for the task of estimating the articulatory parameters from acoustic frames. This studies used the MOCHA database, which is one of the standard datasets used to evaluate tihs task. In our experiments, we found that optimizing k can decrease the MSE significantly, and hence we developed a neural network to estimate the number of neighbors to be taken into building the local model. Three different regression methods were used, simple averaging of the articulatory parameters of the neighbors, weighted averaging using a Gaussian kernel, and second order polynomial fit. It was found that the weighted average method provides the lowest MSE compared to the other two methods. After estimating a frame based parameters, an utterance based parameter vector was smoothed using Maximum Likelihood Trajectory smoothing method, which takes into account the dynamics of the features into account. This smoothing prooved to lower the MSE (0.61) to levels exceeding the state of the art method. [6] D. McIntyre, D. Pollard, R. Smith, and U. of Kansas.. Kansas Geological Survey., Computer programs for automatic contouring. University of Kansas, 1968. [7] K. Tokuda, T. Yoshimura, T. Masuko, T. Kobayashi, and T. Kitamura, Speech parameter generation algorithms for HMM-based speech synthesis, in IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS SPEECH AND SIGNAL PROCESSING, vol. 3. Citeseer, 2000. [8] S. Hiroya and M. Honda, Acoustic-to-articulatory inverse mapping using an HMM-based speech production model, in Seventh International Conference on Spoken Language Processing. ISCA, 2002. References [1] A. Wrench, The MOCHA-TIMIT articulatory database, Queen Margaret University College, Tech. Rep, 1999. [2] C. Atkeson, A. Moore, and S. Schaal, Locally weighted learning, Artificial intelligence review, vol. 11, no. 1, pp. 11 73, 1997. [3] I. Crain and B. Bhattacharyya, Treatment of nonequispaced two dimensional data with a digital computer, Geoexploration, vol. 5, pp. 173 194, 1967. [4] K. Falconer, A general purpose algorithm for contouring over scattered data points, NAC6, National Physical Laboratory, 1971. [5] D. McLain, Drawing contours from arbitrary data points, The Computer Journal, vol. 17, no. 4, p. 318, 1974.