Training LDCRF model on unsegmented sequences using Connectionist Temporal Classification

Similar documents
27: Hybrid Graphical Models and Neural Networks

Latent-Dynamic Discriminative Models for Continuous Gesture Recognition Louis-Philippe Morency, Ariadna Quattoni, and Trevor Darrell

Sequential Emotion Recognition using Latent-Dynamic Conditional Neural Fields

JOINT INTENT DETECTION AND SLOT FILLING USING CONVOLUTIONAL NEURAL NETWORKS. Puyang Xu, Ruhi Sarikaya. Microsoft Corporation

Sequence Classification with Neural Conditional Random Fields

A Comparison of Sequence-Trained Deep Neural Networks and Recurrent Neural Networks Optical Modeling For Handwriting Recognition

Social Interactions: A First-Person Perspective.

Gradient of the lower bound

Proceedings of the International MultiConference of Engineers and Computer Scientists 2018 Vol I IMECS 2018, March 14-16, 2018, Hong Kong

Segmentation-free Vehicle License Plate Recognition using ConvNet-RNN

Handwritten Gurumukhi Character Recognition by using Recurrent Neural Network

Deep Neural Networks for Recognizing Online Handwritten Mathematical Symbols

Unsupervised Feature Learning for Optical Character Recognition

Text Recognition in Videos using a Recurrent Connectionist Approach

Complexity-Aware Assignment of Latent Values in Discriminative Models for Accurate Gesture Recognition

END-TO-END CHINESE TEXT RECOGNITION

Conditional Random Fields and beyond D A N I E L K H A S H A B I C S U I U C,

arxiv: v1 [cs.ne] 6 Oct 2016

Conditional Random Fields as Recurrent Neural Networks

Machine Learning 13. week

Lecture 21 : A Hybrid: Deep Learning and Graphical Models

Speech Recognition with Quaternion Neural Networks

Deep Neural Networks Applications in Handwriting Recognition

Conditional Random Fields for Object Recognition

Long Short-Term Memory Recurrent Neural Network Architectures for Large Scale Acoustic Modeling

A Novel Approach to On-Line Handwriting Recognition Based on Bidirectional Long Short-Term Memory Networks

The Un-normalized Graph p-laplacian based Semi-supervised Learning Method and Speech Recognition Problem

Sign Language Recognition using Dynamic Time Warping and Hand Shape Distance Based on Histogram of Oriented Gradient Features

Structured Learning. Jun Zhu

Conditional Random Fields : Theory and Application

arxiv: v1 [cs.cl] 28 Nov 2016

Bidirectional Recurrent Convolutional Networks for Video Super-Resolution

Real-time Gesture Pattern Classification with IMU Data

Voice command module for Smart Home Automation

Deep Neural Networks Applications in Handwriting Recognition

arxiv: v1 [cs.ai] 14 May 2007

arxiv: v2 [cs.cv] 14 May 2018

Conditional Random Field for tracking user behavior based on his eye s movements 1

Action Recognition by Hierarchical Sequence Summarization

Empirical Evaluation of RNN Architectures on Sentence Classification Task

arxiv: v6 [stat.ml] 15 Jun 2015

Prediction of Pedestrian Trajectories Final Report

arxiv: v1 [cs.cv] 31 Mar 2016

Multi-Dimensional Recurrent Neural Networks

Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling

Lecture 13 Segmentation and Scene Understanding Chris Choy, Ph.D. candidate Stanford Vision and Learning Lab (SVL)

Motivation: Shortcomings of Hidden Markov Model. Ko, Youngjoong. Solution: Maximum Entropy Markov Model (MEMM)

Detecting Fraudulent Behavior Using Recurrent Neural Networks

A FRAMEWORK OF EXTRACTING MULTI-SCALE FEATURES USING MULTIPLE CONVOLUTIONAL NEURAL NETWORKS. Kuan-Chuan Peng and Tsuhan Chen

DEEP LEARNING REVIEW. Yann LeCun, Yoshua Bengio & Geoffrey Hinton Nature Presented by Divya Chitimalla

Clinical Name Entity Recognition using Conditional Random Field with Augmented Features

A Robust Dissimilarity-based Neural Network for Temporal Pattern Recognition

Show, Discriminate, and Tell: A Discriminatory Image Captioning Model with Deep Neural Networks

Learning Diagram Parts with Hidden Random Fields

Conditional Random Fields - A probabilistic graphical model. Yen-Chin Lee 指導老師 : 鮑興國

Energy-Based Learning

RNNs as Directed Graphical Models

Undirected Graphical Models. Raul Queiroz Feitosa

Offline Handwriting Recognition with Multidimensional Recurrent Neural Networks

LSTM and its variants for visual recognition. Xiaodan Liang Sun Yat-sen University

arxiv: v1 [cs.cv] 4 Feb 2018

Deep Learning in Visual Recognition. Thanks Da Zhang for the slides

Learning to Match. Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li

Rolling Bearing Diagnosis Based on CNN-LSTM and Various Condition Dataset

Improving Latent Fingerprint Matching Performance by Orientation Field Estimation using Localized Dictionaries

Lecture 2 Notes. Outline. Neural Networks. The Big Idea. Architecture. Instructors: Parth Shah, Riju Pahwa

Computation-Performance Optimization of Convolutional Neural Networks with Redundant Kernel Removal

Segmentation of Video Sequences using Spatial-temporal Conditional Random Fields

Sum-Product Networks. STAT946 Deep Learning Guest Lecture by Pascal Poupart University of Waterloo October 15, 2015

A Performance Evaluation of HMM and DTW for Gesture Recognition

SEMANTIC COMPUTING. Lecture 8: Introduction to Deep Learning. TU Dresden, 7 December Dagmar Gromann International Center For Computational Logic

Boosting Handwriting Text Recognition in Small Databases with Transfer Learning

Partial Least Squares Regression on Grassmannian Manifold for Emotion Recognition

Robust Face Recognition Based on Convolutional Neural Network

Slide credit from Hung-Yi Lee & Richard Socher

2 OVERVIEW OF RELATED WORK

Deep Boltzmann Machines

Image Captioning with Object Detection and Localization

Supervised Sequence Labelling with Recurrent Neural Networks. Alex Graves

MS-LSTM: a Multi-Scale LSTM Model for BGP Anomaly Detection

Comparisons of Sequence Labeling Algorithms and Extensions

CRFs for Image Classification

Bayesian model ensembling using meta-trained recurrent neural networks

Hidden Markov Models. Slides adapted from Joyce Ho, David Sontag, Geoffrey Hinton, Eric Xing, and Nicholas Ruozzi

Computer Vision Group Prof. Daniel Cremers. 4a. Inference in Graphical Models

Multimodal Gesture Recognition using Multi-stream Recurrent Neural Network

3 : Representation of Undirected GMs

Abstract. 1 Introduction. 2 Motivation. Information and Communication Engineering October 29th 2010

Semantic Segmentation. Zhongang Qi

Particle Filters for Visual Tracking

Temporal HeartNet: Towards Human-Level Automatic Analysis of Fetal Cardiac Screening Video

Semi-supervised Data Representation via Affinity Graph Learning

Structured Attention Networks

EMOTION DETECTION IN SPEECH USING DEEP NETWORKS. Mohamed R. Amer*, Behjat Siddiquie, Colleen Richey, and Ajay Divakaran

Located Hidden Random Fields: Learning Discriminative Parts for Object Detection

Dynamic Facial Expression Recognition Using A Bayesian Temporal Manifold Model

Code Mania Artificial Intelligence: a. Module - 1: Introduction to Artificial intelligence and Python:

Gesture Spotting and Recognition Using Salience Detection and Concatenated Hidden Markov Models

CSC 578 Neural Networks and Deep Learning

HiRF: Hierarchical Random Field for Collective Activity Recognition in Videos

Transcription:

Training LDCRF model on unsegmented sequences using Connectionist Temporal Classification 1 Amir Ahooye Atashin, 2 Kamaledin Ghiasi-Shirazi, 3 Ahad Harati Department of Computer Engineering Ferdowsi University of Mashhad Mashhad, Iran 1 amir.atashin@stu.um.ac.ir 2 k.ghiasi@um.ac.ir 3 a.harati@um.ac.ir Abstract. Many machine learning problems such as speech recognition, gesture recognition, and handwriting recognition are concerned with simultaneous segmentation and labeling of sequence data. Latent-dynamic conditional random field (LDCRF) is a well-known discriminative method that has been successfully used for this task. However, LDCRF can only be trained with pre-segmented data sequences in which the label of each frame is available apriori. In the realm of neural networks, the invention of connectionist temporal classification (CTC) made it possible to train recurrent neural networks on unsegmented sequences with great success. In this paper, we use CTC to train an LDCRF model on unsegmented sequences. Experimental results on two gesture recognition tasks show that the proposed method outperforms LDCRFs, hidden Markov models, and conditional random fields. Keywords: Latent-dynamic conditional random fields; connectionist temporal classification; unsegmented sequences; I. INTRODUCTION Labeling and segmenting of data sequences is a common problem in sequence classification tasks [2]. It arises in many real-world applications such as gesture recognition, speech recognition, and handwriting recognition. Probabilistic graphical models (PGMs) offer a powerful framework for machine learning in which the stochastic relations between several random variables is expressed economically by the use of graphical structures [3]. PGMs are divided into two categories: undirected graphical models or Markov random fields (MRFs) and directed graphical models or Bayesian networks (BNs). Conditional random fields (CRFs) [4], a variant of MRFs, and hidden Markov models (HMMs) [5], a variant of BNs, are two well-known graphical models which are widely used in sequence classification problems. HMM is a generative model which models the joint distribution of input sequences and their corresponding labels, but, CRF is a discriminative model which models the conditional distribution of label given the data sequence. Latent-dynamic conditional random fields (LDCRFs) [6] are a type of CRFs which are used for structured prediction problems. It was proposed to be utilized in the particular case of sequence classification problems in which a label must be assigned to each frame of an unsegmented sequence. Learning procedure for these models is discriminative and carried out on a frame-wise labeled training set. LDCRF model (Figure 1.b) can be described as a generalized form of chain-structured CRF model (Figure 1.a) in which hidden variables are incorporated into the CRF to learn Various structures of class labels. Hidden conditional random fields (HCRFs) [1] are another variant of CRF which includes hidden variables. By incorporating hidden variables, HCRFs can capture intrinsic relations between sub-structures of class labels. In addition, to capturing intrinsic relations, the LDCRF also using those latent variables for learning extrinsic dynamics between class labels. Recent works have shown that LDCRF model and its variants are the proper methods for modeling data sequences. They are mostly applied to the sequence classification problems in the field of human-computer interaction (HCI) such as sign language spotting and human action recognition problems [7-13]. Although LDCRF can be employed for segmenting and tagging of unsegmented sequences, it requires a set of presegmented sequences during training. However, for many realworld data sets in sequence classification tasks, the labeled segmented sequence is not available, and so the LDCRF model cannot be directly trained on them. Graves et all. [14] introduced a method called connectionist temporal classification (CTC) which can be used for training RNNs directly on unsegmented sequences. They used CTC as a layer on top of recurrent neural networks such as LSTM [15] and then applied it to the sequence classification problems such as online/offline unconstrained handwriting recognition. Their experiments showed that their method attained a significant improvement in classification performance over earlier methods [16-19]. In this paper, we use CTC to remove the need of presegmented sequences for training LDCRF and present a method for training LDCRFs directly on unsegmented sequences. The next section reviews some preliminary materials and introduces some definitions. Section III describes the motivations of the proposed method, its mathematical formulation, and its training algorithm. Section IV introduces the datasets used in our experiments for sequence classification. In Section V we experimentally compare the proposed CTC-LDCRF method

(a) In LDCRF, there is a sequence of hidden variables (i. e. h = h, h,, h ) for each sample which not observed in S. Each h is a member of a set H of all possible hidden states for the class label y. In the LDCRF each class label is restricted to have a disjoint set of hidden states from other classes. Also set H defined as a set of all possible hidden states and it is a union of all H (H = H ). Given the above definition, the LDCRF model with the parameter set θ is defined as latent conditional probability model, and for one particular training sample (x, y) it can be written as: P(y x; θ) = P(y h, x; θ)p(h x; θ) h H with LDCRF, CRF, and HMM models. Section VI gives a direction for future works and concludes this paper. II. (b) Figure 1. (a) Conditional random fields (b) Latent-dynamic conditional random fields. Gray nodes are indicative of observed variables. PRELIMINARIES In this section, we first briefly explain the temporal classification task in general, and then review the LDCRF and CTC methods. A. Temporal Classification In temporal classification, each input sequence is associated with a sequence of labels where the alignment between them is unknown [2, 20]. Let S = {(x, y ),, (x n, y )} be a set of training samples where each sample s = (x, y ) consists of an input data sequence x = x, x,, x with length t and its corresponding label sequence y = y, y,, y with length m which is less than or equal to the t (m t ). Each x is a d-dimensional feature vector and each y is a member of set Y which is the set consisting of all possible class labels. The objective of the temporal classification task is to learn the mapping between input data sequences and their corresponding label sequences. B. Latent-dynamic conditional random field (LDCRF) LDCRF is an undirected graphical model that has been developed for the particular case of sequence classification problems where the goal is to predict a label at each frame for a given data sequence (Figure 1.b). Let S be a training set for the LDCRF that consists of n samples. For each training sample in S, the class label sequence is as long as data sequence (i. e. t = u for i = 1 n). = P(h x; θ) h: H (1) where P(h x; θ) is defined by a Gibbs distribution: P(h x; θ) = exp( θ F (h; θ) ) Z(x; θ) and Z is the partition function which is defined by: Z(x; θ) = exp θ F (h; θ) h (2) Also F is a summation over value of f at each point of sequence and written as: x F (h, x) = f (h, h, x, i) (3) Which x is length of data sequence. Each feature function f can be either a state function or pairwise function, the state function related to a single hidden variable in the model while pairwise function related to a pair of hidden variables. C. Connectionist Temporal Classification Connectionist temporal classification (CTC) [14, 20], as its name suggests, is proposed for use in temporal classification tasks. It is used as output layer of RNNs for temporal classification problems. Considering a training set S (as explained in Section II.A), the objective function of CTC is defined as the sum of negative log-likelihood of samples s = (x i, y i ) in S: J = log P(y i x i ) (4) For each frame of an input data sequence x CTC receives the probabilities of all class labels from the unfolded RNN. Hence it receives a table of probabilities with size t Y as

input (where t is the length of the input sequence x i ) and efficiently computes log P(y i x i ) using dynamic programing. Correspondingly, it backproagates a table of errors to the bottom RNN. III. PROPOSED METHOD In the first section, we mentioned that one limitation of LDCRF model is that it can only be applied to sequence classification problems where training sequences are frame-wise labeled. CTC was introduced to train RNNs with unsegmented sequences. In this section, we present a method for training LDCRF directly with unsegmented sequences. In our model, called CTC- LDCRF, we use CTC on top of LDCRF (see Figure 2). In the next subsections, we describe our model, its formulation and learning algorithm. A. Model Overview We propose a discriminative model for unsegmented sequence learning tasks. As shown in Figure 2, it consists of two layers: LDCRF layer and CTC layer. At the first step, LDCRF takes a data sequence as input and computes the label probabilities for each frame, then CTC takes both the LDCRF outputs and the desired label sequence and computes the gradient of its objective function with respect to label probabilities of the output nodes of the LDCRF model. Like the well-known backpropagation algorithm, we use the gradients returned by CTC to update the parameters of the LDCRF model. Finally, the gradient descent technique is used to optimize our model. B. CTC-LDCRF learning algorithm We assume that the training set consists of n pairs (x 1, z 1 ),, (x n, z n ) of data sequences x i = x,, x along with their corresponding label sequences z i = z,, z with unknown alighnment (which is explained in the Section II.A). We use the following objective function to train our model with parameter set θ: J = log P(z i x i ; θ) (5) We use gradient descent for finding the optimal parameter values of the model. Let θ be the parameter that is associated with the state feature function f in the model. By chain rule, for a training sample i we have log P(z i x i ) = log P(z i x i ) θ P y = a x i P y = a x i θ Y (6) As we described in Section II.C, the first term in the inner summation in (6) is computed by CTC, so we only need to calculate the second term which is the marginal conditional probability of (1): Figure 2. Illustration of our proposed method. P y = a x i ; θ = P h = s x i ; θ = h: H H P(h x i ; θ) h P (h x i ; θ) The gradient of (7) with respect to θ can be written as (details omitted for space): P y = a x i θ = P h = b x i ; θ f (j, x i ) H where δ, is defined as: δ, = 1, combining these results, we have: δ, P h = b x i b = k 0, otherwise J J = θ P h = a x i f (j, x i ) H P h = a x i (7) (8) δ, P h = a x i (9) The marginal probabilities P h = a x in (9), can be efficiently computed for all hidden states of the model using

Figure 3. Illustration of the six arm gestures in the ArmGesture database [1]. Green arrows are indicative of the gesture trajectory for each class. belief propagation.the gradient of J with respect to the parameters of the pairwise potentials can be obtained in a similar manner. C. CTC-LDCRF inference For testing, given a new test sequence x, we want to estimate the most probable label sequence y using the formula: y = arg max P(y x, θ ) (10) y where θ is the set of parameter values that have been learned during the training phase. IV. EXPERIMENTS In this section, we describe the data sets and the methodology which were used in our experiments to evaluate our model performance. Avatar-Eye Dataset: This dataset contains sequences of eye gestures gathered from six individual contributors interacting with an avatar. The goal is to discriminate eye gaze aversion gestures from other eye gestures for each frame of unsegmented video. For each video, eye gaze estimated using two view-based appearance models explained in [21], and two-dimensional feature vector obtained from each frame of video. For each video sequence, frames were tagged either as eye gaze aversion or other eye gestures. ArmGesture-Continuous Dataset: This dataset includes unsegmented sequences of arm gestures which are created based on the original ArmGesture dataset. The original dataset contains 724 samples of six types of arm gestures shown in Figure 3, and sequences were collected from 13 contributors with an average of 120 samples per class, where each sample is a sequence of 20-dimensional feature vectors. Each sample of the Arm-Gesture Continuous is generated by serializing 3 to 5 randomly selected samples of six different gesture classes from the original dataset. This dataset contains 182 samples, with a mean length of 92 frames. A. Models We compare three configurations of our proposed method with LDCRF, CRF, and HMM on ArmGesture-Continues and Avatar-Eye data sets. Conditional Random Field: we trained a single chain CRF with its standard objective function and regularized term, we varied window size from 0 to 2. Latent-dynamic Conditional Random Field: we trained a single chain LDCRF model with the objective function described in [6] We are differing the number of hidden states per label from 2 to 6 and the window size from 0 to 2. Hidden Markov Model: we used an HMM for each class. Each HMM was trained with segmented subsequences where the frames of each subsequence all fitted to the corresponding class. This training set contained the same number of frames like the one used for training the CRF and LDCRF models excluding frames is group into subsequences according to their tag. The final model is build by combining HMMs together. CTC-LDCRF: we trained a single chain CTC-LDCRF model with three type of configuration. For the first configuration, we train model with unsegmented label sequence which never done before on these two data sets. For the second configuration, the model trained with framewised tagged sequences similar to LDCRF and CRF training data. For the third configuration, we use a two-step training procedure. At the first, we obtain pre-trained weights by training the model on segmented subsequences (similar to the HMM training set). At the next step, we used the pre-trained weights for initializing the model parameters and trained it on unsegmented data sequences. Within training, we differ the number of hidden states per label for the model from 2 to 6 and window size from 0 to 2. For all of these three configurations, we are adding one more class label into the model to use as non-gesture or blank class label which is required in the CTC layer. Figure 4. ROC curves from the Avatar-Eye experiment. Windows size set to one for all models.

TABLE I. B. Methodology For these two data sets, the experiments are carried out using the K-fold method for comparing models performance. In this method, the dataset is divided into K equal group and one of them chosen for the testing set while the other used for training and validation set. This procedure is repeated K times, and a mean of the K result reported for accuracy. We use K = 5 for the ArmGesture-Continues and K = 2 for the Avatar-Eye dataset. The measure for the model's performance is number of the correct predicted labels for each frame of test sequences: accurency = EXPERIMETS RESULT FOR ARMGESTURE-CONTINUES DATASET Model Configuration Accuracy CTC-LDCRF (w = 1, h = 3) CTC-LDCRF (w = 1, h = 3) CTC-LDCRF (w = 1, h = 2) LDCRF (w = 1, h = 3) Pre training + Train with unsegmented sequence #of correct labels 100 (11) #of sequence frames V. RESULTS AND DISCUSSION 93.9% Train with unsegmented sequence 89.5% Train with segmented sequence 92.0% Train with segmented sequence 92.2% CRF (w = 1) Train with segmented sequence 89.1% HMM - 91.2% We compare the result of the three configurations of our CTC-LDCRF learning algorithm on ArmGesture-Continues and Avatar-Eye data sets with standard LDCRF, Linear CRF, and HMM Models in our experiments. We use (11) to compare test results. The ROC curve for the test on Avatar-Eye dataset (Figure 4) shows that the CTC-LDCRF, which trains with unsegmented sequences, outperforms the other models. It can explain because of data sequences are frame-wise tagged manually by a human agent, and it may not be exactly correct especially on boundaries where label changed in sequence. Since we use unsegmented training sequence and don t use frame wised labels our method achieves better performance compared with other methods, also because of Avatar-Eye is a small dataset with 3-Dimentional feature vector and only 2 class label, the CTC-LDCRF method achieves good performance without needs to pre-trained. Table I. shows the experimental results on ArmGesture- Continues dataset. The third configuration of Our CTC-LDCRF training methods has higher accuracy compare to others experimented methods. VI. CONCLUSION In this paper, we develop a framework for training LDCRF model directly on the unsegmented sequence. We employed CTC approach in our framework to make the LDCRF model able to trains directly with unsegmented sequences. We did experiments on two gesture recognition data sets and showed that our model achieves better performance over other tested models, even though it is trained on unsegmented data sequence and gets less information than other models. For the future works we plan to extend our framework and use a neural network such as LSTM and Convolutional Neural Network (CNN) as a feature extractor and make it possible to apply our framework on more complex, and raw real-world datasets [22-24]. ACKNOWLEDGMENT We would like to thank Mostafa Rafiee, MSc student at the Ferdowsi University of Mashhad, for his noteworthy discussions and useful assist in implementations and experiments. REFERENCES [1] A. Quattoni, S. Wang, L.-P. Morency, M. Collins, and T. Darrell, "Hidden conditional random fields," IEEE Trans. Pattern Anal. Mach. Intell., vol. 29, pp. 1848-1852, 2007. [2] Z. Xing, J. Pei, and E. Keogh, "A brief survey on sequence classification," ACM SIGKDD Explorations Newsletter, vol. 12, pp. 40-48, 2010. [3] D. Koller and N. Friedman, Probabilistic graphical models: principles and techniques: MIT press, 2009. [4] J. Lafferty, A. McCallum, and F. C. Pereira, "Conditional random fields: Probabilistic models for segmenting and labeling sequence data," 2001. [5] L. R. Rabiner, "A tutorial on hidden Markov models and selected applications in speech recognition," Proceedings of the IEEE, vol. 77, pp. 257-286, 1989. [6] L.-P. Morency, A. Quattoni, and T. Darrell, "Latent-dynamic discriminative models for continuous gesture recognition," in 2007 IEEE conference on computer vision and pattern recognition, 2007, pp. 1-8. [7] L.-P. Morency and T. Darrell, "Conditional sequence model for context-based recognition of gaze aversion," in International Workshop on Machine Learning for Multimodal Interaction, 2007, pp. 11-23. [8] H. Ning, W. Xu, Y. Gong, and T. Huang, "Latent pose estimator for continuous action recognition," in European Conference on Computer Vision, 2008, pp. 419-433. [9] M. Elmezain, A. Al-Hamadi, and B. Michaelis, "Discriminative models-based hand gesture recognition," in Machine Vision, 2009. ICMV'09. Second International Conference on, 2009, pp. 123-127.

[10] Y. Song, L.-P. Morency, and R. Davis, "Multi-view latent variable discriminative models for action recognition," in Computer Vision and Pattern Recognition (CVPR), 2012 IEEE Conference on, 2012, pp. 2120-2127. [11] Y. Song, D. Demirdjian, and R. Davis, "Continuous body and hand gesture recognition for natural human-computer interaction," ACM Transactions on Interactive Intelligent Systems (TiiS), vol. 2, p. 5, 2012. [12] J.-C. Lévesque, L.-P. Morency, and C. Gagné, "Sequential Emotion Recognition using Latent-Dynamic Conditional Neural Fields," in Automatic Face and Gesture Recognition (FG), 2013 10th IEEE International Conference and Workshops on, 2013, pp. 1-6. [13] A. T. Schulz and R. Stiefelhagen, "Pedestrian intention recognition using latent-dynamic conditional random fields," in 2015 IEEE Intelligent Vehicles Symposium (IV), 2015, pp. 622-627. [14] A. Graves, S. Fernández, F. Gomez, and J. Schmidhuber, "Connectionist temporal classification: labelling unsegmented sequence data with recurrent neural networks," in Proceedings of the 23rd international conference on Machine learning, 2006, pp. 369-376. [15] S. Hochreiter and J. Schmidhuber, "Long short-term memory," Neural computation, vol. 9, pp. 1735-1780, 1997. [16] A. Graves, M. Liwicki, S. Fernández, R. Bertolami, H. Bunke, and J. Schmidhuber, "A novel connectionist system for unconstrained handwriting recognition," Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol. 31, pp. 855-868, 2009. [17] A. Graves and J. Schmidhuber, "Offline handwriting recognition with multidimensional recurrent neural networks," in Advances in neural information processing systems, 2009, pp. 545-552. [18] A. Graves, A.-r. Mohamed, and G. Hinton, "Speech recognition with deep recurrent neural networks," in Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on, 2013, pp. 6645-6649. [19] A. Graves and N. Jaitly, "Towards End-To-End Speech Recognition with Recurrent Neural Networks," in ICML, 2014, pp. 1764-1772. [20] A. Graves, "Supervised sequence labelling," in Supervised Sequence Labelling with Recurrent Neural Networks, ed: Springer, 2012, pp. 5-13. [21] L.-P. Morency, C. M. Christoudias, and T. Darrell, "Recognizing gaze aversion gestures in embodied conversational discourse," in Proceedings of the 8th international conference on Multimodal interfaces, 2006, pp. 287-294. [22] D. Wu, L. Pigou, P.-J. Kindermans, L. Nam, L. Shao, J. Dambre, et al., "Deep Dynamic Neural Networks for Multimodal Gesture Segmentation and Recognition," 2016. [23] C. Liu, J. Liu, Z. He, Y. Zhai, Q. Hu, and Y. Huang, "Convolutional neural random fields for action recognition," Pattern Recognition, 2016. [24] Z. Huang, W. Xu, and K. Yu, "Bidirectional LSTM-CRF models for sequence tagging," arxiv preprint arxiv:1508.01991, 2015.