Skeleton Based Action Recognition with Convolutional Neural Network

Similar documents
HUMAN action recognition has received increasing attention

Combined Shape Analysis of Human Poses and Motion Units for Action Segmentation and Recognition

Person Identity Recognition on Motion Capture Data Using Label Propagation

Content-Based Image Recovery

Human Action Recognition Using Dynamic Time Warping and Voting Algorithm (1)

Deep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks

arxiv: v1 [cs.cv] 2 May 2017

Human-Object Interaction Recognition by Learning the distances between the Object and the Skeleton Joints

arxiv: v1 [cs.cv] 26 Jun 2017

Sign Language Recognition using Dynamic Time Warping and Hand Shape Distance Based on Histogram of Oriented Gradient Features

Learning to Match. Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li

Machine Learning 13. week

T. Seshagiri 1, S. Varadarajan 2

Temporal-Order Preserving Dynamic Quantization for Human Action Recognition from Multimodal Sensor Streams

Real-Time Continuous Action Detection and Recognition Using Depth Images and Inertial Signals

Skeleton-based Action Recognition Based on Deep Learning and Grassmannian Pyramids

A FRAMEWORK OF EXTRACTING MULTI-SCALE FEATURES USING MULTIPLE CONVOLUTIONAL NEURAL NETWORKS. Kuan-Chuan Peng and Tsuhan Chen

Pose Estimation on Depth Images with Convolutional Neural Network

TOWARDS SIGN LANGUAGE RECOGNITION BASED ON BODY PARTS RELATIONS

Scale Invariant Human Action Detection from Depth Cameras using Class Templates

Deep Learning for Computer Vision II

LEARNING A SPARSE DICTIONARY OF VIDEO STRUCTURE FOR ACTIVITY MODELING. Nandita M. Nayak, Amit K. Roy-Chowdhury. University of California, Riverside

Edge Enhanced Depth Motion Map for Dynamic Hand Gesture Recognition

Deep Learning in Visual Recognition. Thanks Da Zhang for the slides

Convolutional-Recursive Deep Learning for 3D Object Classification

EigenJoints-based Action Recognition Using Naïve-Bayes-Nearest-Neighbor

Bidirectional Recurrent Convolutional Networks for Video Super-Resolution

Rotation Invariance Neural Network

Two-Stream Convolutional Networks for Action Recognition in Videos

Human Action Recognition Using Temporal Hierarchical Pyramid of Depth Motion Map and KECA

1. INTRODUCTION ABSTRACT

Large-scale Continuous Gesture Recognition Using Convolutional Neural Networks

Online RGB-D Gesture Recognition with Extreme Learning Machines

Skeleton-based Gesture Recognition Using Several Fully Connected Layers with Path Signature Features and Temporal Transformer Module

Histogram of 3D Facets: A Characteristic Descriptor for Hand Gesture Recognition

CovP3DJ: Skeleton-parts-based-covariance Descriptor for Human Action Recognition

Chapter 2 Learning Actionlet Ensemble for 3D Human Action Recognition

Classification of objects from Video Data (Group 30)

A Deep Learning Framework for Authorship Classification of Paintings

arxiv: v1 [cs.cv] 1 Feb 2016

An efficient face recognition algorithm based on multi-kernel regularization learning

Arm-hand Action Recognition Based on 3D Skeleton Joints Ling RUI 1, Shi-wei MA 1,a, *, Jia-rui WEN 1 and Li-na LIU 1,2

Human Action Recognition Using a Temporal Hierarchy of Covariance Descriptors on 3D Joint Locations

Weighted Convolutional Neural Network. Ensemble.

Know your data - many types of networks

HUMAN action recognition is a fast developing research

Edge Detection Using Convolutional Neural Network

TRANSPARENT OBJECT DETECTION USING REGIONS WITH CONVOLUTIONAL NEURAL NETWORK

Feature Weighting in Dynamic Time Warping for Gesture Recognition in Depth Data

Channel Locality Block: A Variant of Squeeze-and-Excitation

Cost-alleviative Learning for Deep Convolutional Neural Network-based Facial Part Labeling

Convolutional Neural Network for Facial Expression Recognition

P-CNN: Pose-based CNN Features for Action Recognition. Iman Rezazadeh

Skeleton-based Activity Recognition with Local Order Preserving Match of Linear Patches

LARGE-SCALE PERSON RE-IDENTIFICATION AS RETRIEVAL

Image Question Answering using Convolutional Neural Network with Dynamic Parameter Prediction

Action Recognition from Motion Capture Data using Meta-cognitive RBF Network Classifier

Using Capsule Networks. for Image and Speech Recognition Problems. Yan Xiong

Text Recognition in Videos using a Recurrent Connectionist Approach

Activity recognition with echo state networks using 3D body joints and objects category

Deep Learning For Video Classification. Presented by Natalie Carlebach & Gil Sharon

ACTION BASED FEATURES OF HUMAN ACTIVITY RECOGNITION SYSTEM

Human Body Recognition and Tracking: How the Kinect Works. Kinect RGB-D Camera. What the Kinect Does. How Kinect Works: Overview

I. INTRODUCTION ISSN (PRINT): , (ONLINE): , VOLUME-4, ISSUE-11,

Adaptive Action Detection

Extend the shallow part of Single Shot MultiBox Detector via Convolutional Neural Network

Traffic Signs Recognition using HP and HOG Descriptors Combined to MLP and SVM Classifiers

Human Upper Body Pose Estimation in Static Images

Linear-time Online Action Detection From 3D Skeletal Data Using Bags of Gesturelets

AUTOMATIC 3D HUMAN ACTION RECOGNITION Ajmal Mian Associate Professor Computer Science & Software Engineering

Real-Time Human Pose Recognition in Parts from Single Depth Images

arxiv: v2 [cs.cv] 31 May 2018

Does the Brain do Inverse Graphics?

Human Detection and Tracking for Video Surveillance: A Cognitive Science Approach

Deep Learning. Visualizing and Understanding Convolutional Networks. Christopher Funk. Pennsylvania State University.

A Novel Image Super-resolution Reconstruction Algorithm based on Modified Sparse Representation

DMM-Pyramid Based Deep Architectures for Action Recognition with Depth Cameras

Graph Matching Iris Image Blocks with Local Binary Pattern

Short Survey on Static Hand Gesture Recognition

Non-rigid body Object Tracking using Fuzzy Neural System based on Multiple ROIs and Adaptive Motion Frame Method

Large-scale gesture recognition based on Multimodal data with C3D and TSN

LOCALIZATION OF HUMANS IN IMAGES USING CONVOLUTIONAL NETWORKS

Tri-modal Human Body Segmentation

RSRN: Rich Side-output Residual Network for Medial Axis Detection

REJECTION-BASED CLASSIFICATION FOR ACTION RECOGNITION USING A SPATIO-TEMPORAL DICTIONARY. Stefen Chan Wai Tim, Michele Rombaut, Denis Pellerin

R3DG Features: Relative 3D Geometry-based Skeletal Representations for Human Action Recognition

Background subtraction in people detection framework for RGB-D cameras

Evaluation of Triple-Stream Convolutional Networks for Action Recognition

Shape Preserving RGB-D Depth Map Restoration

Action Recognition with HOG-OF Features

arxiv: v2 [cs.cv] 14 May 2018

Human Activities Recognition Based on Skeleton Information via Sparse Representation

Multilayer and Multimodal Fusion of Deep Neural Networks for Video Classification

Supplementary Materials for Salient Object Detection: A

Multi-Glance Attention Models For Image Classification

Generic Face Alignment Using an Improved Active Shape Model

Human Pose Estimation with Deep Learning. Wei Yang

Classification of RGB-D and Motion Capture Sequences Using Extreme Learning Machine

Intensity-Depth Face Alignment Using Cascade Shape Regression

arxiv: v1 [cs.cv] 22 Feb 2017

Transcription:

B G R Spatial Structure Skeleton Based Action Recognition with Convolutional Neural Network Yong Du,, Yun Fu,,, Liang Wang Center for Research on Intelligent Perception and Computing, CRIPAC Center for Excellence in Brain Science and Intelligence Technology, CEBSIT Nat l Lab of Pattern Recognition, Institute of Automation, Chinese Academy of Sciences College of Engineering, College of Computer and Information Science, Northeastern University, USA {yong.du, wangliang}@nlpr.ia.ac.cn, yunfu@ece.neu.edu Abstract Temporal dynamics of postures over time is crucial for sequence-based action recognition. Human actions can be represented by the corresponding motions of articulated skeleton. Most of the existing approaches for skeleton based action recognition model the spatial-temporal evolution of actions based on hand-crafted features. As a kind of hierarchically adaptive filter banks, Convolutional Neural Network (CNN) performs well in representation learning. In this paper, we propose an end-to-end hierarchical architecture for skeleton based action recognition with CNN. Firstly, we represent a skeleton sequence as a matrix by concatenating the joint coordinates in each instant and arranging those vector representations in a chronological order. Then the matrix is quantified into an image and normalized to handle the variable-length problem. The final image is fed into a CNN model for feature extraction and recognition. For the specific structure of such images, the simple max-pooling plays an important role on spatial feature selection as well as temporal frequency adjustment, which can obtain more discriminative joint information for different actions and meanwhile address the variable-frequency problem. Experimental results demonstrate that our method achieves the state-of-art performance with high computational efficiency, especially surpassing the existing result by more than 15 percentage on the challenging ChaLearn gesture recognition dataset. 1. Introduction Action recognition plays an important role in computer vision and has a wide range of applications, e.g., humancomputer interaction, video surveillance, robotics, game control, and so on [1, 24]. Generally, human body can be regarded as an articulated system with rigid bones and hinged joints, and human actions can be represented as the mox11 x1n y11 y1n z11 z1n x21 x2n y21 y2n z21 z2n x31 x3n y31 y3n z31 z3n x(t-1)1 x(t-1)n y(t-1)1 y(t-1)n z(t-1)1 z(t-1)n xt1 xtn yt1 ytn zt1 ztn Temporal Dynamics Representation Features Filter bank Filter bank Synchronous Pooling Figure 1: An illustrative sketch of the proposed method. Three components of all skeleton joints in each frame are separately concatenated by their physical connections. After arranging the representations of all frames in chronological order, the generated matrix is quantified and normalized into an image, which is fed into the hierarchical spatial-temporal adaptive filter banks model for representation learning and recognition. tions of skeleton [19]. Currently, the cost-effective depth sensor combining with real-time skeleton estimation algorithms [15, 16] can provide relatively reliable joint coordinates. Based on those coordinates, effective and efficient approaches for action recognition have been developed recently [5, 19, 26]. Temporal dynamics of postures over time can be modeled as a time series problem, which is crucial for sequencebased action recognition [4, 8, 9]. As a kind of low-level feature, skeleton joint coordinates can be used to represent human postures and their temporal evolution. Most of the existing skeleton based action recognition approaches model actions based on well-designed hand-crafted local features. Simultaneously, Temporal Pyramids (TPs) and its variants are often employed to capture the local temporal evolution [10, 19, 20]. A following dictionary learning model is employed to generate the representation for the whole sequence [20]. For the restriction from the width of time windows, the TPs methods can only utilize limited contextual information. Moreover, temporal dynamics of 1

Jumping in place Jumping jacks Bending (hands up all the way down) Punching (boxing) Waving two hands Waving right hand Clapping hands Throwing a ball Sit down then stand up Sit down Stand up Figure 2: Image representations obtained on the Berkeley MHAD dataset [12]. sequences is ignored in the global representation based on a learned dictionary. Alternatively, some time series models, mainly Dynamic Time Warpings (DTWs) [19] and Hidden Markov Models (HMMs) [11, 21, 23], are applied to model the global temporal evolution, yet it is very difficult to obtain the temporal aligned sequences and the emission distributions of HMMs. Recently, an end-to-end approach based on Recurrent Neural Network (RNN) was proposed for skeleton based action recognition [1]. Almost all the existing methods for skeleton based action recognition concern that how to utilize the contextual information to capture temporal dynamics in sequences. HMMs and RNNs based approaches directly model the temporal dynamics as a time series problem. However, the temporal dynamics of sequences can be easily transformed as the spatial structure characteristics in images. As an architecture with hierarchical adaptive 2D filter banks, Convolutional Neural Network (CNN) has the advantage of encoding structural information and can be used for representation learning in many tasks. In this paper, we propose a CNN model for skeleton based action recognition. Firstly, we represent each skeleton sequence as a special image, in which the temporal dynamics of the sequence is encoded as changes in rows and the spatial structure of each frame is represented as columns (Fig. 1). A following CNN model is employed for feature learning and recognition. To represent the skeleton structure compactly, we employ the encoding style in [1] that human skeleton is divided into five parts and joints in each part are concatenated according to their physical connections. And three components (x, y, z) of each joint are represented as the corresponding three components (R, G, B) of each pixel. Considering that the length of skeleton sequences are variable, we resize the generated images to an uniform size yet the frequency of actions is changed with different scales. So spatial-temporal synchronous pooling is used to overcome this variable-frequency problem. We evaluate our approach on two benchmark datasets and obtain excellent performance. The main contributions of our work can be summarized as follows. Firstly, we propose an idea to represent action sequences as images as well as preserving the original temporal dynamics and spatial structure information. Secondly, based on such representation, we propose a CNN model for skeleton based action recognition, in which the sample 2D max-pooling plays the role of spatial-temporal synchronous pooling and can overcome the different sequence length and variable-frequency problems. Finally, we demonstrate that our proposed end-to-end model can rapidly handle skeleton based action recognition very well without any sophisticated processing. And this idea can be easily transfered to other time series problems. The remainder of this paper is organized as follows. In Section 2, we introduce our proposed model in detail. Then we provide our experimental results in Section 3. Finally, we conclude the paper in Section 4. 2. Our Method In this section, we first detail how to represent a sequence as an image, and then we introduce our hierarchical model for skeleton based action recognition. Finally, more details about training and testing are provided. 2.1. From Skeleton Sequences to Images How to transform a sequence to an image while preserving its spatial-temporal information is very important. Inspired by [1], all human skeleton joints in each frame are divided into five main parts according to human physical structure, i.e., two arms, two legs and a trunk. For preserving the local motion characteristics, joints in each part are concatenated as a vector by their physical connections, e.g., each arm can be represented as [hand, wrist, elbow, shoulder]. Then the five parts are concatenated as the representation of each frame. Projections on three orthogonal planes are represented separately and treated as the three components of RGB images, i.e., R i = [x i1, x i2,..., x in ], G i = [y i1, y i2,..., y in ], B i = [z i1, z i2,..., z in ], where i denotes the frame index and N indicates the number of frames in a sequence. Finally, representations of all frames are arranged in chronological order to represent the whole sequence (Fig. 2). In this case, spatial distribution of motion characteristics of each part is very clear, and the global discrimination is pretty obvious. Given the variable-length problem, the arranged float matrix is quantified to integral image representation, i.e., ( p c ) min p = floor 255 c max c min (1)

Spatial Structure Adaptive Filter Banks Adaptive Filter Banks Adaptive Filter Banks Adaptive Filter Banks 32 filters 32 filters 64 filters 64 filters Forward Neural Network Temporal Dynamics Figure 3: The framework of our model based on hierarchical spatial-temporal adaptive filter banks. then the image is resized to an uniform size 60 60. The p indicates the pixel value and the floor function means rounding down. The c max and c min are the maximum and minimum of all joint coordinates in the training set, respectively. 2.2. Hierarchical Architecture for Skeleton Based Action Recognition Recognizing actions depends on not only human pose in each time but also its temporal dynamics [1, 4, 8]. Generally, an adaptive filter bank can learn the filter coefficients adaptively and separate the input signal into multiple components. And those components can be analyzed in different domains corresponding to the diverse filter characteristics. Given that the essence of CNN is composed of a range of adaptive 2D filter banks, different from traditional HMMs or other time series models based approaches directly modelling action recognition as a time series problem, we propose a CNN based model to capture the spatialtemporal information in the sequences after transforming them to special image representations. Our proposed model is shown in Fig. 3, which includes 4 cascaded adaptive filter banks. All filter sizes are 3 3 and all strides during convolution are set to 1 for treating CNN as hierarchical adaptive filter banks. Considering that the original action frequencies are changed in different scales when resizing and same actions performed by different subjects may have various frequencies, we adopt the max-pooling strategy following each of the first three filter banks. For the special structure of input images (Fig. 2), the scale-invariance of max-pooling along horizontal axis is transformed as the frequency-invariance of actions. And max-pooling along vertical axis can select more discriminative skeleton joints for different actions. After feature extraction, a feed-forward neural network with two fullyconnected layers is employed for classification. The first fully-connected layer contains 128 neurons, and the number of neurons in the second one is equal to that of actions. ReLU neuron is adopted in all layers. The loss function is as follows [1]: M 1 L(X) = m=0 δ(k r)p(c k x m ) (2) C 1 ln k=0 where M indicates the number of samples in training set X and C denotes the number of categories. δ( ) is the Kronecker function, and p(c k x m ) means the probability that sample x m belongs to action C k. 2.3. Training and Testing Our model is trained by the back propagation algorithm. In order to improve the convergence speed, mean removal is employed for all input images. During training, we randomly select a patch from each original image with size 52 52. Given that temporal dynamics of a skeleton sequence is always embodied in two directions, we randomly flip the training images horizontally to utilize the forward and backward temporal dynamics. During testing, five patches selected form the four corners and the center and their horizontal flips are used. The final recognition is obtained by voting. 3. Experiments In this section, we evaluate our model compared with several recent works on two benchmark datasets: Berkeley Multimodal Human Action Dataset (Berkeley MHAD) [12] and ChaLearn gesture recognition dataset [2]. We also discuss the computational efficiency of our model. 3.1. Evaluation Dataset Berkeley MHAD [12]: It is generated by a multimodal acquisition system and an optical motion capture system is employed to capture the 3D position of skeleton joints with the frequency of 480Hz. There are 659 valid samples in this dataset, which consists of 11 actions performed by 12 subjects with 5 repetitions of each action. And each frame in a

sequence contains 35 joints accurately extracted according to the 3D marker trajectory. ChaLearn Gesture Recognition Dataset [2]: It is the ChaLearn 2013 Multi-model gesture dataset, which contains 23 hours of Kinect data with 27 persons performing 20 Italian gestures. This dataset provides RGB, depth, foreground segmentation and Kinect skeletons. This dataset is split into training, validation and testing sets, and contains total 955 videos, each of which lasts 1-2 minutes and involves 8-20 non-continuous gestures. 3.2. Experimental Results and Analysis Berkeley MHAD: We follow the experimental protocol proposed in [12]. The 384 sequences of the first 7 subjects are used for training while the 275 sequences of the last 5 subjects are used for testing. We compare our proposed approach with Ofli et al. [13], Vantigodi et al. [17], Vantigodi et al. [18], Kapsouras et al. [6] and Du et al. [1]. All the comparative methods on this dataset are directly from their corresponding papers, and the rest likewise. The experimental results are shown in Tab. 1. We can see that our method can achieve the 100% accuracy without any other pre- or post-processing. In contrast to those handcrafted features based approaches [6, 13, 17, 18], our approach, like [1], is an effective end-to-end solution to skeleton based action recognition. Table 1: Experimental results on the Berkeley MHAD [12]. Method Accuracy (%) Ofli et al., 2014 [13] 95.37 Vantigodi et al., 2013 [17] 96.06 Vantigodi et al., 2014 [18] 97.58 Kapsouras et al., 2014 [6] 98.18 Du et al., 2015 [1] 100 Ours 100 ChaLearn Gesture Recognition Dataset: In this more challenging dataset, the ground segments are provided and contain 6850 training samples with 39 frames average length, 3454 validation samples and 3579 test samples. We follow the experimental protocol adopted in [3, 14, 22, 25] and provide precision, recall and F1-score measures on the validation set. We compare our model with Yao et al. [25], Wu et al. [22], Pfister et al. [14], and Fernando et al. [3]. The experimental results are shown in Tab. 2. It is clear that our method significantly surpass the state-of-the-art precision by more than 15 percentage, which demonstrate that it is a great success to transform temporal dynamics in sequences into spatial structure information in images for sequence representation learning. One of the possible reasons for the excellent performance may be that our model can well handle the global temporal dynamics in sequences than the comparative methods. And comparing with those (a) Filters Epoch (a) Convergence Curves Figure 4: Filters and convergence curves on the ChaLearn gesture recognition dataset. recently proposed approaches, our method is simple and straightforward for skeleton based action recognition. Filters and the convergence curves are shown in Fig. 4. Table 2: Experimental results on the ChaLearn gesture recognition dataset [2]. Method Precision Recall F1-score Yao et al., 2014 [25] - - 56.0 Wu et al., 2013 [22] 59.9 59.3 59.6 Pfister et al., 2014 [14] 61.2 62.3 61.7 Fernando et al., 2015 [3] 75.3 75.1 75.2 Ours 91.16 91.25 91.21 3.3. Computational Efficiency Analysis We take the ChaLearn gesture recognition dataset for an example to illustrate the efficiency of our proposed model implemented based on ConvNet [7]. With the implementation on NVIDIA Titan GK110, we spend about 1.95ms per sequence in training and about 2.27ms per sequence (select 5 patches and flip for voting) in testing. 4. Conclusion and Future Work In this paper, we have proposed a simple end-to-end but high-efficiency and high-precision framework for skeleton based action recognition. We first represented human skeleton sequences as images to transform the temporal dynamics of sequences into the spatial structure information in images. Then a hierarchical architecture based on CNN was proposed for feature representation learning and classification. Experimental results on two publicly available datasets demonstrated the excellent performance of the proposed model. Current model classifies actions based on the global spatial and temporal information in skeleton sequences, which requires that the noise distribution in different segments of the same sequence are consistent. That means if data error of local fragments in the input sequences is particularly highlighted, the recognition rate may be cut down. In the future, we will consider the local features as an assistance to overcome this problem.

Acknowledgement This work is jointly supported by National Natural Science Foundation of China (61420106015, 61175003) and National Basic Research Program of China (2012CB316300). Yun Fu is supported by National Science Foundation (NSF) CNS award 1314484. References [1] Y. Du, W. Wang, and L. Wang. Hierarchical recurrent neural network for skeleton based action recognition. In IEEE Conference on Computer Vision and Pattern Recognition, pages 1110 1118, 2015. 1, 2, 3, 4 [2] S. Escalera, J. Gonzàlez, X. Baró, M. Reyes, O. Lopes, I. Guyon, V. Athitsos, and H. Escalante. Multi-modal gesture recognition challenge 2013: Dataset and results. In ACM on International Conference on Multimodal Interaction, pages 445 452. ACM, 2013. 3, 4 [3] B. Fernando, E. Gavves, J. M. Oramas, A. Ghodrati, and T. Tuytelaars. Modeling video evolution for action recognition. In IEEE Conference on Computer Vision and Pattern Recognition, pages 5378 5387, 2015. 4 [4] D. Gong, G. Medioni, and X. Zhao. Structured time series analysis for human action segmentation and recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, VOL. 36, NO. 7, 2014. 1, 3 [5] C. Jia, G. Zhang, and Y. Fu. Low-rank tensor learning with discriminant analysis for action classification and image recovery. In AAAI Conference on Artificial Intelligence, pages 1228 1234, 2014. 1 [6] I. Kapsouras and N. Nikolaidis. Action recognition on motion capture data using a dynemes and forward differences representation. Journal of Visual Communication and Image Representation, 25(6):1432 1445, 2014. 4 [7] A. Krizhevsky, I. Sutskever, and G. E. Hinton. Imagenet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems, pages 1097 1105, 2012. 4 [8] K. Li and Y. Fu. Prediction of human activity by discovering temporal sequence patterns. IEEE Transactions on Pattern Analysis and Machine Intelligence, VOL. 36, NO. 8, 2014. 1, 3 [9] K. Li, J. Hu, and Y. Fu. Modeling complex temporal composition of actionlets for activity prediction. In European Conference on Computer Vision, pages 286 299, 2012. 1 [10] J. Luo, W. Wang, and H. Qi. Group sparsity and geometry constrained dictionary learning for action recognition from depth maps. In IEEE International Conference on Computer Vision, pages 1809 1816. IEEE, 2013. 1 [11] F. Lv and R. Nevatia. Recognition and segmentation of 3- d human action using hmm and multi-class adaboost. In European Conference on Computer Vision, pages 359 372. Springer, 2006. 2 [12] F. Ofli, R. Chaudhry, G. Kurillo, R. Vidal, and R. Bajcsy. Berkeley mhad: A comprehensive multimodal human action database. In IEEE Workshop on Applications of Computer Vision, pages 53 60. IEEE, 2013. 2, 3, 4 [13] F. Ofli, R. Chaudhry, G. Kurillo, R. Vidal, and R. Bajcsy. Sequence of the most informative joints (smij): A new representation for human skeletal action recognition. Journal of Visual Communication and Image Representation, 25(1):24 38, 2014. 4 [14] T. Pfister, J. Charles, and A. Zisserman. Domain-adaptive discriminative one-shot learning of gestures. In European Conference on Computer Vision, pages 814 829. Springer, 2014. 4 [15] J. Shotton, T. Sharp, A. Kipman, A. Fitzgibbon, M. Finocchio, A. Blake, M. Cook, and R. Moore. Real-time human pose recognition in parts from single depth images. Communications of the ACM, 56(1):116 124, 2013. 1 [16] J. Tompson, A. Jain, Y. LeCun, and C. Bregler. Joint training of a convolutional network and a graphical model for human pose estimation. In Advances in Neural Information Processing Systems, pages 1799 1807, 2014. 1 [17] S. Vantigodi and R. V. Babu. Real-time human action recognition from motion capture data. In National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics, pages 1 4. IEEE, 2013. 4 [18] S. Vantigodi and V. B. Radhakrishnan. Action recognition from motion capture data using meta-cognitive rbf network classifier. In International Conference on Intelligent Sensors, Sensor Networks and Information Processing, pages 1 6. IEEE, 2014. 4 [19] R. Vemulapalli, F. Arrate, and R. Chellappa. Human action recognition by representing 3d skeletons as points in a lie group. In IEEE Conference on Computer Vision and Pattern Recognition, pages 588 595. IEEE, 2014. 1, 2 [20] J. Wang, Z. Liu, Y. Wu, and J. Yuan. Mining actionlet ensemble for action recognition with depth cameras. In IEEE Conference on Computer Vision and Pattern Recognition, pages 1290 1297. IEEE, 2012. 1 [21] D. Wu and L. Shao. Leveraging hierarchical parametric networks for skeletal joints based action segmentation and recognition. In IEEE International Conference on Computer Vision, 2014. 2 [22] J. Wu, J. Cheng, C. Zhao, and H. Lu. Fusing multimodal features for gesture recognition. In ACM on International Conference on Multimodal Interaction, pages 453 460. ACM, 2013. 4 [23] L. Xia, C.-C. Chen, and J. Aggarwal. View invariant human action recognition using histograms of 3d joints. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, pages 20 27. IEEE, 2012. 2 [24] X. Yang and Y. Tian. Super normal vector for activity recognition using depth sequences. In IEEE Conference on Computer Vision and Pattern Recognition, pages 804 811. IEEE, 2014. 1 [25] A. Yao, L. Van Gool, and P. Kohli. Gesture recognition portfolios for personalization. In IEEE Conference on Computer Vision and Pattern Recognition, pages 1923 1930. IEEE, 2014. 4 [26] X. Zhao, Y. Liu, and Y. Fu. Exploring discriminative pose sub-patterns for effective action classification. In ACM International Conference on Multimedia, pages 273 282, 2013. 1