Captured Motion Data Processing for Real Time Synthesis of Sign Language

Similar documents
Continuous Control of Lagrangian Data

HySCaS: Hybrid Stereoscopic Calibration Software

ASAP.V2 and ASAP.V3: Sequential optimization of an Algorithm Selector and a Scheduler

Setup of epiphytic assistance systems with SEPIA

Deformetrica: a software for statistical analysis of anatomical shapes

BoxPlot++ Zeina Azmeh, Fady Hamoui, Marianne Huchard. To cite this version: HAL Id: lirmm

Multi-atlas labeling with population-specific template and non-local patch-based label fusion

DANCer: Dynamic Attributed Network with Community Structure Generator

A Methodology for Improving Software Design Lifecycle in Embedded Control Systems

A Database Architecture For Real-Time Motion Retrieval

Service Reconfiguration in the DANAH Assistive System

Comparator: A Tool for Quantifying Behavioural Compatibility

Fuzzy sensor for the perception of colour

Moveability and Collision Analysis for Fully-Parallel Manipulators

Multimedia CTI Services for Telecommunication Systems

Tacked Link List - An Improved Linked List for Advance Resource Reservation

Fault-Tolerant Storage Servers for the Databases of Redundant Web Servers in a Computing Grid

Mokka, main guidelines and future

lambda-min Decoding Algorithm of Regular and Irregular LDPC Codes

How to simulate a volume-controlled flooding with mathematical morphology operators?

The Animation Loop Station: Near Real-Time Animation Production

Structuring the First Steps of Requirements Elicitation

X-Kaapi C programming interface

Robust IP and UDP-lite header recovery for packetized multimedia transmission

Natural Language Based User Interface for On-Demand Service Composition

Lossless and Lossy Minimal Redundancy Pyramidal Decomposition for Scalable Image Compression Technique

KeyGlasses : Semi-transparent keys to optimize text input on virtual keyboard

Real-Time Collision Detection for Dynamic Virtual Environments

Regularization parameter estimation for non-negative hyperspectral image deconvolution:supplementary material

Assisted Policy Management for SPARQL Endpoints Access Control

Scalewelis: a Scalable Query-based Faceted Search System on Top of SPARQL Endpoints

Relabeling nodes according to the structure of the graph

Reverse-engineering of UML 2.0 Sequence Diagrams from Execution Traces

QuickRanking: Fast Algorithm For Sorting And Ranking Data

Branch-and-price algorithms for the Bi-Objective Vehicle Routing Problem with Time Windows

RecordMe: A Smartphone Application for Experimental Collections of Large Amount of Data Respecting Volunteer s Privacy

Comparison of spatial indexes

FIT IoT-LAB: The Largest IoT Open Experimental Testbed

Open Digital Forms. Hiep Le, Thomas Rebele, Fabian Suchanek. HAL Id: hal

Study on Feebly Open Set with Respect to an Ideal Topological Spaces

An FCA Framework for Knowledge Discovery in SPARQL Query Answers

Motion-based obstacle detection and tracking for car driving assistance

Blind Browsing on Hand-Held Devices: Touching the Web... to Understand it Better

Catalogue of architectural patterns characterized by constraint components, Version 1.0

NP versus PSPACE. Frank Vega. To cite this version: HAL Id: hal

Implementing an Automatic Functional Test Pattern Generation for Mixed-Signal Boards in a Maintenance Context

Linux: Understanding Process-Level Power Consumption

The Proportional Colouring Problem: Optimizing Buffers in Radio Mesh Networks

The optimal routing of augmented cubes.

Linked data from your pocket: The Android RDFContentProvider

From medical imaging to numerical simulations

Traffic Grooming in Bidirectional WDM Ring Networks

Primitive roots of bi-periodic infinite pictures

Light field video dataset captured by a R8 Raytrix camera (with disparity maps)

Real-time tracking of multiple persons by Kalman filtering and face pursuit for multimedia applications

LaHC at CLEF 2015 SBS Lab

A Resource Discovery Algorithm in Mobile Grid Computing based on IP-paging Scheme

Malware models for network and service management

A Voronoi-Based Hybrid Meshing Method

Every 3-connected, essentially 11-connected line graph is hamiltonian

Real-time FEM based control of soft surgical robots

SDLS: a Matlab package for solving conic least-squares problems

SIM-Mee - Mobilizing your social network

MUTE: A Peer-to-Peer Web-based Real-time Collaborative Editor

Teaching Encapsulation and Modularity in Object-Oriented Languages with Access Graphs

The Connectivity Order of Links

Change Detection System for the Maintenance of Automated Testing

How to input text on smartwatches? When the screen is tiny, a 3-key keypad is better than a full keyboard

Quality of Service Enhancement by Using an Integer Bloom Filter Based Data Deduplication Mechanism in the Cloud Storage Environment

A N-dimensional Stochastic Control Algorithm for Electricity Asset Management on PC cluster and Blue Gene Supercomputer

Simulations of VANET Scenarios with OPNET and SUMO

Very Tight Coupling between LTE and WiFi: a Practical Analysis

UsiXML Extension for Awareness Support

A Practical Approach for 3D Model Indexing by combining Local and Global Invariants

GDS Resource Record: Generalization of the Delegation Signer Model

YAM++ : A multi-strategy based approach for Ontology matching task

BugMaps-Granger: A Tool for Causality Analysis between Source Code Metrics and Bugs

XML Document Classification using SVM

Privacy-preserving carpooling

Is GPU the future of Scientific Computing?

An Experimental Assessment of the 2D Visibility Complex

Complexity Comparison of Non-Binary LDPC Decoders

Randomized Nonlinear Component Analysis for Dimensionality Reduction of Hyperspectral Images

Formal modelling of ontologies within Event-B

Real-Time and Resilient Intrusion Detection: A Flow-Based Approach

Syrtis: New Perspectives for Semantic Web Adoption

Using a Medical Thesaurus to Predict Query Difficulty

Monitoring Air Quality in Korea s Metropolises on Ultra-High Resolution Wall-Sized Displays

Fuzzy interpolation and level 2 gradual rules

Leveraging ambient applications interactions with their environment to improve services selection relevancy

Graphe-Based Rules For XML Data Conversion to OWL Ontology

The New Territory of Lightweight Security in a Cloud Computing Environment

Acyclic Coloring of Graphs of Maximum Degree

A million pixels, a million polygons: which is heavier?

Representation of Finite Games as Network Congestion Games

Stream Ciphers: A Practical Solution for Efficient Homomorphic-Ciphertext Compression

A case-based reasoning approach for unknown class invoice processing

IntroClassJava: A Benchmark of 297 Small and Buggy Java Programs

Hierarchical Multi-Views Software Architecture

Workspace and joint space analysis of the 3-RPS parallel robot

Transcription:

Captured Motion Data Processing for Real Time Synthesis of Sign Language Alexis Héloir, Sylvie Gibet, Franck Multon, Nicolas Courty To cite this version: Alexis Héloir, Sylvie Gibet, Franck Multon, Nicolas Courty. Captured Motion Data Processing for Real Time Synthesis of Sign Language. Gibet S., Courty N., Kamp J.F. (Revised Selected Papers), May 2005, Vannes, France. Springer, pp.168-171, 2006, Lecture Notes in Computer Science (3881). <hal-00515185> HAL Id: hal-00515185 https://hal.archives-ouvertes.fr/hal-00515185 Submitted on 6 Sep 2010 HAL is a multi-disciplinary open access archive for the deposit and dissemination of scientific research documents, whether they are published or not. The documents may come from teaching and research institutions in France or abroad, or from public or private research centers. L archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d enseignement et de recherche français ou étrangers, des laboratoires publics ou privés.

Captured Motion Data Processing for Real Time Synthesis of Sign Language Alexis Heloir 1, Sylvie Gibet 1, Franck Multon 2, and Nicolas Courty 1 1 Université de Bretagne Sud, Laboratoire Valoria, Bâtiment Yves Coppens, BP 573, 56017 Vannes Cedex, France 2 Université de Rennes 2, Laboratoire de Physiologie et de Biomécanique de l Exercice Musculaire, Av. Charles Tillon CS 24414, 35044 Rennes, France 1 firstname.surname@univ-ubs.fr, 2 franck.multon@uhb.fr Abstract. The work described in this abstract presents a roadmap towards the creation and specification of a virtual humanoid capable of performing expressive gestures in real time. We present a gesture motion data acquisition protocol capable of handling the main articulators involved in human expressive gesture (whole body, fingers and face). We then present the postprocessing of captured data leading to a motion database complying with our motion specification language and capable of feeding data driven animation techniques. Issues. Embodying a virtual humanoid with expressive gestures raises many problems such as computation-cost efficiency, realism and level of expressiveness, or high level specification of expressive gesture [1]. In this abstract, we focus on the acquisition of motion capture data from the main articulators involved in communicative gesture (whole body, face mimics and fingers motion). We then present how acquired data are postprocessed in order to build a database compatible with high level gesture specification and capable of feeding real time data-driven motion synthesis techniques. A recent automatic segmentation algorithm based on Principal Component Analysis (PCA) is then evaluated. Motion acquisition protocol. The motion data acquisition protocol is designed to capture the whole range of articulators involved in order to produce human communicative gestures. This protocol relies on two complementary techniques, as shown in figure 1. The first technique aims at capturing facial and body motions and relies on a set of reflective markers placed on standardized anatomical landmarks and a network of 12 Vicon-MX 1 infrared cameras located all around the subject. The second technique aim to capturing finger motion thanks to pair of Cybergloves 2 measuring finger abduction and flexion. This technique is well suited to finger motion, as it is robust by fingers occlusions that may appear too often during signing performance. The two sets of data acquired are post processed, synchronized and merged offline. 1 http://www.vicon.com/products/viconmx.html 2 http://www.immersion.com/3d/products/cyberglove.php

Fig. 1. motion acquisition protocol The resulting dataset describes the evolution along n frames of a skeleton hierarchy composed by k joints. for each joint i, a set of l i degrees of freedom is defined, 1 l i 3. The size of a posture vector v is then easily computed. size(v) = Table 1 sums up the composition of a posture vector according to our representation model. k i=0 l i segment coord type joints number DOF per joint size of segment subvector body angular 18 1 l 3 54 hand angular 18 1 l 3 25 total angular 36 79 Table 1. detail of a posture vector Processing motion data. One of the early step of data processing consists of segmenting motion into relevant chunks. Extracted chunks must be short enough to guarantee the synthesis of a wide range of new motions while keeping sufficient meaning to comply with high level task oriented specification language [2]. Even though it has been shown that low level motion segmentation can be achieved in a straightforward manner [7][4], Hodgins and al. recently showed that higher level motion segmentation could be efficiently achieved thanks to principal component analysis (PCA) approach. According to the results they

presented [6], PCA segmentation method applied on simple motions representing typical human activities, such as walking, running, sitting, standing idle, etc. achieved very good results: up to 80% precision call for simple body movements. This algorithm is based on the assumption that the intrinsic dimensionality of a motion sequence containing a single behavior should be smaller than the intrinsic dimensionality of a motion sequence containing multiple behaviors. Thus, from one motion sequence to another, the reconstruction error of the frames projected onto the optimal hyperplane of dimension r increases rapidly, for a fixed r. figure 2 illustrates a motion transition detection between two hand configurations. Fig. 2. automatic segmentation using PCA. Cut is performed when derivative of error reaches three standard deviation from the average. Evaluating the PCA approach to hand motion segmentation. We apply the PCA based segmentation algorithm described to a sequence representing a non signer subject finger spelling French dactylologic alphabet [3]. The sequence is 7200 frames long with 120 frames per second. To perform PCA, Decomposition is thus performed on a 7200 25 matrix extracted from the total motion data and representing the right hand motion. According to our experiments, the ratio Er which indicates how much information is retained by projecting the frames on the optimal r-dimensional hyperplane reaches acceptable range [6] when r 3 for all the 20 segments we extracted manually from the alphabet spelling sequence. Further experiments lead us to set the window parameter k originally fixed at 2 seconds to 1.3 seconds, considering that hand motion is much faster than body motion. Fig. 3. segmentation results

Results. In parallel to automated PCA based hand motion segmentation, a human observer segmented manually the finger spelling sequence by identifying probables zones of motion transition. Figure 3 compares how the two methods assigned motion separation zones. The human observer identified 27 zones while PCA based motion segmentation algorithm identified 29 zones. Among those, 22 zones are overlapping. Conclusion. We have presented a motion acquisition framework designed to manage several articulators involved in communicative gesture and in sign language performance. We then rely on the data provided by this framework to evaluate a recent automatic motion segmentation technique based on principal component analysis of hand motion. This method proves to be capable of solving high level segmentation required by our needs. In the near future, we wish to extend this technique to the whole upper body motion. In parallel, we would like to provide a better evaluation framework based on data acquired and annotated by French sign language specialists. Such a framework will provide us with the ground required to perform reliable motion analysis and performance comparisons. Acknowledgements. This work is part of the RobEA HuGEx project and the Signe projet, respectively funded by the department STIC of CNRS and the regional council of Brittany (Ref. B/1042/2004/SIGNE). Motion capture presented in this document has been carried out in collaboration with the following research institutes: the LPBEM based at the University of Rennes 2, the LESP based at the University of Toulon and Var and the LINC based at the University of Paris 8. References 1. Gibet S., Courty N., Multon F., Pelachaud C. and Gorce P.: HUGEX: virtual humanoids embodied with expressive communicative gestures. Proc. of Journées Bilan ROBEA (2005), 233 238, Montpelier, France. 2. Gibet S., Lebourque T., Marteau P. F.: High level Specification and Animation of Communicative Gestures. Journal of Visual Languages and Computing, 12 (2001), 657 687. 3. Moody B., La langue des signes - Tome 1 : Histoire et grammaire. International Visual Theatre (I.V.T.). Éditions Ellipses. 4. Kovar L., Gleicher M. and Pighin F.: Motion Graphs. Proc. of Int. Conf. on Computer graphics and Interactive Techniques (2002), 473 482, San Ãntonio, USA. 5. Keogh F., Palpanas T., Zordan V., Gunopulos D. and Cardle M.: Indexing Large Human-Motion Databases. Proc. of Int. Conf. on Very Large Data Bases (2004), 780 791, Toronto, Canada. 6. Barbič J., Safonova A.,Pan JY, Faloutsos C, Hodgins J.C. and Pollard N.S.: Segmenting Motion Capture Data into Distinct Behaviors. Proc. of Graphics Interface (2004), 185 194, London, Ontario, Canada. 7. A. Fod, M. J. Mataric, and O. C. Jenkins.: Automated derivation of primitives for movement classification. Autonomous Robots, 12 (2002) 39 54,.