Modeling Facial Expressions in 3D Avatars from 2D Images

Similar documents
High-Fidelity Facial and Speech Animation for VR HMDs

Animation of 3D surfaces.

Computer Animation Visualization. Lecture 5. Facial animation

DeepSketch2Face: A Deep Learning Based Sketching System for 3D Face and Caricature Modeling. Part I: User Studies on the Interface

CS 231. Deformation simulation (and faces)

Character animation Christian Miller CS Fall 2011

Advanced Graphics and Animation

Reading. Animation principles. Required:

Advances in Face Recognition Research

CS 231. Deformation simulation (and faces)

Animation of 3D surfaces

Self-supervised Multi-level Face Model Learning for Monocular Reconstruction at over 250 Hz Supplemental Material

Transfer Facial Expressions with Identical Topology

An Introduction to 3D Computer Graphics, Stereoscopic Image, and Animation in OpenGL and C/C++ Fore June

Personal style & NMF-based Exaggerative Expressions of Face. Seongah Chin, Chung-yeon Lee, Jaedong Lee Multimedia Department of Sungkyul University

Face analysis : identity vs. expressions

SYNTHETIC DATA GENERATION FOR AN ALL-IN-ONE DMS. Sagar Bhokre

Human body animation. Computer Animation. Human Body Animation. Skeletal Animation

Online Interactive 4D Character Animation

Animation & AR Modeling Guide. version 3.0

Muscle Based facial Modeling. Wei Xu

Physical based Rigging

Accurate 3D Face and Body Modeling from a Single Fixed Kinect

Real-Time Cutscenes Hand Keyed Animations Joint Only Face Rigs

Facial Motion Capture Editing by Automated Orthogonal Blendshape Construction and Weight Propagation

Deformation Transfer for Triangle Meshes

Kinematics & Motion Capture

Face Tracking. Synonyms. Definition. Main Body Text. Amit K. Roy-Chowdhury and Yilei Xu. Facial Motion Estimation

CrazyTalk 5 Basic Face Fitting and Techniques

Human pose estimation using Active Shape Models

FACIAL ANIMATION FROM SEVERAL IMAGES

Synthesizing Realistic Facial Expressions from Photographs

Animation COM3404. Richard Everson. School of Engineering, Computer Science and Mathematics University of Exeter

LBP Based Facial Expression Recognition Using k-nn Classifier

Selfie Video Stabilization

Tracking facial features using low resolution and low fps cameras under variable light conditions

Robust facial landmark detection and tracking across poses and expressions for in-the-wild monocular video

Unconstrained Realtime Facial Performance Capture

GiD v12 news. GiD Developer Team: Miguel Pasenau, Enrique Escolano, Jorge Suit Pérez, Abel Coll, Adrià Melendo and Anna Monros

FACIAL EXPRESSION USING 3D ANIMATION

Pose Space Deformation A unified Approach to Shape Interpolation and Skeleton-Driven Deformation

An Associate-Predict Model for Face Recognition FIPA Seminar WS 2011/2012

Input Nodes. Surface Input. Surface Input Nodal Motion Nodal Displacement Instance Generator Light Flocking

Facial Animation. Joakim Königsson

FACIAL MOVEMENT BASED PERSON AUTHENTICATION

Animations. Hakan Bilen University of Edinburgh. Computer Graphics Fall Some slides are courtesy of Steve Marschner and Kavita Bala

FACS Based Generating of Facial Expressions

Body Trunk Shape Estimation from Silhouettes by Using Homologous Human Body Model

Facial Processing Projects at the Intelligent Systems Lab

The accuracy and robustness of motion

Modelling Human Perception of Facial Expressions by Discrete Choice Models

Self-supervised CNN for Unconstrained 3D Facial Performance Capture from an RGB-D Camera

Real-time Image-based Reconstruction of Pipes Using Omnidirectional Cameras

Cost-alleviative Learning for Deep Convolutional Neural Network-based Facial Part Labeling

Real time facial expression recognition from image sequences using Support Vector Machines

3D Production Pipeline

A Model-based Approach to Rapid Estimation of Body Shape and Postures Using Low-Cost Depth Cameras

Compu&ng Correspondences in Geometric Datasets. 4.2 Symmetry & Symmetriza/on

Cloth Simulation. Tanja Munz. Master of Science Computer Animation and Visual Effects. CGI Techniques Report

arxiv: v1 [cs.cv] 16 Nov 2015

Feature points based facial animation retargeting

3D Editing System for Captured Real Scenes

Dynamic 2D/3D Registration for the Kinect

Stable Vision-Aided Navigation for Large-Area Augmented Reality

Speech Driven Synthesis of Talking Head Sequences

Data-Driven Face Modeling and Animation

FACIAL EXPRESSION USING 3D ANIMATION TECHNIQUE

FACIAL ANIMATION WITH MOTION CAPTURE BASED ON SURFACE BLENDING

Shape Augmented Regression for 3D Face Alignment

Facial Expression Analysis

Occluded Facial Expression Tracking

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction

A Machine Learning Approach to Automate Facial Expressions from Physical Activity

Sketching Articulation and Pose for Facial Meshes

Moving Object Segmentation Method Based on Motion Information Classification by X-means and Spatial Region Segmentation

Topics in Computer Animation

Automatic Animation of High Resolution Images

Optimal Marker Set for Motion Capture of Dynamical Facial Expressions

Game Programming. Bing-Yu Chen National Taiwan University

Automatic Modelling Image Represented Objects Using a Statistic Based Approach

Emotion Detection System using Facial Action Coding System

IBM Research Report. Automatic Neutral Face Detection Using Location and Shape Features

Black Desert Online. Taking MMO Development to the Next Level. Dongwook Ha Gwanghyeon Go

Light source estimation using feature points from specular highlights and cast shadows

Using the rear projection of the Socibot Desktop robot for creation of applications with facial expressions

CSE452 Computer Graphics

Supplementary Material for Synthesizing Normalized Faces from Facial Identity Features

c Copyright by Zhen Wen, 2004

Classification of Upper and Lower Face Action Units and Facial Expressions using Hybrid Tracking System and Probabilistic Neural Networks

A Comparative Study of Region Matching Based on Shape Descriptors for Coloring Hand-drawn Animation

Meticulously Detailed Eye Model and Its Application to Analysis of Facial Image

Performance Driven Facial Animation using Blendshape Interpolation

Level-of-Detail Triangle Strips for Deforming. meshes

Shape and Expression Space of Real istic Human Faces

Dense 3D Face Alignment from 2D Videos in Real-Time

Recognizing Deformable Shapes. Salvador Ruiz Correa (CSE/EE576 Computer Vision I)

Cloth Simulation on the GPU. Cyril Zeller NVIDIA Corporation

AAM Based Facial Feature Tracking with Kinect

Adding Hand Motion to the Motion Capture Based Character Animation

Chapter 18- Relative Vertex (Shape) Keys

Transcription:

Modeling Facial Expressions in 3D Avatars from 2D Images Emma Sax Division of Science and Mathematics University of Minnesota, Morris Morris, Minnesota, USA 12 November, 2016 Morris, MN Sax (U of Minn, Morris) Modeling Facial Expressions November 2016 Morris, MN 1 / 27

The Big Idea Take a 2D image of a user Use specific facial features on the user as guidelines Render a 3D avatar with the same facial shape and expressions as the user [2] Sax (U of Minn, Morris) Modeling Facial Expressions November 2016 Morris, MN 2 / 27

Outline Outline 1 Introduction 2 3D Shape Regression Tracking 3 Displaced Dynamic Expression (DDE) Regression Tracking 4 Comparison and Conclusions Sax (U of Minn, Morris) Modeling Facial Expressions November 2016 Morris, MN 3 / 27

Introduction Overview Overview Traditionally, animators have found it is easier to model humans and their facial expressions when they use physical humans as models. [1] Sax (U of Minn, Morris) Modeling Facial Expressions November 2016 Morris, MN 4 / 27

Introduction Overview The Basic Process Take a single 2D video frame of a user Track specific landmarks on the user s face Render a 3D virtual image or avatar with the same facial shape and expression Entire process should occur in real time Sax (U of Minn, Morris) Modeling Facial Expressions [3] November 2016 Morris, MN 5 / 27

Introduction Background Ekman s Facial Action Coding System (FACS) Any facial expression can be represented by the contractions and relaxations of specific facial muscles (also known as Action Units or AUs) Categorizes human facial movements by the specific AUs that are used to create the facial expression Any anatomically possible facial expression can therefore be represented through AUs Sax (U of Minn, Morris) Modeling Facial Expressions November 2016 Morris, MN 6 / 27

Introduction Background Linear Blendshape Models A linear blendshape model generates a facial pose as a linear combination of a number of blendshapes A blendshape is a visual approximation of a facial expression in which a single collection of points (mesh) have deformed to a series of fixed vertex positions Each blendshape contains a blend modifier which contains an intensity slider that controls the amount of contraction and relaxation of the AUs [4] Sax (U of Minn, Morris) Modeling Facial Expressions November 2016 Morris, MN 7 / 27

Introduction Background FaceWarehouse Public, online database of 3D facial expression models and blendshapes Composed of 150 individuals Tracked features on each individual s face using FACS, and composed generic blendshape models for each generic facial expression Sax (U of Minn, Morris) Modeling Facial Expressions November 2016 Morris, MN 8 / 27

What is a Regressor? Introduction Background A regressor is an algorithm that is specifically trained to look at input states in order to predict the next output state. The regressor does this by analyzing the relationships between the input data and previously trained output data. Sax (U of Minn, Morris) Modeling Facial Expressions November 2016 Morris, MN 9 / 27

3D Shape Regression Tracking Outline 1 Introduction 2 3D Shape Regression Tracking Gathering and Assembling Data Training the Regressor Runtime Regression Results 3 Displaced Dynamic Expression (DDE) Regression Tracking 4 Comparison and Conclusions Sax (U of Minn, Morris) Modeling Facial Expressions November 2016 Morris, MN 10 / 27

3D Shape Regression Tracking Gathering and Assembling Data Gathering Data and Locating Facial Landmarks 60 images of the user showing pre-defined face positions are captured and categorized into two groups: Various different head poses with neutral facial expression Various different facial expressions A set of 75 facial landmarks are automatically located 60 internal landmarks and 15 contour landmarks All landmarks can be overwritten manually [3] Sax (U of Minn, Morris) Modeling Facial Expressions November 2016 Morris, MN 11 / 27

3D Shape Regression Tracking Gathering and Assembling Data Generating Blendshapes Use a FaceWarehouse generic blendshape model to calculate a user-specific blendshape model. For each image, three adjustments must be made to a generic blendshape: The user s identity: allow the regressor to transform the blendshapes to have the same facial features and shapes as the user The user s expression: in order to provide blendshapes with the same facial expression as the user Transformation: allow the regressor to take account for head turns, rotations, and translations Sax (U of Minn, Morris) Modeling Facial Expressions November 2016 Morris, MN 12 / 27

3D Shape Regression Tracking Training the Regressor Selecting Training Data and Training the Regressor Select training data so that the regressor is prepared to output a 3D facial shape from a single 2D image. The regressor learns a regression function based on the information in the input images. The goal of training the regressor is to teach an effective prediction model through the geometric relationships that a specific user s image data contains. Sax (U of Minn, Morris) Modeling Facial Expressions November 2016 Morris, MN 13 / 27

3D Shape Regression Tracking Runtime Regression Runtime Regression At runtime, the 3D shape regressor tracks the 3D positions of facial landmarks from a 2D video stream. 1 For each video frame, find a facial shape that is similar to the previous frame s facial shape 2 Transform the shape to align the previous frame s shape with the 3D shape space 3 Find a set of shapes in the training data that are similar to the transformed shape 4 Pass each shape in the set through the regressor to find the regressor s output Update the current image s shape with the output 5 Average all of the outputs to make the final facial shape Sax (U of Minn, Morris) Modeling Facial Expressions November 2016 Morris, MN 14 / 27

3D Shape Regression Tracking Runtime Regression Why this Regression Algorithm Works 1 Regression uses a set of similar shapes instead of a single shape to generate the current shape This allows the solution to deal with uncertainty and to avoid error accumulation [3] Sax (U of Minn, Morris) Modeling Facial Expressions November 2016 Morris, MN 15 / 27

3D Shape Regression Tracking Runtime Regression Why this Regression Algorithm Works cont. 2 Regression performs a transformation step before selecting a set of similar shapes This allows the algorithm to account for different head positions and rotations [3] Sax (U of Minn, Morris) Modeling Facial Expressions November 2016 Morris, MN 16 / 27

3D Shape Regression Tracking Results Results of 3D Shape Regression Tracking Implemented on a PC with an Intel Core i7 (3.5GHz) CPU, the overall runtime performance is less than 15 milliseconds The setup and preprocessing steps take less than 45 minutes per user This solution has limitations when there are large occlusions or when there are dramatic changes in lighting Sax (U of Minn, Morris) Modeling Facial Expressions November 2016 Morris, MN 17 / 27

DDE Regression Tracking Outline 1 Introduction 2 3D Shape Regression Tracking 3 Displaced Dynamic Expression (DDE) Regression Tracking DDE Model Preparing Training Data Using the Regressor Results 4 Comparison and Conclusions Sax (U of Minn, Morris) Modeling Facial Expressions November 2016 Morris, MN 18 / 27

DDE Regression Tracking DDE Model DDE Model Designed to allow users to be switched automatically, without a user-specific setup step Designed to represent: 3D facial shape of the user s facial expressions 2D facial landmarks 3D facial shape is represented by a linear combination of expression blendshapes To represent a 2D facial landmark, add a 2D displacement to the projection of the landmark s corresponding vertex on the facial mesh 2D displacement of facial landmark is what accounts for changes in user identity Sax (U of Minn, Morris) Modeling Facial Expressions November 2016 Morris, MN 19 / 27

DDE Regression Tracking Preparing Training Data Preparing Training Data 1 Facial images from FaceWarehouse are used as initial data 73 2D landmarks are labeled on initial data to produce a 2D facial shape 2 Training pairs are created for each image to relate differences in parameters to image features Training pairs simulate cases where input parameters may be inaccurate tive facial motions. The are sent back to the CPR post-processed output, inssion coefficients, can be rive its facial animation. y computes the facial moation and displacements. are a part of the regressor (a) [2] (b) (c) Figure 3: Random identity example. For image (a), we use the identity from (b) but keep the original displacements, resulting in inaccurate 2D landmark positions. Therefore we need to recompute Sax (U of Minn, Morris) Facial Expressions November 2016 Morris, MN learn the DDE regressor the displacements tomodeling get the correct landmarks (c). 20 / 27

DDE Regression Tracking Using the Regressor Training the Regressor and Runtime Regression This approach follows a very similar regression approach to the 3D Shape Regression Training, except this algorithm must account for the use of a DDE model as shape representation. The overall runtime regression algorithm is the same as the algorithm in the 3D Shape Regression Tracking algorithm. Sax (U of Minn, Morris) Modeling Facial Expressions November 2016 Morris, MN 21 / 27

representation withdde the the combined advantages of 3D DEM and Regression Tracking Results 2D landmarks. We consider our approach to be an attractive solution for wide deployment in consumer-level Results of DDE Regression Trackingapplications. As aimplemented video basedontechnique, our approach will fail CPU, to track the a PC with an Intel Core i5 (3.0GHz) and the face entire if many of the takes facial approximately features cannot observedon in average the video approach 20 be milliseconds frames. As shown in Fig. 12, our approach can handle some partial Setup takes about six hours, but this step only occurs once occlusions, but may fail if the face is largely occluded. Moreover, This solution does better than the previous solution when dealing prolonged landmark occlusions during the DEM adaptation period partial occlusions, but it still produces an inaccurate result may with negatively impact the overall accuracy. when there are large occlusions [2] Figure 12: Our approachmodeling can handle some partial occlusions, but Sax (U of Minn, Morris) Facial Expressions November 2016 Morris, MN 22 / 27

Comparison and Conclusions Outline 1 Introduction 2 3D Shape Regression Tracking 3 Displaced Dynamic Expression (DDE) Regression Tracking 4 Comparison and Conclusions Sax (U of Minn, Morris) Modeling Facial Expressions November 2016 Morris, MN 23 / 27

Comparison and Conclusions Comparison of the Two Solutions The DDE Regression Tracking solution is an improvement to the 3D Shape Regression Tracking solution. The improved solution: does not require a data gathering step (even if the initial tracking takes much longer) is not user-specific; users can be switched and swapped during video without a noticeable lag in output results is more robust than the previous solution, especially during dramatic lighting changes Sax (U of Minn, Morris) Modeling Facial Expressions November 2016 Morris, MN 24 / 27

Comparison and Conclusions Conclusions A preliminary solution for rendering 3D virtual images from 2D video frames in real time: 3D Shape Regression Tracking An improved solution that has been shown to be more robust, better at handling lighting changes, handles facial rotations more successfully, and does not requiring user-specific training: DDE Regression Tracking Further research includes solutions driven by both visual and audio data, as well as solutions that work completely online without the use of facial markers or training stages Sax (U of Minn, Morris) Modeling Facial Expressions November 2016 Morris, MN 25 / 27

Comparison and Conclusions Thanks! Thank you for your time and attention! Contact: saxxx027@morris.umn.edu Questions? Sax (U of Minn, Morris) Modeling Facial Expressions November 2016 Morris, MN 26 / 27

References References alice-in wonderland.net. About disney s alice in wonderland 1951 cartoon movie. C. Cao, Q. Hou, and K. Zhou. Displaced dynamic expression regression for real-time facial tracking and animation. ACM Trans. Graph., 33(4):43:1 43:10, July 2014. C. Cao, Y. Weng, S. Lin, and K. Zhou. 3d shape regression for real-time facial animation. ACM Trans. Graph., 32(4):41:1 41:10, July 2013. O. Media. Programming 3d applications with html5 and webgl. See my Modeling Facial Expressions paper for additional references. Sax (U of Minn, Morris) Modeling Facial Expressions November 2016 Morris, MN 27 / 27