Optical-inertial Synchronization of MoCap Suit with Single Camera Setup for Reliable Position Tracking

Size: px
Start display at page:

Download "Optical-inertial Synchronization of MoCap Suit with Single Camera Setup for Reliable Position Tracking"

Transcription

1 Optical-inertial Synchronization of MoCap Suit with Single Camera Setup for Reliable Position Tracking Adam Riečický 2,4, Martin Madaras 1,2,4, Michal Piovarči 3,4 and Roman Ďurikovič 2 1 Institute of Software Technology and Interactive Systems, University of Technology, Vienna, Austria 2 Department of Applied Informatics, Comenius University, Bratislava, Slovakia 3 Saarland University, MMCI Saarbrücken, Germany 4 Skeletex Research, Bratislava, Slovakia Keywords: Abstract: Motion Capture, Inertial Suit, Optical System, Optical-inertial Synchronization, Silhouette Tracking. We propose a method for synchronization of an inertial motion capture suit and a single camera optical setup. Proposed synchronization is based on an iterative optimization of an energy potential in image space, minimizing the error between the camera image and a rendered virtual representation of the scene. For each frame, an input skeleton pose from the mocap suit is used to render a silhouette of a subject. Moreover, the local neighborhood around the last known position is searched by matching the silhouette to the distance transform representation of the camera image based on Chamfer matching. Using the combination of the camera tracking and the inertial motion capture suit, it is possible to retrieve the position of the joints that are hidden from the camera view. Moreover, it is possible to capture the position even if it cannot be captured by the suit sensors. Our system can be used for both real-time tracking and off-line post-processing of already captured mocap data. 1 INTRODUCTION Considering the problem of person tracking and movement analysis, there are a lot of motion capture (mocap) solutions available in both academia and industry. Mocap is a widely used technique for storing and processing movement data of the person. A reliable motion capture and tracking is necessary in the film and games industry, virtual reality, biometrics or even healthcare. Optical-based tracking is more problematic in the tracking area if it has occlusions. Moreover, the tracking area might be very non-convex, and therefore non-coverable by optical-based mocap and tracking systems. In such a case, it is suitable to use non-optical methods for tracking; however, some other limitations appear, such as drifting, calibration and synchronization problems or additional noise in the captured data. Sometimes, these limitations are solved by a post-processing of raw data using complex probabilistic models that have to be trained on reliable training datasets, which might be impossible to obtain in a given situation. Nowadays, there are a variety of mocap systems suitable for recording of the body movements. There are two major groups of mocap systems, opticalbased and inertial-based systems. Each group has its own advantages and limitations. The advantages of inertial systems are the flexible capture area (outdoor capture, water capture), occlusion independence, fast setup time, transferability and the possibility of direct use of the raw output data for a 3D model. The biggest disadvantage is that one can get only rotation data of each skeleton joint. The joint positions in 3D space have to be calculated based on the calibration process and the root position estimation, which have to be approximated by a walking algorithm implemented in the mocap software. On the other hand, the optical systems are limited to indoor use only. They have problems with occlusions and cannot directly return the information about the joint rotation around the bone axis. The individual joint position can be tracked easily; however, the rotations need to be calculated in the next evaluation stage. In order to solve the positioning disadvantage of inertial systems, solutions using radio/nfer or ultrasound positioning have been proposed. However, these systems are in general hard to calibrate and synchronize. Furthermore, in our project, we need to obtain reliable position tracking of the inertial suit using a commodity RGB camera. The whole system should 40 Riečický, A., Madaras, M., Piovarci, M. and Durikovic, R. Optical-inertial Synchronization of MoCap Suit with Single Camera Setup for Reliable Position Tracking. DOI: / In Proceedings of the 13th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISIGRAPP 2018) - Volume 1: GRAPP, pages ISBN: Copyright 2018 by SCITEPRESS Science and Technology Publications, Lda. All rights reserved

2 Optical-inertial Synchronization of MoCap Suit with Single Camera Setup for Reliable Position Tracking be affordable and it should be compatible with both outdoor and indoor usage. Therefore, in our project we propose a hybrid optical-inertial system. In this system, the mocap inertial suit is combined and synchronized with a single camera. Once the camera is calibrated and the relative position and orientation are calculated, it can be used for real-time effortless position estimation. The hybrid system does not require a training phase and has advantages over both optical and inertial motion capture systems. Other hybrid tracking systems need either a complicated setup or are much more expensive. Our system requires only an inertial suit and a single RGB camera. 2 RELATED WORK Inertial Suits. There are several inertial IMU suits available on the market: a 3DSuit by Inertial Labs, an IGS Cobra suit by Synertial, an MVN suit by XSens and a Perception Neuron suit by Noitom. The suits differ in sensor configuration, price and precision. For example, the suits from XSens and Synertial have a higher number of sensors and can stream raw data for all the sensors. The Perception Neuron suit is a cheap and affordable solution for the general public, with a smaller set of sensors. Afterwards, the streamed data available to the reader are interpolated from the raw sensor data. Optical Mocap. Optical systems can be divided into two main groups: systems based on passive retroreflective or active markers, and marker-less mocap systems that are trained on a set of training images. The marker-based systems are able to perform with much higher accuracy. In practice, optical systems and suits with markers are used, e.g. OptiTrack, Vicon. A group of multi-view RGB-only-based mocap systems working without a training stage exists, using shape from silhouette or sums of spatial Gaussians (Cheung et al., 2003; Stoll et al., 2011). These optical systems usually need a complicated setup and multiple cameras. The trained probabilistic marker-less techniques (Wojek et al., 2009; Andriluka et al., 2010) that work on RGB images are not very precise in general. Thus, they are typically used for academic research testing only, or they have to be fused with inertial sensors as in (Pons-Moll et al., 2010). Lately, a helmet with two fisheye RGB cameras was proposed in (Rhodin et al., 2017) for the motion capture of a subject wearing the helmet. The system can only capture motion of the skeleton; this cannot be used for position tracking. Probabilistic optical approaches can be trained directly on depth values obtained by an RGBD camera, for example Kinect. The Kinect is mostly used for real-time pose estimation (Shotton et al., 2013). This probabilistic skeleton estimation is not very precise, but is well suited for the fun real-time applications where Kinect tracking is mostly used. Moreover, the Kinect can be used for both real-time skeleton estimation and surface reconstruction using Kinect Fusion (Izadi et al., 2011). Depth values from the RGBD camera can be used for point cloud reconstruction and the skeleton can be extracted from a point cloud. However, this process is too slow for real-time motion capture. Nevertheless, it can be used for body size estimation and calibration from a single scan. These data can be used to improve motion capture data (Anguelov et al., 2005). Moreover, the probabilistic optical-based systems are trained on RGB or RGBD images and estimate position in 3D space based on probabilistic models (Shotton et al., 2013; Andriluka et al., 2010). An optical flow based on Chamfer matching can be used to track the subject without a training stage (Dimitrijevic et al., 2006; Katz and Aghajan, 2008). These methods can be used directly on the input frames; however, a background subtraction is a necessary preprocessing step to obtain robust tracking results. An extensive comparison of inertial and opticalbased motion capture suits can be found in (Skogstad et al., 2011). Person Tracking. Similarly, as in the case of mocap solutions, the tracking can be optical based or approximated using triangulation of distances to the signal source, e.g. GPS. The lighthouse tracking by Valve is part of HTC Vive and it is based on a measuring of time delay between emitting a flash and sweeping a beam of light across the room. The receivers calculate the time delay between these two light flashes and using a simple trigonometry the 3D position of the receiver can be evaluated in real-time. Hybrid Systems. Several hybrid approaches were published in recent years. The hybrid systems for skeleton and position tracking are based on a fusion of the IMU orientation data and some other sensor. In (Ziegler et al., 2011), the subject in an inertial suit is tracked by a robot with a laser scanner. Such a combination can track the subject s position and trajectory in big areas; however, it might be impossible to use the robot in small interiors and the robot is too expensive tool for common usage. A fusion of multi-view RGB cameras with few IMUs was proposed in (Pons- Moll et al., 2010; von Marcard et al., 2016). These 41

3 GRAPP International Conference on Computer Graphics Theory and Applications approaches for fusion give very good results; however, the fusion needs a scanned template model of the subject and the system needs multiple RGB cameras in order to correctly fit the template into the silhouette. A combination of discriminative and generative tracking using a depth sensor was used in (Helten et al., 2013). The approach also needs a template mesh model and the RGBD camera has a very limited volume where the fusion works precisely enough. In general, the mentioned related hybrid approaches either need a much more complicated and expensive setup (multiple cameras, depth camera, robot), or they have a much more complex tracking pipeline than our approach (template mesh scanning, non-linear energy minimization, training stage). 3 OPTICAL-INERTIAL SYNCHRONIZATION The main idea behind the optical-inertial tracking solution of the suit and the camera is determining the 3D position of the actor from his silhouette in the camera image based on his actual pose. Knowing the actor s skeleton pose from the suit in real-time, we are able to predict the body shape we are looking for within the camera image. First, a base mesh is constructed using the actor s specific parameters such as height or local diameters. This mesh is then used for rendering a shape which is similar to the actor s silhouette in the image. A virtual camera which is used for the base mesh rendering needs to see the scene the same way as a real camera sees the scene with actor; therefore, it needs to be calibrated. The rendered base mesh silhouette is then used to search the local neighborhood of the last known position of the subject in the next image frame. Minimizing the energy composed of spatial integration of Chamfer matching error in the image space, we are able to perform real-time tracking of the subject. During the tracking, a 3D virtual scene is rendered and matched to the camera image; therefore, if it is matched with the precisely calibrated camera setup, we are able to directly estimate the 3D position of the subject in the real world. 3.1 System Overview The whole tracking system is composed of three phases: a calibration phase, a tracking start-up phase and an iterative tracking phase. The first two phases are used for the initial setup only to determine and to correctly represent the real world in the tracking system; therefore, the third phase is the actual tracking stage. The calibration phase needs to be done only once, or when the camera is replaced. This step is required to acquire correct camera parameters. The parameters can be saved and reused before each tracking session. The second stage, the tracking start-up phase, needs to be performed at least once before each session, to synchronize the real-world camera with the virtual camera of a system, and to specify the actor s starting location for the tracking. However, there is a possibility to assign these properties during the iterative tracking phase, seamlessly without the interruption of the tracking procedure. Finally, the tracking phase is iteratively performed during the whole remaining tracking time. Output of this stage is the true 3D position of the actor in both, the virtual scene and the real world. 3.2 Initial Setup and Calibration For the camera calibration, an OpenCV with its builtin modules is used. We used the ChArUco module that uses a hybrid checkerboard for both camera calibration and camera position estimation, as can be seen in Figure 1. Given several pairs of point correspondences of calibration patterns in the real world and in the image space, it is possible to find intrinsic and extrinsic parameters of the camera. Figure 1: The ChArUco calibration board. The ChArUco is a combination of the classical calibration checkerboard and the ChArUco alphabet signs. Next, the body size of the actor needs be measured manually using a ruler, or in an automatic way using a Kinect or calibrated RGB camera. The measured body height and body radii are then used together with the mocap position to construct a base mesh approximating the body of the subject. 3.3 Virtual Scene Creation The main idea behind the synchronization is to create a virtual scene according to the parameters acquired in the real world. This step is called the tracking startup phase, and it must be generally executed before each tracking session, when some scene properties are changed, e.g. camera is moved, camera is changed, starting position is changed, etc. 42

4 Optical-inertial Synchronization of MoCap Suit with Single Camera Setup for Reliable Position Tracking 3.4 Silhouette Image Database Construction The camera image contains objects and subjects which are not important for the system. The goal of the system is to locate and track only the actor dressed in the motion capture suit. Therefore, the pose data from the suit are used to determine this. The reader of the suit rotations is able to stream local transformations for each frame in real-time. Figure 2: A model-view matrix that consists of a rotation and a translation of the ChArUco board into the camera image is obtained in the tracking start-up phase. Thus, the origin of the virtual scene is set into the ChArUco board location and the camera position and orientation are set accordingly. Figure 4: A base mesh is created using the skeleton acquired from the mocap suit. The skeleton is enhanced with measured radii of the actor s body. Note that the base mesh construction process is depicted in 2D only. Figure 3: A virtual scene is constructed based on the camera calibration, calculated scene parameters and current mocap data. Using either the real-time mocap data or stored offline mocap data, a silhouette of the tracked subject is approximated and later used during the tracking phase. In order to get the origin of the virtual space, a ChArUco board marker (see Figure 2) is placed in front of the camera. As an input it takes corresponding points together with board parameters and as an output it produces the rotation and translation transformations that give us a model-view matrix of the checkerboard in world space coordinates. Having the ChArUco marker detected, we are able to estimate the camera position and orientation relative to the origin. The second purpose of placing a ChArUco board into the scene is to define the starting position of the actor (the actor starts in the origin of the virtual scene). This is the position where the tracking starts. The precision of the tracked position relies deeply on the camera calibration and proper virtual scene setup (see Figure 3). If the ChArUco board cannot be used for some reason, the starting position and camera parameters can be always set manually. Firstly, a shape that roughly represents the actor s body is needed. Here, it might be possible to use a broad set of shapes, from primitives roughly approximating the body to a high-detailed 3D scan. However, we choose to create a simple base mesh approximating the body shape mesh from the input skeleton, because it is easily customizable, scalable and can be generated in real-time for any skeleton pose. For this a SQM algorithm (Bærentzen et al., 2012) is used, which is able to generate such a mesh specifying only the skeleton and the radius of a sphere around each skeleton node (see Figure 4). These radii as well as skeleton height are dependant on the actor s body type and need to be measured or approximated manually. Such a specific base mesh is generated only once, and a pose for every frame is created by applying rotations from suit sensors and transforming the base mesh accordingly using a skinning algorithm. With the virtual camera and scene created and successfully calibrated, it is possible to render an image of the base mesh as if it was seen by a real camera. Afterwards, this image is processed to obtain only the silhouette of a rendered image. Finally, a set of silhouette images is rendered, applying several shifts of a base mesh in eight evenly distributed directional vectors. Our set consists of images shifted k times by d in a space around a specified pose as well as one image of a base mesh exactly in the current position, as 43

5 GRAPP International Conference on Computer Graphics Theory and Applications Figure 5: The pipeline of our optical-inertial synchronization. (Top) the camera image is thresholded and transformed into a distance transform image. (Bottom) a base mesh is constructed based on the acquired skeleton and it is rendered in different positions. The base of 8 search vectors is used to render 8 shifted silhouettes. The Chamfer matching of the rendered base mesh and the distance transform image is performed, and the error is evaluated. Finally, the error minimization is used to find the next position in 3D space. can be seen in Figure 5. In our experiments, we used constants d = 10cm and k = 3; therefore, in total a set of 25 images is rendered and stored in the silhouette database. 3.5 Tracking The tracking phase begins after the tracking start-up phase was executed successfully, which properly sets up a camera for rendering the base mesh silhouette database. An actor is located in a specified position defined by the ChArUco board placed in the scene. At this point, the tracking phase is ready to start. This start-up position is considered to be the actor s true position in the first frame. For each next frame, a motion vector is evaluated to evaluate the actor s next position. Figure 6: Tracking of the actor position using Chamfer matching. (Top) an input image and the subtracted background image in grayscale. (Bottom) applied adaptive thresholding and calculated distance transform of the silhouette. First, the captured camera image is pre-processed using background subtraction, thresholded and Canny edge detection is performed so that only the actor s silhouette is obtained. Then, the image is transformed into a distance transform image (see Figure 6). Afterwards, the already pre-rendered database of base mesh silhouette images is used to evaluate the energy e for the optimization. The error potential e is calculated for each silhouette as an integration of the distance transform function DT (x) over the actor silhouette S as e k = DT (x(s, t))dsdt, (1) S where parameters s and t are the parameters of the one-dimensional silhouette curve and a kernel function that is applied over the curve to make the silhouette wider, respectively. The term x(s, t) is a function that maps parameters s and t to the image space, where function DT (x) is evaluated. The integration in discrete form is performed in the image space as a sum of non zero pixels from silhouette image S k and normalized afterwards as e k = m i=0 n j=0 DT (i, j) S k(i, j), (2) count non zero(s k ) where (i, j) refers to a pixel position of the image with dimensions m n, DT is the distance transform image, S k is the binary silhouette mask image and the function count non zero() returns a number of non zero pixels contained in the image. Minimizing the error energy, we are able to evaluate the direction and the magnitude of the subject s movement based on the shift vectors used for the construction of a database image. Adding such a vector to the position of an actor in the last frame, we are able to evaluate the actor 44

6 Optical-inertial Synchronization of MoCap Suit with Single Camera Setup for Reliable Position Tracking Figure 7: Comparison of XSens MVN tracking (top) and our optical-inertial evaluation (bottom). In the second frame, the sliding starts and the mocap system fails to evaluate the position correctly. Images were captured using a Microsoft LifeCam HD 3000 RGB webcam. position in the current frame. In the end of the tracking phase, we have a raw corrected 3D position of the actor. The pipeline of the system until this point can be seen in Figure 5. To enhance the raw data, some suitable post-processing method might be used. 3.6 Post-processing via Gaussian Filtering The change of estimated position of the subject in time might not be continuous. Thus, it is useful to post-process the discontinuities into a continuous movement. In our experiments we have tried two methods of post-processing: a Gaussian-based smoothing and Kalman filtering. We used local Gaussian smoothing in the neighborhood of 20 time steps and the Extended Kalman Filter (EKT) (Julier and Uhlmann, 2004) implemented in OpenCV. Using the EKT filtering, the resulting graphs seem visually smoother, but the overall error was higher. Therefore, in the final results, the Gaussian smoothing was used (see Figure 9). 4 RESULTS We demonstrate the results of our approach on capturing the motion and estimating the position of a subject in a space. The subject wears an XSens suit and the scene is captured using a Microsoft LifeCam HD 3000 RGB webcam and ids 3 ueye monochromatic camera with a fisheye lens. 4.1 Correction of Mocap Suit Data First, the position in 3D space is approximated using the standard walking algorithm usually implemented Figure 8: Comparison of error difference in Euclidean distance between the methods and ground truth obtained using HTC Vive: (Red) original position from inertial suit software and (yellow) our optical-inertial method. Graph is evaluated on a dataset, where the movement was the most problematic for the original method (e.g. jumping, sliding). within the mocap software. In this scenario, the actor starts to run and finishes the running sequence by sliding on the ground. The sliding is the stage, where the inertial mocap suit fails. We use the position estimated by XSens MVN Studio and export it into a bvh file. Second, the position is evaluated using our optical-inertial system. Both estimated positions, from the original method and our camera-based correction, can be seen in Figure Evaluation of Estimated Position In order to evaluate our method by comparing results to the ground truth, we evaluate our correction of position inside a known environment for movement in predefined patterns (see Figure 11). During this evaluation, the subject moves along the defined trajectories with known dimensions. During the evaluation we track three position estimations in time. The subject is tracked by HTC Vive lighthouses (ground truth), and position estimation is done by 45

7 GRAPP International Conference on Computer Graphics Theory and Applications Figure 9: Comparison of error difference with filtered data: (Red) position calculated by original inertial suit software and (blue) post-processed versions of estimated positions using our approach. Another example from the datasets is plotted on the right. In both cases, the smoothed versions of evaluated positions approximate the movement and the real position of the subject much better. MVN XSens Studio and our optical-inertial estimation. Comparison of the error difference in Euclidean distance of the original position from the MVN XSens Studio and our approach can be seen in Figure 8. Furthermore, we tried to smooth the raw results from our approach; the graphs with the smoothed positions of two different datasets are shown in Figure 9. In Figure 10, the setup used for the evaluation is depicted and described. B 1 C Figure 10: The evaluation setup. The subject wearing a mocap suit (C) is captured by a camera (A) and tracked by HTC Vive lighthouses (B). The position estimated by the lighthouses is used as a ground truth when the original mocap suit tracking is compared to the proposed method. 5 LIMITATIONS The main limitation of the proposed solution is the dependency on the static background subtraction; thus we are not able to guarantee robust tracking in scenes with a dynamically changing background. In the case of background changes, there are edges in the image space not related to the actor that may drive the tracking into a local minimum. Another limitation of the system is the predefined set of search directions that produces discretizetion errors. If required, the search space could be sampled more densely at the cost of higher computation time. The rendered base mesh is only a rough approximation of the human body; a B 2 A highly detailed full-body scan could be used for better approximation of the silhouette. However, the base mesh is easy to compute, affordable to acquire and the results are good enough for our applications. 6 CONCLUSION AND FUTURE WORK A system for optical-inertial synchronization of the mocap suit and the camera was implemented and described in this paper. In general, the system can find its utilization in applications such as virtual reality, movement analysis, sports evaluation, and biometrics. Using a hybrid mocap system, drift issues of inertial suits can be solved. Moreover, the lack of positioning capability of inertial mocap was solved, and therefore it can be directly used for subject movement analysis in 3D space, ergonomic work analysis process or virtual reality games. The inertial-optical hybrid system is capable of measuring a subject s position with high precision even if partially or fully occluded, and all the computations can be performed in real-time. These results show promising improvement for inertial suit position tracking, but more extensive evaluation is required in the future. As future work we would like to use the system for an automatic and effortless recalibration of the suit. The correct position and orientation of the joints, evaluated from the camera image, can be used for on-line correction of suit sensors. ACKNOWLEDGMENTS We would like to thank the SAIA Action Austria- Slovakia for financial support of the project and the 46

8 Optical-inertial Synchronization of MoCap Suit with Single Camera Setup for Reliable Position Tracking Figure 11: Evaluation of our method on different trajectories within a grid. The grid has a size of 6 6 meters. The patterns were chosen to fully cover the tracking area in different directions. Moreover, different types of movement were used during the evaluation, such as walking, jumping, sliding etc. Interactive Media Systems group at TU Wien for all the necessary hardware and laboratory space during the project. Our project was also partly supported by Comenius University grant No. UK/293/2017. REFERENCES Andriluka, M., Roth, S., and Schiele, B. (2010). Monocular 3d pose estimation and tracking by detection. In 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pages Anguelov, D., Srinivasan, P., Koller, D., Thrun, S., Rodgers, J., and Davis, J. (2005). Scape: Shape completion and animation of people. ACM Trans. Graph., 24(3): Bærentzen, J., Misztal, M., and Wełnicka, K. (2012). Converting skeletal structures to quad dominant meshes. Computers & Graphics, 36(5): Shape Modeling International (SMI) Conference Cheung, G. K. M., Baker, S., and Kanade, T. (2003). Shapefrom-silhouette of articulated objects and its use for human body kinematics estimation and motion capture. In Proceedings of the 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 03, pages 77 84, Washington, DC, USA. IEEE Computer Society. Dimitrijevic, M., Lepetit, V., and Fua, P. (2006). Human body pose detection using Bayesian spatio-temporal templates. Comput. Vis. Image Underst., 104(2): Helten, T., Muller, M., Seidel, H.-P., and Theobalt, C. (2013). Real-time body tracking with one depth camera and inertial sensors. In The IEEE International Conference on Computer Vision (ICCV). Izadi, S., Kim, D., Hilliges, O., Molyneaux, D., Newcombe, R., Kohli, P., Shotton, J., Hodges, S., Freeman, D., Davison, A., and Fitzgibbon, A. (2011). Kinectfusion: Real-time 3d reconstruction and interaction using a moving depth camera. In Proceedings of the 24th Annual ACM Symposium on User Interface Software and Technology, UIST 11, pages , New York, NY, USA. ACM. Julier, S. J. and Uhlmann, J. K. (2004). Unscented filtering and nonlinear estimation. Proceedings of the IEEE, 92(3): Katz, I. and Aghajan, H. (2008). Multiple camera-based chamfer matching for pedestrian detection. In 2008 Second ACM/IEEE International Conference on Distributed Smart Cameras, pages 1 5. Pons-Moll, G., Baak, A., Helten, T., Müller, M., Seidel, H.- P., and Rosenhahn, B. (2010). Multisensor-fusion for 3d full-body human motion capture. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Rhodin, H., Richardt, C., Casas, D., Insafutdinov, E., Shafiei, M., Seidel, H., Schiele, B., and Theobalt, C. (2017). Egocap: Egocentric marker-less motion capture with two fisheye cameras (extended abstract). CoRR, abs/ Shotton, J., Sharp, T., Kipman, A., Fitzgibbon, A., Finocchio, M., Blake, A., Cook, M., and Moore, R. (2013). Real-time human pose recognition in parts from single depth images. Commun. ACM, 56(1): Skogstad, S. A., Nymoen, K., and Hvin, M. (2011). Comparing inertial and optical mocap technologies for synthesis control. In 2011 IEEE Conference on Systems, Man and Cybernetics. Stoll, C., Hasler, N., Gall, J., Seidel, H.-P., and Theobalt, C. (2011). Fast articulated motion tracking using a sums of gaussians body model. In Computer Vision (ICCV), 2011 IEEE International Conference on, pages IEEE. von Marcard, T., Pons-Moll, G., and Rosenhahn, B. (2016). Human pose estimation from video and imus. Transactions on Pattern Analysis and Machine Intelligence, 38(8): Wojek, C., Walk, S., and Schiele, B. (2009). Multi-cue onboard pedestrian detection. In 2009 IEEE Conference on Computer Vision and Pattern Recognition, pages Ziegler, J., Kretzschmar, H., Stachniss, C., Grisetti, G., and Burgard, W. (2011). Accurate human motion capture in large areas by combining imu- and laser-based people tracking. In IROS, pages IEEE. 47

Motion Capture using Body Mounted Cameras in an Unknown Environment

Motion Capture using Body Mounted Cameras in an Unknown Environment Motion Capture using Body Mounted Cameras in an Unknown Environment Nam Vo Taeyoung Kim Siddharth Choudhary 1. The Problem Motion capture has been recently used to provide much of character motion in several

More information

CS 775: Advanced Computer Graphics. Lecture 17 : Motion Capture

CS 775: Advanced Computer Graphics. Lecture 17 : Motion Capture CS 775: Advanced Computer Graphics Lecture 17 : History Study of human motion Leonardo da Vinci (1452 1519) History Study of human motion Edward J. Muybridge, 1830 1904 http://en.wikipedia.org/wiki/eadweard_muybridge

More information

International Conference on Communication, Media, Technology and Design. ICCMTD May 2012 Istanbul - Turkey

International Conference on Communication, Media, Technology and Design. ICCMTD May 2012 Istanbul - Turkey VISUALIZING TIME COHERENT THREE-DIMENSIONAL CONTENT USING ONE OR MORE MICROSOFT KINECT CAMERAS Naveed Ahmed University of Sharjah Sharjah, United Arab Emirates Abstract Visualizing or digitization of the

More information

Multimodal Motion Capture Dataset TNT15

Multimodal Motion Capture Dataset TNT15 Multimodal Motion Capture Dataset TNT15 Timo v. Marcard, Gerard Pons-Moll, Bodo Rosenhahn January 2016 v1.2 1 Contents 1 Introduction 3 2 Technical Recording Setup 3 2.1 Video Data............................

More information

Articulated Pose Estimation with Flexible Mixtures-of-Parts

Articulated Pose Estimation with Flexible Mixtures-of-Parts Articulated Pose Estimation with Flexible Mixtures-of-Parts PRESENTATION: JESSE DAVIS CS 3710 VISUAL RECOGNITION Outline Modeling Special Cases Inferences Learning Experiments Problem and Relevance Problem:

More information

Personalization and Evaluation of a Real-time Depth-based Full Body Tracker

Personalization and Evaluation of a Real-time Depth-based Full Body Tracker Personalization and Evaluation of a Real-time Depth-based Full Body Tracker Thomas Helten 1 Andreas Baak 1 Gaurav Bharaj 2 Meinard Müller 3 Hans-Peter Seidel 1 Christian Theobalt 1 1 MPI Informatik 2 Harvard

More information

CS 775: Advanced Computer Graphics. Lecture 8 : Motion Capture

CS 775: Advanced Computer Graphics. Lecture 8 : Motion Capture CS 775: Advanced Computer Graphics Lecture 8 : History Study of human motion Leonardo da Vinci (1452 1519) History Study of human motion Edward J. Muybridge, 1830 1904 http://en.wikipedia.org/wiki/eadweard_muybridge

More information

Construction Progress Management and Interior Work Analysis Using Kinect 3D Image Sensors

Construction Progress Management and Interior Work Analysis Using Kinect 3D Image Sensors 33 rd International Symposium on Automation and Robotics in Construction (ISARC 2016) Construction Progress Management and Interior Work Analysis Using Kinect 3D Image Sensors Kosei Ishida 1 1 School of

More information

Walking gait dataset: point clouds, skeletons and silhouettes

Walking gait dataset: point clouds, skeletons and silhouettes Walking gait dataset: point clouds, skeletons and silhouettes Technical Report Number 1379 Trong-Nguyen Nguyen * and Jean Meunier DIRO, University of Montreal, Montreal, QC, Canada September 8, 2018 Abstract

More information

THREE PRE-PROCESSING STEPS TO INCREASE THE QUALITY OF KINECT RANGE DATA

THREE PRE-PROCESSING STEPS TO INCREASE THE QUALITY OF KINECT RANGE DATA THREE PRE-PROCESSING STEPS TO INCREASE THE QUALITY OF KINECT RANGE DATA M. Davoodianidaliki a, *, M. Saadatseresht a a Dept. of Surveying and Geomatics, Faculty of Engineering, University of Tehran, Tehran,

More information

Handheld scanning with ToF sensors and cameras

Handheld scanning with ToF sensors and cameras Handheld scanning with ToF sensors and cameras Enrico Cappelletto, Pietro Zanuttigh, Guido Maria Cortelazzo Dept. of Information Engineering, University of Padova enrico.cappelletto,zanuttigh,corte@dei.unipd.it

More information

A novel approach to motion tracking with wearable sensors based on Probabilistic Graphical Models

A novel approach to motion tracking with wearable sensors based on Probabilistic Graphical Models A novel approach to motion tracking with wearable sensors based on Probabilistic Graphical Models Emanuele Ruffaldi Lorenzo Peppoloni Alessandro Filippeschi Carlo Alberto Avizzano 2014 IEEE International

More information

Dense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm

Dense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm Computer Vision Group Prof. Daniel Cremers Dense Tracking and Mapping for Autonomous Quadrocopters Jürgen Sturm Joint work with Frank Steinbrücker, Jakob Engel, Christian Kerl, Erik Bylow, and Daniel Cremers

More information

Key Developments in Human Pose Estimation for Kinect

Key Developments in Human Pose Estimation for Kinect Key Developments in Human Pose Estimation for Kinect Pushmeet Kohli and Jamie Shotton Abstract The last few years have seen a surge in the development of natural user interfaces. These interfaces do not

More information

A multi-camera positioning system for steering of a THz stand-off scanner

A multi-camera positioning system for steering of a THz stand-off scanner A multi-camera positioning system for steering of a THz stand-off scanner Maria Axelsson, Mikael Karlsson and Staffan Rudner Swedish Defence Research Agency, Box 1165, SE-581 11 Linköping, SWEDEN ABSTRACT

More information

Human Body Recognition and Tracking: How the Kinect Works. Kinect RGB-D Camera. What the Kinect Does. How Kinect Works: Overview

Human Body Recognition and Tracking: How the Kinect Works. Kinect RGB-D Camera. What the Kinect Does. How Kinect Works: Overview Human Body Recognition and Tracking: How the Kinect Works Kinect RGB-D Camera Microsoft Kinect (Nov. 2010) Color video camera + laser-projected IR dot pattern + IR camera $120 (April 2012) Kinect 1.5 due

More information

Accurate 3D Face and Body Modeling from a Single Fixed Kinect

Accurate 3D Face and Body Modeling from a Single Fixed Kinect Accurate 3D Face and Body Modeling from a Single Fixed Kinect Ruizhe Wang*, Matthias Hernandez*, Jongmoo Choi, Gérard Medioni Computer Vision Lab, IRIS University of Southern California Abstract In this

More information

A Model-based Approach to Rapid Estimation of Body Shape and Postures Using Low-Cost Depth Cameras

A Model-based Approach to Rapid Estimation of Body Shape and Postures Using Low-Cost Depth Cameras A Model-based Approach to Rapid Estimation of Body Shape and Postures Using Low-Cost Depth Cameras Abstract Byoung-Keon D. PARK*, Matthew P. REED University of Michigan, Transportation Research Institute,

More information

Mobile Point Fusion. Real-time 3d surface reconstruction out of depth images on a mobile platform

Mobile Point Fusion. Real-time 3d surface reconstruction out of depth images on a mobile platform Mobile Point Fusion Real-time 3d surface reconstruction out of depth images on a mobile platform Aaron Wetzler Presenting: Daniel Ben-Hoda Supervisors: Prof. Ron Kimmel Gal Kamar Yaron Honen Supported

More information

VIRTUAL TRAIL ROOM. South Asian Journal of Engineering and Technology Vol.3, No.5 (2017) 87 96

VIRTUAL TRAIL ROOM. South Asian Journal of Engineering and Technology Vol.3, No.5 (2017) 87 96 VIRTUAL TRAIL ROOM 1 Vipin Paul, 2 Sanju Abel J., 3 Sudharsan S., 4 Praveen M. 1 Vipinpaul95@gmail.com 3 shansudharsan002@gmail.com 2 sanjuabel@gmail.com 4 praveen.pravin6@gmail.com Department of computer

More information

Wide Area Motion Tracking Using Consumer Hardware

Wide Area Motion Tracking Using Consumer Hardware Wide Area Motion Tracking Using Consumer Hardware Christian Schönauer schoenauer@ims.tuwien.ac.at Hannes Kaufmann kaufmann@ims.tuwien.ac.at Interactive Media Systems Group Institute of Software Technology

More information

Real-Time Human Pose Recognition in Parts from Single Depth Images

Real-Time Human Pose Recognition in Parts from Single Depth Images Real-Time Human Pose Recognition in Parts from Single Depth Images Jamie Shotton, Andrew Fitzgibbon, Mat Cook, Toby Sharp, Mark Finocchio, Richard Moore, Alex Kipman, Andrew Blake CVPR 2011 PRESENTER:

More information

SCAPE: Shape Completion and Animation of People

SCAPE: Shape Completion and Animation of People SCAPE: Shape Completion and Animation of People By Dragomir Anguelov, Praveen Srinivasan, Daphne Koller, Sebastian Thrun, Jim Rodgers, James Davis From SIGGRAPH 2005 Presentation for CS468 by Emilio Antúnez

More information

Articulated Gaussian Kernel Correlation for Human Pose Estimation

Articulated Gaussian Kernel Correlation for Human Pose Estimation Articulated Gaussian Kernel Correlation for Human Pose Estimation Meng Ding and Guoliang Fan School of Electrical and Computer Engineering, Oklahoma State University, Stillwater, OK, USA 74074 meng.ding@okstate.edu;

More information

Markerless human motion capture through visual hull and articulated ICP

Markerless human motion capture through visual hull and articulated ICP Markerless human motion capture through visual hull and articulated ICP Lars Mündermann lmuender@stanford.edu Stefano Corazza Stanford, CA 93405 stefanoc@stanford.edu Thomas. P. Andriacchi Bone and Joint

More information

USAGE OF MICROSOFT KINECT FOR AUGMENTED PROTOTYPING SPEED-UP

USAGE OF MICROSOFT KINECT FOR AUGMENTED PROTOTYPING SPEED-UP ACTA UNIVERSITATIS AGRICULTURAE ET SILVICULTURAE MENDELIANAE BRUNENSIS Volume LX 23 Number 2, 2012 USAGE OF MICROSOFT KINECT FOR AUGMENTED PROTOTYPING SPEED-UP J. Landa, D. Procházka Received: November

More information

A Virtual Dressing Room Using Kinect

A Virtual Dressing Room Using Kinect 2017 IJSRST Volume 3 Issue 3 Print ISSN: 2395-6011 Online ISSN: 2395-602X Themed Section: Science and Technology A Virtual Dressing Room Using Kinect Jagtap Prajakta Bansidhar, Bhole Sheetal Hiraman, Mate

More information

Rapid Skin: Estimating the 3D Human Pose and Shape in Real-Time

Rapid Skin: Estimating the 3D Human Pose and Shape in Real-Time Rapid Skin: Estimating the 3D Human Pose and Shape in Real-Time Matthias Straka, Stefan Hauswiesner, Matthias Rüther, and Horst Bischof Institute for Computer Graphics and Vision Graz University of Technology,

More information

Development of a Fall Detection System with Microsoft Kinect

Development of a Fall Detection System with Microsoft Kinect Development of a Fall Detection System with Microsoft Kinect Christopher Kawatsu, Jiaxing Li, and C.J. Chung Department of Mathematics and Computer Science, Lawrence Technological University, 21000 West

More information

3D Reconstruction of Human Bodies with Clothes from Un-calibrated Monocular Video Images

3D Reconstruction of Human Bodies with Clothes from Un-calibrated Monocular Video Images 3D Reconstruction of Human Bodies with Clothes from Un-calibrated Monocular Video Images presented by Tran Cong Thien Qui PhD Candidate School of Computer Engineering & Institute for Media Innovation Supervisor:

More information

Computer Animation and Visualisation. Lecture 3. Motion capture and physically-based animation of characters

Computer Animation and Visualisation. Lecture 3. Motion capture and physically-based animation of characters Computer Animation and Visualisation Lecture 3. Motion capture and physically-based animation of characters Character Animation There are three methods Create them manually Use real human / animal motions

More information

Mesh from Depth Images Using GR 2 T

Mesh from Depth Images Using GR 2 T Mesh from Depth Images Using GR 2 T Mairead Grogan & Rozenn Dahyot School of Computer Science and Statistics Trinity College Dublin Dublin, Ireland mgrogan@tcd.ie, Rozenn.Dahyot@tcd.ie www.scss.tcd.ie/

More information

Research Article Human Model Adaptation for Multiview Markerless Motion Capture

Research Article Human Model Adaptation for Multiview Markerless Motion Capture Mathematical Problems in Engineering Volume 2013, Article ID 564214, 7 pages http://dx.doi.org/10.1155/2013/564214 Research Article Human Model Adaptation for Multiview Markerless Motion Capture Dianyong

More information

Human Pose Estimation in Stereo Images

Human Pose Estimation in Stereo Images Human Pose Estimation in Stereo Images Joe Lallemand 1,2, Magdalena Szczot 1, and Slobodan Ilic 2 1 BMW Group, Munich, Germany {joe.lallemand, magdalena.szczot}@bmw.de http://www.bmw.de 2 Computer Aided

More information

Data-driven Approaches to Simulation (Motion Capture)

Data-driven Approaches to Simulation (Motion Capture) 1 Data-driven Approaches to Simulation (Motion Capture) Ting-Chun Sun tingchun.sun@usc.edu Preface The lecture slides [1] are made by Jessica Hodgins [2], who is a professor in Computer Science Department

More information

HandSonor: A Customizable Vision-based Control Interface for Musical Expression

HandSonor: A Customizable Vision-based Control Interface for Musical Expression HandSonor: A Customizable Vision-based Control Interface for Musical Expression Srinath Sridhar MPI Informatik and Universita t des Saarlandes Campus E1.4, 66123 Saarbru cken, Germany ssridhar@mpi-inf.mpg.de

More information

Motion Capture & Simulation

Motion Capture & Simulation Motion Capture & Simulation Motion Capture Character Reconstructions Joint Angles Need 3 points to compute a rigid body coordinate frame 1 st point gives 3D translation, 2 nd point gives 2 angles, 3 rd

More information

Virtual Production for the Real World Using Autodesk MotionBuilder 2013

Virtual Production for the Real World Using Autodesk MotionBuilder 2013 Virtual Production for the Real World Using Autodesk MotionBuilder 2013 Daryl Obert Autodesk Hein Beute- Xsens DG3148 This class will give attendees a firm understanding of the concepts and workflows involved

More information

Leveraging Two Kinect Sensors for Accurate Full-Body Motion Capture

Leveraging Two Kinect Sensors for Accurate Full-Body Motion Capture Sensors 205, 5, 24297-2437; doi:0.3390/s50924297 OPEN ACCESS sensors ISSN 424-8220 www.mdpi.com/journal/sensors Article Leveraging Two Kinect Sensors for Accurate Full-Body Motion Capture Zhiquan Gao,

More information

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Structured Light II Johannes Köhler Johannes.koehler@dfki.de Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Introduction Previous lecture: Structured Light I Active Scanning Camera/emitter

More information

3D Human Motion Analysis and Manifolds

3D Human Motion Analysis and Manifolds D E P A R T M E N T O F C O M P U T E R S C I E N C E U N I V E R S I T Y O F C O P E N H A G E N 3D Human Motion Analysis and Manifolds Kim Steenstrup Pedersen DIKU Image group and E-Science center Motivation

More information

A Data-Driven Approach for 3D Human Body Pose Reconstruction from a Kinect Sensor

A Data-Driven Approach for 3D Human Body Pose Reconstruction from a Kinect Sensor Journal of Physics: Conference Series PAPER OPEN ACCESS A Data-Driven Approach for 3D Human Body Pose Reconstruction from a Kinect Sensor To cite this article: Li Yao et al 2018 J. Phys.: Conf. Ser. 1098

More information

Data-driven Depth Inference from a Single Still Image

Data-driven Depth Inference from a Single Still Image Data-driven Depth Inference from a Single Still Image Kyunghee Kim Computer Science Department Stanford University kyunghee.kim@stanford.edu Abstract Given an indoor image, how to recover its depth information

More information

PART IV: RS & the Kinect

PART IV: RS & the Kinect Computer Vision on Rolling Shutter Cameras PART IV: RS & the Kinect Per-Erik Forssén, Erik Ringaby, Johan Hedborg Computer Vision Laboratory Dept. of Electrical Engineering Linköping University Tutorial

More information

Virtualized Reality Using Depth Camera Point Clouds

Virtualized Reality Using Depth Camera Point Clouds Virtualized Reality Using Depth Camera Point Clouds Jordan Cazamias Stanford University jaycaz@stanford.edu Abhilash Sunder Raj Stanford University abhisr@stanford.edu Abstract We explored various ways

More information

Background subtraction in people detection framework for RGB-D cameras

Background subtraction in people detection framework for RGB-D cameras Background subtraction in people detection framework for RGB-D cameras Anh-Tuan Nghiem, Francois Bremond INRIA-Sophia Antipolis 2004 Route des Lucioles, 06902 Valbonne, France nghiemtuan@gmail.com, Francois.Bremond@inria.fr

More information

Outdoor Scene Reconstruction from Multiple Image Sequences Captured by a Hand-held Video Camera

Outdoor Scene Reconstruction from Multiple Image Sequences Captured by a Hand-held Video Camera Outdoor Scene Reconstruction from Multiple Image Sequences Captured by a Hand-held Video Camera Tomokazu Sato, Masayuki Kanbara and Naokazu Yokoya Graduate School of Information Science, Nara Institute

More information

CMU Facilities. Motion Capture Lab. Panoptic Studio

CMU Facilities. Motion Capture Lab. Panoptic Studio CMU Facilities Motion Capture Lab The 1700 square foot Motion Capture Lab provides a resource for behavior capture of humans as well as measuring and controlling robot behavior in real time. It includes

More information

Kinect Cursor Control EEE178 Dr. Fethi Belkhouche Christopher Harris Danny Nguyen I. INTRODUCTION

Kinect Cursor Control EEE178 Dr. Fethi Belkhouche Christopher Harris Danny Nguyen I. INTRODUCTION Kinect Cursor Control EEE178 Dr. Fethi Belkhouche Christopher Harris Danny Nguyen Abstract: An XBOX 360 Kinect is used to develop two applications to control the desktop cursor of a Windows computer. Application

More information

Applications. Systems. Motion capture pipeline. Biomechanical analysis. Graphics research

Applications. Systems. Motion capture pipeline. Biomechanical analysis. Graphics research Motion capture Applications Systems Motion capture pipeline Biomechanical analysis Graphics research Applications Computer animation Biomechanics Robotics Cinema Video games Anthropology What is captured?

More information

Lecture 19: Depth Cameras. Visual Computing Systems CMU , Fall 2013

Lecture 19: Depth Cameras. Visual Computing Systems CMU , Fall 2013 Lecture 19: Depth Cameras Visual Computing Systems Continuing theme: computational photography Cameras capture light, then extensive processing produces the desired image Today: - Capturing scene depth

More information

Object Reconstruction

Object Reconstruction B. Scholz Object Reconstruction 1 / 39 MIN-Fakultät Fachbereich Informatik Object Reconstruction Benjamin Scholz Universität Hamburg Fakultät für Mathematik, Informatik und Naturwissenschaften Fachbereich

More information

Intrinsic3D: High-Quality 3D Reconstruction by Joint Appearance and Geometry Optimization with Spatially-Varying Lighting

Intrinsic3D: High-Quality 3D Reconstruction by Joint Appearance and Geometry Optimization with Spatially-Varying Lighting Intrinsic3D: High-Quality 3D Reconstruction by Joint Appearance and Geometry Optimization with Spatially-Varying Lighting R. Maier 1,2, K. Kim 1, D. Cremers 2, J. Kautz 1, M. Nießner 2,3 Fusion Ours 1

More information

Tri-modal Human Body Segmentation

Tri-modal Human Body Segmentation Tri-modal Human Body Segmentation Master of Science Thesis Cristina Palmero Cantariño Advisor: Sergio Escalera Guerrero February 6, 2014 Outline 1 Introduction 2 Tri-modal dataset 3 Proposed baseline 4

More information

Segmentation and Tracking of Partial Planar Templates

Segmentation and Tracking of Partial Planar Templates Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract

More information

FOREGROUND DETECTION ON DEPTH MAPS USING SKELETAL REPRESENTATION OF OBJECT SILHOUETTES

FOREGROUND DETECTION ON DEPTH MAPS USING SKELETAL REPRESENTATION OF OBJECT SILHOUETTES FOREGROUND DETECTION ON DEPTH MAPS USING SKELETAL REPRESENTATION OF OBJECT SILHOUETTES D. Beloborodov a, L. Mestetskiy a a Faculty of Computational Mathematics and Cybernetics, Lomonosov Moscow State University,

More information

Jamie Shotton, Andrew Fitzgibbon, Mat Cook, Toby Sharp, Mark Finocchio, Richard Moore, Alex Kipman, Andrew Blake CVPR 2011

Jamie Shotton, Andrew Fitzgibbon, Mat Cook, Toby Sharp, Mark Finocchio, Richard Moore, Alex Kipman, Andrew Blake CVPR 2011 Jamie Shotton, Andrew Fitzgibbon, Mat Cook, Toby Sharp, Mark Finocchio, Richard Moore, Alex Kipman, Andrew Blake CVPR 2011 Auto-initialize a tracking algorithm & recover from failures All human poses,

More information

High-speed Three-dimensional Mapping by Direct Estimation of a Small Motion Using Range Images

High-speed Three-dimensional Mapping by Direct Estimation of a Small Motion Using Range Images MECATRONICS - REM 2016 June 15-17, 2016 High-speed Three-dimensional Mapping by Direct Estimation of a Small Motion Using Range Images Shinta Nozaki and Masashi Kimura School of Science and Engineering

More information

VNect: Real-time 3D Human Pose Estimation with a Single RGB Camera

VNect: Real-time 3D Human Pose Estimation with a Single RGB Camera VNect: Real-time 3D Human Pose Estimation with a Single RGB Camera By Dushyant Mehta, Srinath Sridhar, Oleksandr Sotnychenko, Helge Rhodin, Mohammad Shafiei, Hans-Peter Seidel, Weipeng Xu, Dan Casas, Christian

More information

Colored Point Cloud Registration Revisited Supplementary Material

Colored Point Cloud Registration Revisited Supplementary Material Colored Point Cloud Registration Revisited Supplementary Material Jaesik Park Qian-Yi Zhou Vladlen Koltun Intel Labs A. RGB-D Image Alignment Section introduced a joint photometric and geometric objective

More information

What have we leaned so far?

What have we leaned so far? What have we leaned so far? Camera structure Eye structure Project 1: High Dynamic Range Imaging What have we learned so far? Image Filtering Image Warping Camera Projection Model Project 2: Panoramic

More information

Motion capture: An evaluation of Kinect V2 body tracking for upper limb motion analysis

Motion capture: An evaluation of Kinect V2 body tracking for upper limb motion analysis Motion capture: An evaluation of Kinect V2 body tracking for upper limb motion analysis Silvio Giancola 1, Andrea Corti 1, Franco Molteni 2, Remo Sala 1 1 Vision Bricks Laboratory, Mechanical Departement,

More information

Human body animation. Computer Animation. Human Body Animation. Skeletal Animation

Human body animation. Computer Animation. Human Body Animation. Skeletal Animation Computer Animation Aitor Rovira March 2010 Human body animation Based on slides by Marco Gillies Human Body Animation Skeletal Animation Skeletal Animation (FK, IK) Motion Capture Motion Editing (retargeting,

More information

Mirror Based Framework for Human Body Measurement

Mirror Based Framework for Human Body Measurement 362 Mirror Based Framework for Human Body Measurement 1 Takeshi Hashimoto, 2 Takayuki Suzuki, 3 András Rövid 1 Dept. of Electrical and Electronics Engineering, Shizuoka University 5-1, 3-chome Johoku,

More information

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University. 3D Computer Vision Structured Light II Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de 1 Introduction

More information

Dynamic Human Shape Description and Characterization

Dynamic Human Shape Description and Characterization Dynamic Human Shape Description and Characterization Z. Cheng*, S. Mosher, Jeanne Smith H. Cheng, and K. Robinette Infoscitex Corporation, Dayton, Ohio, USA 711 th Human Performance Wing, Air Force Research

More information

Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera

Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera Tomokazu Satoy, Masayuki Kanbaray, Naokazu Yokoyay and Haruo Takemuraz ygraduate School of Information

More information

3D Editing System for Captured Real Scenes

3D Editing System for Captured Real Scenes 3D Editing System for Captured Real Scenes Inwoo Ha, Yong Beom Lee and James D.K. Kim Samsung Advanced Institute of Technology, Youngin, South Korea E-mail: {iw.ha, leey, jamesdk.kim}@samsung.com Tel:

More information

When Can We Use KinectFusion for Ground Truth Acquisition?

When Can We Use KinectFusion for Ground Truth Acquisition? When Can We Use KinectFusion for Ground Truth Acquisition? Stephan Meister 1, Shahram Izadi 2, Pushmeet Kohli 3, Martin Hämmerle 4, Carsten Rother 5 and Daniel Kondermann 6 Abstract KinectFusion is a method

More information

Dense 3D Reconstruction from Autonomous Quadrocopters

Dense 3D Reconstruction from Autonomous Quadrocopters Dense 3D Reconstruction from Autonomous Quadrocopters Computer Science & Mathematics TU Munich Martin Oswald, Jakob Engel, Christian Kerl, Frank Steinbrücker, Jan Stühmer & Jürgen Sturm Autonomous Quadrocopters

More information

Model-based Visual Tracking:

Model-based Visual Tracking: Technische Universität München Model-based Visual Tracking: the OpenTL framework Giorgio Panin Technische Universität München Institut für Informatik Lehrstuhl für Echtzeitsysteme und Robotik (Prof. Alois

More information

A Low Power, High Throughput, Fully Event-Based Stereo System: Supplementary Documentation

A Low Power, High Throughput, Fully Event-Based Stereo System: Supplementary Documentation A Low Power, High Throughput, Fully Event-Based Stereo System: Supplementary Documentation Alexander Andreopoulos, Hirak J. Kashyap, Tapan K. Nayak, Arnon Amir, Myron D. Flickner IBM Research March 25,

More information

Visual Perception for Robots

Visual Perception for Robots Visual Perception for Robots Sven Behnke Computer Science Institute VI Autonomous Intelligent Systems Our Cognitive Robots Complete systems for example scenarios Equipped with rich sensors Flying robot

More information

Recovering 6D Object Pose and Predicting Next-Best-View in the Crowd Supplementary Material

Recovering 6D Object Pose and Predicting Next-Best-View in the Crowd Supplementary Material Recovering 6D Object Pose and Predicting Next-Best-View in the Crowd Supplementary Material Andreas Doumanoglou 1,2, Rigas Kouskouridas 1, Sotiris Malassiotis 2, Tae-Kyun Kim 1 1 Imperial College London

More information

Supplemental Material: Detailed, accurate, human shape estimation from clothed 3D scan sequences

Supplemental Material: Detailed, accurate, human shape estimation from clothed 3D scan sequences Supplemental Material: Detailed, accurate, human shape estimation from clothed 3D scan sequences Chao Zhang 1,2, Sergi Pujades 1, Michael Black 1, and Gerard Pons-Moll 1 1 MPI for Intelligent Systems,

More information

Combining PGMs and Discriminative Models for Upper Body Pose Detection

Combining PGMs and Discriminative Models for Upper Body Pose Detection Combining PGMs and Discriminative Models for Upper Body Pose Detection Gedas Bertasius May 30, 2014 1 Introduction In this project, I utilized probabilistic graphical models together with discriminative

More information

Rigid ICP registration with Kinect

Rigid ICP registration with Kinect Rigid ICP registration with Kinect Students: Yoni Choukroun, Elie Semmel Advisor: Yonathan Aflalo 1 Overview.p.3 Development of the project..p.3 Papers p.4 Project algorithm..p.6 Result of the whole body.p.7

More information

ENGN2911I: 3D Photography and Geometry Processing Assignment 1: 3D Photography using Planar Shadows

ENGN2911I: 3D Photography and Geometry Processing Assignment 1: 3D Photography using Planar Shadows ENGN2911I: 3D Photography and Geometry Processing Assignment 1: 3D Photography using Planar Shadows Instructor: Gabriel Taubin Assignment written by: Douglas Lanman 29 January 2009 Figure 1: 3D Photography

More information

CS 231. Inverse Kinematics Intro to Motion Capture

CS 231. Inverse Kinematics Intro to Motion Capture CS 231 Inverse Kinematics Intro to Motion Capture Representation 1) Skeleton Origin (root) Joint centers/ bones lengths 2) Keyframes Pos/Rot Root (x) Joint Angles (q) 3D characters Kinematics study of

More information

Kinect Device. How the Kinect Works. Kinect Device. What the Kinect does 4/27/16. Subhransu Maji Slides credit: Derek Hoiem, University of Illinois

Kinect Device. How the Kinect Works. Kinect Device. What the Kinect does 4/27/16. Subhransu Maji Slides credit: Derek Hoiem, University of Illinois 4/27/16 Kinect Device How the Kinect Works T2 Subhransu Maji Slides credit: Derek Hoiem, University of Illinois Photo frame-grabbed from: http://www.blisteredthumbs.net/2010/11/dance-central-angry-review

More information

Efficient SLAM Scheme Based ICP Matching Algorithm Using Image and Laser Scan Information

Efficient SLAM Scheme Based ICP Matching Algorithm Using Image and Laser Scan Information Proceedings of the World Congress on Electrical Engineering and Computer Systems and Science (EECSS 2015) Barcelona, Spain July 13-14, 2015 Paper No. 335 Efficient SLAM Scheme Based ICP Matching Algorithm

More information

Fully Automatic Endoscope Calibration for Intraoperative Use

Fully Automatic Endoscope Calibration for Intraoperative Use Fully Automatic Endoscope Calibration for Intraoperative Use Christian Wengert, Mireille Reeff, Philippe C. Cattin, Gábor Székely Computer Vision Laboratory, ETH Zurich, 8092 Zurich, Switzerland {wengert,

More information

Pose Estimation on Depth Images with Convolutional Neural Network

Pose Estimation on Depth Images with Convolutional Neural Network Pose Estimation on Depth Images with Convolutional Neural Network Jingwei Huang Stanford University jingweih@stanford.edu David Altamar Stanford University daltamar@stanford.edu Abstract In this paper

More information

Animation. CS 465 Lecture 22

Animation. CS 465 Lecture 22 Animation CS 465 Lecture 22 Animation Industry production process leading up to animation What animation is How animation works (very generally) Artistic process of animation Further topics in how it works

More information

Universiteit Leiden Computer Science

Universiteit Leiden Computer Science Universiteit Leiden Computer Science Optimizing octree updates for visibility determination on dynamic scenes Name: Hans Wortel Student-no: 0607940 Date: 28/07/2011 1st supervisor: Dr. Michael Lew 2nd

More information

Light Field Occlusion Removal

Light Field Occlusion Removal Light Field Occlusion Removal Shannon Kao Stanford University kaos@stanford.edu Figure 1: Occlusion removal pipeline. The input image (left) is part of a focal stack representing a light field. Each image

More information

Research Article Motion Control of Robot by using Kinect Sensor

Research Article Motion Control of Robot by using Kinect Sensor Research Journal of Applied Sciences, Engineering and Technology 8(11): 1384-1388, 2014 DOI:10.19026/rjaset.8.1111 ISSN: 2040-7459; e-issn: 2040-7467 2014 Maxwell Scientific Publication Corp. Submitted:

More information

Real Time Person Detection and Tracking by Mobile Robots using RGB-D Images

Real Time Person Detection and Tracking by Mobile Robots using RGB-D Images Real Time Person Detection and Tracking by Mobile Robots using RGB-D Images Duc My Vo, Lixing Jiang and Andreas Zell Abstract Detecting and tracking humans are key problems for human-robot interaction.

More information

Visual Perception Sensors

Visual Perception Sensors G. Glaser Visual Perception Sensors 1 / 27 MIN Faculty Department of Informatics Visual Perception Sensors Depth Determination Gerrit Glaser University of Hamburg Faculty of Mathematics, Informatics and

More information

Task analysis based on observing hands and objects by vision

Task analysis based on observing hands and objects by vision Task analysis based on observing hands and objects by vision Yoshihiro SATO Keni Bernardin Hiroshi KIMURA Katsushi IKEUCHI Univ. of Electro-Communications Univ. of Karlsruhe Univ. of Tokyo Abstract In

More information

Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview

Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview Arun Das 05/09/2017 Arun Das Waterloo Autonomous Vehicles Lab Introduction What s in a name? Arun Das Waterloo Autonomous

More information

Measurement of 3D Foot Shape Deformation in Motion

Measurement of 3D Foot Shape Deformation in Motion Measurement of 3D Foot Shape Deformation in Motion Makoto Kimura Masaaki Mochimaru Takeo Kanade Digital Human Research Center National Institute of Advanced Industrial Science and Technology, Japan The

More information

Automatic Generation of Animatable 3D Personalized Model Based on Multi-view Images

Automatic Generation of Animatable 3D Personalized Model Based on Multi-view Images Automatic Generation of Animatable 3D Personalized Model Based on Multi-view Images Seong-Jae Lim, Ho-Won Kim, Jin Sung Choi CG Team, Contents Division ETRI Daejeon, South Korea sjlim@etri.re.kr Bon-Ki

More information

Fully Automatic Methodology for Human Action Recognition Incorporating Dynamic Information

Fully Automatic Methodology for Human Action Recognition Incorporating Dynamic Information Fully Automatic Methodology for Human Action Recognition Incorporating Dynamic Information Ana González, Marcos Ortega Hortas, and Manuel G. Penedo University of A Coruña, VARPA group, A Coruña 15071,

More information

3D Modeling of Objects Using Laser Scanning

3D Modeling of Objects Using Laser Scanning 1 3D Modeling of Objects Using Laser Scanning D. Jaya Deepu, LPU University, Punjab, India Email: Jaideepudadi@gmail.com Abstract: In the last few decades, constructing accurate three-dimensional models

More information

CHAPTER 3 DISPARITY AND DEPTH MAP COMPUTATION

CHAPTER 3 DISPARITY AND DEPTH MAP COMPUTATION CHAPTER 3 DISPARITY AND DEPTH MAP COMPUTATION In this chapter we will discuss the process of disparity computation. It plays an important role in our caricature system because all 3D coordinates of nodes

More information

Cloth Animation. CENG 732 Computer Animation. Simple Draping. Simple Draping. Simple Draping. Simple Draping

Cloth Animation. CENG 732 Computer Animation. Simple Draping. Simple Draping. Simple Draping. Simple Draping Cloth Animation CENG 732 Computer Animation Cloth animation in Blender Cloth animation in Maya Spring 2006-2007 Week 9 Animating Cloth Motion Capture Draping will occur as a cloth is hanged from a fixed

More information

Animations. Hakan Bilen University of Edinburgh. Computer Graphics Fall Some slides are courtesy of Steve Marschner and Kavita Bala

Animations. Hakan Bilen University of Edinburgh. Computer Graphics Fall Some slides are courtesy of Steve Marschner and Kavita Bala Animations Hakan Bilen University of Edinburgh Computer Graphics Fall 2017 Some slides are courtesy of Steve Marschner and Kavita Bala Animation Artistic process What are animators trying to do? What tools

More information

Visualization of Temperature Change using RGB-D Camera and Thermal Camera

Visualization of Temperature Change using RGB-D Camera and Thermal Camera 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 Visualization of Temperature

More information

The Kinect Sensor. Luís Carriço FCUL 2014/15

The Kinect Sensor. Luís Carriço FCUL 2014/15 Advanced Interaction Techniques The Kinect Sensor Luís Carriço FCUL 2014/15 Sources: MS Kinect for Xbox 360 John C. Tang. Using Kinect to explore NUI, Ms Research, From Stanford CS247 Shotton et al. Real-Time

More information