A study of a multi-kinect system for human body scanning

Size: px
Start display at page:

Download "A study of a multi-kinect system for human body scanning"

Transcription

1 A study of a multi-kinect system for human body scanning A Seminar Report Submitted in partial fulfillment of requirements for the degree of Master of Technology by Shashwat Rohilla Roll No: under the guidance of Prof. Parag Chaudhuri Department of Computer Science and Engineering IIT Bombay April, 2014

2 Contents 1 Introduction Motivation What is Kinect? Earlier Work Stages of Modeling ICP for Point Cloud Registration Four-Kinect system Calibration Self Calibration Mapping depth to RGB camera Calibration between multiple Kinects Conclusion and Future Scope Limitations Future Work Code Kinect Stand Multi-Kinect Calibration using Hand-Waving Acknowledgement 13 1

3 Abstract 3D models are useful in many areas such as gaming, animation, simulations, virtual worlds etc. Simple objects can be manually created using tools like Blender. But to model a person, data is collected by scanning one. Scanning is not an easy process. 3D scanning techniques like laser scan are fast and accurate but costly, bulky and require expert knowledge. Computer Vision techniques using 2D camera like structure from motion and shapes from shading are not robust. We present a method to scan a person using Microsoft s Kinect. The main reason to choose Kinect is its portability and cheap cost. Currently, we are able to generate 3D model of a person. A person is made to stand in front of the cameras and turn around gradually to have his full body captured. We are getting reasonable output with some noise. But the whole procedure is quite tedious. It is not feasible for a layman to use the setup. This report aims at refining the previous work.

4 Chapter 1 Introduction 1.1 Motivation We aim to build a system that can create a 3D model of a person. Creating such a model in computer is a tough task. Rather, we take structural data of the person (by taking images) and then process them to get the model. There are several techniques for this. One of them is 3D laser scanning [9]. In this, an emitter emits some kind of light which is detected for its reflection or radiation passing through. One of the technique under this is Timeof-flight. In this, the round-trip time of the laser is detected. Knowing the speed of light, c, and the time, one can detect the distance. It is fast, accurate and measures distances on the order of kilometres; but at the cost of price. We don t need such qualities to model a humanoid. Passive techniques [9] are those which does not involve any emission of radiation. They just rely on the natural incident light. Usual 2D camera can be used to take certain images. Stereoscopic systems: It requires two images taken from cameras slightly apart from each other. The depth is calculated by exploiting the slight difference in both the images. Our eye works on the same prinicple. Photometric systems: It uses a single camera but multiple lighting conditions to determine shapes from shading. These techniques are too sensitive to lighting conditions and require lots of computation. To overcome these problems, we propose to use Microsoft s Kinect. Kinect is a 3D camera which calculates depth by using the principle of structured light. The IR projector projects some pattern and the IR camera notices its change from a known pattern to calculate the disparity and hence depth. 1.2 What is Kinect? Kinect 1.1 is a 3D camera that makes use of infra-red (IR) illumination to obtain depth data, color images and sound. It is devised by Microsoft for XBox 360. Later Primesense [6] (the company which designed Kinect) also released open source drivers. 1

5 The system can measure distance with a 1cm accuracy at 2m and has a resolution of 3mm at 2m. The default RGB video stream uses 8-bit VGA resolution ( pixels), but the hardware is capable of resolutions up to (at a lower frame rate). The monochrome depth sensing video stream is in VGA resolution ( pixels) with 11-bit depth, which provides 2,048 levels of sensitivity. Kinect has an IR camera, RGB camera and a laser based IR projector as shown in the figure above. The IR camera and the IR projector forms a stereo set with a baseline of about 7.5cms. The IR projector projects a known pattern which is read by the IR camera. The change in the pattern is observed to calculate the depth of a pixel. As the IR camera does not interfere with the visual spectrum it enables simultaneous capture of texture images. We are using PCL library [5] with libfreenect [8] opensource drivers for Kinect. Figure 1.1: Source: In the next chapter, we will discuss in brief the work done by [2]. We will discuss the process of gathering point clouds and creating model from them. Later, we will talk about calibration because that is something which can be improved considerably. Then, in chapter 4, we wil discuss the limitations of the current system and some ideas to overcome them. 2

6 Chapter 2 Earlier Work Currently, we are able to model a person with good accuracy [1] [2] [3]. Following are the stages of pipeline which we have to go through. 2.1 Stages of Modeling The process of scanning and modeling a human using Kinect has the following stages: Calibrate: The cameras are first calibrated to get their positions and orientations with respect to each other. This will be explained later. Capture: The person stands in front of the cameras and 3D images are taken. It is made sure that the whole body is covered so that there are no holes when the point cloud is generated. Construct Point Cloud: A point cloud is a set of voxels which are captured in the previous stage. To construct it, we back project every pixel in the image. Since we know the depth at each pixel, we know exactly the 3D position of the point. Segmentation: The point cloud constructed above not only contains the person, but also the background. The background is assumed to be quite far so that it becomes easier to segment it out on the basis of its depth value. To make it more rigorous, the person is assumed to have depth values within a certain band. Remove Noise: Depth sensors such as Kinect are prone to measurement errors that lead to sparse outliers which may corrupt the registration process. Hence, these outliers must be removed from the point clouds before the registration process. Register Point Clouds: We now have many point clouds of the same person captured from different viewpoints. These point clouds are registered using ICP algorithm. Meshing and Texture Mapping: Meshes are created out of the generated point clouds to have a smooth surface. After that, texture is mapped from the original point cloud. 3

7 2.2 ICP for Point Cloud Registration Iterative Closest Point (ICP) is an algorithm employed to minimize the difference between two clouds of points. ICP is often used to reconstruct 2D or 3D surfaces from different scans, to localize robots and achieve optimal path planning (especially when wheel odometry is unreliable due to slippery terrain), to co-register bone models, etc. In the algorithm, one point cloud, the reference, or target, is kept fixed, while the other one, the source, is transformed to best match the reference. The algorithm iteratively revises the transformation (combination of translation and rotation) needed to minimize the distance from the source to the reference point cloud. Inputs: reference and source point clouds, initial estimation of the transformation to align the source to the reference (optional), criteria for stopping the iterations. Output: refined transformation. Essentially, the algorithm steps are: 1. For each point in the source point cloud, find the closest point in the reference point cloud. 2. Estimate the combination of rotation and translation using a mean squared error cost function that will best align each source point to its match found in the previous step. 3. Transform the source points using the obtained transformation. 4. Iterate (re-associate the points, and so on). Source: closest point 2.3 Four-Kinect system In the beginning, the idea was to use a single Kinect and move it around the model. Infact, the person can be asked to rotate if scanning process is not continuous. ICP is commonly used to minimize rigid body transformations of two clouds. But the problem with this is that it fails to give a complete and a closed model because of accumulation of errors in the successive frames. Also, ICP is time-consuming. Suppose, we roughly know by how much amount the person has rotated in successive frames. Then we can get a coarse estimation of the initial position of the point cloud. This method gave reasonable results. The problem with the single Kinect is that the point clouds are blurry and of low quality. Hence, we switched to a multi-kinect system [2]. We used four kinect cameras - two pairs on the opposite sides. This four-kinect system 2.1 gives more consistent and smoother models as compared to the previous one. Currently, the setup which we have uses four Kinects as shown below. 4

8 Figure 2.1: Four-Kinect System [2] In our multi-kinect system, one big problem is calibration. We are getting accurate results but the process of calibration is quite complex. In the next chapter, we will discuss different types of calibrations, the related problems and ideas to overcome them. 5

9 Chapter 3 Calibration In the last chapter, we discussed the whole process of human modeling in brief. We realize that calibration process can be improved a lot. Here, we will discuss various types of calibrations and give a light to the improvements which can be done. 3.1 Self Calibration After capturing the image, we backproject the pixels into 3D space to get a point clouds. The backprojection line is known only if we know the intrinsic parameters of the camera. The intrinsic parameters of the camera includes Focal Length f Image origin o x, o y Scale facter s x, s y Skew Factor Lens Distortion Here, we show calculations of only f, o x, o y, s x and s y. Consider the following diagram 3.1. Here, O is the pinhole, p(x, y, z) and P (X, Y, Z) are the points in 2D and 3D respectively. Clearly, x = f X Z ; y = f Y Z...(1) where every quantity is in metres. After this, mapping is done to map camera coordinates (x and y in metres) to pixel coordinates (x p and y p in pixels) respectively. This is done by dividing it by scaling factors (s x, s y ) m pixel. The origin of the camera coordinate system may not coincide with the image center. The offset is denoted by (o x, o y ) pixel. The image pixel now becomes x p = f s x X Z + o x; y p = f s y Y Z + o y; 6

10 Figure 3.1: Representation of a pinhole camera ( camera model) In matrix form, it can be represented as x p y p } z p {{ } p = f s x 0 o x 0 0 f s y o y 0 } 0 0 {{ 1 0 } M p = MP From this, pixel coordinates can be calculated as x p = x p z p ; y p = y p z p X Y Z 1 } {{ } P M (the projection matrix) is solved using least squares by taking at least 6 corresponding features points. Manually marking those points is a tedious task. Also, it leads to high reprojection errors (difference between actual and the calculated pixel positions) because humans are not that accurate. The reprojection error is around pixels. A utility is there which automatically detects good feature points on the checkboard. In this case, good feature points are the corners of its cells. It is observed that the reprojection error drops to around 1-2 pixels. We intend to calibrate both RGB and depth camera simultaneously. The RGB camera is able to distinguish between black and white, thats why the utility can detect the corner of the cells. But to a depth camera, the checkerboard looks like a blank sheet. An idea [10] is to use a checkboard with some hollow cells in between. In this way, the change in depths at the corners can be used to characterize it as a good feature point. So, both RGB and depth camera can be calibrated at once. 3.2 Mapping depth to RGB camera The IR and RGB cameras are separated by a small baseline. So, a stereo calibration algorithm is used to determine the transformation between them. We can use a semi-transparent checkerboard 7

11 so that the shared corners of the cells are visible to depth camera as well. To do this, we first calibrate the individual cameras. Then, with all the internal camera parameters fixed (including focal ( length), we calibrate the external ) transform between the two. Typical translation values are The measured distance between IR and RGB lens centers is about 2.5 cm. The Y and Z offsets are very small. The rotation component of the transform was also very small. Typical offsets were about 0.5 degrees, which translates to a 1 cm offset at 1.5 m. 3.3 Calibration between multiple Kinects Multi-kinect calibration means to calculate the transformations so that coordinate system of a camera can be mapped to the coordinate system of any another camera. This can be done by placing an object in a way visible to every camera. Taking checkerboard is a good idea since features can be detected accurately and automatically. After placing the object, an arbitrary coordinate system is chosen with respect to which the transformation of the cameras are calculated. Its simpler to choose corner (say bottom-left) of the object. Lets say the transformations for i th camera be T ar i. Then, the transformation from i th camera to j th camera would be T ari T ar j. In words, it means to bring the i th camera to the shared coordinate system and then apply T ar j to that. Our case is a bit more complex because there is no position for the checkerboard which is visible to all the four cameras. So we first calibrate camera-1 and camera-2, then camera-3 and camera-4 and then camera-2 and camera-4. Note that these three pairs will have different shared coordinate system. This generates good feature points. To map, say camera-1 to camera-3, we would first map camera-1 to camera-2, then to camera-4 and finally to camera-3. The transformations are refined by repeating the same algorithm again. Point clouds of a person are captured for all the 3 pairs above. Camera-2 and camera-4 are calibrated by putting some objects in a small container. Calibration is done using ICP. The results are good but this step makes the calibration cumbersome. One way to calibrate is by waving the hand. The first two types of calibrations are not so important. They are related to camera intrinsics which are not going to change unless we change the camera. We need to perform the above process only once for a camera. Till now, we have discussed in brief what work has been done so far. We studied the calibration processes and realized that our four-kinect calibration gives good results but is cumbersome. In the next chapter, we will discuss about limitations in the whole process and some ways to overcome them. This will include calibration by hand-waving technique as well. 8

12 Chapter 4 Conclusion and Future Scope We studied the work done by [2] mainly focussing on calibration. We studied in brief the process of creating a human model using Kinect. We got good results but the whole procedure is quite complex. The limitations are listed below. 4.1 Limitations We are able to get point cloud with good accuracy but still the method is not robust. If the person is wearing loose clothes, then the reconstructed point cloud may not be realistic in some parts. The whole process takes a considerable amount of time. The performance is shown below. Process Average computing time (min) Scanning 2 Pre-processing 10 Registration 10 Post-processing 2 Table 4.1: The average computing time of each step with an Intel Core2Duo processor at 2 GHz This can be decreased a lot if we can improve the hardware so as to switch on all the Kinects simultaneously. This might also decrease blurring of the mesh. This is because the frames from different cameras which are captured simutaneously, have a slight delay because the hardware permits only one frame capture at a time. The system assumes that the person doesn t move when the point cloud is being captured. This should be more robust because its not easy for someone to stand still for a long time. Manual editing is required to register different point clouds. Calibration procedure gets tedious to get a finer transformations. We will discuss more about this in the next section. 9

13 4.2 Future Work Code As of now, the whole process is not supposed to be for a layman user. Instructions for a user to get scanned are a bit complex for him. There are lots of executables. openni grabber openni: Captures point clouds to be used for calibration and stores them in the location provided as the argument. There are some keys to start and stop streaming. calibrate2: It takes the above point clouds and aligns them. After that, some manual editing is done to refine the alignment. The transformation parameters are then hard-coded in a function call. openni grabber openni simul fork: Grabs the actual point clouds of the user for scanning. The user rotates from 0 to 90 degrees in 4 steps (effectively 5 frames). In every step, all 4 Kinects scan the person thus capturing 20 point clouds in total. The task of gathering point cloud is over. The subsequent code is to manipulate to get the final mesh. The whole process can be merged into a bash file. The user can just run the file (provided with the path), give point cloud data whenever the program asks, and at the end have his model ready in the system. At some places, key events are not required because their task can be automated. For the rest, the key events can be replaced by speech commands (if needed). Some minor changes were made to the code (like removing some useless infinite loops, changing the hard-coded paths in the makefile etc) Kinect Stand Figure 4.1: Our Setup 10

14 (a) Kinect Stand with sliding bases (b) A scale can be attached to the stand (c) Alternative base: Stable, lighter but does not allow camera to go much lower Figure 4.2: Stand Design The whole apparatus 4.1 to scan the model requires that the Kinects should be steady at their positions. Even the slighest movement can distort the output a lot. If any of the camera is displaced, the whole calibration process would have to be repeated, which takes a lot of time. Currently, the cameras are somehow made stable. We have planned to have a steel stand 4.2 upto a height of 7f t (maximum). It would be able to incorporate 2 Kinects which can slide vertically. Knobs at their back can be tightened to fix their position. A vertical metre scale would be there to measure the distance between Kinects. Two such stands are needed. The main advantage of having this is that two Kinects on a stand will always be rigid and the distance between them is known. Another advantage is that it is portable. We can put the stand anywhere. There would be a proper arrangement of wires Multi-Kinect Calibration using Hand-Waving Our calibration procedure is tedious. First cameras are calibrated using checkerboard. Then, to refine the transformations, point clouds are captured. An idea is to wave an object which is tracked by all the cameras. This paper [7] talks about multi-calibration by moving two LED markers rigid with respect to each other. It assumes that the cameras share a volume space at least pairwise. The markers are tracked to get new voxels with 11

15 time. Since we have pairwise transformations, global calibration is done by constructing a vision graph. Vision graph shows connectivity between the cameras. To calculate the transformation between i th and j th camera, we pick that series of pairwise transformations which has the shortest path (as calculated by Dijkstra s algorithm). Weights assigned to each edge is inversely proportional to the number of corresponding points they share. If there are no common points, there is no edge between the cameras. For this algorithm to work, the graph is needed to be connected. The accuracy of this method depends on many factors one of which is accuracy of marker detection algorithm. This algorithm can be used to track a hand of the user. The main issue which might occur is that the cameras are not synchronized in our case. The user hand position might change while different cameras are capturing point clouds of the same time instance. But one point is that there is a large field volume which is visible to all the four cameras. This should give some advantage. 12

16 Chapter 5 Acknowledgement I thank Prof. Parag Chaudhuri for very helpful discussions and insights into the project, as well as for constructive feedback. 13

17 Bibliography [1] Niket Bagwe, A study of the Microsofts Kinect camera. Technical report, IIT Bombay, 2011 [2] Niket Bagwe, 3D Reconstruction of an Animatable Humanoid using Microsoft Kinect, IIT Bombay, 2012 [3] Jaai Mashalkar, Personalized Animatable Avatars from Depth Data, IIT Bombay, 2013 [4] Microsoft s Kinect, [5] Point Cloud Library(PCL), [6] Prime Sense, [7] Gregorij Kurillo, Zeyu Li, Ruzena Bajcsy, Wide-area external multi camera calibration using vision graphs and virtual calibration objectm University of California, Berkeley [8] Kinect Drivers, [9] Laser Scan, scanner#non-contact active [10] Calibration using Semi-transparent checkerboard, [11] Multi-kinect calibration, 14

3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller

3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller 3D Computer Vision Depth Cameras Prof. Didier Stricker Oliver Wasenmüller Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de

More information

Accurate 3D Face and Body Modeling from a Single Fixed Kinect

Accurate 3D Face and Body Modeling from a Single Fixed Kinect Accurate 3D Face and Body Modeling from a Single Fixed Kinect Ruizhe Wang*, Matthias Hernandez*, Jongmoo Choi, Gérard Medioni Computer Vision Lab, IRIS University of Southern California Abstract In this

More information

Structured light 3D reconstruction

Structured light 3D reconstruction Structured light 3D reconstruction Reconstruction pipeline and industrial applications rodola@dsi.unive.it 11/05/2010 3D Reconstruction 3D reconstruction is the process of capturing the shape and appearance

More information

Rigid ICP registration with Kinect

Rigid ICP registration with Kinect Rigid ICP registration with Kinect Students: Yoni Choukroun, Elie Semmel Advisor: Yonathan Aflalo 1 Overview.p.3 Development of the project..p.3 Papers p.4 Project algorithm..p.6 Result of the whole body.p.7

More information

Lecture 19: Depth Cameras. Visual Computing Systems CMU , Fall 2013

Lecture 19: Depth Cameras. Visual Computing Systems CMU , Fall 2013 Lecture 19: Depth Cameras Visual Computing Systems Continuing theme: computational photography Cameras capture light, then extensive processing produces the desired image Today: - Capturing scene depth

More information

Outline. ETN-FPI Training School on Plenoptic Sensing

Outline. ETN-FPI Training School on Plenoptic Sensing Outline Introduction Part I: Basics of Mathematical Optimization Linear Least Squares Nonlinear Optimization Part II: Basics of Computer Vision Camera Model Multi-Camera Model Multi-Camera Calibration

More information

MERGING POINT CLOUDS FROM MULTIPLE KINECTS. Nishant Rai 13th July, 2016 CARIS Lab University of British Columbia

MERGING POINT CLOUDS FROM MULTIPLE KINECTS. Nishant Rai 13th July, 2016 CARIS Lab University of British Columbia MERGING POINT CLOUDS FROM MULTIPLE KINECTS Nishant Rai 13th July, 2016 CARIS Lab University of British Columbia Introduction What do we want to do? : Use information (point clouds) from multiple (2+) Kinects

More information

Depth Camera for Mobile Devices

Depth Camera for Mobile Devices Depth Camera for Mobile Devices Instructor - Simon Lucey 16-423 - Designing Computer Vision Apps Today Stereo Cameras Structured Light Cameras Time of Flight (ToF) Camera Inferring 3D Points Given we have

More information

Dynamic Geometry Processing

Dynamic Geometry Processing Dynamic Geometry Processing EG 2012 Tutorial Will Chang, Hao Li, Niloy Mitra, Mark Pauly, Michael Wand Tutorial: Dynamic Geometry Processing 1 Articulated Global Registration Introduction and Overview

More information

Design and Calibration of a Network of RGB-D Sensors for Robotic Applications over Large Workspaces

Design and Calibration of a Network of RGB-D Sensors for Robotic Applications over Large Workspaces Design and Calibration of a Network of RGB-D Sensors for Robotic Applications over Large Workspaces By: Rizwan Macknojia A thesis submitted to the Faculty of Graduate and Postdoctoral Studies In partial

More information

LUMS Mine Detector Project

LUMS Mine Detector Project LUMS Mine Detector Project Using visual information to control a robot (Hutchinson et al. 1996). Vision may or may not be used in the feedback loop. Visual (image based) features such as points, lines

More information

ENGN D Photography / Spring 2018 / SYLLABUS

ENGN D Photography / Spring 2018 / SYLLABUS ENGN 2502 3D Photography / Spring 2018 / SYLLABUS Description of the proposed course Over the last decade digital photography has entered the mainstream with inexpensive, miniaturized cameras routinely

More information

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University. 3D Computer Vision Structured Light II Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de 1 Introduction

More information

StereoScan: Dense 3D Reconstruction in Real-time

StereoScan: Dense 3D Reconstruction in Real-time STANFORD UNIVERSITY, COMPUTER SCIENCE, STANFORD CS231A SPRING 2016 StereoScan: Dense 3D Reconstruction in Real-time Peirong Ji, pji@stanford.edu June 7, 2016 1 INTRODUCTION In this project, I am trying

More information

Multiple View Geometry

Multiple View Geometry Multiple View Geometry Martin Quinn with a lot of slides stolen from Steve Seitz and Jianbo Shi 15-463: Computational Photography Alexei Efros, CMU, Fall 2007 Our Goal The Plenoptic Function P(θ,φ,λ,t,V

More information

Introduction to 3D Machine Vision

Introduction to 3D Machine Vision Introduction to 3D Machine Vision 1 Many methods for 3D machine vision Use Triangulation (Geometry) to Determine the Depth of an Object By Different Methods: Single Line Laser Scan Stereo Triangulation

More information

Visual Perception Sensors

Visual Perception Sensors G. Glaser Visual Perception Sensors 1 / 27 MIN Faculty Department of Informatics Visual Perception Sensors Depth Determination Gerrit Glaser University of Hamburg Faculty of Mathematics, Informatics and

More information

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Structured Light II Johannes Köhler Johannes.koehler@dfki.de Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Introduction Previous lecture: Structured Light I Active Scanning Camera/emitter

More information

Miniature faking. In close-up photo, the depth of field is limited.

Miniature faking. In close-up photo, the depth of field is limited. Miniature faking In close-up photo, the depth of field is limited. http://en.wikipedia.org/wiki/file:jodhpur_tilt_shift.jpg Miniature faking Miniature faking http://en.wikipedia.org/wiki/file:oregon_state_beavers_tilt-shift_miniature_greg_keene.jpg

More information

Processing 3D Surface Data

Processing 3D Surface Data Processing 3D Surface Data Computer Animation and Visualisation Lecture 12 Institute for Perception, Action & Behaviour School of Informatics 3D Surfaces 1 3D surface data... where from? Iso-surfacing

More information

A Comparison between Active and Passive 3D Vision Sensors: BumblebeeXB3 and Microsoft Kinect

A Comparison between Active and Passive 3D Vision Sensors: BumblebeeXB3 and Microsoft Kinect A Comparison between Active and Passive 3D Vision Sensors: BumblebeeXB3 and Microsoft Kinect Diana Beltran and Luis Basañez Technical University of Catalonia, Barcelona, Spain {diana.beltran,luis.basanez}@upc.edu

More information

Information page for written examinations at Linköping University TER2

Information page for written examinations at Linköping University TER2 Information page for written examinations at Linköping University Examination date 2016-08-19 Room (1) TER2 Time 8-12 Course code Exam code Course name Exam name Department Number of questions in the examination

More information

A Low Power, High Throughput, Fully Event-Based Stereo System: Supplementary Documentation

A Low Power, High Throughput, Fully Event-Based Stereo System: Supplementary Documentation A Low Power, High Throughput, Fully Event-Based Stereo System: Supplementary Documentation Alexander Andreopoulos, Hirak J. Kashyap, Tapan K. Nayak, Arnon Amir, Myron D. Flickner IBM Research March 25,

More information

Index C, D, E, F I, J

Index C, D, E, F I, J Index A Ambient light, 12 B Blurring algorithm, 68 Brightness thresholding algorithm float testapp::blur, 70 kinect.update(), 69 void testapp::draw(), 70 void testapp::exit(), 70 void testapp::setup(),

More information

3D scanning. 3D scanning is a family of technologies created as a means of automatic measurement of geometric properties of objects.

3D scanning. 3D scanning is a family of technologies created as a means of automatic measurement of geometric properties of objects. Acquiring 3D shape 3D scanning 3D scanning is a family of technologies created as a means of automatic measurement of geometric properties of objects. The produced digital model is formed by geometric

More information

Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision

Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision What Happened Last Time? Human 3D perception (3D cinema) Computational stereo Intuitive explanation of what is meant by disparity Stereo matching

More information

Stereo vision. Many slides adapted from Steve Seitz

Stereo vision. Many slides adapted from Steve Seitz Stereo vision Many slides adapted from Steve Seitz What is stereo vision? Generic problem formulation: given several images of the same object or scene, compute a representation of its 3D shape What is

More information

BIL Computer Vision Apr 16, 2014

BIL Computer Vision Apr 16, 2014 BIL 719 - Computer Vision Apr 16, 2014 Binocular Stereo (cont d.), Structure from Motion Aykut Erdem Dept. of Computer Engineering Hacettepe University Slide credit: S. Lazebnik Basic stereo matching algorithm

More information

L16. Scan Matching and Image Formation

L16. Scan Matching and Image Formation EECS568 Mobile Robotics: Methods and Principles Prof. Edwin Olson L16. Scan Matching and Image Formation Scan Matching Before After 2 Scan Matching Before After 2 Map matching has to be fast 14 robots

More information

3D Perception. CS 4495 Computer Vision K. Hawkins. CS 4495 Computer Vision. 3D Perception. Kelsey Hawkins Robotics

3D Perception. CS 4495 Computer Vision K. Hawkins. CS 4495 Computer Vision. 3D Perception. Kelsey Hawkins Robotics CS 4495 Computer Vision Kelsey Hawkins Robotics Motivation What do animals, people, and robots want to do with vision? Detect and recognize objects/landmarks Find location of objects with respect to themselves

More information

3D Scanning. Qixing Huang Feb. 9 th Slide Credit: Yasutaka Furukawa

3D Scanning. Qixing Huang Feb. 9 th Slide Credit: Yasutaka Furukawa 3D Scanning Qixing Huang Feb. 9 th 2017 Slide Credit: Yasutaka Furukawa Geometry Reconstruction Pipeline This Lecture Depth Sensing ICP for Pair-wise Alignment Next Lecture Global Alignment Pairwise Multiple

More information

A Low Cost and Easy to Use Setup for Foot Scanning

A Low Cost and Easy to Use Setup for Foot Scanning Abstract A Low Cost and Easy to Use Setup for Foot Scanning C. Lovato*, E. Bissolo, N. Lanza, A. Stella, A. Giachetti Dep. of Computer Science, University of Verona, Verona, Italy http://dx.doi.org/10.15221/14.365

More information

PERFORMANCE CAPTURE FROM SPARSE MULTI-VIEW VIDEO

PERFORMANCE CAPTURE FROM SPARSE MULTI-VIEW VIDEO Stefan Krauß, Juliane Hüttl SE, SoSe 2011, HU-Berlin PERFORMANCE CAPTURE FROM SPARSE MULTI-VIEW VIDEO 1 Uses of Motion/Performance Capture movies games, virtual environments biomechanics, sports science,

More information

Processing 3D Surface Data

Processing 3D Surface Data Processing 3D Surface Data Computer Animation and Visualisation Lecture 17 Institute for Perception, Action & Behaviour School of Informatics 3D Surfaces 1 3D surface data... where from? Iso-surfacing

More information

Image Based Reconstruction II

Image Based Reconstruction II Image Based Reconstruction II Qixing Huang Feb. 2 th 2017 Slide Credit: Yasutaka Furukawa Image-Based Geometry Reconstruction Pipeline Last Lecture: Multi-View SFM Multi-View SFM This Lecture: Multi-View

More information

HOW TO RECONSTRUCT DAMAGED PARTS BASED ON PRECISE AND (PARTLY-)AUTOMATISED SCAN METHODS

HOW TO RECONSTRUCT DAMAGED PARTS BASED ON PRECISE AND (PARTLY-)AUTOMATISED SCAN METHODS HOW TO RECONSTRUCT DAMAGED PARTS BASED ON PRECISE AND (PARTLY-)AUTOMATISED SCAN METHODS cirp GmbH 3D models features spare part identification load simulation mesh repair search in data bases machine learning

More information

3D Sensing and Reconstruction Readings: Ch 12: , Ch 13: ,

3D Sensing and Reconstruction Readings: Ch 12: , Ch 13: , 3D Sensing and Reconstruction Readings: Ch 12: 12.5-6, Ch 13: 13.1-3, 13.9.4 Perspective Geometry Camera Model Stereo Triangulation 3D Reconstruction by Space Carving 3D Shape from X means getting 3D coordinates

More information

3D Models from Range Sensors. Gianpaolo Palma

3D Models from Range Sensors. Gianpaolo Palma 3D Models from Range Sensors Gianpaolo Palma Who Gianpaolo Palma Researcher at Visual Computing Laboratory (ISTI-CNR) Expertise: 3D scanning, Mesh Processing, Computer Graphics E-mail: gianpaolo.palma@isti.cnr.it

More information

3D Photography: Active Ranging, Structured Light, ICP

3D Photography: Active Ranging, Structured Light, ICP 3D Photography: Active Ranging, Structured Light, ICP Kalin Kolev, Marc Pollefeys Spring 2013 http://cvg.ethz.ch/teaching/2013spring/3dphoto/ Schedule (tentative) Feb 18 Feb 25 Mar 4 Mar 11 Mar 18 Mar

More information

Other approaches to obtaining 3D structure

Other approaches to obtaining 3D structure Other approaches to obtaining 3D structure Active stereo with structured light Project structured light patterns onto the object simplifies the correspondence problem Allows us to use only one camera camera

More information

Data-driven Approaches to Simulation (Motion Capture)

Data-driven Approaches to Simulation (Motion Capture) 1 Data-driven Approaches to Simulation (Motion Capture) Ting-Chun Sun tingchun.sun@usc.edu Preface The lecture slides [1] are made by Jessica Hodgins [2], who is a professor in Computer Science Department

More information

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Structured Light II Johannes Köhler Johannes.koehler@dfki.de Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Introduction Previous lecture: Structured Light I Active Scanning Camera/emitter

More information

A consumer level 3D object scanning device using Kinect for web-based C2C business

A consumer level 3D object scanning device using Kinect for web-based C2C business A consumer level 3D object scanning device using Kinect for web-based C2C business Geoffrey Poon, Yu Yin Yeung and Wai-Man Pang Caritas Institute of Higher Education Introduction Internet shopping is popular

More information

Range Sensors (time of flight) (1)

Range Sensors (time of flight) (1) Range Sensors (time of flight) (1) Large range distance measurement -> called range sensors Range information: key element for localization and environment modeling Ultrasonic sensors, infra-red sensors

More information

Scene Reconstruction from Uncontrolled Motion using a Low Cost 3D Sensor

Scene Reconstruction from Uncontrolled Motion using a Low Cost 3D Sensor Scene Reconstruction from Uncontrolled Motion using a Low Cost 3D Sensor Pierre Joubert and Willie Brink Applied Mathematics Department of Mathematical Sciences University of Stellenbosch, South Africa

More information

Step-by-Step Model Buidling

Step-by-Step Model Buidling Step-by-Step Model Buidling Review Feature selection Feature selection Feature correspondence Camera Calibration Euclidean Reconstruction Landing Augmented Reality Vision Based Control Sparse Structure

More information

Depth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth

Depth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth Common Classification Tasks Recognition of individual objects/faces Analyze object-specific features (e.g., key points) Train with images from different viewing angles Recognition of object classes Analyze

More information

Motion Capture & Simulation

Motion Capture & Simulation Motion Capture & Simulation Motion Capture Character Reconstructions Joint Angles Need 3 points to compute a rigid body coordinate frame 1 st point gives 3D translation, 2 nd point gives 2 angles, 3 rd

More information

3D Modeling of Objects Using Laser Scanning

3D Modeling of Objects Using Laser Scanning 1 3D Modeling of Objects Using Laser Scanning D. Jaya Deepu, LPU University, Punjab, India Email: Jaideepudadi@gmail.com Abstract: In the last few decades, constructing accurate three-dimensional models

More information

Kinect Cursor Control EEE178 Dr. Fethi Belkhouche Christopher Harris Danny Nguyen I. INTRODUCTION

Kinect Cursor Control EEE178 Dr. Fethi Belkhouche Christopher Harris Danny Nguyen I. INTRODUCTION Kinect Cursor Control EEE178 Dr. Fethi Belkhouche Christopher Harris Danny Nguyen Abstract: An XBOX 360 Kinect is used to develop two applications to control the desktop cursor of a Windows computer. Application

More information

Practical Robotics (PRAC)

Practical Robotics (PRAC) Practical Robotics (PRAC) A Mobile Robot Navigation System (1) - Sensor and Kinematic Modelling Nick Pears University of York, Department of Computer Science December 17, 2014 nep (UoY CS) PRAC Practical

More information

3D Photography: Stereo

3D Photography: Stereo 3D Photography: Stereo Marc Pollefeys, Torsten Sattler Spring 2016 http://www.cvg.ethz.ch/teaching/3dvision/ 3D Modeling with Depth Sensors Today s class Obtaining depth maps / range images unstructured

More information

Stereo Vision A simple system. Dr. Gerhard Roth Winter 2012

Stereo Vision A simple system. Dr. Gerhard Roth Winter 2012 Stereo Vision A simple system Dr. Gerhard Roth Winter 2012 Stereo Stereo Ability to infer information on the 3-D structure and distance of a scene from two or more images taken from different viewpoints

More information

Personal Navigation and Indoor Mapping: Performance Characterization of Kinect Sensor-based Trajectory Recovery

Personal Navigation and Indoor Mapping: Performance Characterization of Kinect Sensor-based Trajectory Recovery Personal Navigation and Indoor Mapping: Performance Characterization of Kinect Sensor-based Trajectory Recovery 1 Charles TOTH, 1 Dorota BRZEZINSKA, USA 2 Allison KEALY, Australia, 3 Guenther RETSCHER,

More information

Sensor Modalities. Sensor modality: Different modalities:

Sensor Modalities. Sensor modality: Different modalities: Sensor Modalities Sensor modality: Sensors which measure same form of energy and process it in similar ways Modality refers to the raw input used by the sensors Different modalities: Sound Pressure Temperature

More information

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation Obviously, this is a very slow process and not suitable for dynamic scenes. To speed things up, we can use a laser that projects a vertical line of light onto the scene. This laser rotates around its vertical

More information

Ultrasonic Multi-Skip Tomography for Pipe Inspection

Ultrasonic Multi-Skip Tomography for Pipe Inspection 18 th World Conference on Non destructive Testing, 16-2 April 212, Durban, South Africa Ultrasonic Multi-Skip Tomography for Pipe Inspection Arno VOLKER 1, Rik VOS 1 Alan HUNTER 1 1 TNO, Stieltjesweg 1,

More information

Dense 3D Reconstruction. Christiano Gava

Dense 3D Reconstruction. Christiano Gava Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Today: dense 3D reconstruction The matching problem

More information

CV: 3D to 2D mathematics. Perspective transformation; camera calibration; stereo computation; and more

CV: 3D to 2D mathematics. Perspective transformation; camera calibration; stereo computation; and more CV: 3D to 2D mathematics Perspective transformation; camera calibration; stereo computation; and more Roadmap of topics n Review perspective transformation n Camera calibration n Stereo methods n Structured

More information

CS4495/6495 Introduction to Computer Vision

CS4495/6495 Introduction to Computer Vision CS4495/6495 Introduction to Computer Vision 9C-L1 3D perception Some slides by Kelsey Hawkins Motivation Why do animals, people & robots need vision? To detect and recognize objects/landmarks Is that a

More information

High-Fidelity Augmented Reality Interactions Hrvoje Benko Researcher, MSR Redmond

High-Fidelity Augmented Reality Interactions Hrvoje Benko Researcher, MSR Redmond High-Fidelity Augmented Reality Interactions Hrvoje Benko Researcher, MSR Redmond New generation of interfaces Instead of interacting through indirect input devices (mice and keyboard), the user is interacting

More information

Computer Vision. Coordinates. Prof. Flávio Cardeal DECOM / CEFET- MG.

Computer Vision. Coordinates. Prof. Flávio Cardeal DECOM / CEFET- MG. Computer Vision Coordinates Prof. Flávio Cardeal DECOM / CEFET- MG cardeal@decom.cefetmg.br Abstract This lecture discusses world coordinates and homogeneous coordinates, as well as provides an overview

More information

Multi-view stereo. Many slides adapted from S. Seitz

Multi-view stereo. Many slides adapted from S. Seitz Multi-view stereo Many slides adapted from S. Seitz Beyond two-view stereo The third eye can be used for verification Multiple-baseline stereo Pick a reference image, and slide the corresponding window

More information

Surface Registration. Gianpaolo Palma

Surface Registration. Gianpaolo Palma Surface Registration Gianpaolo Palma The problem 3D scanning generates multiple range images Each contain 3D points for different parts of the model in the local coordinates of the scanner Find a rigid

More information

Binocular stereo. Given a calibrated binocular stereo pair, fuse it to produce a depth image. Where does the depth information come from?

Binocular stereo. Given a calibrated binocular stereo pair, fuse it to produce a depth image. Where does the depth information come from? Binocular Stereo Binocular stereo Given a calibrated binocular stereo pair, fuse it to produce a depth image Where does the depth information come from? Binocular stereo Given a calibrated binocular stereo

More information

Human body animation. Computer Animation. Human Body Animation. Skeletal Animation

Human body animation. Computer Animation. Human Body Animation. Skeletal Animation Computer Animation Aitor Rovira March 2010 Human body animation Based on slides by Marco Gillies Human Body Animation Skeletal Animation Skeletal Animation (FK, IK) Motion Capture Motion Editing (retargeting,

More information

Advanced Vision Guided Robotics. David Bruce Engineering Manager FANUC America Corporation

Advanced Vision Guided Robotics. David Bruce Engineering Manager FANUC America Corporation Advanced Vision Guided Robotics David Bruce Engineering Manager FANUC America Corporation Traditional Vision vs. Vision based Robot Guidance Traditional Machine Vision Determine if a product passes or

More information

Reconstruction of complete 3D object model from multi-view range images.

Reconstruction of complete 3D object model from multi-view range images. Header for SPIE use Reconstruction of complete 3D object model from multi-view range images. Yi-Ping Hung *, Chu-Song Chen, Ing-Bor Hsieh, Chiou-Shann Fuh Institute of Information Science, Academia Sinica,

More information

Registration of Dynamic Range Images

Registration of Dynamic Range Images Registration of Dynamic Range Images Tan-Chi Ho 1,2 Jung-Hong Chuang 1 Wen-Wei Lin 2 Song-Sun Lin 2 1 Department of Computer Science National Chiao-Tung University 2 Department of Applied Mathematics National

More information

Dense 3D Reconstruction. Christiano Gava

Dense 3D Reconstruction. Christiano Gava Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Wide baseline matching (SIFT) Today: dense 3D reconstruction

More information

10/5/09 1. d = 2. Range Sensors (time of flight) (2) Ultrasonic Sensor (time of flight, sound) (1) Ultrasonic Sensor (time of flight, sound) (2) 4.1.

10/5/09 1. d = 2. Range Sensors (time of flight) (2) Ultrasonic Sensor (time of flight, sound) (1) Ultrasonic Sensor (time of flight, sound) (2) 4.1. Range Sensors (time of flight) (1) Range Sensors (time of flight) (2) arge range distance measurement -> called range sensors Range information: key element for localization and environment modeling Ultrasonic

More information

Chaplin, Modern Times, 1936

Chaplin, Modern Times, 1936 Chaplin, Modern Times, 1936 [A Bucket of Water and a Glass Matte: Special Effects in Modern Times; bonus feature on The Criterion Collection set] Multi-view geometry problems Structure: Given projections

More information

L2 Data Acquisition. Mechanical measurement (CMM) Structured light Range images Shape from shading Other methods

L2 Data Acquisition. Mechanical measurement (CMM) Structured light Range images Shape from shading Other methods L2 Data Acquisition Mechanical measurement (CMM) Structured light Range images Shape from shading Other methods 1 Coordinate Measurement Machine Touch based Slow Sparse Data Complex planning Accurate 2

More information

Complex Sensors: Cameras, Visual Sensing. The Robotics Primer (Ch. 9) ECE 497: Introduction to Mobile Robotics -Visual Sensors

Complex Sensors: Cameras, Visual Sensing. The Robotics Primer (Ch. 9) ECE 497: Introduction to Mobile Robotics -Visual Sensors Complex Sensors: Cameras, Visual Sensing The Robotics Primer (Ch. 9) Bring your laptop and robot everyday DO NOT unplug the network cables from the desktop computers or the walls Tuesday s Quiz is on Visual

More information

Processing 3D Surface Data

Processing 3D Surface Data Processing 3D Surface Data Computer Animation and Visualisation Lecture 15 Institute for Perception, Action & Behaviour School of Informatics 3D Surfaces 1 3D surface data... where from? Iso-surfacing

More information

Overview. Augmented reality and applications Marker-based augmented reality. Camera model. Binary markers Textured planar markers

Overview. Augmented reality and applications Marker-based augmented reality. Camera model. Binary markers Textured planar markers Augmented reality Overview Augmented reality and applications Marker-based augmented reality Binary markers Textured planar markers Camera model Homography Direct Linear Transformation What is augmented

More information

A Qualitative Analysis of 3D Display Technology

A Qualitative Analysis of 3D Display Technology A Qualitative Analysis of 3D Display Technology Nicholas Blackhawk, Shane Nelson, and Mary Scaramuzza Computer Science St. Olaf College 1500 St. Olaf Ave Northfield, MN 55057 scaramum@stolaf.edu Abstract

More information

Project Title: Welding Machine Monitoring System Phase II. Name of PI: Prof. Kenneth K.M. LAM (EIE) Progress / Achievement: (with photos, if any)

Project Title: Welding Machine Monitoring System Phase II. Name of PI: Prof. Kenneth K.M. LAM (EIE) Progress / Achievement: (with photos, if any) Address: Hong Kong Polytechnic University, Phase 8, Hung Hom, Kowloon, Hong Kong. Telephone: (852) 3400 8441 Email: cnerc.steel@polyu.edu.hk Website: https://www.polyu.edu.hk/cnerc-steel/ Project Title:

More information

Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera

Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera Tomokazu Satoy, Masayuki Kanbaray, Naokazu Yokoyay and Haruo Takemuraz ygraduate School of Information

More information

CHAPTER 2: THREE DIMENSIONAL TOPOGRAPHICAL MAPPING SYSTEM. Target Object

CHAPTER 2: THREE DIMENSIONAL TOPOGRAPHICAL MAPPING SYSTEM. Target Object CHAPTER 2: THREE DIMENSIONAL TOPOGRAPHICAL MAPPING SYSTEM 2.1 Theory and Construction Target Object Laser Projector CCD Camera Host Computer / Image Processor Figure 2.1 Block Diagram of 3D Areal Mapper

More information

Stereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman

Stereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman Stereo 11/02/2012 CS129, Brown James Hays Slides by Kristen Grauman Multiple views Multi-view geometry, matching, invariant features, stereo vision Lowe Hartley and Zisserman Why multiple views? Structure

More information

PART IV: RS & the Kinect

PART IV: RS & the Kinect Computer Vision on Rolling Shutter Cameras PART IV: RS & the Kinect Per-Erik Forssén, Erik Ringaby, Johan Hedborg Computer Vision Laboratory Dept. of Electrical Engineering Linköping University Tutorial

More information

Minimizing Noise and Bias in 3D DIC. Correlated Solutions, Inc.

Minimizing Noise and Bias in 3D DIC. Correlated Solutions, Inc. Minimizing Noise and Bias in 3D DIC Correlated Solutions, Inc. Overview Overview of Noise and Bias Digital Image Correlation Background/Tracking Function Minimizing Noise Focus Contrast/Lighting Glare

More information

CS664 Lecture #19: Layers, RANSAC, panoramas, epipolar geometry

CS664 Lecture #19: Layers, RANSAC, panoramas, epipolar geometry CS664 Lecture #19: Layers, RANSAC, panoramas, epipolar geometry Some material taken from: David Lowe, UBC Jiri Matas, CMP Prague http://cmp.felk.cvut.cz/~matas/papers/presentations/matas_beyondransac_cvprac05.ppt

More information

3D Sensing. 3D Shape from X. Perspective Geometry. Camera Model. Camera Calibration. General Stereo Triangulation.

3D Sensing. 3D Shape from X. Perspective Geometry. Camera Model. Camera Calibration. General Stereo Triangulation. 3D Sensing 3D Shape from X Perspective Geometry Camera Model Camera Calibration General Stereo Triangulation 3D Reconstruction 3D Shape from X shading silhouette texture stereo light striping motion mainly

More information

Projector Calibration for Pattern Projection Systems

Projector Calibration for Pattern Projection Systems Projector Calibration for Pattern Projection Systems I. Din *1, H. Anwar 2, I. Syed 1, H. Zafar 3, L. Hasan 3 1 Department of Electronics Engineering, Incheon National University, Incheon, South Korea.

More information

Tecnologie per la ricostruzione di modelli 3D da immagini. Marco Callieri ISTI-CNR, Pisa, Italy

Tecnologie per la ricostruzione di modelli 3D da immagini. Marco Callieri ISTI-CNR, Pisa, Italy Tecnologie per la ricostruzione di modelli 3D da immagini Marco Callieri ISTI-CNR, Pisa, Italy 3D from Photos Our not-so-secret dream: obtain a reliable and precise 3D from simple photos Why? Easier, less

More information

3D Object Representations. COS 526, Fall 2016 Princeton University

3D Object Representations. COS 526, Fall 2016 Princeton University 3D Object Representations COS 526, Fall 2016 Princeton University 3D Object Representations How do we... Represent 3D objects in a computer? Acquire computer representations of 3D objects? Manipulate computer

More information

There are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few...

There are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few... STEREO VISION The slides are from several sources through James Hays (Brown); Srinivasa Narasimhan (CMU); Silvio Savarese (U. of Michigan); Bill Freeman and Antonio Torralba (MIT), including their own

More information

New Sony DepthSense TM ToF Technology

New Sony DepthSense TM ToF Technology ADVANCED MATERIAL HANDLING WITH New Sony DepthSense TM ToF Technology Jenson Chang Product Marketing November 7, 2018 1 3D SENSING APPLICATIONS Pick and Place Drones Collision Detection People Counting

More information

Calibrated Image Acquisition for Multi-view 3D Reconstruction

Calibrated Image Acquisition for Multi-view 3D Reconstruction Calibrated Image Acquisition for Multi-view 3D Reconstruction Sriram Kashyap M S Guide: Prof. Sharat Chandran Indian Institute of Technology, Bombay April 2009 Sriram Kashyap 3D Reconstruction 1/ 42 Motivation

More information

User-assisted Segmentation and 3D Motion Tracking. Michael Fleder Sudeep Pillai Jeremy Scott

User-assisted Segmentation and 3D Motion Tracking. Michael Fleder Sudeep Pillai Jeremy Scott User-assisted Segmentation and 3D Motion Tracking Michael Fleder Sudeep Pillai Jeremy Scott 3D Object Tracking Virtual reality and animation Imitation in robotics Autonomous driving Augmented reality Motivation

More information

Application questions. Theoretical questions

Application questions. Theoretical questions The oral exam will last 30 minutes and will consist of one application question followed by two theoretical questions. Please find below a non exhaustive list of possible application questions. The list

More information

Ceilbot vision and mapping system

Ceilbot vision and mapping system Ceilbot vision and mapping system Provide depth and camera data from the robot's environment Keep a map of the environment based on the received data Keep track of the robot's location on the map Recognize

More information

3D Computer Vision 1

3D Computer Vision 1 3D Computer Vision 1 Multiview Stereo Multiview Stereo Multiview Stereo https://www.youtube.com/watch?v=ugkb7itpnae Shape from silhouette Shape from silhouette Shape from silhouette Shape from silhouette

More information

Autonomous Navigation of Nao using Kinect CS365 : Project Report

Autonomous Navigation of Nao using Kinect CS365 : Project Report Autonomous Navigation of Nao using Kinect CS365 : Project Report Samyak Daga Harshad Sawhney 11633 11297 samyakd@iitk.ac.in harshads@iitk.ac.in Dept. of CSE Dept. of CSE Indian Institute of Technology,

More information

Miniaturized Camera Systems for Microfactories

Miniaturized Camera Systems for Microfactories Miniaturized Camera Systems for Microfactories Timo Prusi, Petri Rokka, and Reijo Tuokko Tampere University of Technology, Department of Production Engineering, Korkeakoulunkatu 6, 33720 Tampere, Finland

More information

Teleimmersion System. Contents. Dr. Gregorij Kurillo. n Introduction. n UCB. n Current problems. n Multi-stereo camera system

Teleimmersion System. Contents. Dr. Gregorij Kurillo. n Introduction. n UCB. n Current problems. n Multi-stereo camera system Teleimmersion System Dr. Gregorij Kurillo Contents n Introduction n Teleimmersion @ UCB n Current problems n Multi-stereo camera system n Networking n Rendering n Conclusion 1 3D scene 3D scene Internet

More information

Stereo Vision Inside Tire

Stereo Vision Inside Tire 1 Stereo Vision Inside Tire P.S. Els C.M. Becker University of Pretoria W911NF-14-1-0590 Final Report. February 2015- August 2015. REPORT DOCUMENTATION PAGE Form Approved OMB No. 0704-0188 Public reporting

More information

Dense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm

Dense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm Computer Vision Group Prof. Daniel Cremers Dense Tracking and Mapping for Autonomous Quadrocopters Jürgen Sturm Joint work with Frank Steinbrücker, Jakob Engel, Christian Kerl, Erik Bylow, and Daniel Cremers

More information