Dense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm

Size: px
Start display at page:

Download "Dense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm"

Transcription

1 Computer Vision Group Prof. Daniel Cremers Dense Tracking and Mapping for Autonomous Quadrocopters Jürgen Sturm Joint work with Frank Steinbrücker, Jakob Engel, Christian Kerl, Erik Bylow, and Daniel Cremers

2 Motivation Imagine you have a flying camera What would you use it for? Dense Tracking and Mapping for Autonomous Quadrocopters 2 Dr.

3 Motivation Imagine you have a flying camera What would you use it for? Dense Tracking and Mapping for Autonomous Quadrocopters 3 Dr.

4 Motivation Imagine you have a flying camera What would you use it for? Dense Tracking and Mapping for Autonomous Quadrocopters 4 Dr.

5 Motivation Imagine you have a flying camera What would you use it for? Dense Tracking and Mapping for Autonomous Quadrocopters 5 Dr.

6 Motivation Our research goal: Enable flying robots to operate autonomously in 3D environments using onboard cameras Use cameras because light weight and rich data Navigation, localization, mapping, exploration, people following,

7 Feature-based Visual Navigation [Engel, Sturm, Cremers, IROS 12]

8 Feature-based Visual Navigation [Engel, Sturm, Cremers, IROS 12] Architecture Quadrocopter Monocular SLAM Extended Kalman Filter PID Control Based on PTAM [Klein and Murray, ISMAR 07] Dense Tracking and Mapping for Autonomous Quadrocopters 8

9 Motivation Video feed from quadrocopter Dense Tracking and Mapping for Autonomous Quadrocopters 9

10 Motivation What PTAM actually sees Dense Tracking and Mapping for Autonomous Quadrocopters 10

11 Motivation Problem: Most approaches only use a small fraction of the available data Keypoint detection Visual features Question: How can we use most/all information to maximize the performance? Dense Tracking and Mapping for Autonomous Quadrocopters 11

12 Outline of the Talk Part 1: Dense tracking Part 2: Dense reconstruction Part 3: Evaluation and benchmarking Dense Tracking and Mapping for Autonomous Quadrocopters 12

13 Related Work on Dense Tracking Lucas and Kanade (IJCAI 81) Lovegrove et al. (IV 11) Newcombe et al. (ICCV 11) Comport/Tykkälä et al. (ICCV 11) Dense Tracking and Mapping for Autonomous Quadrocopters 13

14 Dense Tracking How can we exploit all data of an RGB-D image? Idea Photo-consistency constraint for all pixels Dense Tracking and Mapping for Autonomous Quadrocopters 14

15 How to deal with noise? Photo-consistency constraint will not perfectly hold Sensor noise Pose error Reflections, specular surfaces Dynamic objects (e.g., walking people) Residuals will be non-zero Residual distribution Dense Tracking and Mapping for Autonomous Quadrocopters 15

16 Probability p(r) Residual Distribution Zero-mean, peaked distribution Example: Correct camera pose 0,3 0,25 0,2 0,15 0,1 Correct camera pose 0, Residuals r Dense Tracking and Mapping for Autonomous Quadrocopters 16

17 Probability p(r) Residual Distribution Zero-mean, peaked distribution Example: Wrong camera pose 0,3 0,25 0,2 0,15 0,1 Wrong camera pose 0, Residuals r Dense Tracking and Mapping for Autonomous Quadrocopters 17

18 Probability p(r) Residual Distribution Goal: Find the camera pose that maximizes the observation likelihood 0,3 0,25 0,2 0,15 0,1 0,05 0 Wrong camera pose Correct camera pose Residuals r Dense Tracking and Mapping for Autonomous Quadrocopters 18

19 Motion Estimation Goal: Find the camera pose that maximizes the observation likelihood compute over all pixels Assume pixel-wise residuals are conditionally independent How can we solve this optimization problem? Dense Tracking and Mapping for Autonomous Quadrocopters 19

20 Approach Take negative logarithm Set derivative to zero Dense Tracking and Mapping for Autonomous Quadrocopters 20

21 Approach (cont.d) This can be rewritten as a weighted least squares problem with weights is non-linear in Need to linearize, solve, and iterate Dense Tracking and Mapping for Autonomous Quadrocopters 21

22 Iteratively Reweighted Least Squares Problem: Algorithm: 1. Compute weights 2. Linearize in the camera motion 3. Build and solve normal equations 4. Repeat until convergence Dense Tracking and Mapping for Autonomous Quadrocopters 22

23 Example First input image Second input image Residuals Dense Tracking and Mapping for Autonomous Quadrocopters 23 Image Jacobian for camera motion along x axis

24 What is a Good Model for the Residual Distribution? Dense Tracking and Mapping for Autonomous Quadrocopters 24

25 Weighted Error Dense Tracking and Mapping for Autonomous Quadrocopters 25

26 Example Weights Robust sensor model allows to down-weight outliers (dynamic objects, motion blur, reflections, ) Scene Residuals Weights Dense Tracking and Mapping for Autonomous Quadrocopters 26

27 Coarse-to-Fine Linearization only holds for small motions Coarse-to-fine scheme Image pyramids Dense Tracking and Mapping for Autonomous Quadrocopters 27

28 Dense Tracking: Results Dense Tracking and Mapping for Autonomous Quadrocopters 28

29 Summary: Dense tracking Direct matching of consecutive RGB-D images Pro Super fast, highly accurate (30 Hz on CPU) Low, constant memory consumption Con Accumulates drift over time, sometimes diverges How can we reduce/eliminate drift? Dense Tracking and Mapping for Autonomous Quadrocopters 29

30 Dense Tracking and Mapping Idea: Instead of tracking from frame-to-frame, track frame-to-model to eliminate drift t+1 t vs. Question: Where do we get the model from? Dense Tracking and Mapping for Autonomous Quadrocopters 30

31 Dense Tracking and Mapping Idea: Compute an iterative solution 1. Reconstruct model with known poses 2. Track camera with respect to known model Next question: How to represent the model? Dense Tracking and Mapping for Autonomous Quadrocopters 31

32 Representation of the 3D Model Idea: Instead of representing the cell occupancy, represent the distance of each cell to the surface Occupancy grid maps zero = free space Signed distance function (SDF) negative = outside obj. Dense Tracking and Mapping for Autonomous Quadrocopters 32 z = 1.8 x z = 1.8 x one = occupied x positive = inside obj. x

33 Dense Mapping: 2D Example Camera with known pose Dense Tracking and Mapping for Autonomous Quadrocopters 33

34 Dense Mapping: 2D Example Camera with known pose Grid with signed distance function Dense Tracking and Mapping for Autonomous Quadrocopters 34

35 Dense Mapping: 2D Example For each grid cell, compute its projective distance to the surface projective distance Dense Tracking and Mapping for Autonomous Quadrocopters 35

36 Dense Mapping: 3D Example Generalizes directly to 3D But: memory usage is cubic in side length Dense Tracking and Mapping for Autonomous Quadrocopters 36

37 Dense Tracking Given: SDF with projective distance to surface New depth image with unknown camera pose Wanted: Camera pose Our approach: Optimize camera pose so that the observed depth image minimizes the distance in the SDF Dense Tracking and Mapping for Autonomous Quadrocopters 37

38 Dense Tracking: 2D Example 3D model built from the first k frames Dense Tracking and Mapping for Autonomous Quadrocopters 38

39 Dense Tracking: 2D Example Minimize distance between depth image and SDF Dense Tracking and Mapping for Autonomous Quadrocopters 39

40 Dense Tracking: 2D Example Minimize distance between depth image and SDF Dense Tracking and Mapping for Autonomous Quadrocopters 40

41 Dense Tracking: 2D Example Minimize distance between depth image and SDF Dense Tracking and Mapping for Autonomous Quadrocopters 41

42 3D Example Given: A sequence of depth images Wanted: Accurate and dense 3D model Dense Tracking and Mapping for Autonomous Quadrocopters 42

43 Resulting 3D Model Dense Tracking and Mapping for Autonomous Quadrocopters 43

44 Can We Print These Models in 3D? Dense Tracking and Mapping for Autonomous Quadrocopters 44

45 Summary: Dense 3D Reconstruction Frame-to-model tracking Pro: Reduces drift significantly Outputs nice 3D models Con: Memory intensive (~2 GB for ) Computationally more expensive (~60 Hz on GPU) Dense Tracking and Mapping for Autonomous Quadrocopters 45

46 Evaluation and Benchmarking How can we evaluate such methods? What are good evaluation criteria? Dense Tracking and Mapping for Autonomous Quadrocopters 46

47 Existing Benchmarks Intel dataset: laser + odometry [Haehnel et al., 2004] New College dataset: stereo + omni-directional vision + laser + IMU [Smith et al., IJRR 2009] KITTI Vision benchmark: stereo [Geiger et al., CVPR 12] Our contribution: Dataset for RGB-D evaluation Intel New College KITTI RGB-D Dense Tracking and Mapping for Autonomous Quadrocopters 47

48 Recorded Scenes Different environments (office, industrial hall, ) Large variations in camera speed, camera motion, illumination, number of features, dynamic objects, Handheld and robot-mounted sensor Dense Tracking and Mapping for Autonomous Quadrocopters 48

49 Dataset Acquisition Motion capture system Camera pose (100 Hz) Microsoft Kinect (later: Asus Xtion Pro Live) Color images (30 Hz) Depth images (30 Hz) External video camera (for documentation) Dense Tracking and Mapping for Autonomous Quadrocopters 49

50 Motion Capture System 9 high-speed cameras mounted in room Cameras have active illumination and pre-process image (thresholding) Cameras track positions of retro-reflective markers Dense Tracking and Mapping for Autonomous Quadrocopters 50

51 Calibration Calibration of the overall system is not trivial: 1. Intrinsic calibration (Mocap + Kinect) 2. Extrinsic calibration (Kinect vs. Mocap) 3. Time synchronization (Kinect vs. Mocap) Dense Tracking and Mapping for Autonomous Quadrocopters 51

52 Dense Tracking and Mapping for Autonomous Quadrocopters 52

53 Dense Tracking and Mapping for Autonomous Quadrocopters 53

54 Dense Tracking and Mapping for Autonomous Quadrocopters 54

55 File Formats In total: 69 sequences (33 training, 36 testing) One TGZ archive per sequence, containing Color and depth images (PNG) List of color images (timestamp filename) List of depth images (timestamp filename) List of camera poses (timestamp tx ty tz qx qy qz qw) Dense Tracking and Mapping for Autonomous Quadrocopters 55

56 What Is a Good Evaluation Metric? Visual odometry system outputs Camera trajectory (accumulated) Visual SLAM system outputs Camera trajectory 3D map Ground truth Camera trajectory Dense Tracking and Mapping for Autonomous Quadrocopters 56

57 What Is a Good Evaluation Metric? Trajectory comparison Ground truth trajectory Estimate camera trajectory Two evaluation metrics Drift per second Global consistency Ground truth trajectory Q 1:n Estimated camera trajectory P 1:n Evaluation function Scalar performance index Dense Tracking and Mapping for Autonomous Quadrocopters 57

58 Relative Pose Error (RPE) Measures the (relative) drift between the i-th frame and the (i+δ)-th frame Relative error True motion Estimated motion Ground truth Estimated traj. Relative error Dense Tracking and Mapping for Autonomous Quadrocopters 58

59 Relative Pose Error (RPE) How to choose the time delta Δ? For odometry methods: Δ=1: Drift per frame Δ=30: Drift per second For SLAM methods: Average over all possible deltas Measures the global consistency Dense Tracking and Mapping for Autonomous Quadrocopters 59

60 Absolute Trajectory Error (ATE) Alternative method to evaluate SLAM systems Requires pre-aligned trajectories Absolute error Groundtruth Alignment Estimated Absolute error Ground truth Pre-aligned estimated traj. Dense Tracking and Mapping for Autonomous Quadrocopters 60

61 Evaluation Tools Average over all time steps Evaluation scripts for both evaluation metrics available (Python) Output: RMSE, median, mean Plot to png/pdf (optional) Dense Tracking and Mapping for Autonomous Quadrocopters 61

62 Comparison of RPE and ATE RPE and ATE are strongly related RPE considers additionally rotational errors RPE ATE Dense Tracking and Mapping for Autonomous Quadrocopters 62

63 Example: Online Evaluation Dense Tracking and Mapping for Autonomous Quadrocopters 63

64 Dense Tracking and Mapping for Autonomous Quadrocopters 64

65 Summary TUM RGB-D Benchmark Dataset for the evaluation of RGB-D SLAM systems Ground-truth camera poses Evaluation metrics + tools available Since August 2011: > visitors >4.500 online trajectory evaluations >15 published research papers using the dataset Next steps: Possibility to upload own trajectories/publications Results page and automated ranking Dense Tracking and Mapping for Autonomous Quadrocopters 65

66 Conclusion Dense methods bear a large potential Dense camera tracking Dense 3D reconstruction Use benchmark for the comparison of alternative approaches Release as open-source in preparation Please contact us if you are interested in collaboration! Dense Tracking and Mapping for Autonomous Quadrocopters 66

Visual Navigation for Flying Robots Exploration, Multi-Robot Coordination and Coverage

Visual Navigation for Flying Robots Exploration, Multi-Robot Coordination and Coverage Computer Vision Group Prof. Daniel Cremers Visual Navigation for Flying Robots Exploration, Multi-Robot Coordination and Coverage Dr. Jürgen Sturm Agenda for Today Exploration with a single robot Coordinated

More information

Visual Navigation for Flying Robots

Visual Navigation for Flying Robots Computer Vision Group Prof. Daniel Cremers Visual Navigation for Flying Robots Experimentation, Evaluation and Benchmarking Dr. Jürgen Sturm Agenda for Today Course Evaluation Scientific research: The

More information

Scanning and Printing Objects in 3D

Scanning and Printing Objects in 3D Scanning and Printing Objects in 3D Dr. Jürgen Sturm metaio GmbH (formerly Technical University of Munich) My Research Areas Visual navigation for mobile robots RoboCup Kinematic Learning Articulated Objects

More information

Dense 3D Reconstruction from Autonomous Quadrocopters

Dense 3D Reconstruction from Autonomous Quadrocopters Dense 3D Reconstruction from Autonomous Quadrocopters Computer Science & Mathematics TU Munich Martin Oswald, Jakob Engel, Christian Kerl, Frank Steinbrücker, Jan Stühmer & Jürgen Sturm Autonomous Quadrocopters

More information

Scanning and Printing Objects in 3D Jürgen Sturm

Scanning and Printing Objects in 3D Jürgen Sturm Scanning and Printing Objects in 3D Jürgen Sturm Metaio (formerly Technical University of Munich) My Research Areas Visual navigation for mobile robots RoboCup Kinematic Learning Articulated Objects Quadrocopters

More information

Jakob Engel, Thomas Schöps, Daniel Cremers Technical University Munich. LSD-SLAM: Large-Scale Direct Monocular SLAM

Jakob Engel, Thomas Schöps, Daniel Cremers Technical University Munich. LSD-SLAM: Large-Scale Direct Monocular SLAM Computer Vision Group Technical University of Munich Jakob Engel LSD-SLAM: Large-Scale Direct Monocular SLAM Jakob Engel, Thomas Schöps, Daniel Cremers Technical University Munich Monocular Video Engel,

More information

Semi-Dense Direct SLAM

Semi-Dense Direct SLAM Computer Vision Group Technical University of Munich Jakob Engel Jakob Engel, Daniel Cremers David Caruso, Thomas Schöps, Lukas von Stumberg, Vladyslav Usenko, Jörg Stückler, Jürgen Sturm Technical University

More information

Direct Methods in Visual Odometry

Direct Methods in Visual Odometry Direct Methods in Visual Odometry July 24, 2017 Direct Methods in Visual Odometry July 24, 2017 1 / 47 Motivation for using Visual Odometry Wheel odometry is affected by wheel slip More accurate compared

More information

FLaME: Fast Lightweight Mesh Estimation using Variational Smoothing on Delaunay Graphs

FLaME: Fast Lightweight Mesh Estimation using Variational Smoothing on Delaunay Graphs FLaME: Fast Lightweight Mesh Estimation using Variational Smoothing on Delaunay Graphs W. Nicholas Greene Robust Robotics Group, MIT CSAIL LPM Workshop IROS 2017 September 28, 2017 with Nicholas Roy 1

More information

Autonomous Navigation for Flying Robots

Autonomous Navigation for Flying Robots Computer Vision Group Prof. Daniel Cremers Autonomous Navigation for Flying Robots Lecture 7.1: 2D Motion Estimation in Images Jürgen Sturm Technische Universität München 3D to 2D Perspective Projections

More information

arxiv: v1 [cs.ro] 24 Nov 2018

arxiv: v1 [cs.ro] 24 Nov 2018 BENCHMARKING AND COMPARING POPULAR VISUAL SLAM ALGORITHMS arxiv:1811.09895v1 [cs.ro] 24 Nov 2018 Amey Kasar Department of Electronics and Telecommunication Engineering Pune Institute of Computer Technology

More information

Colored Point Cloud Registration Revisited Supplementary Material

Colored Point Cloud Registration Revisited Supplementary Material Colored Point Cloud Registration Revisited Supplementary Material Jaesik Park Qian-Yi Zhou Vladlen Koltun Intel Labs A. RGB-D Image Alignment Section introduced a joint photometric and geometric objective

More information

Autonomous Navigation for Flying Robots

Autonomous Navigation for Flying Robots Computer Vision Group Prof. Daniel Cremers Autonomous Navigation for Flying Robots Lecture 7.2: Visual Odometry Jürgen Sturm Technische Universität München Cascaded Control Robot Trajectory 0.1 Hz Visual

More information

Intrinsic3D: High-Quality 3D Reconstruction by Joint Appearance and Geometry Optimization with Spatially-Varying Lighting

Intrinsic3D: High-Quality 3D Reconstruction by Joint Appearance and Geometry Optimization with Spatially-Varying Lighting Intrinsic3D: High-Quality 3D Reconstruction by Joint Appearance and Geometry Optimization with Spatially-Varying Lighting R. Maier 1,2, K. Kim 1, D. Cremers 2, J. Kautz 1, M. Nießner 2,3 Fusion Ours 1

More information

Super-Resolution Keyframe Fusion for 3D Modeling with High-Quality Textures

Super-Resolution Keyframe Fusion for 3D Modeling with High-Quality Textures Super-Resolution Keyframe Fusion for 3D Modeling with High-Quality Textures Robert Maier, Jörg Stückler, Daniel Cremers International Conference on 3D Vision (3DV) October 2015, Lyon, France Motivation

More information

Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview

Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview Arun Das 05/09/2017 Arun Das Waterloo Autonomous Vehicles Lab Introduction What s in a name? Arun Das Waterloo Autonomous

More information

Vision-based Mobile Robot Localization and Mapping using Scale-Invariant Features

Vision-based Mobile Robot Localization and Mapping using Scale-Invariant Features Vision-based Mobile Robot Localization and Mapping using Scale-Invariant Features Stephen Se, David Lowe, Jim Little Department of Computer Science University of British Columbia Presented by Adam Bickett

More information

Deep Incremental Scene Understanding. Federico Tombari & Christian Rupprecht Technical University of Munich, Germany

Deep Incremental Scene Understanding. Federico Tombari & Christian Rupprecht Technical University of Munich, Germany Deep Incremental Scene Understanding Federico Tombari & Christian Rupprecht Technical University of Munich, Germany C. Couprie et al. "Toward Real-time Indoor Semantic Segmentation Using Depth Information"

More information

Simultaneous Localization and Mapping (SLAM)

Simultaneous Localization and Mapping (SLAM) Simultaneous Localization and Mapping (SLAM) RSS Lecture 16 April 8, 2013 Prof. Teller Text: Siegwart and Nourbakhsh S. 5.8 SLAM Problem Statement Inputs: No external coordinate reference Time series of

More information

Hybrids Mixed Approaches

Hybrids Mixed Approaches Hybrids Mixed Approaches Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline Why mixing? Parallel Tracking and Mapping Benefits

More information

Visual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

Visual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech Visual Odometry Features, Tracking, Essential Matrix, and RANSAC Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline The

More information

3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller

3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller 3D Computer Vision Depth Cameras Prof. Didier Stricker Oliver Wasenmüller Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de

More information

Geometric Reconstruction Dense reconstruction of scene geometry

Geometric Reconstruction Dense reconstruction of scene geometry Lecture 5. Dense Reconstruction and Tracking with Real-Time Applications Part 2: Geometric Reconstruction Dr Richard Newcombe and Dr Steven Lovegrove Slide content developed from: [Newcombe, Dense Visual

More information

Dealing with Scale. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

Dealing with Scale. Stephan Weiss Computer Vision Group NASA-JPL / CalTech Dealing with Scale Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline Why care about size? The IMU as scale provider: The

More information

Fitting (LMedS, RANSAC)

Fitting (LMedS, RANSAC) Fitting (LMedS, RANSAC) Thursday, 23/03/2017 Antonis Argyros e-mail: argyros@csd.uoc.gr LMedS and RANSAC What if we have very many outliers? 2 1 Least Median of Squares ri : Residuals Least Squares n 2

More information

Lecture 19: Depth Cameras. Visual Computing Systems CMU , Fall 2013

Lecture 19: Depth Cameras. Visual Computing Systems CMU , Fall 2013 Lecture 19: Depth Cameras Visual Computing Systems Continuing theme: computational photography Cameras capture light, then extensive processing produces the desired image Today: - Capturing scene depth

More information

CSE 527: Introduction to Computer Vision

CSE 527: Introduction to Computer Vision CSE 527: Introduction to Computer Vision Week 10 Class 2: Visual Odometry November 2nd, 2017 Today Visual Odometry Intro Algorithm SLAM Visual Odometry Input Output Images, Video Camera trajectory, motion

More information

Simultaneous Localization

Simultaneous Localization Simultaneous Localization and Mapping (SLAM) RSS Technical Lecture 16 April 9, 2012 Prof. Teller Text: Siegwart and Nourbakhsh S. 5.8 Navigation Overview Where am I? Where am I going? Localization Assumed

More information

Accurate 3D Face and Body Modeling from a Single Fixed Kinect

Accurate 3D Face and Body Modeling from a Single Fixed Kinect Accurate 3D Face and Body Modeling from a Single Fixed Kinect Ruizhe Wang*, Matthias Hernandez*, Jongmoo Choi, Gérard Medioni Computer Vision Lab, IRIS University of Southern California Abstract In this

More information

Visual SLAM for small Unmanned Aerial Vehicles

Visual SLAM for small Unmanned Aerial Vehicles Visual SLAM for small Unmanned Aerial Vehicles Margarita Chli Autonomous Systems Lab, ETH Zurich Simultaneous Localization And Mapping How can a body navigate in a previously unknown environment while

More information

EE795: Computer Vision and Intelligent Systems

EE795: Computer Vision and Intelligent Systems EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 11 140311 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Motion Analysis Motivation Differential Motion Optical

More information

Live Metric 3D Reconstruction on Mobile Phones ICCV 2013

Live Metric 3D Reconstruction on Mobile Phones ICCV 2013 Live Metric 3D Reconstruction on Mobile Phones ICCV 2013 Main Contents 1. Target & Related Work 2. Main Features of This System 3. System Overview & Workflow 4. Detail of This System 5. Experiments 6.

More information

EE795: Computer Vision and Intelligent Systems

EE795: Computer Vision and Intelligent Systems EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 14 130307 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Stereo Dense Motion Estimation Translational

More information

Robust Online 3D Reconstruction Combining a Depth Sensor and Sparse Feature Points

Robust Online 3D Reconstruction Combining a Depth Sensor and Sparse Feature Points 2016 23rd International Conference on Pattern Recognition (ICPR) Cancún Center, Cancún, México, December 4-8, 2016 Robust Online 3D Reconstruction Combining a Depth Sensor and Sparse Feature Points Erik

More information

Design-Space Exploration for Dense 3D Scene Understanding: Exploring ElasticFusion

Design-Space Exploration for Dense 3D Scene Understanding: Exploring ElasticFusion Design-Space Exploration for Dense 3D Scene Understanding: Exploring ElasticFusion Imperial College of Science, Technology and Medicine Department of Computing Author: Denise Carroll Supervisor: Prof.

More information

PL-SVO: Semi-Direct Monocular Visual Odometry by Combining Points and Line Segments

PL-SVO: Semi-Direct Monocular Visual Odometry by Combining Points and Line Segments 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) Daejeon Convention Center October 9-14, 2016, Daejeon, Korea PL-SVO: Semi-Direct Monocular Visual Odometry by Combining Points

More information

3D Scene Reconstruction with a Mobile Camera

3D Scene Reconstruction with a Mobile Camera 3D Scene Reconstruction with a Mobile Camera 1 Introduction Robert Carrera and Rohan Khanna Stanford University: CS 231A Autonomous supernumerary arms, or "third arms", while still unconventional, hold

More information

OpenStreetSLAM: Global Vehicle Localization using OpenStreetMaps

OpenStreetSLAM: Global Vehicle Localization using OpenStreetMaps OpenStreetSLAM: Global Vehicle Localization using OpenStreetMaps Georgios Floros, Benito van der Zander and Bastian Leibe RWTH Aachen University, Germany http://www.vision.rwth-aachen.de floros@vision.rwth-aachen.de

More information

視覚情報処理論. (Visual Information Processing ) 開講所属 : 学際情報学府水 (Wed)5 [16:50-18:35]

視覚情報処理論. (Visual Information Processing ) 開講所属 : 学際情報学府水 (Wed)5 [16:50-18:35] 視覚情報処理論 (Visual Information Processing ) 開講所属 : 学際情報学府水 (Wed)5 [16:50-18:35] Computer Vision Design algorithms to implement the function of human vision 3D reconstruction from 2D image (retinal image)

More information

Human Body Recognition and Tracking: How the Kinect Works. Kinect RGB-D Camera. What the Kinect Does. How Kinect Works: Overview

Human Body Recognition and Tracking: How the Kinect Works. Kinect RGB-D Camera. What the Kinect Does. How Kinect Works: Overview Human Body Recognition and Tracking: How the Kinect Works Kinect RGB-D Camera Microsoft Kinect (Nov. 2010) Color video camera + laser-projected IR dot pattern + IR camera $120 (April 2012) Kinect 1.5 due

More information

Targetless Calibration of a Lidar - Perspective Camera Pair. Levente Tamás, Zoltán Kató

Targetless Calibration of a Lidar - Perspective Camera Pair. Levente Tamás, Zoltán Kató Targetless Calibration of a Lidar - Perspective Camera Pair Levente Tamás, Zoltán Kató Content Introduction Region-based calibration framework Evaluation on synthetic data Real data experiments Conclusions

More information

arxiv: v1 [cs.cv] 28 Sep 2018

arxiv: v1 [cs.cv] 28 Sep 2018 Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,

More information

Motion Capture using Body Mounted Cameras in an Unknown Environment

Motion Capture using Body Mounted Cameras in an Unknown Environment Motion Capture using Body Mounted Cameras in an Unknown Environment Nam Vo Taeyoung Kim Siddharth Choudhary 1. The Problem Motion capture has been recently used to provide much of character motion in several

More information

Augmented Reality, Advanced SLAM, Applications

Augmented Reality, Advanced SLAM, Applications Augmented Reality, Advanced SLAM, Applications Prof. Didier Stricker & Dr. Alain Pagani alain.pagani@dfki.de Lecture 3D Computer Vision AR, SLAM, Applications 1 Introduction Previous lectures: Basics (camera,

More information

Lecture 10 Multi-view Stereo (3D Dense Reconstruction) Davide Scaramuzza

Lecture 10 Multi-view Stereo (3D Dense Reconstruction) Davide Scaramuzza Lecture 10 Multi-view Stereo (3D Dense Reconstruction) Davide Scaramuzza REMODE: Probabilistic, Monocular Dense Reconstruction in Real Time, ICRA 14, by Pizzoli, Forster, Scaramuzza [M. Pizzoli, C. Forster,

More information

Deep Supervision with Shape Concepts for Occlusion-Aware 3D Object Parsing

Deep Supervision with Shape Concepts for Occlusion-Aware 3D Object Parsing Deep Supervision with Shape Concepts for Occlusion-Aware 3D Object Parsing Supplementary Material Introduction In this supplementary material, Section 2 details the 3D annotation for CAD models and real

More information

W4. Perception & Situation Awareness & Decision making

W4. Perception & Situation Awareness & Decision making W4. Perception & Situation Awareness & Decision making Robot Perception for Dynamic environments: Outline & DP-Grids concept Dynamic Probabilistic Grids Bayesian Occupancy Filter concept Dynamic Probabilistic

More information

Segmentation and Tracking of Partial Planar Templates

Segmentation and Tracking of Partial Planar Templates Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract

More information

3D Modeling from Range Images

3D Modeling from Range Images 1 3D Modeling from Range Images A Comprehensive System for 3D Modeling from Range Images Acquired from a 3D ToF Sensor Dipl.-Inf. March 22th, 2007 Sensor and Motivation 2 3D sensor PMD 1k-S time-of-flight

More information

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University. 3D Computer Vision Structured Light II Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de 1 Introduction

More information

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Structured Light II Johannes Köhler Johannes.koehler@dfki.de Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Introduction Previous lecture: Structured Light I Active Scanning Camera/emitter

More information

Feature Tracking and Optical Flow

Feature Tracking and Optical Flow Feature Tracking and Optical Flow Prof. D. Stricker Doz. G. Bleser Many slides adapted from James Hays, Derek Hoeim, Lana Lazebnik, Silvio Saverse, who 1 in turn adapted slides from Steve Seitz, Rick Szeliski,

More information

Application questions. Theoretical questions

Application questions. Theoretical questions The oral exam will last 30 minutes and will consist of one application question followed by two theoretical questions. Please find below a non exhaustive list of possible application questions. The list

More information

Outline. 1 Why we re interested in Real-Time tracking and mapping. 3 Kinect Fusion System Overview. 4 Real-time Surface Mapping

Outline. 1 Why we re interested in Real-Time tracking and mapping. 3 Kinect Fusion System Overview. 4 Real-time Surface Mapping Outline CSE 576 KinectFusion: Real-Time Dense Surface Mapping and Tracking PhD. work from Imperial College, London Microsoft Research, Cambridge May 6, 2013 1 Why we re interested in Real-Time tracking

More information

Stable Vision-Aided Navigation for Large-Area Augmented Reality

Stable Vision-Aided Navigation for Large-Area Augmented Reality Stable Vision-Aided Navigation for Large-Area Augmented Reality Taragay Oskiper, Han-Pang Chiu, Zhiwei Zhu Supun Samarasekera, Rakesh Teddy Kumar Vision and Robotics Laboratory SRI-International Sarnoff,

More information

Lecture: Autonomous micro aerial vehicles

Lecture: Autonomous micro aerial vehicles Lecture: Autonomous micro aerial vehicles Friedrich Fraundorfer Remote Sensing Technology TU München 1/41 Autonomous operation@eth Zürich Start 2/41 Autonomous operation@eth Zürich 3/41 Outline MAV system

More information

Stereo Rig Final Report

Stereo Rig Final Report Stereo Rig Final Report Yifei Zhang Abstract The ability to generate 3D images for the underwater environment offers researchers better insights by enabling them to record scenes for future analysis. The

More information

A Benchmark for RGB-D Visual Odometry, 3D Reconstruction and SLAM

A Benchmark for RGB-D Visual Odometry, 3D Reconstruction and SLAM 2014 IEEE International Conference on Robotics & Automation (ICRA) Hong Kong Convention and Exhibition Center May 31 - June 7, 2014. Hong Kong, China A Benchmark for RGB-D Visual Odometry, 3D Reconstruction

More information

Visual-Inertial Localization and Mapping for Robot Navigation

Visual-Inertial Localization and Mapping for Robot Navigation Visual-Inertial Localization and Mapping for Robot Navigation Dr. Guillermo Gallego Robotics & Perception Group University of Zurich Davide Scaramuzza University of Zurich - http://rpg.ifi.uzh.ch Mocular,

More information

Computational Optical Imaging - Optique Numerique. -- Single and Multiple View Geometry, Stereo matching --

Computational Optical Imaging - Optique Numerique. -- Single and Multiple View Geometry, Stereo matching -- Computational Optical Imaging - Optique Numerique -- Single and Multiple View Geometry, Stereo matching -- Autumn 2015 Ivo Ihrke with slides by Thorsten Thormaehlen Reminder: Feature Detection and Matching

More information

Lecture 10 Dense 3D Reconstruction

Lecture 10 Dense 3D Reconstruction Institute of Informatics Institute of Neuroinformatics Lecture 10 Dense 3D Reconstruction Davide Scaramuzza 1 REMODE: Probabilistic, Monocular Dense Reconstruction in Real Time M. Pizzoli, C. Forster,

More information

Efficient Online Surface Correction for Real-time Large-Scale 3D Reconstruction

Efficient Online Surface Correction for Real-time Large-Scale 3D Reconstruction MAIER ET AL.: EFFICIENT LARGE-SCALE ONLINE SURFACE CORRECTION 1 Efficient Online Surface Correction for Real-time Large-Scale 3D Reconstruction Robert Maier maierr@in.tum.de Raphael Schaller schaller@in.tum.de

More information

Deep Supervision with Shape Concepts for Occlusion-Aware 3D Object Parsing Supplementary Material

Deep Supervision with Shape Concepts for Occlusion-Aware 3D Object Parsing Supplementary Material Deep Supervision with Shape Concepts for Occlusion-Aware 3D Object Parsing Supplementary Material Chi Li, M. Zeeshan Zia 2, Quoc-Huy Tran 2, Xiang Yu 2, Gregory D. Hager, and Manmohan Chandraker 2 Johns

More information

Data Association for SLAM

Data Association for SLAM CALIFORNIA INSTITUTE OF TECHNOLOGY ME/CS 132a, Winter 2011 Lab #2 Due: Mar 10th, 2011 Part I Data Association for SLAM 1 Introduction For this part, you will experiment with a simulation of an EKF SLAM

More information

Improving Initial Estimations for Structure from Motion Methods

Improving Initial Estimations for Structure from Motion Methods Improving Initial Estimations for Structure from Motion Methods University of Bonn Outline Motivation Computer-Vision Basics Stereo Vision Bundle Adjustment Feature Matching Global Initial Estimation Component

More information

KinectFusion: Real-Time Dense Surface Mapping and Tracking

KinectFusion: Real-Time Dense Surface Mapping and Tracking KinectFusion: Real-Time Dense Surface Mapping and Tracking Gabriele Bleser Thanks to Richard Newcombe for providing the ISMAR slides Overview General: scientific papers (structure, category) KinectFusion:

More information

CVPR 2014 Visual SLAM Tutorial Kintinuous

CVPR 2014 Visual SLAM Tutorial Kintinuous CVPR 2014 Visual SLAM Tutorial Kintinuous kaess@cmu.edu The Robotics Institute Carnegie Mellon University Recap: KinectFusion [Newcombe et al., ISMAR 2011] RGB-D camera GPU 3D/color model RGB TSDF (volumetric

More information

Topological Mapping. Discrete Bayes Filter

Topological Mapping. Discrete Bayes Filter Topological Mapping Discrete Bayes Filter Vision Based Localization Given a image(s) acquired by moving camera determine the robot s location and pose? Towards localization without odometry What can be

More information

Introduction to Mobile Robotics Techniques for 3D Mapping

Introduction to Mobile Robotics Techniques for 3D Mapping Introduction to Mobile Robotics Techniques for 3D Mapping Wolfram Burgard, Michael Ruhnke, Bastian Steder 1 Why 3D Representations Robots live in the 3D world. 2D maps have been applied successfully for

More information

Autonomous Navigation in Complex Indoor and Outdoor Environments with Micro Aerial Vehicles

Autonomous Navigation in Complex Indoor and Outdoor Environments with Micro Aerial Vehicles Autonomous Navigation in Complex Indoor and Outdoor Environments with Micro Aerial Vehicles Shaojie Shen Dept. of Electrical and Systems Engineering & GRASP Lab, University of Pennsylvania Committee: Daniel

More information

Automatic Tracking of Moving Objects in Video for Surveillance Applications

Automatic Tracking of Moving Objects in Video for Surveillance Applications Automatic Tracking of Moving Objects in Video for Surveillance Applications Manjunath Narayana Committee: Dr. Donna Haverkamp (Chair) Dr. Arvin Agah Dr. James Miller Department of Electrical Engineering

More information

Monocular Visual Odometry

Monocular Visual Odometry Elective in Robotics coordinator: Prof. Giuseppe Oriolo Monocular Visual Odometry (slides prepared by Luca Ricci) Monocular vs. Stereo: eamples from Nature Predator Predators eyes face forward. The field

More information

DNA-SLAM: Dense Noise Aware SLAM for ToF RGB-D Cameras

DNA-SLAM: Dense Noise Aware SLAM for ToF RGB-D Cameras DNA-SLAM: Dense Noise Aware SLAM for ToF RGB-D Cameras Oliver Wasenmüller, Mohammad Dawud Ansari, Didier Stricker German Research Center for Artificial Intelligence (DFKI) University of Kaiserslautern

More information

Autonomous navigation in industrial cluttered environments using embedded stereo-vision

Autonomous navigation in industrial cluttered environments using embedded stereo-vision Autonomous navigation in industrial cluttered environments using embedded stereo-vision Julien Marzat ONERA Palaiseau Aerial Robotics workshop, Paris, 8-9 March 2017 1 Copernic Lab (ONERA Palaiseau) Research

More information

ME/CS 132: Introduction to Vision-based Robot Navigation! Low-level Image Processing" Larry Matthies"

ME/CS 132: Introduction to Vision-based Robot Navigation! Low-level Image Processing Larry Matthies ME/CS 132: Introduction to Vision-based Robot Navigation! Low-level Image Processing" Larry Matthies" lhm@jpl.nasa.gov, 818-354-3722" Announcements" First homework grading is done! Second homework is due

More information

Exploiting Depth Camera for 3D Spatial Relationship Interpretation

Exploiting Depth Camera for 3D Spatial Relationship Interpretation Exploiting Depth Camera for 3D Spatial Relationship Interpretation Jun Ye Kien A. Hua Data Systems Group, University of Central Florida Mar 1, 2013 Jun Ye and Kien A. Hua (UCF) 3D directional spatial relationships

More information

Deep learning for dense per-pixel prediction. Chunhua Shen The University of Adelaide, Australia

Deep learning for dense per-pixel prediction. Chunhua Shen The University of Adelaide, Australia Deep learning for dense per-pixel prediction Chunhua Shen The University of Adelaide, Australia Image understanding Classification error Convolution Neural Networks 0.3 0.2 0.1 Image Classification [Krizhevsky

More information

Step-by-Step Model Buidling

Step-by-Step Model Buidling Step-by-Step Model Buidling Review Feature selection Feature selection Feature correspondence Camera Calibration Euclidean Reconstruction Landing Augmented Reality Vision Based Control Sparse Structure

More information

RGB-D Camera-Based Parallel Tracking and Meshing

RGB-D Camera-Based Parallel Tracking and Meshing RGB-D Camera-Based Parallel Tracking and Meshing Sebastian Lieberknecht Andrea Huber Slobodan Ilic Selim Benhimane metaio GmbH metaio GmbH TUM metaio GmbH Figure 1: A real-time tracking method based on

More information

Lost! Leveraging the Crowd for Probabilistic Visual Self-Localization

Lost! Leveraging the Crowd for Probabilistic Visual Self-Localization Lost! Leveraging the Crowd for Probabilistic Visual Self-Localization Marcus A. Brubaker (Toyota Technological Institute at Chicago) Andreas Geiger (Karlsruhe Institute of Technology & MPI Tübingen) Raquel

More information

ROBUST SCALE ESTIMATION FOR MONOCULAR VISUAL ODOMETRY USING STRUCTURE FROM MOTION AND VANISHING POINTS

ROBUST SCALE ESTIMATION FOR MONOCULAR VISUAL ODOMETRY USING STRUCTURE FROM MOTION AND VANISHING POINTS ROBUST SCALE ESTIMATION FOR MONOCULAR VISUAL ODOMETRY USING STRUCTURE FROM MOTION AND VANISHING POINTS Johannes Gräter 1, Tobias Schwarze 1, Martin Lauer 1 August 20, 2015 Abstract While monocular visual

More information

Accurate Figure Flying with a Quadrocopter Using Onboard Visual and Inertial Sensing

Accurate Figure Flying with a Quadrocopter Using Onboard Visual and Inertial Sensing Accurate Figure Flying with a Quadrocopter Using Onboard Visual and Inertial Sensing Jakob Engel, Jürgen Sturm, Daniel Cremers Abstract We present an approach that enables a low-cost quadrocopter to accurately

More information

SLAM II: SLAM for robotic vision-based perception

SLAM II: SLAM for robotic vision-based perception SLAM II: SLAM for robotic vision-based perception Margarita Chli Martin Rufli, Roland Siegwart Margarita Chli, Martin Rufli, Roland Siegwart SLAM II today s lecture Last time: how to do SLAM? Today: what

More information

A Photometrically Calibrated Benchmark For Monocular Visual Odometry

A Photometrically Calibrated Benchmark For Monocular Visual Odometry A Photometrically Calibrated Benchmark For Monocular Visual Odometry Jakob Engel and Vladyslav Usenko and Daniel Cremers Technical University Munich Abstract arxiv:67.2555v2 [cs.cv] 8 Oct 26 We present

More information

SLAM II: SLAM for robotic vision-based perception

SLAM II: SLAM for robotic vision-based perception SLAM II: SLAM for robotic vision-based perception Margarita Chli Martin Rufli, Roland Siegwart Margarita Chli, Martin Rufli, Roland Siegwart SLAM II today s lecture Last time: how to do SLAM? Today: what

More information

15 Years of Visual SLAM

15 Years of Visual SLAM 15 Years of Visual SLAM Andrew Davison Robot Vision Group and Dyson Robotics Laboratory Department of Computing Imperial College London www.google.com/+andrewdavison December 18, 2015 What Has Defined

More information

arxiv: v2 [cs.cv] 26 Aug 2018

arxiv: v2 [cs.cv] 26 Aug 2018 A Synchronized Stereo and Plenoptic Visual Odometry Dataset Niclas Zeller 1,2, Franz Quint 2, and Uwe Stilla 1 1 Technische Universität München niclas.zeller@tum.de, stilla@tum.de 2 Karlsruhe University

More information

Deep Learning for Virtual Shopping. Dr. Jürgen Sturm Group Leader RGB-D

Deep Learning for Virtual Shopping. Dr. Jürgen Sturm Group Leader RGB-D Deep Learning for Virtual Shopping Dr. Jürgen Sturm Group Leader RGB-D metaio GmbH Augmented Reality with the Metaio SDK: IKEA Catalogue App Metaio: Augmented Reality Metaio SDK for ios, Android and Windows

More information

Project Updates Short lecture Volumetric Modeling +2 papers

Project Updates Short lecture Volumetric Modeling +2 papers Volumetric Modeling Schedule (tentative) Feb 20 Feb 27 Mar 5 Introduction Lecture: Geometry, Camera Model, Calibration Lecture: Features, Tracking/Matching Mar 12 Mar 19 Mar 26 Apr 2 Apr 9 Apr 16 Apr 23

More information

Long-term motion estimation from images

Long-term motion estimation from images Long-term motion estimation from images Dennis Strelow 1 and Sanjiv Singh 2 1 Google, Mountain View, CA, strelow@google.com 2 Carnegie Mellon University, Pittsburgh, PA, ssingh@cmu.edu Summary. Cameras

More information

Robot Localization based on Geo-referenced Images and G raphic Methods

Robot Localization based on Geo-referenced Images and G raphic Methods Robot Localization based on Geo-referenced Images and G raphic Methods Sid Ahmed Berrabah Mechanical Department, Royal Military School, Belgium, sidahmed.berrabah@rma.ac.be Janusz Bedkowski, Łukasz Lubasiński,

More information

Camera calibration. Robotic vision. Ville Kyrki

Camera calibration. Robotic vision. Ville Kyrki Camera calibration Robotic vision 19.1.2017 Where are we? Images, imaging Image enhancement Feature extraction and matching Image-based tracking Camera models and calibration Pose estimation Motion analysis

More information

Integrating Depth and Color Cues for Dense Multi-Resolution Scene Mapping Using RGB-D Cameras

Integrating Depth and Color Cues for Dense Multi-Resolution Scene Mapping Using RGB-D Cameras In Proc. of the IEEE Int. Conf. on Multisensor Fusion and Information Integration (MFI), Hamburg, Germany, 2012 Integrating Depth and Color Cues for Dense Multi-Resolution Scene Mapping Using RGB-D Cameras

More information

Peripheral drift illusion

Peripheral drift illusion Peripheral drift illusion Does it work on other animals? Computer Vision Motion and Optical Flow Many slides adapted from J. Hays, S. Seitz, R. Szeliski, M. Pollefeys, K. Grauman and others Video A video

More information

Estimating Camera Position And Posture by Using Feature Landmark Database

Estimating Camera Position And Posture by Using Feature Landmark Database Estimating Camera Position And Posture by Using Feature Landmark Database Motoko Oe 1, Tomokazu Sato 2 and Naokazu Yokoya 2 1 IBM Japan 2 Nara Institute of Science and Technology, Japan Abstract. Estimating

More information

Outline. ETN-FPI Training School on Plenoptic Sensing

Outline. ETN-FPI Training School on Plenoptic Sensing Outline Introduction Part I: Basics of Mathematical Optimization Linear Least Squares Nonlinear Optimization Part II: Basics of Computer Vision Camera Model Multi-Camera Model Multi-Camera Calibration

More information

Direct Multichannel Tracking

Direct Multichannel Tracking MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Direct Multichannel Tracking Jaramillo, C.; Taguchi, Y.; Feng, C. TR2017-146 October 2017 Abstract We present direct multichannel tracking,

More information

Personal Navigation and Indoor Mapping: Performance Characterization of Kinect Sensor-based Trajectory Recovery

Personal Navigation and Indoor Mapping: Performance Characterization of Kinect Sensor-based Trajectory Recovery Personal Navigation and Indoor Mapping: Performance Characterization of Kinect Sensor-based Trajectory Recovery 1 Charles TOTH, 1 Dorota BRZEZINSKA, USA 2 Allison KEALY, Australia, 3 Guenther RETSCHER,

More information

Structured light 3D reconstruction

Structured light 3D reconstruction Structured light 3D reconstruction Reconstruction pipeline and industrial applications rodola@dsi.unive.it 11/05/2010 3D Reconstruction 3D reconstruction is the process of capturing the shape and appearance

More information

Eppur si muove ( And yet it moves )

Eppur si muove ( And yet it moves ) Eppur si muove ( And yet it moves ) - Galileo Galilei University of Texas at Arlington Tracking of Image Features CSE 4392-5369 Vision-based Robot Sensing, Localization and Control Dr. Gian Luca Mariottini,

More information