Dense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm
|
|
- Johnathan Bennett
- 6 years ago
- Views:
Transcription
1 Computer Vision Group Prof. Daniel Cremers Dense Tracking and Mapping for Autonomous Quadrocopters Jürgen Sturm Joint work with Frank Steinbrücker, Jakob Engel, Christian Kerl, Erik Bylow, and Daniel Cremers
2 Motivation Imagine you have a flying camera What would you use it for? Dense Tracking and Mapping for Autonomous Quadrocopters 2 Dr.
3 Motivation Imagine you have a flying camera What would you use it for? Dense Tracking and Mapping for Autonomous Quadrocopters 3 Dr.
4 Motivation Imagine you have a flying camera What would you use it for? Dense Tracking and Mapping for Autonomous Quadrocopters 4 Dr.
5 Motivation Imagine you have a flying camera What would you use it for? Dense Tracking and Mapping for Autonomous Quadrocopters 5 Dr.
6 Motivation Our research goal: Enable flying robots to operate autonomously in 3D environments using onboard cameras Use cameras because light weight and rich data Navigation, localization, mapping, exploration, people following,
7 Feature-based Visual Navigation [Engel, Sturm, Cremers, IROS 12]
8 Feature-based Visual Navigation [Engel, Sturm, Cremers, IROS 12] Architecture Quadrocopter Monocular SLAM Extended Kalman Filter PID Control Based on PTAM [Klein and Murray, ISMAR 07] Dense Tracking and Mapping for Autonomous Quadrocopters 8
9 Motivation Video feed from quadrocopter Dense Tracking and Mapping for Autonomous Quadrocopters 9
10 Motivation What PTAM actually sees Dense Tracking and Mapping for Autonomous Quadrocopters 10
11 Motivation Problem: Most approaches only use a small fraction of the available data Keypoint detection Visual features Question: How can we use most/all information to maximize the performance? Dense Tracking and Mapping for Autonomous Quadrocopters 11
12 Outline of the Talk Part 1: Dense tracking Part 2: Dense reconstruction Part 3: Evaluation and benchmarking Dense Tracking and Mapping for Autonomous Quadrocopters 12
13 Related Work on Dense Tracking Lucas and Kanade (IJCAI 81) Lovegrove et al. (IV 11) Newcombe et al. (ICCV 11) Comport/Tykkälä et al. (ICCV 11) Dense Tracking and Mapping for Autonomous Quadrocopters 13
14 Dense Tracking How can we exploit all data of an RGB-D image? Idea Photo-consistency constraint for all pixels Dense Tracking and Mapping for Autonomous Quadrocopters 14
15 How to deal with noise? Photo-consistency constraint will not perfectly hold Sensor noise Pose error Reflections, specular surfaces Dynamic objects (e.g., walking people) Residuals will be non-zero Residual distribution Dense Tracking and Mapping for Autonomous Quadrocopters 15
16 Probability p(r) Residual Distribution Zero-mean, peaked distribution Example: Correct camera pose 0,3 0,25 0,2 0,15 0,1 Correct camera pose 0, Residuals r Dense Tracking and Mapping for Autonomous Quadrocopters 16
17 Probability p(r) Residual Distribution Zero-mean, peaked distribution Example: Wrong camera pose 0,3 0,25 0,2 0,15 0,1 Wrong camera pose 0, Residuals r Dense Tracking and Mapping for Autonomous Quadrocopters 17
18 Probability p(r) Residual Distribution Goal: Find the camera pose that maximizes the observation likelihood 0,3 0,25 0,2 0,15 0,1 0,05 0 Wrong camera pose Correct camera pose Residuals r Dense Tracking and Mapping for Autonomous Quadrocopters 18
19 Motion Estimation Goal: Find the camera pose that maximizes the observation likelihood compute over all pixels Assume pixel-wise residuals are conditionally independent How can we solve this optimization problem? Dense Tracking and Mapping for Autonomous Quadrocopters 19
20 Approach Take negative logarithm Set derivative to zero Dense Tracking and Mapping for Autonomous Quadrocopters 20
21 Approach (cont.d) This can be rewritten as a weighted least squares problem with weights is non-linear in Need to linearize, solve, and iterate Dense Tracking and Mapping for Autonomous Quadrocopters 21
22 Iteratively Reweighted Least Squares Problem: Algorithm: 1. Compute weights 2. Linearize in the camera motion 3. Build and solve normal equations 4. Repeat until convergence Dense Tracking and Mapping for Autonomous Quadrocopters 22
23 Example First input image Second input image Residuals Dense Tracking and Mapping for Autonomous Quadrocopters 23 Image Jacobian for camera motion along x axis
24 What is a Good Model for the Residual Distribution? Dense Tracking and Mapping for Autonomous Quadrocopters 24
25 Weighted Error Dense Tracking and Mapping for Autonomous Quadrocopters 25
26 Example Weights Robust sensor model allows to down-weight outliers (dynamic objects, motion blur, reflections, ) Scene Residuals Weights Dense Tracking and Mapping for Autonomous Quadrocopters 26
27 Coarse-to-Fine Linearization only holds for small motions Coarse-to-fine scheme Image pyramids Dense Tracking and Mapping for Autonomous Quadrocopters 27
28 Dense Tracking: Results Dense Tracking and Mapping for Autonomous Quadrocopters 28
29 Summary: Dense tracking Direct matching of consecutive RGB-D images Pro Super fast, highly accurate (30 Hz on CPU) Low, constant memory consumption Con Accumulates drift over time, sometimes diverges How can we reduce/eliminate drift? Dense Tracking and Mapping for Autonomous Quadrocopters 29
30 Dense Tracking and Mapping Idea: Instead of tracking from frame-to-frame, track frame-to-model to eliminate drift t+1 t vs. Question: Where do we get the model from? Dense Tracking and Mapping for Autonomous Quadrocopters 30
31 Dense Tracking and Mapping Idea: Compute an iterative solution 1. Reconstruct model with known poses 2. Track camera with respect to known model Next question: How to represent the model? Dense Tracking and Mapping for Autonomous Quadrocopters 31
32 Representation of the 3D Model Idea: Instead of representing the cell occupancy, represent the distance of each cell to the surface Occupancy grid maps zero = free space Signed distance function (SDF) negative = outside obj. Dense Tracking and Mapping for Autonomous Quadrocopters 32 z = 1.8 x z = 1.8 x one = occupied x positive = inside obj. x
33 Dense Mapping: 2D Example Camera with known pose Dense Tracking and Mapping for Autonomous Quadrocopters 33
34 Dense Mapping: 2D Example Camera with known pose Grid with signed distance function Dense Tracking and Mapping for Autonomous Quadrocopters 34
35 Dense Mapping: 2D Example For each grid cell, compute its projective distance to the surface projective distance Dense Tracking and Mapping for Autonomous Quadrocopters 35
36 Dense Mapping: 3D Example Generalizes directly to 3D But: memory usage is cubic in side length Dense Tracking and Mapping for Autonomous Quadrocopters 36
37 Dense Tracking Given: SDF with projective distance to surface New depth image with unknown camera pose Wanted: Camera pose Our approach: Optimize camera pose so that the observed depth image minimizes the distance in the SDF Dense Tracking and Mapping for Autonomous Quadrocopters 37
38 Dense Tracking: 2D Example 3D model built from the first k frames Dense Tracking and Mapping for Autonomous Quadrocopters 38
39 Dense Tracking: 2D Example Minimize distance between depth image and SDF Dense Tracking and Mapping for Autonomous Quadrocopters 39
40 Dense Tracking: 2D Example Minimize distance between depth image and SDF Dense Tracking and Mapping for Autonomous Quadrocopters 40
41 Dense Tracking: 2D Example Minimize distance between depth image and SDF Dense Tracking and Mapping for Autonomous Quadrocopters 41
42 3D Example Given: A sequence of depth images Wanted: Accurate and dense 3D model Dense Tracking and Mapping for Autonomous Quadrocopters 42
43 Resulting 3D Model Dense Tracking and Mapping for Autonomous Quadrocopters 43
44 Can We Print These Models in 3D? Dense Tracking and Mapping for Autonomous Quadrocopters 44
45 Summary: Dense 3D Reconstruction Frame-to-model tracking Pro: Reduces drift significantly Outputs nice 3D models Con: Memory intensive (~2 GB for ) Computationally more expensive (~60 Hz on GPU) Dense Tracking and Mapping for Autonomous Quadrocopters 45
46 Evaluation and Benchmarking How can we evaluate such methods? What are good evaluation criteria? Dense Tracking and Mapping for Autonomous Quadrocopters 46
47 Existing Benchmarks Intel dataset: laser + odometry [Haehnel et al., 2004] New College dataset: stereo + omni-directional vision + laser + IMU [Smith et al., IJRR 2009] KITTI Vision benchmark: stereo [Geiger et al., CVPR 12] Our contribution: Dataset for RGB-D evaluation Intel New College KITTI RGB-D Dense Tracking and Mapping for Autonomous Quadrocopters 47
48 Recorded Scenes Different environments (office, industrial hall, ) Large variations in camera speed, camera motion, illumination, number of features, dynamic objects, Handheld and robot-mounted sensor Dense Tracking and Mapping for Autonomous Quadrocopters 48
49 Dataset Acquisition Motion capture system Camera pose (100 Hz) Microsoft Kinect (later: Asus Xtion Pro Live) Color images (30 Hz) Depth images (30 Hz) External video camera (for documentation) Dense Tracking and Mapping for Autonomous Quadrocopters 49
50 Motion Capture System 9 high-speed cameras mounted in room Cameras have active illumination and pre-process image (thresholding) Cameras track positions of retro-reflective markers Dense Tracking and Mapping for Autonomous Quadrocopters 50
51 Calibration Calibration of the overall system is not trivial: 1. Intrinsic calibration (Mocap + Kinect) 2. Extrinsic calibration (Kinect vs. Mocap) 3. Time synchronization (Kinect vs. Mocap) Dense Tracking and Mapping for Autonomous Quadrocopters 51
52 Dense Tracking and Mapping for Autonomous Quadrocopters 52
53 Dense Tracking and Mapping for Autonomous Quadrocopters 53
54 Dense Tracking and Mapping for Autonomous Quadrocopters 54
55 File Formats In total: 69 sequences (33 training, 36 testing) One TGZ archive per sequence, containing Color and depth images (PNG) List of color images (timestamp filename) List of depth images (timestamp filename) List of camera poses (timestamp tx ty tz qx qy qz qw) Dense Tracking and Mapping for Autonomous Quadrocopters 55
56 What Is a Good Evaluation Metric? Visual odometry system outputs Camera trajectory (accumulated) Visual SLAM system outputs Camera trajectory 3D map Ground truth Camera trajectory Dense Tracking and Mapping for Autonomous Quadrocopters 56
57 What Is a Good Evaluation Metric? Trajectory comparison Ground truth trajectory Estimate camera trajectory Two evaluation metrics Drift per second Global consistency Ground truth trajectory Q 1:n Estimated camera trajectory P 1:n Evaluation function Scalar performance index Dense Tracking and Mapping for Autonomous Quadrocopters 57
58 Relative Pose Error (RPE) Measures the (relative) drift between the i-th frame and the (i+δ)-th frame Relative error True motion Estimated motion Ground truth Estimated traj. Relative error Dense Tracking and Mapping for Autonomous Quadrocopters 58
59 Relative Pose Error (RPE) How to choose the time delta Δ? For odometry methods: Δ=1: Drift per frame Δ=30: Drift per second For SLAM methods: Average over all possible deltas Measures the global consistency Dense Tracking and Mapping for Autonomous Quadrocopters 59
60 Absolute Trajectory Error (ATE) Alternative method to evaluate SLAM systems Requires pre-aligned trajectories Absolute error Groundtruth Alignment Estimated Absolute error Ground truth Pre-aligned estimated traj. Dense Tracking and Mapping for Autonomous Quadrocopters 60
61 Evaluation Tools Average over all time steps Evaluation scripts for both evaluation metrics available (Python) Output: RMSE, median, mean Plot to png/pdf (optional) Dense Tracking and Mapping for Autonomous Quadrocopters 61
62 Comparison of RPE and ATE RPE and ATE are strongly related RPE considers additionally rotational errors RPE ATE Dense Tracking and Mapping for Autonomous Quadrocopters 62
63 Example: Online Evaluation Dense Tracking and Mapping for Autonomous Quadrocopters 63
64 Dense Tracking and Mapping for Autonomous Quadrocopters 64
65 Summary TUM RGB-D Benchmark Dataset for the evaluation of RGB-D SLAM systems Ground-truth camera poses Evaluation metrics + tools available Since August 2011: > visitors >4.500 online trajectory evaluations >15 published research papers using the dataset Next steps: Possibility to upload own trajectories/publications Results page and automated ranking Dense Tracking and Mapping for Autonomous Quadrocopters 65
66 Conclusion Dense methods bear a large potential Dense camera tracking Dense 3D reconstruction Use benchmark for the comparison of alternative approaches Release as open-source in preparation Please contact us if you are interested in collaboration! Dense Tracking and Mapping for Autonomous Quadrocopters 66
Visual Navigation for Flying Robots Exploration, Multi-Robot Coordination and Coverage
Computer Vision Group Prof. Daniel Cremers Visual Navigation for Flying Robots Exploration, Multi-Robot Coordination and Coverage Dr. Jürgen Sturm Agenda for Today Exploration with a single robot Coordinated
More informationVisual Navigation for Flying Robots
Computer Vision Group Prof. Daniel Cremers Visual Navigation for Flying Robots Experimentation, Evaluation and Benchmarking Dr. Jürgen Sturm Agenda for Today Course Evaluation Scientific research: The
More informationScanning and Printing Objects in 3D
Scanning and Printing Objects in 3D Dr. Jürgen Sturm metaio GmbH (formerly Technical University of Munich) My Research Areas Visual navigation for mobile robots RoboCup Kinematic Learning Articulated Objects
More informationDense 3D Reconstruction from Autonomous Quadrocopters
Dense 3D Reconstruction from Autonomous Quadrocopters Computer Science & Mathematics TU Munich Martin Oswald, Jakob Engel, Christian Kerl, Frank Steinbrücker, Jan Stühmer & Jürgen Sturm Autonomous Quadrocopters
More informationScanning and Printing Objects in 3D Jürgen Sturm
Scanning and Printing Objects in 3D Jürgen Sturm Metaio (formerly Technical University of Munich) My Research Areas Visual navigation for mobile robots RoboCup Kinematic Learning Articulated Objects Quadrocopters
More informationJakob Engel, Thomas Schöps, Daniel Cremers Technical University Munich. LSD-SLAM: Large-Scale Direct Monocular SLAM
Computer Vision Group Technical University of Munich Jakob Engel LSD-SLAM: Large-Scale Direct Monocular SLAM Jakob Engel, Thomas Schöps, Daniel Cremers Technical University Munich Monocular Video Engel,
More informationSemi-Dense Direct SLAM
Computer Vision Group Technical University of Munich Jakob Engel Jakob Engel, Daniel Cremers David Caruso, Thomas Schöps, Lukas von Stumberg, Vladyslav Usenko, Jörg Stückler, Jürgen Sturm Technical University
More informationDirect Methods in Visual Odometry
Direct Methods in Visual Odometry July 24, 2017 Direct Methods in Visual Odometry July 24, 2017 1 / 47 Motivation for using Visual Odometry Wheel odometry is affected by wheel slip More accurate compared
More informationFLaME: Fast Lightweight Mesh Estimation using Variational Smoothing on Delaunay Graphs
FLaME: Fast Lightweight Mesh Estimation using Variational Smoothing on Delaunay Graphs W. Nicholas Greene Robust Robotics Group, MIT CSAIL LPM Workshop IROS 2017 September 28, 2017 with Nicholas Roy 1
More informationAutonomous Navigation for Flying Robots
Computer Vision Group Prof. Daniel Cremers Autonomous Navigation for Flying Robots Lecture 7.1: 2D Motion Estimation in Images Jürgen Sturm Technische Universität München 3D to 2D Perspective Projections
More informationarxiv: v1 [cs.ro] 24 Nov 2018
BENCHMARKING AND COMPARING POPULAR VISUAL SLAM ALGORITHMS arxiv:1811.09895v1 [cs.ro] 24 Nov 2018 Amey Kasar Department of Electronics and Telecommunication Engineering Pune Institute of Computer Technology
More informationColored Point Cloud Registration Revisited Supplementary Material
Colored Point Cloud Registration Revisited Supplementary Material Jaesik Park Qian-Yi Zhou Vladlen Koltun Intel Labs A. RGB-D Image Alignment Section introduced a joint photometric and geometric objective
More informationAutonomous Navigation for Flying Robots
Computer Vision Group Prof. Daniel Cremers Autonomous Navigation for Flying Robots Lecture 7.2: Visual Odometry Jürgen Sturm Technische Universität München Cascaded Control Robot Trajectory 0.1 Hz Visual
More informationIntrinsic3D: High-Quality 3D Reconstruction by Joint Appearance and Geometry Optimization with Spatially-Varying Lighting
Intrinsic3D: High-Quality 3D Reconstruction by Joint Appearance and Geometry Optimization with Spatially-Varying Lighting R. Maier 1,2, K. Kim 1, D. Cremers 2, J. Kautz 1, M. Nießner 2,3 Fusion Ours 1
More informationSuper-Resolution Keyframe Fusion for 3D Modeling with High-Quality Textures
Super-Resolution Keyframe Fusion for 3D Modeling with High-Quality Textures Robert Maier, Jörg Stückler, Daniel Cremers International Conference on 3D Vision (3DV) October 2015, Lyon, France Motivation
More informationNonlinear State Estimation for Robotics and Computer Vision Applications: An Overview
Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview Arun Das 05/09/2017 Arun Das Waterloo Autonomous Vehicles Lab Introduction What s in a name? Arun Das Waterloo Autonomous
More informationVision-based Mobile Robot Localization and Mapping using Scale-Invariant Features
Vision-based Mobile Robot Localization and Mapping using Scale-Invariant Features Stephen Se, David Lowe, Jim Little Department of Computer Science University of British Columbia Presented by Adam Bickett
More informationDeep Incremental Scene Understanding. Federico Tombari & Christian Rupprecht Technical University of Munich, Germany
Deep Incremental Scene Understanding Federico Tombari & Christian Rupprecht Technical University of Munich, Germany C. Couprie et al. "Toward Real-time Indoor Semantic Segmentation Using Depth Information"
More informationSimultaneous Localization and Mapping (SLAM)
Simultaneous Localization and Mapping (SLAM) RSS Lecture 16 April 8, 2013 Prof. Teller Text: Siegwart and Nourbakhsh S. 5.8 SLAM Problem Statement Inputs: No external coordinate reference Time series of
More informationHybrids Mixed Approaches
Hybrids Mixed Approaches Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline Why mixing? Parallel Tracking and Mapping Benefits
More informationVisual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech
Visual Odometry Features, Tracking, Essential Matrix, and RANSAC Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline The
More information3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller
3D Computer Vision Depth Cameras Prof. Didier Stricker Oliver Wasenmüller Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de
More informationGeometric Reconstruction Dense reconstruction of scene geometry
Lecture 5. Dense Reconstruction and Tracking with Real-Time Applications Part 2: Geometric Reconstruction Dr Richard Newcombe and Dr Steven Lovegrove Slide content developed from: [Newcombe, Dense Visual
More informationDealing with Scale. Stephan Weiss Computer Vision Group NASA-JPL / CalTech
Dealing with Scale Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline Why care about size? The IMU as scale provider: The
More informationFitting (LMedS, RANSAC)
Fitting (LMedS, RANSAC) Thursday, 23/03/2017 Antonis Argyros e-mail: argyros@csd.uoc.gr LMedS and RANSAC What if we have very many outliers? 2 1 Least Median of Squares ri : Residuals Least Squares n 2
More informationLecture 19: Depth Cameras. Visual Computing Systems CMU , Fall 2013
Lecture 19: Depth Cameras Visual Computing Systems Continuing theme: computational photography Cameras capture light, then extensive processing produces the desired image Today: - Capturing scene depth
More informationCSE 527: Introduction to Computer Vision
CSE 527: Introduction to Computer Vision Week 10 Class 2: Visual Odometry November 2nd, 2017 Today Visual Odometry Intro Algorithm SLAM Visual Odometry Input Output Images, Video Camera trajectory, motion
More informationSimultaneous Localization
Simultaneous Localization and Mapping (SLAM) RSS Technical Lecture 16 April 9, 2012 Prof. Teller Text: Siegwart and Nourbakhsh S. 5.8 Navigation Overview Where am I? Where am I going? Localization Assumed
More informationAccurate 3D Face and Body Modeling from a Single Fixed Kinect
Accurate 3D Face and Body Modeling from a Single Fixed Kinect Ruizhe Wang*, Matthias Hernandez*, Jongmoo Choi, Gérard Medioni Computer Vision Lab, IRIS University of Southern California Abstract In this
More informationVisual SLAM for small Unmanned Aerial Vehicles
Visual SLAM for small Unmanned Aerial Vehicles Margarita Chli Autonomous Systems Lab, ETH Zurich Simultaneous Localization And Mapping How can a body navigate in a previously unknown environment while
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 11 140311 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Motion Analysis Motivation Differential Motion Optical
More informationLive Metric 3D Reconstruction on Mobile Phones ICCV 2013
Live Metric 3D Reconstruction on Mobile Phones ICCV 2013 Main Contents 1. Target & Related Work 2. Main Features of This System 3. System Overview & Workflow 4. Detail of This System 5. Experiments 6.
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 14 130307 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Stereo Dense Motion Estimation Translational
More informationRobust Online 3D Reconstruction Combining a Depth Sensor and Sparse Feature Points
2016 23rd International Conference on Pattern Recognition (ICPR) Cancún Center, Cancún, México, December 4-8, 2016 Robust Online 3D Reconstruction Combining a Depth Sensor and Sparse Feature Points Erik
More informationDesign-Space Exploration for Dense 3D Scene Understanding: Exploring ElasticFusion
Design-Space Exploration for Dense 3D Scene Understanding: Exploring ElasticFusion Imperial College of Science, Technology and Medicine Department of Computing Author: Denise Carroll Supervisor: Prof.
More informationPL-SVO: Semi-Direct Monocular Visual Odometry by Combining Points and Line Segments
2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) Daejeon Convention Center October 9-14, 2016, Daejeon, Korea PL-SVO: Semi-Direct Monocular Visual Odometry by Combining Points
More information3D Scene Reconstruction with a Mobile Camera
3D Scene Reconstruction with a Mobile Camera 1 Introduction Robert Carrera and Rohan Khanna Stanford University: CS 231A Autonomous supernumerary arms, or "third arms", while still unconventional, hold
More informationOpenStreetSLAM: Global Vehicle Localization using OpenStreetMaps
OpenStreetSLAM: Global Vehicle Localization using OpenStreetMaps Georgios Floros, Benito van der Zander and Bastian Leibe RWTH Aachen University, Germany http://www.vision.rwth-aachen.de floros@vision.rwth-aachen.de
More information視覚情報処理論. (Visual Information Processing ) 開講所属 : 学際情報学府水 (Wed)5 [16:50-18:35]
視覚情報処理論 (Visual Information Processing ) 開講所属 : 学際情報学府水 (Wed)5 [16:50-18:35] Computer Vision Design algorithms to implement the function of human vision 3D reconstruction from 2D image (retinal image)
More informationHuman Body Recognition and Tracking: How the Kinect Works. Kinect RGB-D Camera. What the Kinect Does. How Kinect Works: Overview
Human Body Recognition and Tracking: How the Kinect Works Kinect RGB-D Camera Microsoft Kinect (Nov. 2010) Color video camera + laser-projected IR dot pattern + IR camera $120 (April 2012) Kinect 1.5 due
More informationTargetless Calibration of a Lidar - Perspective Camera Pair. Levente Tamás, Zoltán Kató
Targetless Calibration of a Lidar - Perspective Camera Pair Levente Tamás, Zoltán Kató Content Introduction Region-based calibration framework Evaluation on synthetic data Real data experiments Conclusions
More informationarxiv: v1 [cs.cv] 28 Sep 2018
Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,
More informationMotion Capture using Body Mounted Cameras in an Unknown Environment
Motion Capture using Body Mounted Cameras in an Unknown Environment Nam Vo Taeyoung Kim Siddharth Choudhary 1. The Problem Motion capture has been recently used to provide much of character motion in several
More informationAugmented Reality, Advanced SLAM, Applications
Augmented Reality, Advanced SLAM, Applications Prof. Didier Stricker & Dr. Alain Pagani alain.pagani@dfki.de Lecture 3D Computer Vision AR, SLAM, Applications 1 Introduction Previous lectures: Basics (camera,
More informationLecture 10 Multi-view Stereo (3D Dense Reconstruction) Davide Scaramuzza
Lecture 10 Multi-view Stereo (3D Dense Reconstruction) Davide Scaramuzza REMODE: Probabilistic, Monocular Dense Reconstruction in Real Time, ICRA 14, by Pizzoli, Forster, Scaramuzza [M. Pizzoli, C. Forster,
More informationDeep Supervision with Shape Concepts for Occlusion-Aware 3D Object Parsing
Deep Supervision with Shape Concepts for Occlusion-Aware 3D Object Parsing Supplementary Material Introduction In this supplementary material, Section 2 details the 3D annotation for CAD models and real
More informationW4. Perception & Situation Awareness & Decision making
W4. Perception & Situation Awareness & Decision making Robot Perception for Dynamic environments: Outline & DP-Grids concept Dynamic Probabilistic Grids Bayesian Occupancy Filter concept Dynamic Probabilistic
More informationSegmentation and Tracking of Partial Planar Templates
Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract
More information3D Modeling from Range Images
1 3D Modeling from Range Images A Comprehensive System for 3D Modeling from Range Images Acquired from a 3D ToF Sensor Dipl.-Inf. March 22th, 2007 Sensor and Motivation 2 3D sensor PMD 1k-S time-of-flight
More information3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.
3D Computer Vision Structured Light II Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de 1 Introduction
More informationStructured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov
Structured Light II Johannes Köhler Johannes.koehler@dfki.de Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Introduction Previous lecture: Structured Light I Active Scanning Camera/emitter
More informationFeature Tracking and Optical Flow
Feature Tracking and Optical Flow Prof. D. Stricker Doz. G. Bleser Many slides adapted from James Hays, Derek Hoeim, Lana Lazebnik, Silvio Saverse, who 1 in turn adapted slides from Steve Seitz, Rick Szeliski,
More informationApplication questions. Theoretical questions
The oral exam will last 30 minutes and will consist of one application question followed by two theoretical questions. Please find below a non exhaustive list of possible application questions. The list
More informationOutline. 1 Why we re interested in Real-Time tracking and mapping. 3 Kinect Fusion System Overview. 4 Real-time Surface Mapping
Outline CSE 576 KinectFusion: Real-Time Dense Surface Mapping and Tracking PhD. work from Imperial College, London Microsoft Research, Cambridge May 6, 2013 1 Why we re interested in Real-Time tracking
More informationStable Vision-Aided Navigation for Large-Area Augmented Reality
Stable Vision-Aided Navigation for Large-Area Augmented Reality Taragay Oskiper, Han-Pang Chiu, Zhiwei Zhu Supun Samarasekera, Rakesh Teddy Kumar Vision and Robotics Laboratory SRI-International Sarnoff,
More informationLecture: Autonomous micro aerial vehicles
Lecture: Autonomous micro aerial vehicles Friedrich Fraundorfer Remote Sensing Technology TU München 1/41 Autonomous operation@eth Zürich Start 2/41 Autonomous operation@eth Zürich 3/41 Outline MAV system
More informationStereo Rig Final Report
Stereo Rig Final Report Yifei Zhang Abstract The ability to generate 3D images for the underwater environment offers researchers better insights by enabling them to record scenes for future analysis. The
More informationA Benchmark for RGB-D Visual Odometry, 3D Reconstruction and SLAM
2014 IEEE International Conference on Robotics & Automation (ICRA) Hong Kong Convention and Exhibition Center May 31 - June 7, 2014. Hong Kong, China A Benchmark for RGB-D Visual Odometry, 3D Reconstruction
More informationVisual-Inertial Localization and Mapping for Robot Navigation
Visual-Inertial Localization and Mapping for Robot Navigation Dr. Guillermo Gallego Robotics & Perception Group University of Zurich Davide Scaramuzza University of Zurich - http://rpg.ifi.uzh.ch Mocular,
More informationComputational Optical Imaging - Optique Numerique. -- Single and Multiple View Geometry, Stereo matching --
Computational Optical Imaging - Optique Numerique -- Single and Multiple View Geometry, Stereo matching -- Autumn 2015 Ivo Ihrke with slides by Thorsten Thormaehlen Reminder: Feature Detection and Matching
More informationLecture 10 Dense 3D Reconstruction
Institute of Informatics Institute of Neuroinformatics Lecture 10 Dense 3D Reconstruction Davide Scaramuzza 1 REMODE: Probabilistic, Monocular Dense Reconstruction in Real Time M. Pizzoli, C. Forster,
More informationEfficient Online Surface Correction for Real-time Large-Scale 3D Reconstruction
MAIER ET AL.: EFFICIENT LARGE-SCALE ONLINE SURFACE CORRECTION 1 Efficient Online Surface Correction for Real-time Large-Scale 3D Reconstruction Robert Maier maierr@in.tum.de Raphael Schaller schaller@in.tum.de
More informationDeep Supervision with Shape Concepts for Occlusion-Aware 3D Object Parsing Supplementary Material
Deep Supervision with Shape Concepts for Occlusion-Aware 3D Object Parsing Supplementary Material Chi Li, M. Zeeshan Zia 2, Quoc-Huy Tran 2, Xiang Yu 2, Gregory D. Hager, and Manmohan Chandraker 2 Johns
More informationData Association for SLAM
CALIFORNIA INSTITUTE OF TECHNOLOGY ME/CS 132a, Winter 2011 Lab #2 Due: Mar 10th, 2011 Part I Data Association for SLAM 1 Introduction For this part, you will experiment with a simulation of an EKF SLAM
More informationImproving Initial Estimations for Structure from Motion Methods
Improving Initial Estimations for Structure from Motion Methods University of Bonn Outline Motivation Computer-Vision Basics Stereo Vision Bundle Adjustment Feature Matching Global Initial Estimation Component
More informationKinectFusion: Real-Time Dense Surface Mapping and Tracking
KinectFusion: Real-Time Dense Surface Mapping and Tracking Gabriele Bleser Thanks to Richard Newcombe for providing the ISMAR slides Overview General: scientific papers (structure, category) KinectFusion:
More informationCVPR 2014 Visual SLAM Tutorial Kintinuous
CVPR 2014 Visual SLAM Tutorial Kintinuous kaess@cmu.edu The Robotics Institute Carnegie Mellon University Recap: KinectFusion [Newcombe et al., ISMAR 2011] RGB-D camera GPU 3D/color model RGB TSDF (volumetric
More informationTopological Mapping. Discrete Bayes Filter
Topological Mapping Discrete Bayes Filter Vision Based Localization Given a image(s) acquired by moving camera determine the robot s location and pose? Towards localization without odometry What can be
More informationIntroduction to Mobile Robotics Techniques for 3D Mapping
Introduction to Mobile Robotics Techniques for 3D Mapping Wolfram Burgard, Michael Ruhnke, Bastian Steder 1 Why 3D Representations Robots live in the 3D world. 2D maps have been applied successfully for
More informationAutonomous Navigation in Complex Indoor and Outdoor Environments with Micro Aerial Vehicles
Autonomous Navigation in Complex Indoor and Outdoor Environments with Micro Aerial Vehicles Shaojie Shen Dept. of Electrical and Systems Engineering & GRASP Lab, University of Pennsylvania Committee: Daniel
More informationAutomatic Tracking of Moving Objects in Video for Surveillance Applications
Automatic Tracking of Moving Objects in Video for Surveillance Applications Manjunath Narayana Committee: Dr. Donna Haverkamp (Chair) Dr. Arvin Agah Dr. James Miller Department of Electrical Engineering
More informationMonocular Visual Odometry
Elective in Robotics coordinator: Prof. Giuseppe Oriolo Monocular Visual Odometry (slides prepared by Luca Ricci) Monocular vs. Stereo: eamples from Nature Predator Predators eyes face forward. The field
More informationDNA-SLAM: Dense Noise Aware SLAM for ToF RGB-D Cameras
DNA-SLAM: Dense Noise Aware SLAM for ToF RGB-D Cameras Oliver Wasenmüller, Mohammad Dawud Ansari, Didier Stricker German Research Center for Artificial Intelligence (DFKI) University of Kaiserslautern
More informationAutonomous navigation in industrial cluttered environments using embedded stereo-vision
Autonomous navigation in industrial cluttered environments using embedded stereo-vision Julien Marzat ONERA Palaiseau Aerial Robotics workshop, Paris, 8-9 March 2017 1 Copernic Lab (ONERA Palaiseau) Research
More informationME/CS 132: Introduction to Vision-based Robot Navigation! Low-level Image Processing" Larry Matthies"
ME/CS 132: Introduction to Vision-based Robot Navigation! Low-level Image Processing" Larry Matthies" lhm@jpl.nasa.gov, 818-354-3722" Announcements" First homework grading is done! Second homework is due
More informationExploiting Depth Camera for 3D Spatial Relationship Interpretation
Exploiting Depth Camera for 3D Spatial Relationship Interpretation Jun Ye Kien A. Hua Data Systems Group, University of Central Florida Mar 1, 2013 Jun Ye and Kien A. Hua (UCF) 3D directional spatial relationships
More informationDeep learning for dense per-pixel prediction. Chunhua Shen The University of Adelaide, Australia
Deep learning for dense per-pixel prediction Chunhua Shen The University of Adelaide, Australia Image understanding Classification error Convolution Neural Networks 0.3 0.2 0.1 Image Classification [Krizhevsky
More informationStep-by-Step Model Buidling
Step-by-Step Model Buidling Review Feature selection Feature selection Feature correspondence Camera Calibration Euclidean Reconstruction Landing Augmented Reality Vision Based Control Sparse Structure
More informationRGB-D Camera-Based Parallel Tracking and Meshing
RGB-D Camera-Based Parallel Tracking and Meshing Sebastian Lieberknecht Andrea Huber Slobodan Ilic Selim Benhimane metaio GmbH metaio GmbH TUM metaio GmbH Figure 1: A real-time tracking method based on
More informationLost! Leveraging the Crowd for Probabilistic Visual Self-Localization
Lost! Leveraging the Crowd for Probabilistic Visual Self-Localization Marcus A. Brubaker (Toyota Technological Institute at Chicago) Andreas Geiger (Karlsruhe Institute of Technology & MPI Tübingen) Raquel
More informationROBUST SCALE ESTIMATION FOR MONOCULAR VISUAL ODOMETRY USING STRUCTURE FROM MOTION AND VANISHING POINTS
ROBUST SCALE ESTIMATION FOR MONOCULAR VISUAL ODOMETRY USING STRUCTURE FROM MOTION AND VANISHING POINTS Johannes Gräter 1, Tobias Schwarze 1, Martin Lauer 1 August 20, 2015 Abstract While monocular visual
More informationAccurate Figure Flying with a Quadrocopter Using Onboard Visual and Inertial Sensing
Accurate Figure Flying with a Quadrocopter Using Onboard Visual and Inertial Sensing Jakob Engel, Jürgen Sturm, Daniel Cremers Abstract We present an approach that enables a low-cost quadrocopter to accurately
More informationSLAM II: SLAM for robotic vision-based perception
SLAM II: SLAM for robotic vision-based perception Margarita Chli Martin Rufli, Roland Siegwart Margarita Chli, Martin Rufli, Roland Siegwart SLAM II today s lecture Last time: how to do SLAM? Today: what
More informationA Photometrically Calibrated Benchmark For Monocular Visual Odometry
A Photometrically Calibrated Benchmark For Monocular Visual Odometry Jakob Engel and Vladyslav Usenko and Daniel Cremers Technical University Munich Abstract arxiv:67.2555v2 [cs.cv] 8 Oct 26 We present
More informationSLAM II: SLAM for robotic vision-based perception
SLAM II: SLAM for robotic vision-based perception Margarita Chli Martin Rufli, Roland Siegwart Margarita Chli, Martin Rufli, Roland Siegwart SLAM II today s lecture Last time: how to do SLAM? Today: what
More information15 Years of Visual SLAM
15 Years of Visual SLAM Andrew Davison Robot Vision Group and Dyson Robotics Laboratory Department of Computing Imperial College London www.google.com/+andrewdavison December 18, 2015 What Has Defined
More informationarxiv: v2 [cs.cv] 26 Aug 2018
A Synchronized Stereo and Plenoptic Visual Odometry Dataset Niclas Zeller 1,2, Franz Quint 2, and Uwe Stilla 1 1 Technische Universität München niclas.zeller@tum.de, stilla@tum.de 2 Karlsruhe University
More informationDeep Learning for Virtual Shopping. Dr. Jürgen Sturm Group Leader RGB-D
Deep Learning for Virtual Shopping Dr. Jürgen Sturm Group Leader RGB-D metaio GmbH Augmented Reality with the Metaio SDK: IKEA Catalogue App Metaio: Augmented Reality Metaio SDK for ios, Android and Windows
More informationProject Updates Short lecture Volumetric Modeling +2 papers
Volumetric Modeling Schedule (tentative) Feb 20 Feb 27 Mar 5 Introduction Lecture: Geometry, Camera Model, Calibration Lecture: Features, Tracking/Matching Mar 12 Mar 19 Mar 26 Apr 2 Apr 9 Apr 16 Apr 23
More informationLong-term motion estimation from images
Long-term motion estimation from images Dennis Strelow 1 and Sanjiv Singh 2 1 Google, Mountain View, CA, strelow@google.com 2 Carnegie Mellon University, Pittsburgh, PA, ssingh@cmu.edu Summary. Cameras
More informationRobot Localization based on Geo-referenced Images and G raphic Methods
Robot Localization based on Geo-referenced Images and G raphic Methods Sid Ahmed Berrabah Mechanical Department, Royal Military School, Belgium, sidahmed.berrabah@rma.ac.be Janusz Bedkowski, Łukasz Lubasiński,
More informationCamera calibration. Robotic vision. Ville Kyrki
Camera calibration Robotic vision 19.1.2017 Where are we? Images, imaging Image enhancement Feature extraction and matching Image-based tracking Camera models and calibration Pose estimation Motion analysis
More informationIntegrating Depth and Color Cues for Dense Multi-Resolution Scene Mapping Using RGB-D Cameras
In Proc. of the IEEE Int. Conf. on Multisensor Fusion and Information Integration (MFI), Hamburg, Germany, 2012 Integrating Depth and Color Cues for Dense Multi-Resolution Scene Mapping Using RGB-D Cameras
More informationPeripheral drift illusion
Peripheral drift illusion Does it work on other animals? Computer Vision Motion and Optical Flow Many slides adapted from J. Hays, S. Seitz, R. Szeliski, M. Pollefeys, K. Grauman and others Video A video
More informationEstimating Camera Position And Posture by Using Feature Landmark Database
Estimating Camera Position And Posture by Using Feature Landmark Database Motoko Oe 1, Tomokazu Sato 2 and Naokazu Yokoya 2 1 IBM Japan 2 Nara Institute of Science and Technology, Japan Abstract. Estimating
More informationOutline. ETN-FPI Training School on Plenoptic Sensing
Outline Introduction Part I: Basics of Mathematical Optimization Linear Least Squares Nonlinear Optimization Part II: Basics of Computer Vision Camera Model Multi-Camera Model Multi-Camera Calibration
More informationDirect Multichannel Tracking
MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Direct Multichannel Tracking Jaramillo, C.; Taguchi, Y.; Feng, C. TR2017-146 October 2017 Abstract We present direct multichannel tracking,
More informationPersonal Navigation and Indoor Mapping: Performance Characterization of Kinect Sensor-based Trajectory Recovery
Personal Navigation and Indoor Mapping: Performance Characterization of Kinect Sensor-based Trajectory Recovery 1 Charles TOTH, 1 Dorota BRZEZINSKA, USA 2 Allison KEALY, Australia, 3 Guenther RETSCHER,
More informationStructured light 3D reconstruction
Structured light 3D reconstruction Reconstruction pipeline and industrial applications rodola@dsi.unive.it 11/05/2010 3D Reconstruction 3D reconstruction is the process of capturing the shape and appearance
More informationEppur si muove ( And yet it moves )
Eppur si muove ( And yet it moves ) - Galileo Galilei University of Texas at Arlington Tracking of Image Features CSE 4392-5369 Vision-based Robot Sensing, Localization and Control Dr. Gian Luca Mariottini,
More information