Visual-Inertial Localization and Mapping for Robot Navigation

Similar documents
Vol agile avec des micro-robots volants contrôlés par vision

Lecture 10 Multi-view Stereo (3D Dense Reconstruction) Davide Scaramuzza

Lecture 10 Dense 3D Reconstruction

Autonomous Navigation for Flying Robots

FLaME: Fast Lightweight Mesh Estimation using Variational Smoothing on Delaunay Graphs

Jakob Engel, Thomas Schöps, Daniel Cremers Technical University Munich. LSD-SLAM: Large-Scale Direct Monocular SLAM

Direct Methods in Visual Odometry

Lecture 13 Visual Inertial Fusion

Computationally Efficient Visual-inertial Sensor Fusion for GPS-denied Navigation on a Small Quadrotor

Visual SLAM for small Unmanned Aerial Vehicles

Live Metric 3D Reconstruction on Mobile Phones ICCV 2013

Hybrids Mixed Approaches

Semi-Dense Direct SLAM

4D Crop Analysis for Plant Geometry Estimation in Precision Agriculture

Graph-based SLAM (Simultaneous Localization And Mapping) for Bridge Inspection Using UAV (Unmanned Aerial Vehicle)

Hybrid, Frame and Event based Visual Inertial Odometry for Robust, Autonomous Navigation of Quadrotors

Lecture: Autonomous micro aerial vehicles

SLAM II: SLAM for robotic vision-based perception

Dealing with Scale. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

Autonomous navigation in industrial cluttered environments using embedded stereo-vision

PL-SVO: Semi-Direct Monocular Visual Odometry by Combining Points and Line Segments

SLAM II: SLAM for robotic vision-based perception

Autonomous Navigation in Complex Indoor and Outdoor Environments with Micro Aerial Vehicles

Monocular Visual-Inertial SLAM. Shaojie Shen Assistant Professor, HKUST Director, HKUST-DJI Joint Innovation Laboratory

Lost! Leveraging the Crowd for Probabilistic Visual Self-Localization

Miniaturized embedded event based vision for high-speed robots

3D Scene Reconstruction with a Mobile Camera

Dense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm

Visual SLAM. An Overview. L. Freda. ALCOR Lab DIAG University of Rome La Sapienza. May 3, 2016

Dense 3D Reconstruction from Autonomous Quadrocopters

Davide Scaramuzza. University of Zurich

Robust Visual Simultaneous Localization and Mapping for MAV Using Smooth Variable Structure Filter

Vision Algorithms for Mobile Robotics. Lecture 01 Introduction

Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview

FLaME: Fast Lightweight Mesh Estimation using Variational Smoothing on Delaunay Graphs

Citation for the original published paper (version of record):

Application of Vision-based Particle Filter and Visual Odometry for UAV Localization

Visual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

Monocular Visual-Inertial State Estimation for Micro Aerial Vehicles

Efficient Multi-Camera Visual-Inertial SLAM for Micro Aerial Vehicles

STANDARD frame-based CMOS cameras operate at fixed

Reconstruction, Motion Estimation and SLAM from Events

Visual Navigation for Micro Air Vehicles

Egomotion Estimation by Point-Cloud Back-Mapping

Scanning and Printing Objects in 3D Jürgen Sturm

An Evaluation of Robust Cost Functions for RGB Direct Mapping

Image-based Visual Perception and Representation for Collision Avoidance

W4. Perception & Situation Awareness & Decision making

Guidance: A Visual Sensing Platform For Robotic Applications

Marker Based Localization of a Quadrotor. Akshat Agarwal & Siddharth Tanwar

Autonomous 3D Reconstruction Using a MAV

Zürich. Roland Siegwart Margarita Chli Martin Rufli Davide Scaramuzza. ETH Master Course: L Autonomous Mobile Robots Summary

Event-based, 6-DOF Camera Tracking from Photometric Depth Maps

15 Years of Visual SLAM

THE task of estimating a sensor s ego-motion has important

IEEE/CAA JOURNAL OF AUTOMATICA SINICA, VOL. 2, NO. 1, JANUARY Robust and Accurate Monocular Visual Navigation Combining IMU for a Quadrotor

Creating Affordable and Reliable Autonomous Vehicle Systems

3-Point RANSAC for Fast Vision based Rotation Estimation using GPU Technology

Direct Visual Odometry for a Fisheye-Stereo Camera

Deep Incremental Scene Understanding. Federico Tombari & Christian Rupprecht Technical University of Munich, Germany

Parallel Tracking, Depth Estimation, and Image Reconstruction with an Event Camera

GTSAM 4.0 Tutorial Theory, Programming, and Applications

CSE 527: Introduction to Computer Vision

Application questions. Theoretical questions

AUTONOMOUS DRONE NAVIGATION WITH DEEP LEARNING

Guidance: A Visual Sensing Platform For Robotic Applications

Duo-VIO: Fast, Light-weight, Stereo Inertial Odometry

Accurate Figure Flying with a Quadrocopter Using Onboard Visual and Inertial Sensing

Geometric Reconstruction Dense reconstruction of scene geometry

UAV Autonomous Navigation in a GPS-limited Urban Environment

Direct Stereo Visual Odometry Based on Lines

7630 Autonomous Robotics Probabilities for Robotics

Real-Time Vision-Based State Estimation and (Dense) Mapping

Monocular Localization of a moving person onboard a Quadrotor MAV

Monocular Visual Odometry: Sparse Joint Optimisation or Dense Alternation?

Feature Detection and Tracking with the Dynamic and Active-pixel Vision Sensor (DAVIS)

Camera Drones Lecture 3 3D data generation

A Loosely-Coupled Approach for Metric Scale Estimation in Monocular Vision-Inertial Systems

Aerial and Ground-based Collaborative Mapping: An Experimental Study

Quadrotor tracking and pose estimation using a dynamic vision sensor

Perception for a River Mapping Robot

Unscented Kalman Filter for Vision Based Target Localisation with a Quadrotor

View Planning and Navigation Algorithms for Autonomous Bridge Inspection with UAVs

Semi-Direct EKF-based Monocular Visual-Inertial Odometry

IEEE ROBOTICS AND AUTOMATION LETTERS. PREPRINT VERSION. ACCEPTED DECEMBER, Active Autonomous Aerial Exploration for Ground Robot Path Planning

LEARNING RIGIDITY IN DYNAMIC SCENES FOR SCENE FLOW ESTIMATION

Monocular Image Space Tracking on a Computationally Limited MAV

A Benchmark Comparison of Monocular Visual-Inertial Odometry Algorithms for Flying Robots

ROBUST SCALE ESTIMATION FOR MONOCULAR VISUAL ODOMETRY USING STRUCTURE FROM MOTION AND VANISHING POINTS

A Real-Time RGB-D Registration and Mapping Approach by Heuristically Switching Between Photometric And Geometric Information

3D Reconstruction with Tango. Ivan Dryanovski, Google Inc.

1-Point-based Monocular Motion Estimation for Computationally-Limited Micro Aerial Vehicles

IEEE TRANSACTIONS ON ROBOTICS. PREPRINT VERSION. ACCEPTED JUNE, Continuous-Time Visual-Inertial Odometry for Event Cameras

OpenStreetSLAM: Global Vehicle Localization using OpenStreetMaps

Dynamically Feasible Motion Planning for Micro Air Vehicles using an Egocylinder

Aerial Robotic Autonomous Exploration & Mapping in Degraded Visual Environments. Kostas Alexis Autonomous Robots Lab, University of Nevada, Reno

Loosely Coupled Stereo Inertial Odometry on Low-cost System

Distributed Vision-Aided Cooperative Navigation Based on Three-View Geometry

DPPTAM: Dense Piecewise Planar Tracking and Mapping from a Monocular Sequence

Gaussian Process-based Visual Servoing Framework for an Aerial Parallel Manipulator

Transcription:

Visual-Inertial Localization and Mapping for Robot Navigation Dr. Guillermo Gallego Robotics & Perception Group University of Zurich Davide Scaramuzza University of Zurich - http://rpg.ifi.uzh.ch

Mocular, Vision-inertial Navigation of Mobile Robots https://www.youtube.com/watch?v=vhpw8zc7-jq [Scaramuzza, Achtelik, Weiss, Fraundorfer, et al., Vision-Controlled Micro Flying Robots: from System Design to Autonomous Davide Scaramuzza Navigation and University Mapping of in Zurich GPS-denied - http://rpg.ifi.uzh.ch Environments, IEEE RAM, 2014]

Keyframe-based Visual Odometry Keyframe 1 Keyframe 2 Current frame New keyframe Initial pointcloud New triangulated points PTAM (Parallel Tracking & Mapping) [Klein, ISMAR 08]

Feature-based methods 1. Extract & match features (+RANSAC) 2. Minimize Reprojection error minimization u i T k,k 1 =? u i T k,k 1 = arg min T i u i π p i Σ 2 p i Direct methods T k,k 1 1. Minimize photometric error T k,k 1 = arg min T i I k u i I k 1 (u i ) σ 2 I k 1 u i I k u i where u i = π T π 1 u i d d i [Jin,Favaro,Soatto 03] [Silveira, Malis, Rives, TRO 08], [Newcombe et al., ICCV 11], [Engel et al., ECCV 14], [Forster et al., ICRA 14] p i 4

Feature-based methods 1. Extract & match features (+RANSAC) 2. Minimize Reprojection error minimization Large frame-to-frame motions Slow due to costly feature extraction and matching Matching Outliers (RANSAC) T k,k 1 = arg min T i u i π p i Σ 2 Direct methods 1. Minimize photometric error T k,k 1 = arg min T where u i = π T π 1 u i i I k u i I k 1 (u i ) σ 2 d All information in the image can be exploited (precision, robustness) Increasing camera frame-rate reduces computational cost per frame Limited frame-to-frame motion [Jin,Favaro,Soatto 03] [Silveira, Malis, Rives, TRO 08], [Newcombe et al., ICCV 11], [Engel et al., ECCV 14], [Forster et al., ICRA 14] 5

Feature-based methods 1. Extract & match features (+RANSAC) 2. Minimize Reprojection error minimization Large frame-to-frame motions Slow due to costly feature extraction and matching Matching Outliers (RANSAC) T k,k 1 = arg min T u i π p i Σ 2 i SVO: Semi-direct Visual Odometry [ICRA 14] Direct methods Combines feature-based and direct methods 1. Minimize photometric error T k,k 1 = arg min T where u i = π T π 1 u i i I k u i I k 1 (u i ) σ 2 d Our solution: All information in the image can be exploited (precision, robustness) Increasing camera frame-rate reduces computational cost per frame Limited frame-to-frame motion [Jin,Favaro,Soatto 03] [Silveira, Malis, Rives, TRO 08], [Newcombe et al., ICCV 11], [Engel et al., ECCV 14], [Forster et al., ICRA 14] 6

SVO Results: Fast and Abrupt Motions https://www.youtube.com/watch?v=2ynimfw6bjy Open Source [Forster, Pizzoli, Scaramuzza, «SVO: Semi Direct Visual Odometry», ICRA 14]

Processing Times of SVO Laptop (Intel i7, 2.8 GHz) 400 frames per second Embedded ARM Cortex-A9, 1.7 GHz Up to 70 frames per second

Probabilistic Depth Estimation Depth-Filter: Depth Filter for every feature Recursive Bayesian depth estimation Mixture of Gaussian + Uniform distribution [Forster, Pizzoli, Scaramuzza, SVO: Semi Direct Visual Odometry, IEEE ICRA 14]

Visual-Inertial Fusion via Optimization [RSS 15] Fusion solved as a non-linear optimization problem (no Kalman filter): Increased accuracy over filtering methods IMU residuals Reprojection residuals Forster, Carlone, Dellaert, Scaramuzza, IMU Preintegration on Manifold for efficient Visual-Inertial Maximum-a- Posteriori Estimation, Robotics Science and Systens 15, Best Paper Award finalist

Comparison with Previous Works https://www.youtube.com/watch?v=csjkci5lfco Open Source Accuracy: 0.1% of the travel distance Proposed Google Tango ASLAM Forster, Carlone, Dellaert, Scaramuzza, IMU Preintegration on Manifold for efficient Visual-Inertial Maximum-a-Posteriori Estimation, Robotics Science and Systens 15

Integration on a Quadrotor Platform

Quadrotor System Odroid U3 Computer Quad Core Odroid (ARM Cortex A-9) used in Samsung Galaxy S4 phones Runs Linux Ubuntu and ROS 450 grams Global-Shutter Camera 752x480 pixels High dynamic range 90 fps

Quadrotor System Odroid U3 Computer Quad Core Odroid (ARM Cortex A-9) used in Samsung Galaxy S4 phones Runs Linux Ubuntu and ROS 450 grams Global-Shutter Camera 752x480 pixels High dynamic range 90 fps

Indoors and outdoors experiments https://www.youtube.com/watch?v=4x6voft4z_0 RMS error: 5 mm, height: 1.5 m Down-looking camera Speed: 4 m/s, height: 1.5 m Down-looking camera https://www.youtube.com/watch?v=3mny9-dsudk Faessler, Fontana, Forster, Mueggler, Pizzoli, Scaramuzza, Autonomous, Vision-based Flight and Live Dense 3D Mapping with a Quadrotor Micro Aerial Vehicle, Journal of Field Robotics, 2015.

Visual-Inertial Odometry - Results https://www.youtube.com/watch?v=6kxboprgar0 Open Source Forster, Carlone, Dellaert, Scaramuzza, IMU Preintegration on Manifold for efficient Visual-Inertial Maximum-a- Posteriori Estimation, Robotics Science and Systens 15, Best Paper Award Finalist

Application: Autonomous Inspection of Bridges and Power Masts Project with Parrot: Autonomous vision-based navigation https://www.youtube.com/watch?v=bxvoalrkj4u Albris drone

From Sparse to Dense 3D Models [M. Pizzoli, C. Forster, D. Scaramuzza, REMODE: Probabilistic, Monocular Dense Reconstruction in Real Time, ICRA 14]

Dense Reconstruction Pipeline Local methods Estimate depth for every pixel independently using photometric cost aggregation Global methods Refine the depth surface as a whole by enforcing smoothness constraint ( Regularization ) E d = E d d + λe s (d) Data term Regularization term: penalizes non-smooth surfaces [Newcombe et al. 2011]

REMODE: Probabilistic Monocular Dense Reconstruction [ICRA 14] Track independently every pixel using the same recursive Bayesian depth estimation of SVO A regularized depth map F(u) is computed from the noisy depth map D(u) as where Minimization is done using [Chambolle & Pock, 2011] [M. Pizzoli, C. Forster, D. Scaramuzza, REMODE: Probabilistic, Monocular Dense Reconstruction in Real Time, ICRA 14]

REMODE: Probabilistic Monocular Dense Reconstruction [ICRA 14] https://www.youtube.com/watch?v=qtkd5uwcg0q Running at 50 Hz on GPU on a Lenovo W530, i7 Open Source [M. Pizzoli, C. Forster, D. Scaramuzza, REMODE: Probabilistic, Monocular Dense Reconstruction in Real Time, ICRA 14]

Autonomus, Flying 3D Scanning https://www.youtube.com/watch?v=7-kpiwafyac Sensing, control, state estimation run onboard at 50 Hz (Odroid U3, ARM Cortex A9) Dense reconstruction runs live on video streamed to laptop (Lenovo W530, i7) 2x Faessler, Fontana, Forster, Mueggler, Pizzoli, Scaramuzza, Autonomous, Vision-based Flight and Live Dense 3D Mapping with a Quadrotor Micro Aerial Vehicle, Journal of Field Robotics, 2015.

Autonomus, Flying 3D Scanning [ JFR 15] https://www.youtube.com/watch?v=7-kpiwafyac Sensing, control, state estimation run onboard at 50 Hz (Odroid U3, ARM Cortex A9) Dense reconstruction runs live on video streamed to laptop (Lenovo W530, i7) Faessler, Fontana, Forster, Mueggler, Pizzoli, Scaramuzza, Autonomous, Vision-based Flight and Live Dense 3D Mapping with a Quadrotor Micro Aerial Vehicle, Journal of Field Robotics, 2015.

Autonomous Landing-Spot Detection and Landing https://www.youtube.com/watch?v=phabkfwfcj4 Forster, Faessler, Fontana, Werlberger, Scaramuzza, Continuous On-Board Monocular-Vision-based Elevation Mapping Applied to Autonomous Landing of Micro Aerial Vehicles, ICRA 15.

Outlook

To go faster, we need faster sensors! At the current state, the agility of a robot is limited by the latency and temporal discretization of its sensing pipeline. Currently, the average robot-vision algorithms have latencies of 50-200 ms. This puts a hard bound on the agility of the platform. frame next frame time computation command command latency temporal discretization Can we create a low-latency, low-discretization perception pipeline? - Yes, if we use event-based cameras [Censi & Scaramuzza, «Low Latency, Event-based Visual Odometry», ICRA 14]

Human Vision System 130 million photoreceptors But only 2 million axons!

Dynamic Vision Sensor (DVS) Event-based camera developed by Tobi Delbruck s group (ETH & UZH). Temporal resolution: 1 μs High dynamic range: 120 db Low power: 20 mw Cost: 2,500 EUR Image of the solar eclipse (March 15) captured by a DVS (courtesy of IniLabs) [Lichtsteiner, Posch, Delbruck. A 128x128 120 db 15µs Latency Asynchronous Temporal Contrast Vision Sensor. 2008]

Camera vs Dynamic Vision Sensor Video with DVS explanation https://www.youtube.com/watch?v=lauq6lwtkxm

Camera vs Dynamic Vision Sensor Video with DVS explanation https://www.youtube.com/watch?v=lauq6lwtkxm

High-speed State Estimation https://www.youtube.com/watch?v=izz77f-hwzs [Event-based Camera Pose Tracking using a Generative Event Model, Under review. Available at arxiv] [Censi & Scaramuzza, Low Latency, Event-based Visual Odometry, ICRA 14]

Conclusions Visual-inertial state estimation is crucial for GPS-denied navigation. More accurate than GPS, DGPS, and RTK-GPS. Pending challenges: robustness to changing illumination and highspeed motion. Event cameras are revolutionary visual sensors that can address such challenges where standard cameras fail.