Lost! Leveraging the Crowd for Probabilistic Visual Self-Localization

Similar documents
OpenStreetSLAM: Global Vehicle Localization using OpenStreetMaps

W4. Perception & Situation Awareness & Decision making

Collaborative Mapping with Streetlevel Images in the Wild. Yubin Kuang Co-founder and Computer Vision Lead

MULTI-MODAL MAPPING. Robotics Day, 31 Mar Frank Mascarich, Shehryar Khattak, Tung Dang

Self Driving. DNN * * Reinforcement * Unsupervised *

StereoScan: Dense 3D Reconstruction in Real-time

Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview

Stereo Epipolar Geometry for General Cameras. Sanja Fidler CSC420: Intro to Image Understanding 1 / 33

Visual-Inertial Localization and Mapping for Robot Navigation

VINet: Visual-Inertial Odometry as a Sequence-to-Sequence Learning Problem

Vision-based Mobile Robot Localization and Mapping using Scale-Invariant Features

A novel georeferenced dataset for stereo visual odometry

arxiv: v1 [cs.ro] 23 Feb 2018

PML-SLAM: a solution for localization in large-scale urban environments

Towards Fully-automated Driving. tue-mps.org. Challenges and Potential Solutions. Dr. Gijs Dubbelman Mobile Perception Systems EE-SPS/VCA

Dense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm

Dealing with Scale. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

ROBUST SCALE ESTIMATION FOR MONOCULAR VISUAL ODOMETRY USING STRUCTURE FROM MOTION AND VANISHING POINTS

A Whole New World of Mapping and Sensing: Uses from Asset to Management to In Vehicle Sensing for Collision Avoidance

STEREO IMAGE POINT CLOUD AND LIDAR POINT CLOUD FUSION FOR THE 3D STREET MAPPING

Supplementary Material for A Multi-View Stereo Benchmark with High-Resolution Images and Multi-Camera Videos

Measuring the World: Designing Robust Vehicle Localization for Autonomous Driving. Frank Schuster, Dr. Martin Haueis

Probabilistic Robotics. FastSLAM

Evaluating the Performance of a Vehicle Pose Measurement System

Probabilistic Robotics

USING 3D DATA FOR MONTE CARLO LOCALIZATION IN COMPLEX INDOOR ENVIRONMENTS. Oliver Wulf, Bernardo Wagner

Pattern Recognition for Autonomous. Pattern Recognition for Autonomous. Driving. Freie Universität t Berlin. Raul Rojas

A Generative Model for 3D Urban Scene Understanding from Movable Platforms

OFERTA O120410PA CURRENT DATE 10/04//2012 VALID UNTIL 10/05/2012 SUMMIT XL

VISION FOR AUTOMOTIVE DRIVING

Particle Filters. CSE-571 Probabilistic Robotics. Dependencies. Particle Filter Algorithm. Fast-SLAM Mapping

Practical Course WS12/13 Introduction to Monte Carlo Localization

Where s the Boss? : Monte Carlo Localization for an Autonomous Ground Vehicle using an Aerial Lidar Map

ECE276A: Sensing & Estimation in Robotics Lecture 11: Simultaneous Localization and Mapping using a Particle Filter

Topometric Localization on a Road Network

Lecture: Autonomous micro aerial vehicles

Sensor Modalities. Sensor modality: Different modalities:

Monocular Heading Estimation in Non-stationary Urban Environment

08 An Introduction to Dense Continuous Robotic Mapping

Introduction to Mobile Robotics SLAM Grid-based FastSLAM. Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Diego Tipaldi, Luciano Spinello

3D LIDAR Point Cloud based Intersection Recognition for Autonomous Driving

Vehicle Localization. Hannah Rae Kerner 21 April 2015

IROS 05 Tutorial. MCL: Global Localization (Sonar) Monte-Carlo Localization. Particle Filters. Rao-Blackwellized Particle Filters and Loop Closing

Matching Evaluation of 2D Laser Scan Points using Observed Probability in Unstable Measurement Environment

Training models for road scene understanding with automated ground truth Dan Levi

On Road Vehicle Detection using Shadows

Motion-Based Calibration of Multimodal Sensor Arrays

Localization and Map Building

Acoustic/Lidar Sensor Fusion for Car Tracking in City Traffic Scenarios

OpenStreetSLAM: Global Vehicle Localization Using OpenStreetMaps

Direct Methods in Visual Odometry

Robotics. Chapter 25. Chapter 25 1

arxiv: v2 [cs.cv] 14 May 2018

Egomotion Estimation by Point-Cloud Back-Mapping

Sensor Fusion: Potential, Challenges and Applications. Presented by KVH Industries and Geodetics, Inc. December 2016

Robotics. Haslum COMP3620/6320

High-precision, consistent EKF-based visual-inertial odometry

Camera and Inertial Sensor Fusion

Graph-based SLAM (Simultaneous Localization And Mapping) for Bridge Inspection Using UAV (Unmanned Aerial Vehicle)

Stereo Vision Based Traversable Region Detection for Mobile Robots Using U-V-Disparity

Autonomous Navigation for Flying Robots

Analysis of the CMU Localization Algorithm Under Varied Conditions

Joint Vanishing Point Extraction and Tracking. 9. June 2015 CVPR 2015 Till Kroeger, Dengxin Dai, Luc Van Gool, Computer Vision ETH Zürich

Precision Roadway Feature Mapping Jay A. Farrell, University of California-Riverside James A. Arnold, Department of Transportation

Long-term motion estimation from images

Mapping with Dynamic-Object Probabilities Calculated from Single 3D Range Scans

DD2426 Robotics and Autonomous Systems Lecture 4: Robot Sensors and Perception

AN INCREMENTAL SLAM ALGORITHM FOR INDOOR AUTONOMOUS NAVIGATION

Robotics. Lecture 7: Simultaneous Localisation and Mapping (SLAM)

Simultaneous Localization and Mapping (SLAM)

Probabilistic Tracking and Reconstruction of 3D Human Motion in Monocular Video Sequences

Visual Bearing-Only Simultaneous Localization and Mapping with Improved Feature Matching

Revising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History

Davide Scaramuzza. University of Zurich

Servosila Robotic Heads

Public Sensing Using Your Mobile Phone for Crowd Sourcing

Robot Mapping. A Short Introduction to the Bayes Filter and Related Models. Gian Diego Tipaldi, Wolfram Burgard

ActiveStereoNet: End-to-End Self-Supervised Learning for Active Stereo Systems (Supplementary Materials)

PROBE: Predictive Robust Estimation for Visual-Inertial Navigation

Robotic Perception and Action: Vehicle SLAM Assignment

Stable Vision-Aided Navigation for Large-Area Augmented Reality

Driving dataset in the wild: Various driving scenes

Probabilistic Matching for 3D Scan Registration

arxiv: v1 [cs.cv] 2 Jan 2018

Zürich. Roland Siegwart Margarita Chli Martin Rufli Davide Scaramuzza. ETH Master Course: L Autonomous Mobile Robots Summary

LOAM: LiDAR Odometry and Mapping in Real Time

Robot Mapping. SLAM Front-Ends. Cyrill Stachniss. Partial image courtesy: Edwin Olson 1

IP-S2 HD HD IP-S2. 3D Mobile Mapping System. 3D Mobile Mapping System

What is the SLAM problem?

Accurate Motion Estimation and High-Precision 3D Reconstruction by Sensor Fusion

Dense 3D Reconstruction from Autonomous Quadrocopters

Efficient Large-Scale Stereo Matching

arxiv: v3 [cs.cv] 20 Mar 2018

3-Point RANSAC for Fast Vision based Rotation Estimation using GPU Technology

Today MAPS AND MAPPING. Features. process of creating maps. More likely features are things that can be extracted from images:

A New Omnidirectional Vision Sensor for Monte-Carlo Localization

Simultaneous Localization

Visual Recognition and Search April 18, 2008 Joo Hyun Kim

Computationally Efficient Visual-inertial Sensor Fusion for GPS-denied Navigation on a Small Quadrotor

Visual vehicle tracking through noise and occlusions using crowd-sourced maps

Transcription:

Lost! Leveraging the Crowd for Probabilistic Visual Self-Localization Marcus A. Brubaker (Toyota Technological Institute at Chicago) Andreas Geiger (Karlsruhe Institute of Technology & MPI Tübingen) Raquel Urtasun (Toyota Technological Institute at Chicago) 1 1

Introduction Localization is crucial for autonomous systems 2 2

Introduction GPS has limitations in terms of reliability and availability Place recognition techniques use image features and a database of previously collected images [Dellaert et al, ICRA 1999; Thrun et al, AI 2001; Hays and Efros, CVPR 2008; Schindler et al, CVPR 2008; Crandall et al, WWW 2009; Kalogerakis et al, ICCV 2009] We develop an inexpensive technique for localizing to ~3m in unseen regions 3 3

Introduction Humans are able to use a map, combined with visual input and exploration, to localize effectively Detailed, community developed maps are freely available (OpenStreetMap) How can we exploit maps, combined with visual cues, to localize a vehicle? 4 4

Probabilistic Localization using Visual Odometry Visual odometry provides a strong source of information for localization Visual odometry has some issues Over short time periods it can be noisy and highly ambiguous Over long time periods it drifts when integrated We adopt a probabilistic approach to represent and maintain this uncertainty [Geiger et al, IV 2011] 5 5

Probabilistic Localization using Visual Odometry Maps can be considered as a graph SINK Nodes of the graph represent street segments 1 6 5 6 5 Edges represent intersections and allowed transitions between these segments 2 3 4 1 2 3 4 Position is defined by the current street and the distance travelled, d, and orientation,, on that street θ θ d 6 6

Probabilistic Localization using Visual Odometry The complete state includes the current street segment, and the current and previous position and orientation on the street segment, s t =(d t, θ t,d t 1, θ t 1 ) Odometry observations y 1:t =(y 1,...,y t ) Localization is formulated as posterior inference p(u t, s t y 1:t ) u t θ d State Transition Previous Posterior p(y t u t, s t ) p(u t u t 1, s t 1 )p(s t u t,u t 1, s t 1 )p(u t 1, s t 1 y 1:t 1 )ds t 1 u t 1 Likelihood Street Segment Transition Pose Transition 7 7

Probabilistic Localization using Visual Odometry Likelihood: p(y t u t, s t ) y t = Ms t + η η N(0, Σ y ) Pose transition: p(s t u t,u t 1, s t 1 ) s t = As t 1 + b + ζ ζ N(0, Σ s ) Street segment transition: p(ut ut 1, st 1) Length of street segment u t 1 Parameters (e.g., variances) estimated from data 1 9 8 7 6 5 4 3 2 1 0 d t 1 Model is nearly Gauss-Linear which we exploit to derive a custom inference algorithm 8 8

Probabilistic Localization using Visual Odometry To represent the posterior we factorize it p(u t, s t y 1:t )=p(s t u t, y 1:t )p(u t y 1:t ) The posterior over pose is represented with a Mixture of Gaussians for each street segment p(s t u t, y 1:t )= Posterior over pose, given the street segment N ut i=1 π (i) u t N Discrete distribution over street segments s t µ (i) u t, Σ (i) u t We ve derived a general algorithm for simplifying mixture models and do this periodically to reduce computation 9 9

1 Experiments Vision meets Robotics: The KITTI Dataset Andreas Geiger, Philip Lenz, Christoph Stiller and Raquel Urtasun Abstract We present a novel dataset captured from a VW We used the Visual station Odometry sequences from the wagon for use in mobile robotics and autonomous driving research. In total, we recorded 6 hours of traffic scenarios at 10-100et Hz al, using CVPR a variety of sensor modalities such as highkitti dataset [Geiger 2012] resolution color and grayscale stereo cameras, a Velodyne 3D laser scanner and a high-precision GPS/IMU inertial navigation system. The scenarios are diverse, capturing real-world traffic situations and range from freeways over rural areas to innercity scenes with many static and dynamic objects. Our data is calibrated, synchronized and timestamped, and we provide the rectified and raw image sequences. Our dataset also contains object labels in the form of 3D tracklets and we provide online benchmarks for stereo, optical flow, object detection and other tasks. This paper describes our recording platform, the data format and the utilities that we provide. Video captured from car-mounted cameras 11 sequences captured in a variety of settings (e.g., urban, highway, rural, etc)! " " #$$#!! % &'( Index Terms dataset, autonomous driving, mobile robotics, Monocular and Stereo visual odometry field robotics, computer vision, cameras, laser, GPS, benchmarks, stereo, optical flow, SLAM, object detection, tracking, KITTI computed using LIBVISO2 [Geiger et al, IV 2011] I. I NTRODUCTION The KITTI dataset has been recorded from a moving plat Errors computed in position and heading form (Fig. 1) while driving in and around Karlsruhe, Germany (Fig. 2). It includes camera images, laser scans, high-precision GPS measurements and IMU accelerations from a combined GPS/IMU system. The main purpose of this dataset is to push forward the development of computer vision and robotic algorithms targeted to autonomous driving [1] [7]. While our introductory paper [8] mainly focuses on the benchmarks, their creation and use for evaluating state-of-the-art computer vision methods, here we complement this information by providing technical details on the raw data itself. We give precise instructions on how to access the data and comment on sensor limitations and common pitfalls. The dataset can be downloaded from http://www.cvlibs.net/datasets/kitti. For a review on related work, we refer the reader to [8]. Fig. 1. Recording Platform. Our VW Passat station wagon is equipped with four video cameras (two color and two grayscale cameras), a rotating 3D laser scanner and a combined GPS/IMU inertial navigation system. RT3003 inertial and GPS navigation system, Parallelized implementation runs at frame rate on16 axis,oxts 100 Hz, L1/L2 RTK, resolution: 0.02m / 0.1 Note that the color cameras lack in terms of resolution due average on 16 cores to the Bayer pattern interpolation process and are less sensitive II. S ENSOR S ETUP Our sensor setup is illustrated in Fig. 3: 2 PointGray Flea2 grayscale cameras (FL2-14S3M-C), 1.4 Megapixels, 1/2 Sony ICX267 CCD, global shutter to light. This is the reason why we use two stereo camera rigs, one for grayscale and one for color. The baseline of both stereo camera rigs is approximately 54 cm. The trunk of our vehicle houses a PC with two six-core Intel XEON X5650 processors and a shock-absorbed RAID 5 hard disk storage with a capacity of 4 Terabytes. Our computer runs Ubuntu Linux (64 bit) and a real-time database [9] to store the incoming data streams. III. DATASET The raw data described in this paper can be accessed from 10 10

Experiments Accuracy: 2.1m/1.2 Localization Time: 21s 11 11

Experiments Accuracy: 2.4m/1.2 Localization Time: 81s 12 12

Experiments: Quantitative Results Average Stereo Odometry Monocular Odometry Map Projection Position Error Heading Error Localization Time 3.1m 18.4m 1.4m 1.3 3.6-36s 62s - 13 13

Experiments: Ambiguous Sequences 14 14

Experiments: Initial Map Size 2.0 10.0 Initial Map Size (km of road) 50.0 15 15

Experiments: Full City Maps 2,150km of road 16 16

Experiments: Full City Maps 2,150km of road 17 17

Conclusions Fast, accurate map-based localization using only visual odometry Accuracy of 3.1m/1.3 in 36 seconds of driving time on average Highly parallelizable, runs at real-time on average w/ 16 cores Code will be available: http://www.cs.toronto.edu/~mbrubake Future work Exploiting other map information, e.g., landmarks, speed limits Integration of other sensors, e.g., accelerometers 18 18

19 19