Planetary Rover Absolute Localization by Combining Visual Odometry with Orbital Image Measurements

Similar documents
Visual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

Accurate 3D Face and Body Modeling from a Single Fixed Kinect

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy

Live Metric 3D Reconstruction on Mobile Phones ICCV 2013

Davide Scaramuzza. University of Zurich

PERFORMANCE CAPTURE FROM SPARSE MULTI-VIEW VIDEO

Depth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth

Stereo Vision. MAN-522 Computer Vision

Camera Drones Lecture 3 3D data generation

Simultaneous Localization and Mapping (SLAM)

AUTONOMOUS LONG-RANGE ROVER NAVIGATION EXPERIMENTAL RESULTS

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Yu Tao, Panos Sidiropoulos, Jan-Peter Muller Imaging group, Mullard Space Science Laboratory, University College London BiDS 17 March 2016

Simultaneous Localization

Tightly-Integrated Visual and Inertial Navigation for Pinpoint Landing on Rugged Terrains

Local features: detection and description. Local invariant features

Supplementary Note to Detecting Building-level Changes of a City Using Street Images and a 2D City Map

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation

Direct Methods in Visual Odometry

Feature Tracking and Optical Flow

Feature Tracking and Optical Flow

Structured light 3D reconstruction

A Low Power, High Throughput, Fully Event-Based Stereo System: Supplementary Documentation

CSE 527: Introduction to Computer Vision

Building a Panorama. Matching features. Matching with Features. How do we build a panorama? Computational Photography, 6.882

arxiv: v1 [cs.cv] 28 Sep 2018

Motion Analysis. Motion analysis. Now we will talk about. Differential Motion Analysis. Motion analysis. Difference Pictures

Stable Vision-Aided Navigation for Large-Area Augmented Reality

Dense 3D Reconstruction. Christiano Gava

3D Point Cloud Processing

Mosaics. Today s Readings

FPGA-Based Feature Detection

Improving Vision-based Topological Localization by Combining Local and Global Image Features

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

Local Feature Detectors

SIMPLE ROOM SHAPE MODELING WITH SPARSE 3D POINT INFORMATION USING PHOTOGRAMMETRY AND APPLICATION SOFTWARE

Dense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm

Multi-View Stereo for Static and Dynamic Scenes

3D Modeling of Objects Using Laser Scanning

Peripheral drift illusion

Dynamic Time Warping for Binocular Hand Tracking and Reconstruction

Long-term motion estimation from images

Towards a visual perception system for LNG pipe inspection

Application questions. Theoretical questions

Multiple View Geometry

Autonomous Navigation for Flying Robots

Dense 3D Reconstruction. Christiano Gava

Image correspondences and structure from motion

Local features: detection and description May 12 th, 2015

Feature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies

CS5670: Computer Vision

Midterm Wed. Local features: detection and description. Today. Last time. Local features: main components. Goal: interest operator repeatability

OpenStreetSLAM: Global Vehicle Localization using OpenStreetMaps

Data Representation in Visualisation

Automatic Image Alignment (feature-based)

What is Computer Vision? Introduction. We all make mistakes. Why is this hard? What was happening. What do you see? Intro Computer Vision

Pipe Mapping with Monocular Fisheye Imagery

Stereo and Monocular Vision Applied to Computer Aided Navigation for Aerial and Ground Vehicles

Lecture 19: Depth Cameras. Visual Computing Systems CMU , Fall 2013

Segmentation and Tracking of Partial Planar Templates

Motion Tracking and Event Understanding in Video Sequences

Data Association for SLAM

Target Shape Identification for Nanosatellites using Monocular Point Cloud Techniques

LOCAL AND GLOBAL DESCRIPTORS FOR PLACE RECOGNITION IN ROBOTICS

CNES robotics activities : Towards long distance on-board decision-making navigation

Intrinsic3D: High-Quality 3D Reconstruction by Joint Appearance and Geometry Optimization with Spatially-Varying Lighting

MULTI-MODAL MAPPING. Robotics Day, 31 Mar Frank Mascarich, Shehryar Khattak, Tung Dang

3D from Photographs: Automatic Matching of Images. Dr Francesco Banterle

Motion in 2D image sequences

Visual SLAM. An Overview. L. Freda. ALCOR Lab DIAG University of Rome La Sapienza. May 3, 2016

Large-Scale. Point Cloud Processing Tutorial. Application: Mobile Mapping

Safe Long-Range Travel for Planetary Rovers through Forward Sensing

Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision

Stereo Vision Based Traversable Region Detection for Mobile Robots Using U-V-Disparity

Midterm Examination CS 534: Computational Photography

CS4733 Class Notes, Computer Vision

Files Used in this Tutorial

Image Features. Work on project 1. All is Vanity, by C. Allan Gilbert,

From Structure-from-Motion Point Clouds to Fast Location Recognition

3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller

Estimation of Camera Pose with Respect to Terrestrial LiDAR Data

Lecture: Autonomous micro aerial vehicles

Camera Parameters Estimation from Hand-labelled Sun Sositions in Image Sequences

High Definition Modeling of Calw, Badstrasse and its Google Earth Integration

L2 Data Acquisition. Mechanical measurement (CMM) Structured light Range images Shape from shading Other methods

5 member consortium o University of Strathclyde o Wideblue o National Nuclear Laboratory o Sellafield Ltd o Inspectahire

FAST REGISTRATION OF TERRESTRIAL LIDAR POINT CLOUD AND SEQUENCE IMAGES

Image Features: Detection, Description, and Matching and their Applications

Motion and Optical Flow. Slides from Ce Liu, Steve Seitz, Larry Zitnick, Ali Farhadi

Outline of Presentation. Introduction to Overwatch Geospatial Software Feature Analyst and LIDAR Analyst Software

Master Automática y Robótica. Técnicas Avanzadas de Vision: Visual Odometry. by Pascual Campoy Computer Vision Group

Requirements for region detection

An Evaluation of Volumetric Interest Points

LAUROPE Six Legged Walking Robot for Planetary Exploration participating in the SpaceBot Cup

ORB SLAM 2 : an Open-Source SLAM System for Monocular, Stereo and RGB-D Cameras

+50,000 Archived GCPs. The Most Comprehensive Ground Control Points Solution. Make Geospatial Data More Accurate

Patch-based Object Recognition. Basic Idea

Harder case. Image matching. Even harder case. Harder still? by Diva Sian. by swashford

Harder case. Image matching. Even harder case. Harder still? by Diva Sian. by swashford

Transcription:

Planetary Rover Absolute Localization by Combining Visual Odometry with Orbital Image Measurements M. Lourakis and E. Hourdakis Institute of Computer Science Foundation for Research and Technology Hellas Heraklion, Crete, Greece ASTRA 2015, 11-13 May 2015, Noordwijk, The Netherlands

Introduction Planetary exploration increasingly depends on planetary rovers Accurate localization is essential to planetary rovers This talk is concerned with increasing the accuracy of rover localization by employing images collected on the ground and from orbit Reported work was pursued in SEXTANT, an ESA-funded activity aiming to develop visual navigation algorithms suitable for use by ExoMars Martian rovers 2012/11/28 Page 2

Background Visual odometry (VO) is a technology that has been used with considerable success by Martian rovers (e.g., MERs and MSL) VO works by tracking natural features in successive images VO provides only relative pose estimates and suffers from drift that accumulates without bound over time Countering drift with techniques such as visual SLAM or local bundle adjustment, challenges the limited computational capacity of current rovers Non-visual sensors (e.g., IMU) can only reduce the rate of drift but do not eliminate it completely 2012/11/28 Page 3

Using prior knowledge Localization accuracy can be increased by using prior knowledge regarding the environment (e.g., maps) Orthorectified aerial images of planetary surface (aka orbital images) are nowadays available, e.g. Mars HiRISE / MRO Several map -based localization techniques tailored to terrestrial environments exist, employing features such roads, line segments and building footprints or sensors such as GPS only available on Earth Here we choose to use as features boulders which abound on Mars 2012/11/28 Page 4

Sample images Orbital and ground images differ dramatically in appearance! 2012/11/28 Page 5

Using boulders for localization We seek to extract boulders visible in the orbital images and match them to those seen in the ground images Two boulder detection pipelines, i.e. orbital ground Boulder matching is the primary challenge; cannot be based on appearance cues such as local patch descriptors The relative geometric arrangement of boulders remains the same, thus their matching is based on geometry and strives to align constellations of ground and orbital boulders This provides absolute localization as the orbital boulder locations are determined offline in a cartographic system 2012/11/28 Page 6

Overview of the approach Boulders are extracted offline from orbital images and their locations are recorded Boulders are detected online in the ground stereo images acquired by the rover Ground boulders are combined with VO to build a local ground map with relative boulder positions The rover periodically matches its local ground map against the map originating from the orbital images Successful matching permits absolute pose refinement, thus accounting for drift 2012/11/28 Page 7

Visual odometry VO estimates the inter-frame camera motion Binocular camera system t+1 Harris corner detection SIFT descriptor extraction and matching t Sparse stereo triangulation Pose determination from two matched 3D point clouds Blue: projection rays, Green: spatial matches, Red: temporal matches SEXTANT TConf1 2012/09/24 Page 8

Boulder detection Boulders have to be detected both in ground and orbital images Boulders differ with respect to their morphology, size, texture, color, etc. Shadows, changes in illumination and occlusions complicate things further Literature reports attempts to develop boulder detectors for ground images; these are not yet very reliable Boulder detection in orbital images is somewhat easier 2012/11/28 Page 9

Boulder detection: orbital images Image binarization with adaptive thresholding Noise reduction with median filtering Detection of boulders via connected components labeling Size filtering for eliminating very small components Boulder locations are computed from the centroids of detected regions; pixel coordinates then converted to cartographic coordinates Centroids are a coarse approximation of true boulder locations; this nevertheless suffices, as will be shown later 2012/11/28 Page 10

Boulder detection in orbital images example Part of an orbital image (left) and the boulders detected in it 2012/11/28 Page 11

Boulder detection: ground images Ground images segmentation into boulders & ground Formation of boulder point clouds via 3D reconstruction and temporal matching Refinement of boulder point clouds via clustering Computation of boulder point cloud centroids Transformation of centroids to world origin using VO estimates Down-projection by dropping Z coordinates Local map update with the centroids of detected boulders 2012/11/28 Page 12

Ground boulder detection: segmentation Segmentation is based on the mean shift algorithm Small regions are ignored Performs reliably, suffers from over-segmentation Post-processing: neighboring regions are merged when they have similar intensity properties Triangulated points are also merged in 3D (more later) 2012/11/28 Page 13

Ground boulder detection: sparse reconstruction Features extracted from segmented regions and matched in stereo are reconstructed in 3D to form point clouds 2012/11/28 Page 14

Ground boulder detection: temporal matching Features extracted from segmented regions are also tracked over time and used to temporally match boulder regions Points reconstructed from matching regions are presumed to be on the same boulder and are assigned to the same point cloud 2012/11/28 Page 15

Ground boulder detection: clustering Point clouds refined with agglomerative clustering 2012/11/28 Page 16

Ground orbital boulder matching Key assumption: the rover knows its approximate location (i.e. the VO estimate can be trusted) Boulders extracted locally from the ground images are matched with the ICP algorithm against the overhead boulders located in the vicinity of the presumed rover location A byproduct of matching is the 2D transformation (Z-axis rotation and XY-axes translation) aligning the ground with the overhead boulders The VO estimate is refined with the 2D aligning transformation VO continues to estimate the relative displacement from the refined estimate The VO refinement is executed every few meters 2012/11/28 Page 17

Dataset for evaluation Graphically rendered images of a synthetic Mars-like environment (courtesy GMV) A stereo image sequence with accurate ground truth poses that corresponds to an S-shaped, 100m long traverse Ground images are 512 384 pixels An orbital image with size 15360 15360, corresponding to an area of size 76.8 76.8 m 2 (resolution: 0.005 m/pixel) Orbital image consists of 236 Mpixels Coarser resolutions generated by sub-sampling 2012/11/28 Page 18

Results: boulders extracted & matched Ground boulders along the entire trajectory: 3D view (left) superimposed on orbital image (right) 2012/11/28 Page 19

Results: positional accuracy Resolutions 15360 x 15360 11520 x 11520 7680 x 7680 3840 x 3840 Red is the plain VO positional error in all frames of a 100m traverse Green, blue, cyan and magenta are resp. the results of the proposed method in orbital images of the original, 25%, 50% and 75% lower resolutions 2012/11/28 Page 20

Results: rotational accuracy Resolutions 15360 x 15360 11520 x 11520 7680 x 7680 3840 x 3840 Red is the plain VO rotational error in all frames of a 100m traverse Green, blue, cyan and magenta are resp. the results of the proposed method in orbital images of the original, 25%, 50% and 75% lower resolutions 2012/11/28 Page 21

Summary & conclusions Improving VO localization using boulders is a promising idea A very terse representation consisting of only approximate 2D locations was shown to bring significant accuracy improvements Same approach can potentially solve the kidnapped robot problem (substituting ICP with geometric hashing for matching) More work on boulder detection and matching is needed Testing on more datasets is also necessary Accuracy of localization corrections is limited by the resolution of the orbital image; they actually degrade performance at the beginning of the sequence Vertical displacements and out-of-plane rotations are not corrected; full 3D info necessary SEXTANT TConf1 2012/09/24 Page 22

Thank you Any questions? 23

Schematic of the approach 2012/11/28 Page 24

Ground orbital boulder matching example Ground boulders (blue) matched against ground ones (red) 2012/11/28 Page 25