Vision-based Mobile Robot Localization and Mapping using Scale-Invariant Features

Similar documents
SLAM with SIFT (aka Mobile Robot Localization and Mapping with Uncertainty using Scale-Invariant Visual Landmarks ) Se, Lowe, and Little

Object Recognition with Invariant Features

Global Localization using Distinctive Visual Features

Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview

Visual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

The SIFT (Scale Invariant Feature

Dense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

Global localization from a single feature correspondence

Autonomous Navigation for Flying Robots

Building a Panorama. Matching features. Matching with Features. How do we build a panorama? Computational Photography, 6.882

Revising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History

EECS150 - Digital Design Lecture 14 FIFO 2 and SIFT. Recap and Outline

Scale Invariant Feature Transform

LUMS Mine Detector Project

A High Dynamic Range Vision Approach to Outdoor Localization

Large-Scale 3D Point Cloud Processing Tutorial 2013

Stable Vision-Aided Navigation for Large-Area Augmented Reality

Ceilbot vision and mapping system

Towards a visual perception system for LNG pipe inspection

Scale Invariant Feature Transform

CS231A Section 6: Problem Set 3

Basics of Localization, Mapping and SLAM. Jari Saarinen Aalto University Department of Automation and systems Technology

Data Association for SLAM

School of Computing University of Utah

REMOTELY controlled mobile robots have been a subject

CSE 527: Introduction to Computer Vision

Long-term motion estimation from images

Introduction to SLAM Part II. Paul Robertson

Structure from Motion. Introduction to Computer Vision CSE 152 Lecture 10

CS231A Midterm Review. Friday 5/6/2016

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014

3D Computer Vision. Structure from Motion. Prof. Didier Stricker

Combining Appearance and Topology for Wide

Stereo and Epipolar geometry

Local Features: Detection, Description & Matching

CS4670: Computer Vision

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy

Personal Navigation and Indoor Mapping: Performance Characterization of Kinect Sensor-based Trajectory Recovery

Deep Learning: Image Registration. Steven Chen and Ty Nguyen

Feature descriptors and matching

Vision-Based 3D Trajectory Tracking for Unknown Environments

Outline 7/2/201011/6/

EXAM SOLUTIONS. Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006,

CS 231A Computer Vision (Fall 2012) Problem Set 3

ME132 February 3, 2011

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation

Localization and Map Building

Dense 3D Reconstruction. Christiano Gava

Robotics. Lecture 7: Simultaneous Localisation and Mapping (SLAM)

Department of Electrical and Computer Systems Engineering

Overview. EECS 124, UC Berkeley, Spring 2008 Lecture 23: Localization and Mapping. Statistical Models

Local Feature Detectors

L15. POSE-GRAPH SLAM. NA568 Mobile Robotics: Methods & Algorithms

CS 4495 Computer Vision A. Bobick. CS 4495 Computer Vision. Features 2 SIFT descriptor. Aaron Bobick School of Interactive Computing

StereoScan: Dense 3D Reconstruction in Real-time

Image processing and features

A Rapid Automatic Image Registration Method Based on Improved SIFT

Cooperative Robot Localization with Vision-based Mapping *

Optical flow and tracking

Visual Tracking (1) Tracking of Feature Points and Planar Rigid Objects

3D Scene Reconstruction with a Mobile Camera

CS 532: 3D Computer Vision 7 th Set of Notes

Uncertainties: Representation and Propagation & Line Extraction from Range data

Instance-level recognition

EE368 Project Report CD Cover Recognition Using Modified SIFT Algorithm

Master Automática y Robótica. Técnicas Avanzadas de Vision: Visual Odometry. by Pascual Campoy Computer Vision Group

Visual Bearing-Only Simultaneous Localization and Mapping with Improved Feature Matching

Instance-level recognition

CS 223B Computer Vision Problem Set 3

Instance-level recognition part 2

CEE598 - Visual Sensing for Civil Infrastructure Eng. & Mgmt.

CSE 252B: Computer Vision II

CS201: Computer Vision Introduction to Tracking

Monocular SLAM for a Small-Size Humanoid Robot

SCALE INVARIANT FEATURE TRANSFORM (SIFT)

Attentional Landmarks and Active Gaze Control for Visual SLAM

Lost! Leveraging the Crowd for Probabilistic Visual Self-Localization

Scale Invariant Feature Transform by David Lowe

Stereo Matching.

Week 2: Two-View Geometry. Padua Summer 08 Frank Dellaert

calibrated coordinates Linear transformation pixel coordinates

Fast, Unconstrained Camera Motion Estimation from Stereo without Tracking and Robust Statistics

Requirements for region detection

Lecture 19: Depth Cameras. Visual Computing Systems CMU , Fall 2013

Features Points. Andrea Torsello DAIS Università Ca Foscari via Torino 155, Mestre (VE)

Corner Detection. GV12/3072 Image Processing.

Introduction. Introduction. Related Research. SIFT method. SIFT method. Distinctive Image Features from Scale-Invariant. Scale.

Multiple View Geometry. Frank Dellaert

Patch-based Object Recognition. Basic Idea

PERFORMANCE CAPTURE FROM SPARSE MULTI-VIEW VIDEO

NAVIGATION SYSTEM OF AN OUTDOOR SERVICE ROBOT WITH HYBRID LOCOMOTION SYSTEM

Eppur si muove ( And yet it moves )

Dense 3D Reconstruction. Christiano Gava

Image correspondences and structure from motion

ECE276A: Sensing & Estimation in Robotics Lecture 11: Simultaneous Localization and Mapping using a Particle Filter

Synchronized Ego-Motion Recovery of Two Face-to-Face Cameras

Localization, Where am I?

Segmentation and Tracking of Partial Planar Templates

Transcription:

Vision-based Mobile Robot Localization and Mapping using Scale-Invariant Features Stephen Se, David Lowe, Jim Little Department of Computer Science University of British Columbia Presented by Adam Bickett Objective: SLAMB Simultaneous Localization and Map Building Need localization to build a map Need a map to accurately localize Critical for successful robot navigation in a large environment 1

Previous Approaches Different sensor modalities: Sonar Lasers Vision Many early approaches require beacons or visual patterns present in the environment How do we use vision techniques in an unmarked environment? Previous Approaches Most previous approaches allow significant drift over time DROID system uses image corner features, tracked with Kalman filters. [Harris] [Murray et al] track features with a 2D occupancy map, but localization is not updated, so odometry errors accumulate. 2

Previous Approaches Other approaches use probabilistic modeling Batch EM approach which estimates the most likely map by considering all previous locations of sonar scans [Thrun et al] Monte Carlo localization [Dellaert et al] updates probability distribution of current pose given new sensor input Most of these use laser or sonar. Using SIFT for SLAMB Much of the previous work uses laser or sonar based sensors, why not use vision? More processor intensive Stable vision features are hard to extract Approach: Build a 3D map of SIFT features, constructed using trinocular stereo rig mounted on the robot 3

SIFT features SIFT feature attributes make them well-suited for the purpose Invariant to image translation, scaling, rotation Partially invariant to illumination and affine/3d projections Keypoints are at maxima/minima of DoG applied at different scales. Using SIFT features Matching between SIFT features in the three views provides stable landmarks for the environment 4

Stereo Matching with SIFT The right camera is used as a reference camera Left camera is 10 cm left, top camera 10 cm up Constraints: Epipolar constraint Disparity constraint SIFT scale/orientation constraint If a feature has more than one match, it is discarded as ambiguous SIFT Stereo Matching From the final matching set, the 3D world coordinates (X,Y,Z) can be computed Each match can be stored with coordinates and averaged SIFT orientation/scale. Relaxing some constraints doesn t necessarily increase matches because ambiguous points are discarded 5

Example Matches Ego-motion Estimation For each matched feature, calculate: (r,c): measured image coordinates in reference camera (s,o,d) : scale, orientation, disparity for each feature (X,Y,Z) : estimated real-world coordinates To build a map, movement measurements between scenes is needed Odometry data is only a rough estimate Use point correspondences in least-squares estimation to calculate a more accurate localization 6

Limiting the feature-matching search Searching the whole database for SIFT feature matches is too expensive Use the odometry data to compute the expected location and scale of features in the new scene The search for SIFT feature matches is constrained to a 5x5 pixel window around this estimate. Results Matching across movement a) 10 cm forward movement b) 5 degree rotation a) b) 7

Least-Squares Error Minimization To refine the initial pose (R,T) estimates between the matched points, use Newton s method: p i :parameters at step I x : terms to subtract to minimize error J: Jacobian matrix encoding derivative of error wrt change in parameter One observed error e i should be equal to the sum of changes of that error w.r.t changes in the parameters x Solve the resulting system of linear equations Pose Estimation Results Odometry measurements are used to initialize the minimization SIFT features give very high percentage of inliers; so outliers are discarded (residual error >3 pixels) We re estimating the 6 d.o.f. in (T,R) 8

Landmark Tracking Need a database to store observed SIFT landmarks for subsequent matching Entry includes position w.r.t camera, orientation, scale of SIFT feature, and l, consecutive frames feature has been missed Features not seen for N (20) frames, and expected to be in view, are pruned. New features are added after 3 frames Matched features values are averaged, and these features are used to update pose estimation. Experimental Results Estimated position after a (~8m) tour of the lab Change in height is forced to 0 Error minimization terms are limited; odometry between frames is fairly accurate 40-60 matches per frame 3590 SIFT features in database at the end Runs at 2Hz on a PIII 9

SIFT feature map Enhancements Permanent Landmarks In a volatile environment, we need a way to know which features to keep Keep features which meet a certain reliability: percentage of frames they are matched out of the total frames they are expected Create a static database of reliable features while environment is clear. These features won t be pruned when occluded 10

Enhancements Viewpoint Variation Even though SIFT features are reasonably invariant, they can t accommodate large changes in viewpoint Store view angle with SIFT feature If current view angle of a matched feature is larger than a threshold (20 ), add the SIFT feature from new viewpoint to the same landmark entry Questions? 11