Personal Navigation and Indoor Mapping: Performance Characterization of Kinect Sensor-based Trajectory Recovery

Similar documents
Srikumar Ramalingam. Review. 3D Reconstruction. Pose Estimation Revisited. School of Computing University of Utah

Srikumar Ramalingam. Review. 3D Reconstruction. Pose Estimation Revisited. School of Computing University of Utah

Homographies and RANSAC

3D Reconstruction of a Hopkins Landmark

3D Environment Reconstruction

Runway Centerline Deviation Estimation from Point Clouds using LiDAR imagery

3D object recognition used by team robotto

Exterior Orientation Parameters

CSE 527: Introduction to Computer Vision

Vision-based Mobile Robot Localization and Mapping using Scale-Invariant Features

Tightly-Integrated Visual and Inertial Navigation for Pinpoint Landing on Rugged Terrains

Reference Data Set for Performance Evaluation of MEMS-based Integrated Navigation Solutions

Presented at the FIG Congress 2018, May 6-11, 2018 in Istanbul, Turkey

A Whole New World of Mapping and Sensing: Uses from Asset to Management to In Vehicle Sensing for Collision Avoidance

Visual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

OUTDOOR AND INDOOR NAVIGATION WITH MICROSOFT KINECT

Stable Vision-Aided Navigation for Large-Area Augmented Reality

3D Visualization through Planar Pattern Based Augmented Reality

Object Reconstruction

CS 231A Computer Vision (Winter 2014) Problem Set 3

Instance-level recognition

Computational Optical Imaging - Optique Numerique. -- Single and Multiple View Geometry, Stereo matching --

Augmenting Reality, Naturally:

3D Photography: Stereo

Miniature faking. In close-up photo, the depth of field is limited.

Instance-level recognition

Motion Tracking and Event Understanding in Video Sequences

Camera Registration in a 3D City Model. Min Ding CS294-6 Final Presentation Dec 13, 2006

Computational Optical Imaging - Optique Numerique. -- Multiple View Geometry and Stereo --

Feature descriptors and matching

PHOTOGRAMMETRIC SOLUTIONS OF NON-STANDARD PHOTOGRAMMETRIC BLOCKS INTRODUCTION

3D Photography: Active Ranging, Structured Light, ICP

FAST REGISTRATION OF TERRESTRIAL LIDAR POINT CLOUD AND SEQUENCE IMAGES

Dense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

ME132 February 3, 2011

Chapter 3 Image Registration. Chapter 3 Image Registration

Sensor Fusion: Potential, Challenges and Applications. Presented by KVH Industries and Geodetics, Inc. December 2016

Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview

Motion Estimation and Optical Flow Tracking

Feature Matching and Robust Fitting

Overview. Augmented reality and applications Marker-based augmented reality. Camera model. Binary markers Textured planar markers

CS 4758: Automated Semantic Mapping of Environment

Dealing with Scale. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

Stereo and Epipolar geometry

Object Recognition with Invariant Features

Structured light 3D reconstruction

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.

Geometric Approach to Position Determination in Space: Advantages and Limitations

MULTI-MODAL MAPPING. Robotics Day, 31 Mar Frank Mascarich, Shehryar Khattak, Tung Dang

Chaplin, Modern Times, 1936

Chapters 1 9: Overview

TRAINING MATERIAL HOW TO OPTIMIZE ACCURACY WITH CORRELATOR3D

10/03/11. Model Fitting. Computer Vision CS 143, Brown. James Hays. Slides from Silvio Savarese, Svetlana Lazebnik, and Derek Hoiem

A Comparison of SIFT, PCA-SIFT and SURF

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy

Image Features: Local Descriptors. Sanja Fidler CSC420: Intro to Image Understanding 1/ 58

3D Models from Range Sensors. Gianpaolo Palma

Local Feature Detectors

EVOLUTION OF POINT CLOUD

Feature Based Registration - Image Alignment

BUILDING POINT GROUPING USING VIEW-GEOMETRY RELATIONS INTRODUCTION

Towards a visual perception system for LNG pipe inspection

Stereo Epipolar Geometry for General Cameras. Sanja Fidler CSC420: Intro to Image Understanding 1 / 33

LOCAL AND GLOBAL DESCRIPTORS FOR PLACE RECOGNITION IN ROBOTICS

Calibration of IRS-1C PAN-camera

Il colore: acquisizione e visualizzazione. Lezione 17: 11 Maggio 2012

Journal of Asian Scientific Research AN APPLICATION OF STEREO MATCHING ALGORITHM FOR WASTE BIN LEVEL ESTIMATION

Egomotion Estimation by Point-Cloud Back-Mapping

Live Metric 3D Reconstruction on Mobile Phones ICCV 2013

Perspective Sensing for Inertial Stabilization

Efficient SLAM Scheme Based ICP Matching Algorithm Using Image and Laser Scan Information

Building a Panorama. Matching features. Matching with Features. How do we build a panorama? Computational Photography, 6.882

Accurate Motion Estimation and High-Precision 3D Reconstruction by Sensor Fusion

Outline. Introduction System Overview Camera Calibration Marker Tracking Pose Estimation of Markers Conclusion. Media IC & System Lab Po-Chen Wu 2

Instance-level recognition part 2

Topological Mapping. Discrete Bayes Filter

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014

POSITIONING A PIXEL IN A COORDINATE SYSTEM

INFO0948 Fitting and Shape Matching

PS3 Review Session. Kuan Fang CS231A 02/16/2018

III. VERVIEW OF THE METHODS

An Automatic Method for Adjustment of a Camera Calibration Room

User-assisted Segmentation and 3D Motion Tracking. Michael Fleder Sudeep Pillai Jeremy Scott

arxiv: v1 [cs.cv] 28 Sep 2018

Instance-level recognition II.

3D Digitization of Human Foot Based on Computer Stereo Vision Combined with KINECT Sensor Hai-Qing YANG a,*, Li HE b, Geng-Xin GUO c and Yong-Jun XU d

Cluster-based 3D Reconstruction of Aerial Video

HISTOGRAMS OF ORIENTATIO N GRADIENTS

Structure from Motion. Introduction to Computer Vision CSE 152 Lecture 10

Vision par ordinateur

arxiv: v1 [cs.cv] 28 Sep 2018

Image Stitching. Slides from Rick Szeliski, Steve Seitz, Derek Hoiem, Ira Kemelmacher, Ali Farhadi

Research on an Adaptive Terrain Reconstruction of Sequence Images in Deep Space Exploration

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 263

Video Georegistration: Key Challenges. Steve Blask Harris Corporation GCSD Melbourne, FL 32934

3D Perception. CS 4495 Computer Vision K. Hawkins. CS 4495 Computer Vision. 3D Perception. Kelsey Hawkins Robotics

Data-driven Depth Inference from a Single Still Image

Large Scale 3D Reconstruction by Structure from Motion

ifp Universität Stuttgart Performance of IGI AEROcontrol-IId GPS/Inertial System Final Report

Transcription:

Personal Navigation and Indoor Mapping: Performance Characterization of Kinect Sensor-based Trajectory Recovery 1 Charles TOTH, 1 Dorota BRZEZINSKA, USA 2 Allison KEALY, Australia, 3 Guenther RETSCHER, Austria 1 Department of Civil, Environmental and Geodetic Engineering Satellite Positioning and Inertial Navigation (SPIN) Lab The Ohio State University 2 Melbourne School of Engineering 3 Vienna University of Technology toth.2@osu.edu XXV FIG Congress 2014 Kuala Lumpur, Malaysia 16-21 June 2014 1 Outline Introduction FIG 5.5 Ubiquitous Positioning System (indoor, collab-nav) 2D/3D optical ranging and tracking Concept Feature extraction and matching SIFT (2D feature based) ICP (3D shape based) Kinect sensor Characteristics, test configuration Data collection Performance results and evaluation 2D image-based trajectory reconstruction 3D image-based trajectory reconstruction Combining 2D and 3D imagery Conclusions 2 1

Introduction Navigation and mapping trends: GPS-level indoor navigation performance is expected Sensor integration (multisensory systems) Collaborative navigation Indoor mapping and personal navigation (growing demand) Objectives To use low-cost sensors for indoor navigation and mapping purposes To assess the performance of using a low-cost 2D/3D imaging sensor, performance evaluation of a component of a multisensory system (error budget formation) To perform simultaneous indoor navigation and mapping of unknown environment (without a priori information of the surveyed environment) 3 2D/3D optical ranging and tracking Sensor integration-based trajectory estimation: GPS,, IMU EKF Navigation solution,,, Image image i+2 image i image i+1 Basic image-based trajectory estimation:,,,,, 0,,, 0,,, 0 } 1 } } Matching Matching Transformation estimation with outlier removal Transformation estimation with outlier removal 4,,, 2

Notes: Images can be 2D and/or 3D, typical matching combinations 2D image to 2D image (image-based navigation) 3D image to 3D image (terrain-based navigation) Combined 2D and 3D image matching Features can be points, linear features, surfaces, volumes, etc., typically characterized by a higher dimensional feature vector Matched features can be 2D/3D optical ranging and tracking Tie or conjugate features Landmarks, targets; position information of the feature is known in some frame Full transformation estimation from 2D images is not possible, as the scale is unknown; note using stereo camera configuration, the scale is known Matched features must be filtered for outliers Navigation solution (estimate) allows for limiting search space for feature matching (also, helps GPS processing after an outage, etc.) There are many ways to include image-based information in EKF, including full or partial changes between epochs, using landmark coordinates in the EKF state vector, etc. 5 SIFT SIFT Feature (4 parameter vector) 1. (X) Location 2. (Y) Location 3. Orientation 4. Scale SIFT Descriptor (128 parameter vector) - Sum of Magnitudes in region of the SIFT Feature - 4x4 bins reflect sum in each region - Taken at 8 orientations - 4x4x8 = 128 parameter descriptor D. G. Lowe, 'Distinctive image features from scale-invariant key points,' IJCV, vol. 2, no. 60, pp. 91-110, 2004. 6 3

SIFT matching of various imagery Left image: 199 keypoints 35 matched points Right image: 224 keypoints Circles represent location, orientation, and size of descriptor Yellow circles Unmatched features Red circles Matched features 7 Kinect and Casio video image matching Kinect 2D SIFT matching between Kinect and Casio video images is reliable Different scale, focal length Different sensor orientation 77 of 183 matches are inliers after RANSAC with 1-pixel threshold value. Casio P&S 8 4

Iterative Closest Point method (ICP) The ICP algorithm finds the best correspondence between two surfaces (point sets, point clouds) in 3D by iteratively determining the translations and rotation parameters of a 3D rigid body transformation The ICP algorithm works in three phases: 1) Establish correspondence between pairs of features based on proximity (for each point in D compute the closest point in M) 2) Estimate the rigid transformation that best maps the first member of the pair onto the second and then min M i ( RDi + T 2 ( R, T ) ) i http://en.wikipedia.org/wiki/point_set_registration where R is a 3*3 rotation matrix, T is a 3*1 translation vector and subscript i refer to the corresponding points of the sets M (model) and D (data). 3) Apply that transformation to all features in the first structure, repeat steps 1-2 until convergence is reached 9 Kinect sensor RGB-D camera: passive RGB + active IR Properties Interface 2D sensor (RGB camera) 3D sensor (structured IR) Frame rate Operating range Range resolution Parameters USB (12 VDC) VGA. 640x480; SXGA. 1280x1024 640x480 and 320x240 30 Hz 0.8 to 4 m a 12 bits FOV 57 x 43 Software tools (Microsoft) SDK. Windows 7. Visual Studio 2010 Enterprise. and DirectX Open source a It can be extended by ambiguity resolution up to 10 m Large user group. variety of drivers and tools in C++ and Matlab 10 5

Kinect data RGB image Depth image IR camera has a smaller field of view than RGB camera Depth image has voided areas (where 3D reconstruction failed due to object characteristics and range) Ranging accuracy is around a 1% of the range, rather stable Both sensor are calibrated (individual and inter sensor) 11 Test area Kinect sensor Data acq laptop I Approaches to sensor Kinect trajectory reconstruction: ICP matching of 3D images Matching 2D SIFT features from 2D images Combining 2D and 3D methods Based on sensor trajectory reconstructing object space (colored point cloud) 12 6

Data acquisitions Nominal data acquisition frequency: 5/30Hz Images resolution (RGB and D): 640 x 480 UWB network used as a reference Several runs by different persons Office room and hallway scenarios Generally straight path with several turns including U-turns Test 1: Average rate: 4.5 FPS Frames analyzed: 460 Significant gaps in image coverage Test 1: Average rate: 26.8 FPS Frames analyzed: 3416 No significant gaps in image coverage 13 Image characterization Image-to-image feature matching using SIFT from the Kinect s 2D camera Low quality sensor Small number of SIFT features (~42) Features may be unreliable or incorrect The distribution of SIFT features may provide bad geometry Incorrect/unstable calibration can lead to unreliable SIFT points Image-to-image matching using ICP based on the Kinect s 3D camera Modest quality sensor About 70% of image matrix contains a corresponding 3D point (640 x 480) The distribution of 3D points varies Areas with high levels of ambient IR caused by large windows in the hallway have poor reconstruction 14 7

3D image based trajectory reconstruction (ICP) Iterative closest point (ICP) algorithm was used for determining rotation and translation (3+3) parameters 2D view 3D view 15 3D image based trajectory reconstruction (ICP) 16 8

2D image based trajectory reconstruction (SIFT) Control points for iand i+1 initial images Exterior orientation i 2D matching points for i, i+1, i+2 images Space Intersection Tie points i, i+1, i+2 Exterior orientation i+1 Tie points i+1, i+2, i+3 Space Intersection 2D matching points for i+1, i+2, i+3images Exterior orientation i+2 2D matching points for i+2, i+3, i+4images Space Intersection Tie points i+2, i+3, i+4, Exterior orientation i+3, 17 Reference on slide 119 Image orientation and performance of 2D imaging Matched SIFT features in different scenarios Forward-looking camera orientation (side-looking would give better geometry) 18 9

Fusing 2D and 3D imagery 2D image Reconstructed 2D image Images can be reconstructed from point cloud (3D) on pixel-based correspondence Reconstruction has some issues at gaps (max range, etc.) and bright objects Method 1. Apply the SIFT process to reconstructed images 2. Keypoints are extracted from reconstructed 2D images 3. In addition, SIFT was used on all bands (RGB) to find more features 4. Robust estimation of 3D transformation parameters 5. Use the 3D parameters as initial values for ICP 19 Extended ICP method Use RGB information, photo registration, and previous frames TR = identity Push the first cloud into S stack Pop the last 5 known clouds from stack S, it is model M. Transform D into first frame CS with TR Read D point cloud from sensor Initial parameters from SIFT based matching Calculate rotation and transformation parameters by minimizing min,,,,,,,, algorithmic step initialization data Transform D into M. Error budget analysis Push down D into S stack 20 10

Image reconstruction Each point in the point cloud (on the depth image) has an unique corresponding point on RGB image Matching 3D points in consecutive frames can be found using 2D matching techniques, e.g. SIFT, SURF 2D matching applied on reconstructed images created from colored point clouds No features descriptors are returned for blank image parts (different from using the original 2D images) 21 SIFT-based matching in RGB SIFT keypoints on two consecutive frames R channel 21 points Gray image 29 points G channel 23 points R + G + B channels 66 points B channel 23 points R + G + B + Gray 87 points 22 11

Similarity transformation Outlier removal Applying filtering mask to maintain even (as possible) distribution of keypoints over image area 6 transformation parameters (scale equal to 1) estimated on the basis of 3D matching points Outliers filtered during transformation parameters estimation (robust) outliers get weight close to 0, 1, where: weight of point in k step of iteration, point coordinate residual in k-1 step of iteration, and a, b, s damping function parameters, empirically chosen as 1, 2, and 0.01, respectively. 23 Trajectory reconstruction performance Mismatch removal Point weighting Iterative solution Black: SIFT solution Red: Outliers removed Accuracy of 3D matching - example 232 number of 3D matching points (232 x 3 residuals are shown) only two iterations needed in the robust estimation process insignificant weights for outliers 3D matching accuracy: 7.8cm 24 12

Estimated accuracy 25 Error identification Low image variability scenario: insufficient number of inliers uneven distribution of inliers Rapid turn and variation in the frame rate: small overlapped area low number of inliers uneven distribution of inliers 26 13

3D Solutions in 2D Projections ICP 3D solution Modified ICP 3D solution 27 Final Trajectory Reconstruction Accuracy assessment: UWB reference accuracy: ~ 30 cm per coordinate For the properly reconstructed sections of the trajectory: ~ 50 cm per coordinate Providing fixes 28 14

Reconstructed trajectory and stitched point cloud Jumps Drift due to incremental approach Turn and straight path reconstructed properly 29 Conclusion Transformation parameters between consecutive frames are properly estimated for most of the survey, producing reliable trajectory and stitched point cloud Errors are mainly caused by unusual navigation behavior (U-turns) and frame rate variations; note, these situations can be easily detected Integration with other sensor data, e.g. IMU may/will improve quality of trajectory reconstruction and mapping Drift effects can be reduced introducing key-frames and applying Kalman Filter 30 15