Localisation using Automatically Selected Landmarks from Panoramic Images

Similar documents
Revising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History

Behavior Learning for a Mobile Robot with Omnidirectional Vision Enhanced by an Active Zoom Mechanism

Matching Evaluation of 2D Laser Scan Points using Observed Probability in Unstable Measurement Environment

3D Environment Measurement Using Binocular Stereo and Motion Stereo by Mobile Robot with Omnidirectional Stereo Camera

STEREO-VISION SYSTEM PERFORMANCE ANALYSIS

3D Terrain Sensing System using Laser Range Finder with Arm-Type Movable Unit

Dept. of Adaptive Machine Systems, Graduate School of Engineering Osaka University, Suita, Osaka , Japan

Canny Edge Based Self-localization of a RoboCup Middle-sized League Robot

CONCLUSION ACKNOWLEDGMENTS REFERENCES

Learning Image-Based Landmarks for Wayfinding using Neural Networks

Depth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth

A deformable model driven method for handling clothes

Localization and Map Building

Proc. 14th Int. Conf. on Intelligent Autonomous Systems (IAS-14), 2016

Estimation of Camera Motion with Feature Flow Model for 3D Environment Modeling by Using Omni-Directional Camera

A SIMPLE VISUAL COMPASS WITH LEARNED PIXEL WEIGHTS. Andrew Vardy

(a) (b) (c) Fig. 1. Omnidirectional camera: (a) principle; (b) physical construction; (c) captured. of a local vision system is more challenging than

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation

Today MAPS AND MAPPING. Features. process of creating maps. More likely features are things that can be extracted from images:

A Symmetry Operator and Its Application to the RoboCup

Real-time Generation and Presentation of View-dependent Binocular Stereo Images Using a Sequence of Omnidirectional Images

Realtime Omnidirectional Stereo for Obstacle Detection and Tracking in Dynamic Environments

Bearing only visual servo control of a non-holonomic mobile robot. Robert Mahony

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

Detecting Salient Contours Using Orientation Energy Distribution. Part I: Thresholding Based on. Response Distribution

ADVANCES IN ROBOT VISION: MECHANISMS AND ALGORITHMS

Active noise control at a moving location in a modally dense three-dimensional sound field using virtual sensing

Interactive Collision Detection for Engineering Plants based on Large-Scale Point-Clouds

DEVELOPMENT OF POSITION MEASUREMENT SYSTEM FOR CONSTRUCTION PILE USING LASER RANGE FINDER

Fisheye Camera s Intrinsic Parameter Estimation Using Trajectories of Feature Points Obtained from Camera Rotation

Estimating Camera Position And Posture by Using Feature Landmark Database

Real-Time Human Detection using Relational Depth Similarity Features

Domain Adaptation For Mobile Robot Navigation

Object Detection through Forward Movement

Chapter 9 Object Tracking an Overview

An Algorithm to Determine the Chromaticity Under Non-uniform Illuminant

Path-planning by Tessellation of Obstacles

Fast Local Planner for Autonomous Helicopter

Kaijen Hsiao. Part A: Topics of Fascination

ECE 172A: Introduction to Intelligent Systems: Machine Vision, Fall Midterm Examination

An Efficient Need-Based Vision System in Variable Illumination Environment of Middle Size RoboCup

A threshold decision of the object image by using the smart tag

A Visual Landmark Recognition System for Autonomous Robot Navigation

CS4758: Moving Person Avoider

HOG-Based Person Following and Autonomous Returning Using Generated Map by Mobile Robot Equipped with Camera and Laser Range Finder

BROWN UNIVERSITY Department of Computer Science Master's Project CS-95-M17

Robotics. Lecture 5: Monte Carlo Localisation. See course website for up to date information.

Flexible Visual Inspection. IAS-13 Industrial Forum Horizon 2020 Dr. Eng. Stefano Tonello - CEO

Epipolar geometry-based ego-localization using an in-vehicle monocular camera

Change detection using joint intensity histogram

Intelligent Cutting of the Bird Shoulder Joint

A Vision System for Automatic State Determination of Grid Based Board Games

Vision-based Mobile Robot Localization and Mapping using Scale-Invariant Features

Map-Based Strategies for Robot Navigation in Unknown Environments

LEARNING NAVIGATION MAPS BY LOOKING AT PEOPLE

Robert Collins CSE486, Penn State Lecture 08: Introduction to Stereo

Depth Camera for Mobile Devices

Centre for Digital Image Measurement and Analysis, School of Engineering, City University, Northampton Square, London, ECIV OHB

A Simple Interface for Mobile Robot Equipped with Single Camera using Motion Stereo Vision

CS 4758: Automated Semantic Mapping of Environment

Localization, Where am I?

Continuous Localization Using Evidence Grids *

Saliency Extraction for Gaze-Contingent Displays

Dominant plane detection using optical flow and Independent Component Analysis

Expanding gait identification methods from straight to curved trajectories

Task analysis based on observing hands and objects by vision

Autonomous Robot Navigation: Using Multiple Semi-supervised Models for Obstacle Detection

AR Cultural Heritage Reconstruction Based on Feature Landmark Database Constructed by Using Omnidirectional Range Sensor

One category of visual tracking. Computer Science SURJ. Michael Fischer

Enabling a Robot to Open Doors Andrei Iancu, Ellen Klingbeil, Justin Pearson Stanford University - CS Fall 2007

Image-Based Memory of Environment. homing uses a similar idea that the agent memorizes. [Hong 91]. However, the agent nds diculties in arranging its

Detection of Small-Waving Hand by Distributed Camera System

NERC Gazebo simulation implementation

A Model of Dynamic Visual Attention for Object Tracking in Natural Image Sequences

Automatic Enhancement of Correspondence Detection in an Object Tracking System

Ensemble of Bayesian Filters for Loop Closure Detection

Path Planning for Robot Navigation using View Sequences

126 ACTION AND PERCEPTION

DEVELOPMENT OF A TRACKING AND GUIDANCE SYSTEM FOR A FIELD ROBOT

CIRCULAR MOIRÉ PATTERNS IN 3D COMPUTER VISION APPLICATIONS

Development of a video-rate range finder using dynamic threshold method for characteristic point detection

Case Study: Attempts at Parametric Reduction

DETECTION OF 3D POINTS ON MOVING OBJECTS FROM POINT CLOUD DATA FOR 3D MODELING OF OUTDOOR ENVIRONMENTS

A 3-D Scanner Capturing Range and Color for the Robotics Applications

Postprint.

Vision-based endoscope tracking for 3D ultrasound image-guided surgical navigation [Yang et al. 2014, Comp Med Imaging and Graphics]

3D Audio Perception System for Humanoid Robots

Simultaneous surface texture classification and illumination tilt angle prediction

Decoding the Human Motor Cortex

Using temporal seeding to constrain the disparity search range in stereo matching

A High Dynamic Range Vision Approach to Outdoor Localization

Mobile Human Detection Systems based on Sliding Windows Approach-A Review

Automatic Generation of Indoor VR-Models by a Mobile Robot with a Laser Range Finder and a Color Camera

using an omnidirectional camera, sufficient information for controlled play can be collected. Another example for the use of omnidirectional cameras i

A Low Power, High Throughput, Fully Event-Based Stereo System: Supplementary Documentation

Binocular cues to depth PSY 310 Greg Francis. Lecture 21. Depth perception

Acquisition of Qualitative Spatial Representation by Visual Observation

Vehicle Localization. Hannah Rae Kerner 21 April 2015

Marcel Worring Intelligent Sensory Information Systems

Localization of Multiple Robots with Simple Sensors

Transcription:

Localisation using Automatically Selected Landmarks from Panoramic Images Simon Thompson, Toshihiro Matsui and Alexander Zelinsky Abstract Intelligent Systems Division Electro-Technical Laboratory --4 Umezono, Tsukuba, Ibaraki 35 Japan Robotic Systems Laboratory Department of Systems Engineering, RSISE The Australian National University Canberra, ACT Australia [simon alex]@syseng.anu.edu.au, matsui@etl.go.jp The use of visual landmarks for robot localisation is a promising field. It is apparent that the success of localisation by visual landmarks depends on the landmarks chosen. Good landmarks are those which remain reliable over time and through changes in position and orientation. This paper describes a system which learns places by automatically selecting landmarks from panoramic images and uses them for localisation tasks. An adaption of the biologically inspired Turn Back and Look behaviour is used to evaluate potential landmarks. Normalised correlation is used to overcome the affects of changes in illumination in the environment. Results from real robot experiments are reported, showing successful localisation from up to one meter away from the learnt position. Introduction Visual localisation is one of the key problems in making successful autonomous robots. Vision as a sensor is the richest source of information about a mobile agent s environment and as such contains information vital to solving localisation problems. One limit to visual localisation is the narrow field of view offered by normal monocular camera systems. Obviously the greater the field of view of the camera the more information about the environment can be extracted through visual processing. Figure : A panoramic image Another problem is that vision provides us with too much information. In fact one of the main tasks of vision science is to work out which parts of the information is necessary for the given job and to discard or put aside that which is not.. Panoramic Imaging for Localisation Panoramic imaging can solve the first problem mentioned above [Yagi et al., 994]. By the use of convex mirror shapes greater fields of view can be achieved. Panoramic mirrors give the full 36 degrees of horizontal visual field and can get over 4 degrees in the vertical field as well as shown in figure. This increase in the visual field comes at the cost of resolution in some parts of the picture however, as it must be captured with a camera with a normal field of view. With the full 36 degree horizontal field of view there is more information from which to achieve the navigation tasks of localisation, path planning and path traversal. [Matsumoto et al., 997] use panoramic pictures to navigate a robot down corridors using a view sequence ap-

proach. The robot memorises a sequence of panoramic images along a route, acquiring new images once the current view drops below a set correlation threshold with the last image. This works well in the corridor environment, with the environment only changing sufficiently to warrant a new view being memorised every -2 meters. In a non-regular environment however the rate of acquisition would increase dramatically as the panoramic image would change dramatically over small distances..2 Natural Landmarks for Navigation An attractive solution to this problem is the use of landmarks for localisation tasks. Using landmarks in localisation requires only parts of an environment to be stored and can contribute to accurate localisation with the aid of triangulation techniques. The use of landmarks would overcome the acquisition rate problem by not requiring the entire visual field to remain constant, therefore maximising the area in which one set of landmarks can be used for localisation. Obviously the success of localisation using landmarks depends on the choice of landmarks [Thrun, 996]. [Bianco and Zelinsky, 999] describe a monocular system where landmarks are chosen on the basis of their reliability. To be selected, landmarks must display uniqueness in the immediate surroundings and the ability to remain reliable as the robot moves through the environment. To this end the system monitors the reliability of landmarks through a series of small arcs. This turn back and look behaviour is inspired from observations of honey bee flight patterns [Collet and Zeil, 996] [Lehrer, 993]. The applicability of panoramic camera systems to visual localisation using landmarks has been investigated [Thompson et al., 999]. In light of the increased field of view available to the robot, both the landmark selection strategy and the localisation procedures need to be adjusted. Like Bianco and Zelinsky, Thompson et al. describe systems which have the goals of:. Learning places in an environment associating areas with sets of appropriate visual landmarks. 2. Localising mobile robots by locating sets of landmarks which denote a particular place. 3. To ultimately use these landmarks to home to places, and to navigate between places. This paper details the implementation of such a system and reports on some initial results from localisation experiments on a real robot. 2 The System 2. System Setup This system is implemented on a Nomad mobile robot platform from Nomadic Technologies Inc. A spherical mirror is mounted on an upwards pointing camera on top of the robot to achieve panoramic vision. Images are captures via a ACVC capture card and unwarped in software to produce a 32 2 greyscale image. All image processing is done in software, taking advantage of Intel s MMX Technology to speed correlation computations. 2.2 Learning a Place The process of landmark selection is aimed at increasing the localisation and navigation abilities of a mobile robot in subsequent exposure to the environment. This means that landmarks must be reliable, strongly identifiable, they must be distributed throughout the image to minimise the error in navigation calculations. They also must be able to withstand distortions due to temporal and translational distortions. To this end Bianco and Zelinsky [999] selects a number of landmarks from different sections of the static environment based on their reliability. This reliability is then tested in a dynamic environment by moving the robot about the original position. Static Reliability of Landmarks The static reliability of a landmark in Bianco and Zelinsky s [999] model, is determined by the uniqueness of the landmark within its local region of the image. Local uniqueness is defined as the degree to which the landmark template differs from the area of the image immediately surrounding the landmark. This approach is based on the The Valley Method proposed by Mori et al [Mori et al., 995] to generate attention tokens in a scene, which in turn appears to be an instance of a Moravec interest operator [Moravec, 977] applied to feature tracking. Bianco and Zelinsky adapt this method for the present task of automatic selection of landmarks as shown in figure 2. The figure shows correlation results obtained by matching a 6 6 template on a 32 32 search window centered on the original template. The valley in the image is the template matched with itself, thus having a high correlation (low distortion). Comparing the minimum matching distortion resulting from the match of the landmark and the sixteen surrounding values, shown by the grey square, with the value at the bottom of the valley, the local uniqueness of the landmark is found. More formally: r = g/g where r is the reliability of the landmark, g is the distortion of the landmark matched with itself, and g is the minimum matching distortion from the surrounding circle of pixels. Given that g should only result in distortion due to noise, then the higher the distortion of the minimum of the surrounding templates, the steeper

Figure 2: An example of the valley in the distortion matrix caused by a reliable landmark. Figure from Bianco and Zelinksy 999 the valley in the distortion matrix and subsequently the more unique the local template should be (see figure 2). The panoramic image is divided up into 4 sectors corresponding roughly to forward, back, left and right of the robot, and the best four landmarks in each sector are selected. By dividing up the visual field in such a way an equal distribution of landmarks can be ensured. The sixteen best landmarks are then evaluated on their dynamic reliability. Dynamic Reliability of Landmarks To be good navigational cues, landmarks need to be reliable in a changing environment. They need to remain strongly identifiable under small changes in lighting and shifts in perspective. To this end Bianco and Zelinsky test the reliability of the statically selected landmarks throughout a series of small arcs. This Turn Back and Look (TBL) phase is inspired by observations of wasps and bees on flights away from the hive [Collet and Zeil, 996] [Lehrer, 993]. Figure 3 shows the path taken by Bianco and Zelinsky s [999] robot on the TBL phase. The robot moves further away from the original, or goal, position while keeping the camera oriented towards it. In a panoramic scene containing potential landmarks in any direction, a TBL phase is needed which tests the reliability of the landmarks throughout movements in all directions. At this stage a basic TBL movement is implemented in the shape of a cross as shown in figure 4. This covers movement in two planes and it is yet to be determined if a more elaborate movement would have an effect on the localisation performance. Potential landmarks are tracked along this path and their reliability measures are evaluated at approximately 4 steps along the way. Each landmark s dynamic reliability is given by the average of the reliability measures taken over the entire path. The landmarks with the highest dynamic reliability measure from each sector are chosen to represent the place. Figure 3: Bianco and Zelinsky s TBL phase for dynamic landmark selection Figure 4: Modified TBL phase for dynamic landmark selection with panoramic vision 2.3 Localisation Localisation is simply a matter of matching sets of landmarks, and their associated place, to the current visual scene. A brute force search of landmarks throughout the entire image is undertaken for each set of landmarks. The robot is assumed to be in the place associated with the set of landmarks which have the highest average correlation in the current scene. 3 Localisation Experiments Localisation experiments were carried out to determine the performance of the system under a variety of conditions described below. All experiments took place in the Electro-Technical Laboratory, Intelligent Systems Division. The environment is a semi-structured corridor about 2.5 meters in width and over 5 meters in length. All experiments consisted of an initial phase of learning a place by guiding the robot to the desired place and initiating the automatic landmark selection software. A subsequent phase of guiding the robot to several positions and attempting to locate the learnt landmarks gave the system s localisation performance results. The results are correlation values from to, with representing perfect correlation. The desired localisation performance is to achieve high correlation results over the largest possible area, while still being able to discriminate between seperate places. By having landmarks which can cover a wide area, the system can cut back on the amount of storage and processing time needed for localisation.

Static Landmarks Dynamic Landmarks.8.8.6.4.6.4.2.2 5 5 5 5 5 5 5 5 Figure 6: Sample panoramic image captured at 5: Figure 5: Localisation performance of Statically vs Dynamically selected landmarks High correlation within that area can allow more accurate positioning by landmark triangualtion and can lead to navigation behaviours such as homing and moving between places. 3. Static Landmarks vs Dynamic Landmarks First to establish the usefulness of the Turn Back and Look phase in automatic landmark selection, a comparision of localisation performance for a given place was made using landmarks selected by either soley static landmark selection or by both static and dynamic selection. In the place learning phase the most reliable static landmarks are stored in addition to the best dynamic landmarks. Both were used to localise within a 2 2 meter section of the corridor environment, centered on the place that was learnt. Measurements were taken at 2cm intervals. The results of using static landmarks for localisation are shown in the first plot of figure 5. A sharp peak is evident near the center, peaking at.79, but falling to around.5, just 4cm from the peak, and maintaining this to the edges of the graph. The second plot of figure 5 shows the results when using dynamic landmarks. Again there is a peak near the center (.86) but it is not nearly so sharp and drops less rapidly. At about 6cm from the peak with correlation values around.7, the slope of the graph decreases further and eventually falls to approximately.63 at the edges of the graph. Comparing the two graphs shows that the use of dynamic landmarks for localisation results in higher correlation measures over a greater area around the learnt place than when using static landmarks. The two different slopes observed in the dynamic graph can be attributed to the higher distortion of landmarks located on the sides of the corridor (closer to robot), when compared to those at the end of the corridor (further away). Figure 7: Sample panoramic image captured at 2: Obviously the distances measured were not enough to say for certain when the graph of the dynamic landmark localisation would approach the.5 value. This should be observable in the results from the place discrimination experiments. 3.2 Localisation under Illumination Changes Next the affect of changes of illumination on the system s performance was investigated. A place was learnt at 5:, and the selected landmarks stored. The localisation phase was carried out immediately after learning and again at 2: that evening, using the same set of learnt landmarks. Again the results given are for a 2 2 meter section of the corridor centered on the learnt place, and the measurements taken at 2cm intervals. Sample images from the robot during the 5: run and the 2: run are shown in figures 6 and 7 respectively to demonstrate the variance in illumination between trials. Figure 8 shows the results from the localisation experiments conducted with differing levels of illumination. Both peak at the same position in the graph with similar values (.86,.87 respectively) and both follow the same two step slope described in the previous experiment, with values of around.63 at the edges of the graph. Figure 9 shows a good example of the robustness of the normalised template matching routines. Landmarks are tracked successfully despite rapidly diminishing illumination.

.9 Place Discrimination in Localisation Localisation at 5: Localisation at 2:.85 landmark set landmark set 2 landmark set 3.8.75.8.8.6.4.2.6.4.2.7.65.6 5 5.55.5 5 5 5 5 5 5.45 2 3 4 5 6 7 Path Image Number (over 4m) Figure 8: Localisation performance with differing light levels: tests performed at 3: and 2: Figure : Place discrimination using learnt landmarks. 3 sets of landmarks associated with 3 learnt places are used to localise a robot along a 4 meter path. Places are learnt at the positions associated with image numbers 8 (landmark set ), 27 (set 2) and 58 (set 3). 3.3 Place Discrimination Finally the ability of the system to correctly discriminate between places was tested. Three places were learnt at sites roughly.5 meters apart down the center of the corridor. Localisation measures were then taken continuously for each of the learnt places as the robot moved in a straight path of 4 meters in length down the middle of the corridor. The correlation results for each set of landmarks (and their associated place) are given in figure. As the robot progresses down the path the landmark set associated with the nearest learnt place has the highest correlation. The deterioration in the performance from landmark set to set 3 can be attributed to the robot wandering from the center of the corridor. For the majority of the four meter traverse the robot knows approximately where it is. It achieves this from purely visual clues, not relying on any past observations or odometry. Clearly the process could benefit from knowledge of previous positioning and a probablistic model could be added to the system to enhance performance and robustness. Also an expectation of encountering a place due to past observations could significantly reduce the search time for evaluating potential places and add to the scalelability of the system. Figure 9: Landmark tracking under changing illumination 4 Conclusions and Further Work This system described in this paper demonstrates the use of automatically selected landmarks from panoramic images in mobile robot localisation. The Turn Back and

Look phase of landmark selection results in landmarks which are more reliable and can be recognised from a greater area in the robot s environment. These landmarks can also be located despite dramatic changes in illumination, due to the normalised correlation routines implemented. Finally the system was demonstrated by the robot localising between a number of learnt places along a four meter trajectory. An obvious extension of this work, as mentioned above, is to introduce a probabilistic model for localisation, with past observations influencing the outcome of the next observation. The association of landmarks with measures of reliability and correlation to the observed scene should lend well to this approach. The number and positioning of landmarks that are desirable to represent a place is also an interesting question. While in these experiments a minimal number of landmarks were used, today s hardware allows a much greater number of landmarks to be tracked at frame rate. The physical position of landmarks in relation to the robot also poses a problem. For gross localisation, landmarks which are far away and will not change under translation, or even move in the visual field, are desirable. For accurate localisation and navigation tasks, however, landmarks that do not move, or move very slightly in the visual field in response to ego motion, cannot provide the robot reliable information about its relative position. So landmarks which are very far away or lie along the direction of motion of the robot are not as useful. The desirable number and positioning of landmarks for localisation and navigation is currently being investigated. Navigation algorithms for homing to places and navigation between places are also being developed. [Moravec, 977] H. Moravec. Towards automatic visual obstacle avoidance. In Proceedings of the 5th International Joint Conference on Artificial Intelligence), volume Vision-, page 584, 977. [Mori et al., 995] T. Mori, Y.Matsumoto, T. Shibata, M. Inaba, and H. Inoue. Trackable attention point generate based on classification of correlation value distribution. In JSME Annual Conf. on Robotics and Mechatronics (ROBOMEC 95)., pages 76 79, 995. [Thompson et al., 999] S. Thompson, A. Zelinsky, and M.V.Srinivasan. Automatic landmark selection for navigation with panoramic vision. In Proceedings of Australian Conference on Robotics and Automation (ACRA 99), 999. [Thrun, 996] S. Thrun. A bayesian approach to landmark discovery and active perception in mobile robot navigation. Technical report, School of Computer Science Carnegie Mellon University, 996. [Yagi et al., 994] Y. Yagi, S. Kawato, and S. Tsuji. Real-time omnidirectional image sensor (copis) for vision-guided navigation. IEEE Transactions on Robotics and Automation, (): 2, 994. References [Bianco and Zelinsky, 999] G. Bianco and A. Zelinsky. Biologically inspired visual landmark learning and navigation for mobile robots. In Proceedings of the 999 IEEE/RSJ International Conference on Intelligent Robotic Systems (IROS 99), volume 2, pages 67 676, 999. [Collet and Zeil, 996] T.S. Collet and J. Zeil. Flights of learning. Journal of the American Psychological Society, pages 49 55, 996. [Lehrer, 993] Lehrer. Why do bees turn back and look? Journal of Comparative Physiology, 72(A):549 563, 993. [Matsumoto et al., 997] Y. Matsumoto, M. Inaba, and H. Inoue. Memory-based navigation using omni-view sequence. In Int. Conf. of Field and Service Robotics, pages 84 9, 997.