Mobile robot localization using laser range scanner and omnicamera
|
|
- Lynette Cox
- 5 years ago
- Views:
Transcription
1 Mobile robot localization using laser range scanner and omnicamera Mariusz Olszewski Barbara Siemiatkowska, * Rafal Chojecki Piotr Marcinkiewicz Piotr Trojanek 2 Marek Majchrowski 2 * Institute of Fundamental Technological Research, Polish Academy of Sciences, Swietokrzyska 21, Warsaw, Poland bsiem@.ippt.gov.pl Warsaw University of Technology Institute of Automatic Control and Robotics sw. Andrzeja Boboli 8, Warsaw, Poland Warsaw University of Technology Institute of Control and Computation Engineering Nowowiejska, 15/19, Warsaw, Poland Abstract. In this paper a method of mobile robot localization in an unknown indoor environment is presented. The robot is equipped with a single omni-directional camera and a laser range nder. The data of the 2D scanner is used to detect walls in the robot environment. The images taken from omnicamera allow to detect texture on the walls. The robot's position can easily be estimated by combining information taken from the camera and from the laser scanner and this data can be used to follow the robot's path. The method has been tested with the use of mobile robot ELEKTRON R1 in a real oce environment. 1 Introduction Autonomous mobile robot has to be able to determine its position and orientation while moving and the most fundamental part of navigation systems is self-localization. The pose of a robot is denoted by a triple (x, y, θ), where (x,y) is the position of the robot and θ is its orientation. Until now numerous localization methods have been proposed[17][18][4][6][7][9]. The most widely used is odometry but error in determining the position of the robot increases proportionally to the distance traveled by the vehicle. In order to improve localization additional methods should be used. The methods belong to one of two classes: Relative - the current pose p t of the robot is determined relative to the previous one p t 1. Global - the position and orientation of the robot is computed within a priori given exact map or a known scene. 1
2 In this paper relative localization technique is considered but proposed techniques can be used for absolute and topological localization. Usually, an on-board laser range nder is used in order to trace the position of a robot. If the robot moves along a corridor the orientation of the robot and its displacement in the direction perpendicular to the wall can be compute very accurately but it is dicult to determine the displacement of the vehicle paralelly to the walls. In order to improve the localization technique additional sensor should be used. In our experiments the mobile platform is equipped with the omni-directional camera with hyperbolic mirror. The main advantage of omni-directional sensors is that they give 360 degrees view of the robot's environment and they simplify the interpretation of a scene. Based on images taken from an omnicamera the angular information about position of the object relatively to the robot is available immediately but the distance can be estimated after calibration of the sensor is performed. In our approach the calibration is performed using the method described in [15]. Figures 1a), 1b) present the readouts of the laser range nder taken in the corridor in two dierent positions. The displacement of the robot equals 40cm. Figure 2a), 2b) present the images taken from the omnicamera. Analysis of the images allows to determine the displacement of the robot pararelly to the walls. Figure 1. Laser readings The method presented in this paper is a development of the algorithm proposed in [15] and consists of following steps: 1. The omnicamera sensor is calibrated. The method of calibration is described in [15]; 2. Readings from laser range nder and omni-camera images are collected; 3. 2D map of an environment based on laser readings is build. The map is represented as a set of segments. 4. The images taken from omni-camera sensor are analysed. The images of the walls are transformed. 5. The displacement along the wall is computed. 2
3 Figure 2. Image of the wall taken from omnicamera 2 Wall detection The method of building 2D map of environment is described in [17]. The method of walls detection based on laser scanner readings consists of two stages: rstly the modication of Hough transform is used in order to built sets of collinear points. A line segment is described using normal notation: x cosα + y sinα = c (2.1) where c is the distance from the origin to this line computed along a normal, and α is the orientation of the normal with respect to the X axis. For points (x i, y i ), which belong to the same line, the parameters (α, c) are the same. In order to improve the precision of computed values of parameters we look for c and α which minimize the error: δ = N (R i cos(α φ i ) c) 2 (2.2) i=0 where N is the number of collinear readings, R i is the distance to the nearest obstacles in direction φ i. The straight line is indicated by the pair of parameters (c, α ) for which the value of δ is minimal. The method described above allows of determining the orientation of the robot and its displacement in direction perpendicular to the wall. When the robot acts in the corridor the error of detrming the robots orientation do not increase 0.3 o and the error of determining the displacement in the direction perpendicular to the wall 0.2cm. The error does not increase in time. 3 Cosine transform The discrete cosine transform (DTF) is a technique for converting a signal into elementary components. It is widely used in image compression. It was developed by Ahmed, 3
4 Natarajan and Rao [2]. Chen and Pratt [8] started to use DTF for images compression. 1D sequence of length N is dened by equation: N 1 C(u) = α(u) x=0 f(x)cos( π(2x + 1)u ) (3.1) 2N for u = 0, 1,2,...,N-1. α(u) is dened as: 1 α(u) = N for u = 0 (3.2) 2 N for u 0 The inverted cosine transform is dened by: f(x) = N 1 n=0 α(u)c(n)cos( π(2n + 1)x ) (3.3) 2N When cosine transform is applied to an input signal, it yields a vector of weighted values corresponding to how much of each basis function is present in the original signal. For most cases, much of the signal energy lies at low frequencies, so higher frequencies can be neglected with little distortion. 4 Detecting displacement along the wall The positions of observed walls are acquired by the laser scanner. Inverse perspective transform [1] can be used to nd texture of wall based on omnidirectional image. The angular image coordinates (ψ, β) are transformed to wall coordinates (w, h). where w - horizontal position, h - vertical position. Transformation is given by w = d (4.1) tg(ψ) tg(β) h = (4.2) d 2 + w 2 where d is distance to wall measured by laser scanner. Figure 4a) presents the fragment of the image taken from the omnicamera. Figure 4b) presents transformed image of the wall. The features from the walls are projected onto horizontal edges. For each vertical line of the wall the 1D cosine transform is performed. The weighted values of the cosine transform describe a texture of a line and they are the features which are projected. The lines of similar texture correspond to similar weighted values. In our experiment 8 values are used. Considering transformed images of the walls for two dierent positions of the robot we can determine the robot displacement along the wall by comparing the position of the areas of similar texture. 4
5 To gure out best t movement along the wall, one should nd minimum of the error given by following formulae: E(x, y) = [i 1 (ˆx) i 2 (x + ˆx)] 2 (4.3) M 1 ˆx=0 i - is a vector of feature of a line ˆx. M is the number of vertical lines in the image of the wall. When omnicamera sensor is used the resolution depends on the distance between the center of the camera and the observed region. In order to reduce an impact of errors we are searching for minimum of a function: E(x) = M 1 ˆx=0 a i [i 1 (ˆx) i 2 (x + ˆx)] 2 (4.4) where coecient a i qualify reliability of camera readings. The values are computed during calibration process. In our approach the value of a i is described by formulae a i = 1 + Ri 2 (4.5) Where R i is the distance between corresponding point on the wall and the robot position. a) Figure 3. Transormation of the image 5 Experimental results The robot ELECTRON 1 is used in our experiments 5. The robot was developed in the Institute of Automatic Control and Robotics and Institute of Control and Computation Engineering of Warsaw University of Technology. The six wheels mobile platform is equipped with on-board Biscuit PC computer (Celeron 650 MHz). Communication 5
6 between the robot and the host computer is done via wireless Ethernet. The robot is powered by two reversible DC motors. It can move with an approximate maximum speed of 0,2m/s. The vehicle is equipped with odometry sensors, 18 optical distance sensors, SICK LMS 200 laser scanner and a panoramic camera. The omni-directional sensor is composed of C-MOS color camera pointed upwards at the vertex of the hyperbolical mirror. The optical axis of the camera and the optical axis of the mirror are aligned. Figure 4. The mobile robot ELECTRON 1 1) catadioptric sensor, 2) camera, 3) laser scanner, 4) robot Experiments were performed in the corridor. The error of determining the robot displacement along the wall depends not only on the distance between the robot and the wall but also on the texture of the wall. The method is useless when there are not any characteristic area on the wall. The result depends also on the distance between the robot and characteristic places on the wall. Figures 5-6 present the result of our experiment. The task for the robot is to determined its distance to the center of the door. Error of determining robots position was computed for three dierent paths of the robot. In the rst situation approximated distance between the robot and the wall equals 0.5m, in the second one the distance equals 1m and in the last one 1.5m. 6 Conclusion The main motivation of the work presented in this article was to built the system for mobile robot localization. The environment is represented as 2D map of walls. Each wall is described using linear cosine transform. Values of coecients of cosine transform are 6
7 Figure 5. The path of the robot Figure 6. Error of determining the position of the robot the features which are traced. The performed experiments proved the eciency of the proposed method as a tool for local navigation. Presented method can be used not only as a localization technique but also as a tool for 3D map building. Bibliography [1] Adorni G., Cagnoni S., Mordonini M., Sgorbissa A.: Omnidirectional stereo systems for robot navigation. I EEE Workshop on Omnidirectional Vision (in conjunction with CVPR), June 2003 [2] Ahmed N., Natarajan T., Rao K. R. On image processing and a discrete cosine transform. I EEE Transaction on Computers C-23(1), pp [3] Aihira H., Iwasa N., et. al., Memory-based self-localization using omnidirectional images, I nternational Conference on Pattern Recognition, Silver Spring, pp
8 [4] Andersen C.S. Jones S. and Crowley J.L., Appearance Based Processes for Visual Navigation, Proc. of Symposium on Intelligent Robotics Systems, [5] Baker S., Nayar S., A Theory of Catadioptric Image Formation, I CCV'98, pp [6] Betke M. and Gurvits L. Mobile Robot Localization using Landmarks, I EEE Transactions on Robotics and Automation, 1997 [7] Borges G.A., Aldon M. J. Robustied estimation algorithms for mobile robot localization based on geometrical environment maps, Robotics and Autonomous Systems Volume: 45, pp [8] Chen W. H., Pratt W. K. Scene adaptive coder, I EEE Transaction on Communication, pp , 1984 [9] Davison A.J. and Murray D. W. (1998). Mobile Robot Localization Using Active Vision, P roc. of European Conference on Computer Vision (ECCV) [10] Gaspar J., Winters N., Santos-Victor J., Vision-based navigation and environmental representation with an omnidirectional camera, I EEE Transactions on Robotics and Automation, 16, pp , [11] Geyer C., Catadioptric Projective Geometry: Theory And Applications, C hristopher Geyer, PhD thesis, 2003 [12] Meyer S. L., Data Analysis for Scientist and Engineers, J ohn Wiley & Sons, 1997 [13] Mennegatti, E, Zoccarato M., Pagell E., Ishiguro H. Hierarchical image-based localization for mobile robot with Monte-Carlo Localisation, E CMR, pp , [14] Siemiatkowska B., Dubrawski, Cellular Neural Network for Navigation of a Mobile Robot, RSTC'98, Springer, pp , 1998 [15] Siemiatkowska B., Chojecki R., Mobile Robot Localization Based on Omincamera I AV04, 2004 [16] Svoboda T., Central Panoramic Cameras Design, Geometry, Egomotion T omás Svoboda, PhD thesis, [17] Thrun S. and Fox D. and Burgard W., Probabilistic Approach to Concurrent Mapping and Localization for Mobile Robots, M achine Learning vol. 31, pp [18] Thrun S. and Fox D. and Burgard W. Probabilistic mapping of an environment by a mobile robot. I EEE ICRA, pp ,
Catadioptric camera model with conic mirror
LÓPEZ-NICOLÁS, SAGÜÉS: CATADIOPTRIC CAMERA MODEL WITH CONIC MIRROR Catadioptric camera model with conic mirror G. López-Nicolás gonlopez@unizar.es C. Sagüés csagues@unizar.es Instituto de Investigación
More informationDominant plane detection using optical flow and Independent Component Analysis
Dominant plane detection using optical flow and Independent Component Analysis Naoya OHNISHI 1 and Atsushi IMIYA 2 1 School of Science and Technology, Chiba University, Japan Yayoicho 1-33, Inage-ku, 263-8522,
More informationVol. 21 No. 6, pp ,
Vol. 21 No. 6, pp.69 696, 23 69 3 3 3 Map Generation of a Mobile Robot by Integrating Omnidirectional Stereo and Laser Range Finder Yoshiro Negishi 3, Jun Miura 3 and Yoshiaki Shirai 3 This paper describes
More informationAUTONOMOUS SYSTEMS. LOCALIZATION, MAPPING & SIMULTANEOUS LOCALIZATION AND MAPPING Part V Mapping & Occupancy Grid Mapping
AUTONOMOUS SYSTEMS LOCALIZATION, MAPPING & SIMULTANEOUS LOCALIZATION AND MAPPING Part V Mapping & Occupancy Grid Mapping Maria Isabel Ribeiro Pedro Lima with revisions introduced by Rodrigo Ventura Instituto
More informationSensor Modalities. Sensor modality: Different modalities:
Sensor Modalities Sensor modality: Sensors which measure same form of energy and process it in similar ways Modality refers to the raw input used by the sensors Different modalities: Sound Pressure Temperature
More informationOmni Flow. Libor Spacek Department of Computer Science University of Essex, Colchester, CO4 3SQ, UK. Abstract. 1. Introduction
Omni Flow Libor Spacek Department of Computer Science University of Essex, Colchester, CO4 3SQ, UK. Abstract Catadioptric omnidirectional sensors (catadioptric cameras) capture instantaneous images with
More information3D Environment Measurement Using Binocular Stereo and Motion Stereo by Mobile Robot with Omnidirectional Stereo Camera
3D Environment Measurement Using Binocular Stereo and Motion Stereo by Mobile Robot with Omnidirectional Stereo Camera Shinichi GOTO Department of Mechanical Engineering Shizuoka University 3-5-1 Johoku,
More informationRobotics. Lecture 8: Simultaneous Localisation and Mapping (SLAM)
Robotics Lecture 8: Simultaneous Localisation and Mapping (SLAM) See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information. Andrew Davison Department of Computing Imperial College
More informationCOMPARING GLOBAL MEASURES OF IMAGE SIMILARITY FOR USE IN TOPOLOGICAL LOCALIZATION OF MOBILE ROBOTS
COMPARING GLOBAL MEASURES OF IMAGE SIMILARITY FOR USE IN TOPOLOGICAL LOCALIZATION OF MOBILE ROBOTS Syeda Nusrat Ferdaus, Andrew Vardy, George Mann and Ray Gosine Intelligent Systems Lab, Faculty of Engineering
More informationPrecise Omnidirectional Camera Calibration
Precise Omnidirectional Camera Calibration Dennis Strelow, Jeffrey Mishler, David Koes, and Sanjiv Singh Carnegie Mellon University {dstrelow, jmishler, dkoes, ssingh}@cs.cmu.edu Abstract Recent omnidirectional
More informationAn Extended Line Tracking Algorithm
An Extended Line Tracking Algorithm Leonardo Romero Muñoz Facultad de Ingeniería Eléctrica UMSNH Morelia, Mich., Mexico Email: lromero@umich.mx Moises García Villanueva Facultad de Ingeniería Eléctrica
More informationEstimation of Camera Motion with Feature Flow Model for 3D Environment Modeling by Using Omni-Directional Camera
Estimation of Camera Motion with Feature Flow Model for 3D Environment Modeling by Using Omni-Directional Camera Ryosuke Kawanishi, Atsushi Yamashita and Toru Kaneko Abstract Map information is important
More informationComputer Vision. Coordinates. Prof. Flávio Cardeal DECOM / CEFET- MG.
Computer Vision Coordinates Prof. Flávio Cardeal DECOM / CEFET- MG cardeal@decom.cefetmg.br Abstract This lecture discusses world coordinates and homogeneous coordinates, as well as provides an overview
More informationFlexible Calibration of a Portable Structured Light System through Surface Plane
Vol. 34, No. 11 ACTA AUTOMATICA SINICA November, 2008 Flexible Calibration of a Portable Structured Light System through Surface Plane GAO Wei 1 WANG Liang 1 HU Zhan-Yi 1 Abstract For a portable structured
More informationTopological Mobile Robot Localization Using Fast Vision Techniques
Proceedings of the 2002 IEEE International Conference on Robotics 8 Automation Washington, DC May 2002 Topological Mobile Robot Localization Using Fast Vision Techniques Paul Blaer and Peter Allen Department
More informationMOBILE ROBOT LOCALIZATION. REVISITING THE TRIANGULATION METHODS. Josep Maria Font, Joaquim A. Batlle
MOBILE ROBOT LOCALIZATION. REVISITING THE TRIANGULATION METHODS Josep Maria Font, Joaquim A. Batlle Department of Mechanical Engineering Technical University of Catalonia (UC) Avda. Diagonal 647, 08028
More informationAutomatic Generation of Indoor VR-Models by a Mobile Robot with a Laser Range Finder and a Color Camera
Automatic Generation of Indoor VR-Models by a Mobile Robot with a Laser Range Finder and a Color Camera Christian Weiss and Andreas Zell Universität Tübingen, Wilhelm-Schickard-Institut für Informatik,
More informationStructure from Small Baseline Motion with Central Panoramic Cameras
Structure from Small Baseline Motion with Central Panoramic Cameras Omid Shakernia René Vidal Shankar Sastry Department of Electrical Engineering & Computer Sciences, UC Berkeley {omids,rvidal,sastry}@eecs.berkeley.edu
More informationCOMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION
COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION Mr.V.SRINIVASA RAO 1 Prof.A.SATYA KALYAN 2 DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING PRASAD V POTLURI SIDDHARTHA
More informationOmni Stereo Vision of Cooperative Mobile Robots
Omni Stereo Vision of Cooperative Mobile Robots Zhigang Zhu*, Jizhong Xiao** *Department of Computer Science **Department of Electrical Engineering The City College of the City University of New York (CUNY)
More informationRobotics. Lecture 7: Simultaneous Localisation and Mapping (SLAM)
Robotics Lecture 7: Simultaneous Localisation and Mapping (SLAM) See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information. Andrew Davison Department of Computing Imperial College
More informationVisually Augmented POMDP for Indoor Robot Navigation
Visually Augmented POMDP for Indoor obot Navigation LÓPEZ M.E., BAEA., BEGASA L.M., ESCUDEO M.S. Electronics Department University of Alcalá Campus Universitario. 28871 Alcalá de Henares (Madrid) SPAIN
More informationUSING 3D DATA FOR MONTE CARLO LOCALIZATION IN COMPLEX INDOOR ENVIRONMENTS. Oliver Wulf, Bernardo Wagner
USING 3D DATA FOR MONTE CARLO LOCALIZATION IN COMPLEX INDOOR ENVIRONMENTS Oliver Wulf, Bernardo Wagner Institute for Systems Engineering (RTS/ISE), University of Hannover, Germany Mohamed Khalaf-Allah
More informationIndoor Positioning System Based on Distributed Camera Sensor Networks for Mobile Robot
Indoor Positioning System Based on Distributed Camera Sensor Networks for Mobile Robot Yonghoon Ji 1, Atsushi Yamashita 1, and Hajime Asama 1 School of Engineering, The University of Tokyo, Japan, t{ji,
More informationLocalization and Map Building
Localization and Map Building Noise and aliasing; odometric position estimation To localize or not to localize Belief representation Map representation Probabilistic map-based localization Other examples
More informationRegion matching for omnidirectional images using virtual camera planes
Computer Vision Winter Workshop 2006, Ondřej Chum, Vojtěch Franc (eds.) Telč, Czech Republic, February 6 8 Czech Pattern Recognition Society Region matching for omnidirectional images using virtual camera
More informationImage-Based Memory of Environment. homing uses a similar idea that the agent memorizes. [Hong 91]. However, the agent nds diculties in arranging its
Image-Based Memory of Environment Hiroshi ISHIGURO Department of Information Science Kyoto University Kyoto 606-01, Japan E-mail: ishiguro@kuis.kyoto-u.ac.jp Saburo TSUJI Faculty of Systems Engineering
More informationA New Omnidirectional Vision Sensor for Monte-Carlo Localization
A New Omnidirectional Vision Sensor for Monte-Carlo Localization E. Menegatti 1, A. Pretto 1, and E. Pagello 12 1 Intelligent Autonomous Systems Laboratory Department of Information Engineering The University
More informationA Computer Vision Sensor for Panoramic Depth Perception
A Computer Vision Sensor for Panoramic Depth Perception Radu Orghidan 1, El Mustapha Mouaddib 2, and Joaquim Salvi 1 1 Institute of Informatics and Applications, Computer Vision and Robotics Group University
More informationDept. of Adaptive Machine Systems, Graduate School of Engineering Osaka University, Suita, Osaka , Japan
An Application of Vision-Based Learning for a Real Robot in RoboCup - A Goal Keeping Behavior for a Robot with an Omnidirectional Vision and an Embedded Servoing - Sho ji Suzuki 1, Tatsunori Kato 1, Hiroshi
More informationTopological Mapping. Discrete Bayes Filter
Topological Mapping Discrete Bayes Filter Vision Based Localization Given a image(s) acquired by moving camera determine the robot s location and pose? Towards localization without odometry What can be
More informationMathematics of a Multiple Omni-Directional System
Mathematics of a Multiple Omni-Directional System A. Torii A. Sugimoto A. Imiya, School of Science and National Institute of Institute of Media and Technology, Informatics, Information Technology, Chiba
More informationMonte Carlo Localization using 3D Texture Maps
Monte Carlo Localization using 3D Texture Maps Yu Fu, Stephen Tully, George Kantor, and Howie Choset Abstract This paper uses KLD-based (Kullback-Leibler Divergence) Monte Carlo Localization (MCL) to localize
More informationMatching Evaluation of 2D Laser Scan Points using Observed Probability in Unstable Measurement Environment
Matching Evaluation of D Laser Scan Points using Observed Probability in Unstable Measurement Environment Taichi Yamada, and Akihisa Ohya Abstract In the real environment such as urban areas sidewalk,
More information3-D MAP GENERATION BY A MOBILE ROBOT EQUIPPED WITH A LASER RANGE FINDER. Takumi Nakamoto, Atsushi Yamashita, and Toru Kaneko
3-D AP GENERATION BY A OBILE ROBOT EQUIPPED WITH A LAER RANGE FINDER Takumi Nakamoto, Atsushi Yamashita, and Toru Kaneko Department of echanical Engineering, hizuoka Univerty 3-5-1 Johoku, Hamamatsu-shi,
More informationPartial Calibration and Mirror Shape Recovery for Non-Central Catadioptric Systems
Partial Calibration and Mirror Shape Recovery for Non-Central Catadioptric Systems Abstract In this paper we present a method for mirror shape recovery and partial calibration for non-central catadioptric
More informationPanoramic 3D Reconstruction Using Rotational Stereo Camera with Simple Epipolar Constraints
Panoramic 3D Reconstruction Using Rotational Stereo Camera with Simple Epipolar Constraints Wei Jiang Japan Science and Technology Agency 4-1-8, Honcho, Kawaguchi-shi, Saitama, Japan jiang@anken.go.jp
More informationBehavior Learning for a Mobile Robot with Omnidirectional Vision Enhanced by an Active Zoom Mechanism
Behavior Learning for a Mobile Robot with Omnidirectional Vision Enhanced by an Active Zoom Mechanism Sho ji Suzuki, Tatsunori Kato, Minoru Asada, and Koh Hosoda Dept. of Adaptive Machine Systems, Graduate
More informationSemantics in Human Localization and Mapping
Semantics in Human Localization and Mapping Aidos Sarsembayev, Antonio Sgorbissa University of Genova, Dept. DIBRIS Via Opera Pia 13, 16145 Genova, Italy aidos.sarsembayev@edu.unige.it, antonio.sgorbissa@unige.it
More informationCollecting outdoor datasets for benchmarking vision based robot localization
Collecting outdoor datasets for benchmarking vision based robot localization Emanuele Frontoni*, Andrea Ascani, Adriano Mancini, Primo Zingaretti Department of Ingegneria Infromatica, Gestionale e dell
More informationRobot Localization based on Geo-referenced Images and G raphic Methods
Robot Localization based on Geo-referenced Images and G raphic Methods Sid Ahmed Berrabah Mechanical Department, Royal Military School, Belgium, sidahmed.berrabah@rma.ac.be Janusz Bedkowski, Łukasz Lubasiński,
More informationMonitoring surrounding areas of truck-trailer combinations
Monitoring surrounding areas of truck-trailer combinations Tobias Ehlgen 1 and Tomas Pajdla 2 1 Daimler-Chrysler Research and Technology, Ulm tobias.ehlgen@daimlerchrysler.com 2 Center of Machine Perception,
More informationAn Omnidirectional Camera Simulation for the USARSim World
An Omnidirectional Camera Simulation for the USARSim World Tijn Schmits and Arnoud Visser Universiteit van Amsterdam, 1098 SJ Amsterdam, the Netherlands [tschmits,arnoud]@science.uva.nl Abstract. Omnidirectional
More informationReal-time Generation and Presentation of View-dependent Binocular Stereo Images Using a Sequence of Omnidirectional Images
Real-time Generation and Presentation of View-dependent Binocular Stereo Images Using a Sequence of Omnidirectional Images Abstract This paper presents a new method to generate and present arbitrarily
More informationIntroduction to Robotics
Introduction to Robotics Ph.D. Antonio Marin-Hernandez Artificial Intelligence Department Universidad Veracruzana Sebastian Camacho # 5 Xalapa, Veracruz Robotics Action and Perception LAAS-CNRS 7, av du
More informationComputer Vision cmput 428/615
Computer Vision cmput 428/615 Basic 2D and 3D geometry and Camera models Martin Jagersand The equation of projection Intuitively: How do we develop a consistent mathematical framework for projection calculations?
More informationA Simple Interface for Mobile Robot Equipped with Single Camera using Motion Stereo Vision
A Simple Interface for Mobile Robot Equipped with Single Camera using Motion Stereo Vision Stephen Karungaru, Atsushi Ishitani, Takuya Shiraishi, and Minoru Fukumi Abstract Recently, robot technology has
More informationOMNIDIRECTIONAL STEREOVISION SYSTEM WITH TWO-LOBE HYPERBOLIC MIRROR FOR ROBOT NAVIGATION
Proceedings of COBEM 005 Copyright 005 by ABCM 18th International Congress of Mechanical Engineering November 6-11, 005, Ouro Preto, MG OMNIDIRECTIONAL STEREOVISION SYSTEM WITH TWO-LOBE HYPERBOLIC MIRROR
More informationDEAD RECKONING FOR MOBILE ROBOTS USING TWO OPTICAL MICE
DEAD RECKONING FOR MOBILE ROBOTS USING TWO OPTICAL MICE Andrea Bonarini Matteo Matteucci Marcello Restelli Department of Electronics and Information Politecnico di Milano Piazza Leonardo da Vinci, I-20133,
More informationDEALING WITH SENSOR ERRORS IN SCAN MATCHING FOR SIMULTANEOUS LOCALIZATION AND MAPPING
Inženýrská MECHANIKA, roč. 15, 2008, č. 5, s. 337 344 337 DEALING WITH SENSOR ERRORS IN SCAN MATCHING FOR SIMULTANEOUS LOCALIZATION AND MAPPING Jiří Krejsa, Stanislav Věchet* The paper presents Potential-Based
More informationImprovements for an appearance-based SLAM-Approach for large-scale environments
1 Improvements for an appearance-based SLAM-Approach for large-scale environments Alexander Koenig Jens Kessler Horst-Michael Gross Neuroinformatics and Cognitive Robotics Lab, Ilmenau University of Technology,
More informationSimulation of a mobile robot with a LRF in a 2D environment and map building
Simulation of a mobile robot with a LRF in a 2D environment and map building Teslić L. 1, Klančar G. 2, and Škrjanc I. 3 1 Faculty of Electrical Engineering, University of Ljubljana, Tržaška 25, 1000 Ljubljana,
More informationProbabilistic Matching for 3D Scan Registration
Probabilistic Matching for 3D Scan Registration Dirk Hähnel Wolfram Burgard Department of Computer Science, University of Freiburg, 79110 Freiburg, Germany Abstract In this paper we consider the problem
More informationSimuntaneous Localisation and Mapping with a Single Camera. Abhishek Aneja and Zhichao Chen
Simuntaneous Localisation and Mapping with a Single Camera Abhishek Aneja and Zhichao Chen 3 December, Simuntaneous Localisation and Mapping with asinglecamera 1 Abstract Image reconstruction is common
More informationTransactions on Information and Communications Technologies vol 16, 1996 WIT Press, ISSN
ransactions on Information and Communications echnologies vol 6, 996 WI Press, www.witpress.com, ISSN 743-357 Obstacle detection using stereo without correspondence L. X. Zhou & W. K. Gu Institute of Information
More informationMOTION. Feature Matching/Tracking. Control Signal Generation REFERENCE IMAGE
Head-Eye Coordination: A Closed-Form Solution M. Xie School of Mechanical & Production Engineering Nanyang Technological University, Singapore 639798 Email: mmxie@ntuix.ntu.ac.sg ABSTRACT In this paper,
More informationA MOBILE ROBOT MAPPING SYSTEM WITH AN INFORMATION-BASED EXPLORATION STRATEGY
A MOBILE ROBOT MAPPING SYSTEM WITH AN INFORMATION-BASED EXPLORATION STRATEGY Francesco Amigoni, Vincenzo Caglioti, Umberto Galtarossa Dipartimento di Elettronica e Informazione, Politecnico di Milano Piazza
More informationMobile Robotics. Mathematics, Models, and Methods. HI Cambridge. Alonzo Kelly. Carnegie Mellon University UNIVERSITY PRESS
Mobile Robotics Mathematics, Models, and Methods Alonzo Kelly Carnegie Mellon University HI Cambridge UNIVERSITY PRESS Contents Preface page xiii 1 Introduction 1 1.1 Applications of Mobile Robots 2 1.2
More informationHigh-speed Three-dimensional Mapping by Direct Estimation of a Small Motion Using Range Images
MECATRONICS - REM 2016 June 15-17, 2016 High-speed Three-dimensional Mapping by Direct Estimation of a Small Motion Using Range Images Shinta Nozaki and Masashi Kimura School of Science and Engineering
More informationReal-time Security Monitoring around a Video Surveillance Vehicle with a Pair of Two-camera Omni-imaging Devices
Real-time Security Monitoring around a Video Surveillance Vehicle with a Pair of Two-camera Omni-imaging Devices Pei-Hsuan Yuan, Kuo-Feng Yang and Wen-Hsiang Tsai, Senior Member, IEEE Abstract A pair of
More informationLocalization, Where am I?
5.1 Localization, Where am I?? position Position Update (Estimation?) Encoder Prediction of Position (e.g. odometry) YES matched observations Map data base predicted position Matching Odometry, Dead Reckoning
More informationLocalization of Multiple Robots with Simple Sensors
Proceedings of the IEEE International Conference on Mechatronics & Automation Niagara Falls, Canada July 2005 Localization of Multiple Robots with Simple Sensors Mike Peasgood and Christopher Clark Lab
More informationData Association for SLAM
CALIFORNIA INSTITUTE OF TECHNOLOGY ME/CS 132a, Winter 2011 Lab #2 Due: Mar 10th, 2011 Part I Data Association for SLAM 1 Introduction For this part, you will experiment with a simulation of an EKF SLAM
More informationarxiv: v1 [cs.cv] 28 Sep 2018
Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,
More informationS-SHAPED ONE TRAIL PARALLEL PARKING OF A CAR-LIKE MOBILE ROBOT
S-SHAPED ONE TRAIL PARALLEL PARKING OF A CAR-LIKE MOBILE ROBOT 1 SOE YU MAUNG MAUNG, 2 NU NU WIN, 3 MYINT HTAY 1,2,3 Mechatronic Engineering Department, Mandalay Technological University, The Republic
More informationA Fast Linear Registration Framework for Multi-Camera GIS Coordination
A Fast Linear Registration Framework for Multi-Camera GIS Coordination Karthik Sankaranarayanan James W. Davis Dept. of Computer Science and Engineering Ohio State University Columbus, OH 4320 USA {sankaran,jwdavis}@cse.ohio-state.edu
More informationSpatial Localization Method with Omnidirectional Vision
Spatial Localization Method with Omnidirectional Vision Cyril Cauchois, Eric Brassart, Laurent Delahoche, Cyril Drocourt CREA (Center of Robotic, Electrotechnic and Automatic) Institut Universitaire de
More informationMulti-resolution SLAM for Real World Navigation
Proceedings of the International Symposium of Research Robotics Siena, Italy, October 2003 Multi-resolution SLAM for Real World Navigation Agostino Martinelli, Adriana Tapus, Kai Olivier Arras, and Roland
More informationPerception. Autonomous Mobile Robots. Sensors Vision Uncertainties, Line extraction from laser scans. Autonomous Systems Lab. Zürich.
Autonomous Mobile Robots Localization "Position" Global Map Cognition Environment Model Local Map Path Perception Real World Environment Motion Control Perception Sensors Vision Uncertainties, Line extraction
More informationThis chapter explains two techniques which are frequently used throughout
Chapter 2 Basic Techniques This chapter explains two techniques which are frequently used throughout this thesis. First, we will introduce the concept of particle filters. A particle filter is a recursive
More informationLocalization, Mapping and Exploration with Multiple Robots. Dr. Daisy Tang
Localization, Mapping and Exploration with Multiple Robots Dr. Daisy Tang Two Presentations A real-time algorithm for mobile robot mapping with applications to multi-robot and 3D mapping, by Thrun, Burgard
More informationImage Formation. Antonino Furnari. Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania
Image Formation Antonino Furnari Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania furnari@dmi.unict.it 18/03/2014 Outline Introduction; Geometric Primitives
More informationMobile Robot Navigation With Use of Semantic Map Constructed From 3D Laser Range Scans
Intelligent Information Systems 9999 ISBN 666-666-666, pages 114 Mobile Robot Navigation With Use of Semantic Map Constructed From 3D Laser Range Scans Barbara Siemi tkowska 21, Jacek Szklarski 1, and
More informationEnvironment Identification by Comparing Maps of Landmarks
Environment Identification by Comparing Maps of Landmarks Jens-Steffen Gutmann Masaki Fukuchi Kohtaro Sabe Digital Creatures Laboratory Sony Corporation -- Kitashinagawa, Shinagawa-ku Tokyo, 4- Japan Email:
More informationCOMPUTER VISION. Dr. Sukhendu Das Deptt. of Computer Science and Engg., IIT Madras, Chennai
COMPUTER VISION Dr. Sukhendu Das Deptt. of Computer Science and Engg., IIT Madras, Chennai 600036. Email: sdas@iitm.ac.in URL: //www.cs.iitm.ernet.in/~sdas 1 INTRODUCTION 2 Human Vision System (HVS) Vs.
More informationA Sparse Hybrid Map for Vision-Guided Mobile Robots
A Sparse Hybrid Map for Vision-Guided Mobile Robots Feras Dayoub Grzegorz Cielniak Tom Duckett Department of Computing and Informatics, University of Lincoln, Lincoln, UK {fdayoub,gcielniak,tduckett}@lincoln.ac.uk
More informationThree-Dimensional Measurement of Objects in Liquid with an Unknown Refractive Index Using Fisheye Stereo Camera
Three-Dimensional Measurement of Objects in Liquid with an Unknown Refractive Index Using Fisheye Stereo Camera Kazuki Sakamoto, Alessandro Moro, Hiromitsu Fujii, Atsushi Yamashita, and Hajime Asama Abstract
More informationDegeneracy of the Linear Seventeen-Point Algorithm for Generalized Essential Matrix
J Math Imaging Vis 00 37: 40-48 DOI 0007/s085-00-09-9 Authors s version The final publication is available at wwwspringerlinkcom Degeneracy of the Linear Seventeen-Point Algorithm for Generalized Essential
More informationPlanar pattern for automatic camera calibration
Planar pattern for automatic camera calibration Beiwei Zhang Y. F. Li City University of Hong Kong Department of Manufacturing Engineering and Engineering Management Kowloon, Hong Kong Fu-Chao Wu Institute
More informationLocal Image Registration: An Adaptive Filtering Framework
Local Image Registration: An Adaptive Filtering Framework Gulcin Caner a,a.murattekalp a,b, Gaurav Sharma a and Wendi Heinzelman a a Electrical and Computer Engineering Dept.,University of Rochester, Rochester,
More informationCalibration of a rotating multi-beam Lidar
The 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems October 18-22, 2010, Taipei, Taiwan Calibration of a rotating multi-beam Lidar Naveed Muhammad 1,2 and Simon Lacroix 1,2 Abstract
More informationReview Article A State-of-the-Art Review on Mapping and Localization of Mobile Robots Using Omnidirectional Vision Sensors
Hindawi Journal of Sensors Volume 2017, Article ID 3497650, 20 pages https://doi.org/10.1155/2017/3497650 Review Article A State-of-the-Art Review on Mapping and Localization of Mobile Robots Using Omnidirectional
More informationMobile Robot Mapping and Localization in Non-Static Environments
Mobile Robot Mapping and Localization in Non-Static Environments Cyrill Stachniss Wolfram Burgard University of Freiburg, Department of Computer Science, D-790 Freiburg, Germany {stachnis burgard @informatik.uni-freiburg.de}
More informationCamera Calibration for a Robust Omni-directional Photogrammetry System
Camera Calibration for a Robust Omni-directional Photogrammetry System Fuad Khan 1, Michael Chapman 2, Jonathan Li 3 1 Immersive Media Corporation Calgary, Alberta, Canada 2 Ryerson University Toronto,
More informationLEARNING NAVIGATION MAPS BY LOOKING AT PEOPLE
LEARNING NAVIGATION MAPS BY LOOKING AT PEOPLE Roger Freitas,1 José Santos-Victor Mário Sarcinelli-Filho Teodiano Bastos-Filho Departamento de Engenharia Elétrica, Universidade Federal do Espírito Santo,
More informationVision-Based Localization for Mobile Platforms
Vision-Based Localization for Mobile Platforms Josep M. Porta and Ben J.A. Kröse IAS Group, Informatics Institute, University of Amsterdam, Kruislaan 403, 1098SJ, Amsterdam, The Netherlands {porta,krose}@science.uva.nl
More informationA 3-D Scanner Capturing Range and Color for the Robotics Applications
J.Haverinen & J.Röning, A 3-D Scanner Capturing Range and Color for the Robotics Applications, 24th Workshop of the AAPR - Applications of 3D-Imaging and Graph-based Modeling, May 25-26, Villach, Carinthia,
More informationUsing Feature Scale Change for Robot Localization along a Route
Using Feature Scale Change for Robot Localization along a Route Andrew Vardy Department of Computer Science Faculty of Engineering & Applied Science Memorial University of Newfoundland St. John s, Canada
More informationHumanoid Robotics. Monte Carlo Localization. Maren Bennewitz
Humanoid Robotics Monte Carlo Localization Maren Bennewitz 1 Basis Probability Rules (1) If x and y are independent: Bayes rule: Often written as: The denominator is a normalizing constant that ensures
More informationDevelopment of Low-Cost Compact Omnidirectional Vision Sensors and their applications
Development of Low-Cost Compact Omnidirectional Vision Sensors and their applications Hiroshi ISHIGURO Department of Electrical & Computer Engineering, University of California, San Diego (9500 Gilman
More informationPROGRAMA DE CURSO. Robotics, Sensing and Autonomous Systems. SCT Auxiliar. Personal
PROGRAMA DE CURSO Código Nombre EL7031 Robotics, Sensing and Autonomous Systems Nombre en Inglés Robotics, Sensing and Autonomous Systems es Horas de Horas Docencia Horas de Trabajo SCT Docentes Cátedra
More informationare now opportunities for applying stereo ranging to problems in mobile robot navigation. We
A Multiresolution Stereo Vision System for Mobile Robots Luca Iocchi Dipartimento di Informatica e Sistemistica Universita di Roma \La Sapienza", Italy iocchi@dis.uniroma1.it Kurt Konolige Articial Intelligence
More informationEE565:Mobile Robotics Lecture 2
EE565:Mobile Robotics Lecture 2 Welcome Dr. Ing. Ahmad Kamal Nasir Organization Lab Course Lab grading policy (40%) Attendance = 10 % In-Lab tasks = 30 % Lab assignment + viva = 60 % Make a group Either
More informationStereo camera de-calibration detection based on observing kinematic attributes of detected objects and the camera rig
Technical University of Dortmund Stereo camera de-calibration detection based on observing kinematic attributes of detected objects and the camera rig by Venkata Rama Prasad Donda A thesis submitted in
More informationHeterogeneous Multi-Robot Localization in Unknown 3D Space
Heterogeneous Multi-Robot Localization in Unknown 3D Space Yi Feng Department of Computer Science The Graduate Center, CUNY New York, NY 10016 Email: yfeng@gc.cuny.edu Zhigang Zhu Department of Computer
More informationAppearance-based Visual Localisation in Outdoor Environments with an Omnidirectional Camera
52. Internationales Wissenschaftliches Kolloquium Technische Universität Ilmenau 10. - 13. September 2007 M. Schenderlein / K. Debes / A. Koenig / H.-M. Gross Appearance-based Visual Localisation in Outdoor
More informationURBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES
URBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES An Undergraduate Research Scholars Thesis by RUI LIU Submitted to Honors and Undergraduate Research Texas A&M University in partial fulfillment
More informationExam in DD2426 Robotics and Autonomous Systems
Exam in DD2426 Robotics and Autonomous Systems Lecturer: Patric Jensfelt KTH, March 16, 2010, 9-12 No aids are allowed on the exam, i.e. no notes, no books, no calculators, etc. You need a minimum of 20
More informationCamera Pose Measurement from 2D-3D Correspondences of Three Z Shaped Lines
International Journal of Intelligent Engineering & Systems http://www.inass.org/ Camera Pose Measurement from 2D-3D Correspondences of Three Z Shaped Lines Chang Liu 1,2,3,4, Feng Zhu 1,4, Jinjun Ou 1,4,
More informationRevising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History
Revising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History Simon Thompson and Satoshi Kagami Digital Human Research Center National Institute of Advanced
More information