Application of Robot Navigation Technology for Civil Robot Development

Size: px
Start display at page:

Download "Application of Robot Navigation Technology for Civil Robot Development"

Transcription

1 Safety, Reliability and Risk of Structures, Infrastructures and Engineering Systems Furuta, Frangopol & Shinozuka (eds) 2010 Taylor & Francis Group, London, ISBN Application of Robot Navigation Technology for Civil Robot Development Hyun Myung Dept. of Civil and Environmental Engineering, KAIST, Korea, Seungmok Lee Dept. of Civil and Environmental Engineering, KAIST, Korea, Haemin Jeon Dept. of Civil and Environmental Engineering, KAIST, Korea, Bum-Joo Lee Dept. of Electrical Engineering, KAIST, Korea, ABSTRACT: In implementing autonomous navigation systems, localization and map building of mobile robots are key technologies. Owing to many efforts in the past decade it has proved that simultaneous localization and map building (SLAM) is possible. Features or landmarks obtained from sensor measurements are registered and associated in SLAM if they are sustainable and correctly matched ones. To have such features it is essential to use a sensor system that provides precise range measurements. Typical sensors for range measurement include the laser range finder (LRF), ultra sonic, or vision sensor. A LRF provide very accurate and long range measurements while measurements from ultrasonic sensors are shorter and coarse due to cone shaped beam patterns. In this paper, a structured light system which combines laser and vision is considered for use with structural displacement measurement. We propose multiple dual laser-vision system to be used as a structural displacement measurement. The proposed dual laser-vision system is quite similar to the structured light system but it solves accurate relative pose instead of range map of the environment. Some preliminary simulation was done to show the effectiveness of the system for the long span structural displacement monitoring. Another good application example such as LBS (Location-Based Service) for U-City (ubiquitous city) will be introduced in the later part. Without initially specifying the locations of ubiquitous sensors, SLAM technique simultaneously estimates the locations of beacons, robot, and even humans. 1. INTRODUCTION As civil infrastructures are exposed to various external loads such as traffic, earthquakes, gusts, and possible wave loads, Structural health monitoring has become an important research topic for continuous assessment and evaluation of structural safety and integrity. Conventional approaches use accelerometer, strain gauge, PZT sensors, etc. The displacement measurement, though important, was not popular due to inaccessibility and the huge size of the civil infrastructures. One of the most frequently employed approaches uses Linear Variable Differential Transformer (LVDT), a contact-type sensor for displacement measurement that needs stable reference points underneath, which is not so practical to be used. Thus the development of structural health monitoring system that measures the displacement of the structure using vision or laser-based mobile robot system can be very helpful. In solving this, we can use the modified structured light system which is widely used in robot navigation. The problem of Simultaneous Localization And Map-building, known as SLAM, has gained great attention in the mobile robotics research area. SLAM focuses on the problem of self-localization while building a map of an unknown environment from a sequence of noisy sensor measurements obtained from a moving robot. SLAM is considered to be a core capability for autonomous mobile robots operating in unknown environments. The dominant approach to the SLAM problem 2552

2 was introduced as a form of Extended Kalman Filter (EKF) Smith et al., 1990, Dissanayake et al., In the last decade, this approach has found widespread acceptance in field robotics. An alternative approach to the SLAM problem has been introduced that factors the SLAM posterior into a localization problem and K independent landmark estimation problems conditioned on the robot pose estimate. This algorithm, called Fast- SLAM Montemerlo et al., 2002, uses a modified particle filter for estimating posterior over the robot path. Each particle possesses K independent Kalman filters that estimate the landmark locations conditioned on the particle s path. FastSLAM has been demonstrated with up to 100,000 landmarks, problems far beyond the reach of the EKF. 2D laser range finders, sonar sensors and cameras are the most popular sensors in indoor SLAM implementations. They can be used separately (Gutmann and Schlegel, 1996) or in combination (Castellanos and Tardos, 1999). Time of flight laser sensors such as SICK laser provide very long and accurate range measurements. Although some error models of TOF type laser sensors were proposed the range measurements are accurate enough for segmentation without preprocessing. On the contrary sonar sensors have wide beam patterns hence inducing large uncertainty. In the SLAM system developed earlier, we used a structured light range sensor system that consists of a B/W camera and a line laser projector (Myung et al., 2004). The sensor provides medium range measurements (up to 4 meters). To use the structured light like system in a long distant observation, it is recommended to put the camera in the observation area not in the projecting area. By formulating this kind of configuration, we will introduce the dual laser-vision system for displacement measurement. In Section 2, the structured light range sensor is introduced for SLAM purpose. Section 3 describes the structural health monitoring system that estimates the 6-DOF displacement using laser and vision sensors. Some preliminary simulation results will be shown to demonstrate the applicability of the proposed system. In Section 4, the application of robot navigation technique to LBS (Location-Based Service) for ubiquitous city will be introduced. Without initially specifying the locations of beacons or ubiquitous sensors, SLAM technique can simultaneously estimate the locations of beacons and robot. Moreover the locations of humans can be estimated with the help of SLAM techniques. 2. STRUCTURED LIGHT SYSTEM FOR SLAM 2.1 Obtaining 2D range profiles In this section we discuss the characteristics of the SL sensor implemented in our SLAM system. As shown in Figure 1, several steps are involved in obtaining a 2D range profile from the laser pattern captured by camera. The first step is to correct the camera image that has been distorted by the lens. The OpenCV libraries of Zhang s method (Zhang, 2000) provide camera s intrinsic and extrinsic parameters as well as lens distortion model coefficients. Scale factors of the intrinsic parameters and all of the extrinsic parameters are used later in mapping the laser regions in the image to range profiles in 2D plane. The following equation is the relationship between real and distorted image coordinates in the model: ˆx = x + x[k 1 r 2 + k 2 r 4 ]+[2p 1 xy + p 2 (r 2 +2x 2 )], ŷ = y + y[k 1 r 2 + k 2 r 4 ]+[2p 1 xy + p 2 (r 2 +2y 2 )], where x and y are ideal, and ˆx and ŷ are real distorted image physical coordinates. The radial distance r is defined as r 2 = x 2 + y 2. Hence the distortion model is characterized by the four coefficients k 1, k 2, p 1 and p 2. The corrected image is then convolved with a kernel to determine the laser region. The kernel is designed in consideration of the thickness of the laser pattern in the image. After the convolution only those pixels whose convolution values are above a threshold remain as the candidate laser regions. Then the image is looked into column by column so that only one laser region is selected in each column. If there are more than one region in a column, the largest region is selected. Finally, in each column, the region is averaged in row values to be mapped in 2D plane. Figure 2 shows the relationship between the camera frame Σ XC Y C Z C and that of the laser projector Σ XP Y P Z P. Camera longitudinal displacement along Y p is annotated with P Y, and angular 2553

3 Image Corrected Image Correction of lens distortion Image processing (detection of laser region) laser region Triangulation for 2D mapping 2D range map Figure 1. Process to obtain 2D range profiles Figure 2. Configuration of camera and laser projector displacement along X p is with α. The following equations are used to map the laser pattern in the image to the 2D plane. x p y p y c z p [ ] x c = = y p = cos α sin α 0 sin α cos α [ ] xs sx f z y s c sy f x c y c z c + 0 P Y 0 is the most well known to find corners from a 2D range scan (Tardos, 2002). One of the difficulties in using the algorithm is to find a right threshold with which the segmentation is done correctly. If the threshold is large, the process will terminate prematurely, and if it is small segmentation will occur even at small difference in the range profile. To find the correct threshold, the sensor characteristics must be considered. Figure 3 shows the nonlinear and discrete characteristics of structured light range sensor. The range resolution decreases with range nonlinearly as shown in the figure. Range resolution (mm) In the equation z p is the longitudinal distance to the laser pattern while x p is the lateral distance with respect to the origin of Σ XP Y P Z P. Laser pattern coordinates [x s y s ] T are the coordinates in the image frame (in pixels) subtracted by the image center. Intrinsic parameters such as image center coordinates and scale factors along image x ( sx ) f and y ( sy ) axes are obtained from the camera f calibration stage. 2.2 Corner Point Extraction from 2D Range Profiles For SLAM it is necessary to find features in a space with which the mapping and localization is possible. In our SLAM algorithm we use corner points as the features. The split and merge method Figure 3. Range resolution along range distance (mm) The threshold in our algorithm is set proportional to the range resolution at each range. Hence, in addition to the 2D range profiles, the range resolution at each range is calculated together in the range calculation of previous subsection, and the following triplets are used for corner point detection: [x p (k) z p (k) r(k)] T (1) where x p (k), z p (k) are the lateral and longitudinal 2554

4 distance to the laser pattern found at column k of the image, and r(k) is the range resolution at that point. (k is the index of image column and in our system k [0 639].) In addition to the location of corners, the start and end angles that make the corner are also calculated to ease the data association of the SLAM routine. If occlusion occurs and only one angle is available then the other will be annotated with a wildcard symbol. Hence each feature point consists of the following quadruplets. [x p z p θ 1 θ 2 ] T (2) 3. STRUCTURAL HEALTH MONITORING SYSTEM 3.1 Dual laser-vision system Recently structural health monitoring has become an important research topic for continuous assessment and evaluation of structural safety and integrity. The development of structural health monitoring system that measures displacement of the structure using mobile robot system can be very helpful. A possible candidate for the displacement measurement of the long structure is illustrated in Figure 4. Figure 5 shows the configuration of the proposed dual laser-vision system which corresponds one module in the overall system described in Figure 4. Laser sensor A composed of at least two point lasers projects its parallel beam to the distant screen B. The distance can be quite large such as 100 meters or so. In screen B, the image of two points are captured by camera B, and we can obtain the two image points. The laser sensor B also projects its two points to the distant screen A. We can also observe two distinct image points in screen A. It is also possible to place the camera at the front side of the screen instead of the backside. We observe overall four different image points in 2D screens and can obtain 8 scalar values. Using this observed values, the relative displacement, i.e., the translation x, y, z, and rotation angle θ, φ, ψ around x-axis, y-axis, and z-axis, are to be estimated. The kinematics defines the geometric relationship between the observed data M = [ A O, A Y, B O, B Y ] T and estimated pose P = [x, y, z, θ, φ, ψ] T. To derive the kinematics, the transformation matrices A T B and B T A can be used. As can be seen in equation (4), the transformation matrix consists of rotation about x, y, z-axis and translation in x, y, z-axis. A T B = T (x, y, z)r(x, θ)r(y, φ)r(z,ψ) = x y 0 cos θ sin θ z 0 sin θ cos θ cos φ 0 sin φ sin φ 0 cos φ cos ψ sin ψ 0 0 sin ψ cos ψ Using the transformation matrix A T B, we can obtain the kinematic equation after some mathematical manipulations. In each screen, we assume the laser sensor is installed a distance of Y S away in y-axis direction. If we consider the laser projected from B to screen A, the screen coordinate A O can be obtained from the following equation: A O = A T B 0 Y S Z AB 1 (4) As z =0on the screen A in coordinate frame A, we let z component of A O vector to zero to obtain Z AB. By substituting the obtained Z AB to equation (4), we can obtain the actual screen coordinate A O of the projected laser beam from B. Inthe same way, we can easily obtain A Y, B O, and B Y. This yields the following kinematic equation (5): By using the steepest descent method, the estimation of P can be obtained as follows: ˆP (k +1)= ˆP (k)+j + p (M(k) ˆM(k)) (6) where J p = M is the Jacobian of the kinematic p equation and J p + is the pseudo-inverse of the Jacobian. 3.2 Simulation results This section describes the simulation result. In the simulation, we set the true values for (x, y, z, θ, φ, ψ) = (0.1, 1.02, 100, 0.001, 0.002, 0.003). Using the distance of z = 100 meters, the estimated value converged (3). 2555

5 Figure 4. Displacement monitoring using multiple dual laser-vision system Figure 5. One unit of dual laser-vision system M = ( cos θ sin ψ sin θ sin φ cos ψ z sin φ + x cos θ cos φ)/(cos θ cos φ) (cos ψ + z sin θ + y cos θ)/ cos θ 1/10( 9 cos θ sin ψ 9 sin θ sin φ cos ψ 10z sin φ +10x cos θ cos φ)/(cos θ cos φ) 1/10(9 cos ψ +10z sin θ +10y cos θ)/ cos θ (x cos θ cos ψ + cos φ sin ψ x sin θ sin ψ sin φ + y sin ψ cos φ)/(cos θ cos φ) (x cos θ sin ψ cos ψ cos φ + x sin θ cos ψ sin φ y cos ψ cos φ)/(cos θ cos φ) 1/10(10x cos θ cos ψ + 9 cos φ sin ψ 10x sin θ sin ψ sin φ +10y sin ψ cos φ)/(cos θ cos φ) 1/10(10x cos θ sin ψ 9 cos ψ cos φ +10x sin θ cos ψ sin φ 10y cos ψ cos φ)/(cos θ/ cos φ) (5) to true values very rapidly. As it can be seen in Figure 6, it converged only after 10 iterations. When there exists a random unform measurement noise, the estimation also quickly converged to the true value as can be seen in Figure 7. We also tested the effect of dynamic disturbance in x-axis direction by applying x = 0.01 sin(0.1t) with the same amount of random uniform measurement 2556

6 Figure 7. Simulation result for random uniform measurement noise of [ 0.001, 0.001]. Figure 6. Simulation result for (x, y, z, θ, φ, ψ) = (0.1, 1.02, 100, 0.001, 0.002, 0.003) noise as in the previous simulation. As can be seen in Figure 8, x direction estimation quickly tracks the dynamic displacement. 4. LBS USING SLAM 4.1 Conventional beacon-based LBS The conventional beacon-based localization techniques usually incorporates trilateration like the same principle used by GPS (Global Positioning System). The most common methods are Timeof-Arrival (ToA) and Time-Difference-of-Arrival (TDoA). The example system is shown in Figure 9. To perform those kind of method, the locations of each beacon should be known prior to localization of a tag. But it would be very hard to determine the beacon location if the number of beacons grows rapidly for use in large scale. The accuracy of localization in this case would be determined by the accuracy of beacon locations. The SLAM (Simultaneous Localization and Mapping) technique used in robot navigation can be used to solve this problem. The autonomous mapping capability of SLAM can estimate the beacon locations without the need to initially specifying the exact beacon locations. The structure of SLAM will be discussed in detail. 4.2 SLAM-based LBS Since only the range data are used in SLAM for LBS, researchers sometimes call it a range-only SLAM technique. Some implementations of rangeonly SLAM can be found in Djugash et al., 2005 and Kurth, The structure of particle filter based SLAM is depicted in Fig.10 (Myung et al., 2004). The basic SLAM structure is the same as that of FastSLAM (Montemerlo et al., 2002) which is also known as a Rao-Blackwellized particle filter. The reader is recommended to refer to the original FastSLAM paper for detailed explanation (Montemerlo, 2003). Here a per-particle maximum likelihood data association combined with Monte Carlo data association is used. An observation (range data from a beacon) is associated with a beacon of the same ID. It should be noted that the term landmark and beacon location are interchangeable throughout the paper since the beacon based SLAM is considered. Initially, the beacon location is initialized with the aid of dead reckoning of the robot. The robot installed with a tag, moves and receives the range data from each beacon. And as the distance traveled can be calculated using the odometry of robot wheels, the beacon location can be initialized using trilateration method. The importance weight of each particle defined by the equation (7) is related to the Mahalanobis 2557

7 Figure 8. Simulation result for dynamic disturbance of x = 0.01 sin(0.1t) with random uniform measurement noise of [ 0.001, 0.001]. Figure 10. SLAM structure Figure 9. Localization of robot using trilateration. Image taken from DustBot Project(EU). with the same ID previously known. Finally, the resampling stage of the FastSLAM is performed whenever the effective number of particles N eff defined by equation (8) is below some threshold such as half of the number of particles. The importance weight is used as selection probability of each particle. N eff = 1 Ni=1 w 2 i N 2 (8) Once the exact location of each beacon is obtained by using SLAM, the particle filter can be used to estimate the location of each human with a unique tag. It should be noted that in simple environment that needs only several number of beacons, EKFbased SLAM can also be used instead of Fast- SLAM. distance between estimated and observed ranges. 5. CONCLUSIONS 1 w t = 2π Zn,t exp{ 1 2 (z t ẑ n,t ) T [Z n,t ] 1 (z t ẑ n,t )} When there exists a measurement noise, the use of extended Kalman filter (EKF) can be more (7) effective. In this case, as the estimated instantaneous translation and rotation are assumed to be where z t is the measured range, ẑ n,t is the estimated range, and Z n,t is the covariance of range stationary, the system function F is an identity data. Since each beacon can transmit it s unique matrix. And the observation matrix is the same as ID, landmark management routine is much simpler the one derived earlier. We don t need to solve the than that of the landmark initialization algorithm inverse of Jacobian in this EKF scheme. Instead, suggested in Dissanayake et al., The range Kalman gain plays the role of inverse of Jacobian. data can simply be associated with the beacon The advantage of EKF is that we can predict 2558

8 the error covariance and it can quickly track the dynamic behavior. We believe that a lot of robot technology can have wide range of application in civil engineering as well as in home environment. For example, these robot techniques can be used to the monitoring of civil structures such as concrete crack width measurement using structured light range sensor as well as the displacement measurement using laser and vision sensors. Another good application example such as LBS (Location-Based Service) for ubiquitous city was also discussed. Without initially specifying the locations of beacons, SLAM technique simultaneously estimates the locations of beacons, robot, and even the humans. dissertation, tech. report CMU-RI-TR-03-28, Robotics Institute, Carnegie Mellon University. Myung, H., Jung, M.-J., Hong, S.-G., Park, D., Lee, H.- K., and Bang, S Evolutionary simultaneous localization and map-building (SLAM) in home environments using structured light 2d range finder. In Proc. of Simulated Evolution And Learning (SEAL). Busan, Korea. on CD-ROM. Smith, R., Self, M., and Cheeseman, P Estimating uncertain spatial relationships in robotics. Autonomous Robot Vehnicles. Tardos, J. D Data association in SLAM. presentation slides at the KTH Summer School on SLAM. Zhang, Z A flexible new technique for camera calibration. IEEE Transactions on Pattern Analysis and Machine Intelligence, 22(11): ACKNOWLEDGMENT This research was supported by a grant (07High Tech A01) from High tech Urban Development Program funded by Ministry of Land, Transportation and Maritime Affairs of Korean government. REFERENCES Castellanos, J. A. and Tardos, J. D Mobile Robot Localization and Map Building: A Multisensor Fusion Approach. Kluwer Academic Publishers. Dissanayake, G., Newman, P., Clark, S., Durrant-Whyte, H.F., and M. Csorba A solution to the simultaneous localisation and map building (SLAM) problem IEEE Transactions of Robotics and Automation, 17(3): Djugash, J., Singh, S., and P.I. Corke, July, Further Results with Localization and Mapping using Range from Radio In Proc. of International Conference on Field and Service Robotics (FSR 05). Gutmann, J.-S. and Schlegel, C Amos: Comparison of scan matching approaches for self-localization in indoor environments. In Proceedings of the 1st Euromicro Workshop on Advanced Mobile Robots. IEEE Computer Society Press. Kurth,D., May Range-only robot localization and SLAM with radio Masters thesis, Robotics Institute, Carnegie Mellon University, Pittsburgh, PA, tech. report CMU-RI-TR Montemerlo, M., Thrun, S., Koller, D., Wegbreit, B FastSLAM: A Factored Solution to the Simultaneous Localization and Mapping Problem. In Proceedings of the Eighteenth National Conference on Artificial Intelligence, AAAI Press. Montemerlo, M. July FastSLAM: A Factored Solution to the Simultaneous Localization and Mapping Problem with Unknown Data Association Doctoral 2559

INCREMENTAL DISPLACEMENT ESTIMATION METHOD FOR VISUALLY SERVOED PARIED STRUCTURED LIGHT SYSTEM (ViSP)

INCREMENTAL DISPLACEMENT ESTIMATION METHOD FOR VISUALLY SERVOED PARIED STRUCTURED LIGHT SYSTEM (ViSP) Blucher Mechanical Engineering Proceedings May 2014, vol. 1, num. 1 www.proceedings.blucher.com.br/evento/10wccm INCREMENAL DISPLACEMEN ESIMAION MEHOD FOR VISUALLY SERVOED PARIED SRUCURED LIGH SYSEM (ViSP)

More information

Geometrical Feature Extraction Using 2D Range Scanner

Geometrical Feature Extraction Using 2D Range Scanner Geometrical Feature Extraction Using 2D Range Scanner Sen Zhang Lihua Xie Martin Adams Fan Tang BLK S2, School of Electrical and Electronic Engineering Nanyang Technological University, Singapore 639798

More information

High Accuracy Navigation Using Laser Range Sensors in Outdoor Applications

High Accuracy Navigation Using Laser Range Sensors in Outdoor Applications Proceedings of the 2000 IEEE International Conference on Robotics & Automation San Francisco, CA April 2000 High Accuracy Navigation Using Laser Range Sensors in Outdoor Applications Jose Guivant, Eduardo

More information

Autonomous Mobile Robot Design

Autonomous Mobile Robot Design Autonomous Mobile Robot Design Topic: EKF-based SLAM Dr. Kostas Alexis (CSE) These slides have partially relied on the course of C. Stachniss, Robot Mapping - WS 2013/14 Autonomous Robot Challenges Where

More information

Probabilistic Robotics

Probabilistic Robotics Probabilistic Robotics FastSLAM Sebastian Thrun (abridged and adapted by Rodrigo Ventura in Oct-2008) The SLAM Problem SLAM stands for simultaneous localization and mapping The task of building a map while

More information

PROGRAMA DE CURSO. Robotics, Sensing and Autonomous Systems. SCT Auxiliar. Personal

PROGRAMA DE CURSO. Robotics, Sensing and Autonomous Systems. SCT Auxiliar. Personal PROGRAMA DE CURSO Código Nombre EL7031 Robotics, Sensing and Autonomous Systems Nombre en Inglés Robotics, Sensing and Autonomous Systems es Horas de Horas Docencia Horas de Trabajo SCT Docentes Cátedra

More information

DEALING WITH SENSOR ERRORS IN SCAN MATCHING FOR SIMULTANEOUS LOCALIZATION AND MAPPING

DEALING WITH SENSOR ERRORS IN SCAN MATCHING FOR SIMULTANEOUS LOCALIZATION AND MAPPING Inženýrská MECHANIKA, roč. 15, 2008, č. 5, s. 337 344 337 DEALING WITH SENSOR ERRORS IN SCAN MATCHING FOR SIMULTANEOUS LOCALIZATION AND MAPPING Jiří Krejsa, Stanislav Věchet* The paper presents Potential-Based

More information

Mobile Robotics. Mathematics, Models, and Methods. HI Cambridge. Alonzo Kelly. Carnegie Mellon University UNIVERSITY PRESS

Mobile Robotics. Mathematics, Models, and Methods. HI Cambridge. Alonzo Kelly. Carnegie Mellon University UNIVERSITY PRESS Mobile Robotics Mathematics, Models, and Methods Alonzo Kelly Carnegie Mellon University HI Cambridge UNIVERSITY PRESS Contents Preface page xiii 1 Introduction 1 1.1 Applications of Mobile Robots 2 1.2

More information

Mobile robot localisation and navigation using multi-sensor fusion via interval analysis and UKF

Mobile robot localisation and navigation using multi-sensor fusion via interval analysis and UKF Mobile robot localisation and navigation using multi-sensor fusion via interval analysis and UKF Immanuel Ashokaraj, Antonios Tsourdos, Peter Silson and Brian White. Department of Aerospace, Power and

More information

A Single-Step Maximum A Posteriori Update for Bearing-Only SLAM

A Single-Step Maximum A Posteriori Update for Bearing-Only SLAM A Single-Step Maximum A Posteriori Update for Bearing-Only SLAM Stephen ully Department of Electrical and Computer Engineering Carnegie Mellon University, Pittsburgh, PA stully@ece.cmu.edu George Kantor

More information

Particle Filters. CSE-571 Probabilistic Robotics. Dependencies. Particle Filter Algorithm. Fast-SLAM Mapping

Particle Filters. CSE-571 Probabilistic Robotics. Dependencies. Particle Filter Algorithm. Fast-SLAM Mapping CSE-571 Probabilistic Robotics Fast-SLAM Mapping Particle Filters Represent belief by random samples Estimation of non-gaussian, nonlinear processes Sampling Importance Resampling (SIR) principle Draw

More information

EE565:Mobile Robotics Lecture 3

EE565:Mobile Robotics Lecture 3 EE565:Mobile Robotics Lecture 3 Welcome Dr. Ahmad Kamal Nasir Today s Objectives Motion Models Velocity based model (Dead-Reckoning) Odometry based model (Wheel Encoders) Sensor Models Beam model of range

More information

Localization and Map Building

Localization and Map Building Localization and Map Building Noise and aliasing; odometric position estimation To localize or not to localize Belief representation Map representation Probabilistic map-based localization Other examples

More information

IN recent years, NASA s Mars Exploration Rovers (MERs)

IN recent years, NASA s Mars Exploration Rovers (MERs) Robust Landmark Estimation for SLAM in Dynamic Outdoor Environment Atsushi SAKAI, Teppei SAITOH and Yoji KURODA Meiji University, Department of Mechanical Engineering, 1-1-1 Higashimita, Tama-ku, Kawasaki,

More information

Appearance-based Concurrent Map Building and Localization

Appearance-based Concurrent Map Building and Localization Appearance-based Concurrent Map Building and Localization Josep M. Porta and Ben J.A. Kröse IAS Group, Informatics Institute, University of Amsterdam Kruislaan 403, 1098SJ, Amsterdam, The Netherlands {porta,krose}@science.uva.nl

More information

IROS 05 Tutorial. MCL: Global Localization (Sonar) Monte-Carlo Localization. Particle Filters. Rao-Blackwellized Particle Filters and Loop Closing

IROS 05 Tutorial. MCL: Global Localization (Sonar) Monte-Carlo Localization. Particle Filters. Rao-Blackwellized Particle Filters and Loop Closing IROS 05 Tutorial SLAM - Getting it Working in Real World Applications Rao-Blackwellized Particle Filters and Loop Closing Cyrill Stachniss and Wolfram Burgard University of Freiburg, Dept. of Computer

More information

Implementation of Odometry with EKF for Localization of Hector SLAM Method

Implementation of Odometry with EKF for Localization of Hector SLAM Method Implementation of Odometry with EKF for Localization of Hector SLAM Method Kao-Shing Hwang 1 Wei-Cheng Jiang 2 Zuo-Syuan Wang 3 Department of Electrical Engineering, National Sun Yat-sen University, Kaohsiung,

More information

EKF Localization and EKF SLAM incorporating prior information

EKF Localization and EKF SLAM incorporating prior information EKF Localization and EKF SLAM incorporating prior information Final Report ME- Samuel Castaneda ID: 113155 1. Abstract In the context of mobile robotics, before any motion planning or navigation algorithm

More information

Visual Bearing-Only Simultaneous Localization and Mapping with Improved Feature Matching

Visual Bearing-Only Simultaneous Localization and Mapping with Improved Feature Matching Visual Bearing-Only Simultaneous Localization and Mapping with Improved Feature Matching Hauke Strasdat, Cyrill Stachniss, Maren Bennewitz, and Wolfram Burgard Computer Science Institute, University of

More information

SLAM: Robotic Simultaneous Location and Mapping

SLAM: Robotic Simultaneous Location and Mapping SLAM: Robotic Simultaneous Location and Mapping William Regli Department of Computer Science (and Departments of ECE and MEM) Drexel University Acknowledgments to Sebastian Thrun & others SLAM Lecture

More information

A Stochastic Environment Modeling Method for Mobile Robot by using 2-D Laser scanner Young D. Kwon,Jin.S Lee Department of Electrical Engineering, Poh

A Stochastic Environment Modeling Method for Mobile Robot by using 2-D Laser scanner Young D. Kwon,Jin.S Lee Department of Electrical Engineering, Poh A Stochastic Environment Modeling Method for Mobile Robot by using -D Laser scanner Young D. Kwon,Jin.S Lee Department of Electrical Engineering, Pohang University of Science and Technology, e-mail: jsoo@vision.postech.ac.kr

More information

Robotic Perception and Action: Vehicle SLAM Assignment

Robotic Perception and Action: Vehicle SLAM Assignment Robotic Perception and Action: Vehicle SLAM Assignment Mariolino De Cecco Mariolino De Cecco, Mattia Tavernini 1 CONTENTS Vehicle SLAM Assignment Contents Assignment Scenario 3 Odometry Localization...........................................

More information

Sensor Modalities. Sensor modality: Different modalities:

Sensor Modalities. Sensor modality: Different modalities: Sensor Modalities Sensor modality: Sensors which measure same form of energy and process it in similar ways Modality refers to the raw input used by the sensors Different modalities: Sound Pressure Temperature

More information

Exam in DD2426 Robotics and Autonomous Systems

Exam in DD2426 Robotics and Autonomous Systems Exam in DD2426 Robotics and Autonomous Systems Lecturer: Patric Jensfelt KTH, March 16, 2010, 9-12 No aids are allowed on the exam, i.e. no notes, no books, no calculators, etc. You need a minimum of 20

More information

Particle Filter in Brief. Robot Mapping. FastSLAM Feature-based SLAM with Particle Filters. Particle Representation. Particle Filter Algorithm

Particle Filter in Brief. Robot Mapping. FastSLAM Feature-based SLAM with Particle Filters. Particle Representation. Particle Filter Algorithm Robot Mapping FastSLAM Feature-based SLAM with Particle Filters Cyrill Stachniss Particle Filter in Brief! Non-parametric, recursive Bayes filter! Posterior is represented by a set of weighted samples!

More information

Long-term motion estimation from images

Long-term motion estimation from images Long-term motion estimation from images Dennis Strelow 1 and Sanjiv Singh 2 1 Google, Mountain View, CA, strelow@google.com 2 Carnegie Mellon University, Pittsburgh, PA, ssingh@cmu.edu Summary. Cameras

More information

Probabilistic Robotics. FastSLAM

Probabilistic Robotics. FastSLAM Probabilistic Robotics FastSLAM The SLAM Problem SLAM stands for simultaneous localization and mapping The task of building a map while estimating the pose of the robot relative to this map Why is SLAM

More information

A FastSLAM Algorithm for Omnivision

A FastSLAM Algorithm for Omnivision A FastSLAM Algorithm for Omnivision Cristina Gamallo, Manuel Mucientes and Carlos V. Regueiro Abstract Omnidirectional cameras have a wide field of view, which makes them specially suitable for Simultaneous

More information

NAVIGATION SYSTEM OF AN OUTDOOR SERVICE ROBOT WITH HYBRID LOCOMOTION SYSTEM

NAVIGATION SYSTEM OF AN OUTDOOR SERVICE ROBOT WITH HYBRID LOCOMOTION SYSTEM NAVIGATION SYSTEM OF AN OUTDOOR SERVICE ROBOT WITH HYBRID LOCOMOTION SYSTEM Jorma Selkäinaho, Aarne Halme and Janne Paanajärvi Automation Technology Laboratory, Helsinki University of Technology, Espoo,

More information

D-SLAM: Decoupled Localization and Mapping for Autonomous Robots

D-SLAM: Decoupled Localization and Mapping for Autonomous Robots D-SLAM: Decoupled Localization and Mapping for Autonomous Robots Zhan Wang, Shoudong Huang, and Gamini Dissanayake ARC Centre of Excellence for Autonomous Systems (CAS), Faculty of Engineering, University

More information

This chapter explains two techniques which are frequently used throughout

This chapter explains two techniques which are frequently used throughout Chapter 2 Basic Techniques This chapter explains two techniques which are frequently used throughout this thesis. First, we will introduce the concept of particle filters. A particle filter is a recursive

More information

D-SLAM: Decoupled Localization and Mapping for Autonomous Robots

D-SLAM: Decoupled Localization and Mapping for Autonomous Robots D-SLAM: Decoupled Localization and Mapping for Autonomous Robots Zhan Wang, Shoudong Huang, and Gamini Dissanayake ARC Centre of Excellence for Autonomous Systems (CAS) Faculty of Engineering, University

More information

DYNAMIC ROBOT LOCALIZATION AND MAPPING USING UNCERTAINTY SETS. M. Di Marco A. Garulli A. Giannitrapani A. Vicino

DYNAMIC ROBOT LOCALIZATION AND MAPPING USING UNCERTAINTY SETS. M. Di Marco A. Garulli A. Giannitrapani A. Vicino DYNAMIC ROBOT LOCALIZATION AND MAPPING USING UNCERTAINTY SETS M. Di Marco A. Garulli A. Giannitrapani A. Vicino Dipartimento di Ingegneria dell Informazione, Università di Siena, Via Roma, - 1 Siena, Italia

More information

DYNAMIC POSITIONING OF A MOBILE ROBOT USING A LASER-BASED GONIOMETER. Joaquim A. Batlle*, Josep Maria Font*, Josep Escoda**

DYNAMIC POSITIONING OF A MOBILE ROBOT USING A LASER-BASED GONIOMETER. Joaquim A. Batlle*, Josep Maria Font*, Josep Escoda** DYNAMIC POSITIONING OF A MOBILE ROBOT USING A LASER-BASED GONIOMETER Joaquim A. Batlle*, Josep Maria Font*, Josep Escoda** * Department of Mechanical Engineering Technical University of Catalonia (UPC)

More information

Localization, Where am I?

Localization, Where am I? 5.1 Localization, Where am I?? position Position Update (Estimation?) Encoder Prediction of Position (e.g. odometry) YES matched observations Map data base predicted position Matching Odometry, Dead Reckoning

More information

Efficient SLAM Scheme Based ICP Matching Algorithm Using Image and Laser Scan Information

Efficient SLAM Scheme Based ICP Matching Algorithm Using Image and Laser Scan Information Proceedings of the World Congress on Electrical Engineering and Computer Systems and Science (EECSS 2015) Barcelona, Spain July 13-14, 2015 Paper No. 335 Efficient SLAM Scheme Based ICP Matching Algorithm

More information

Localization algorithm using a virtual label for a mobile robot in indoor and outdoor environments

Localization algorithm using a virtual label for a mobile robot in indoor and outdoor environments Artif Life Robotics (2011) 16:361 365 ISAROB 2011 DOI 10.1007/s10015-011-0951-7 ORIGINAL ARTICLE Ki Ho Yu Min Cheol Lee Jung Hun Heo Youn Geun Moon Localization algorithm using a virtual label for a mobile

More information

Localization, Mapping and Exploration with Multiple Robots. Dr. Daisy Tang

Localization, Mapping and Exploration with Multiple Robots. Dr. Daisy Tang Localization, Mapping and Exploration with Multiple Robots Dr. Daisy Tang Two Presentations A real-time algorithm for mobile robot mapping with applications to multi-robot and 3D mapping, by Thrun, Burgard

More information

Revising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History

Revising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History Revising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History Simon Thompson and Satoshi Kagami Digital Human Research Center National Institute of Advanced

More information

Localization of Multiple Robots with Simple Sensors

Localization of Multiple Robots with Simple Sensors Proceedings of the IEEE International Conference on Mechatronics & Automation Niagara Falls, Canada July 2005 Localization of Multiple Robots with Simple Sensors Mike Peasgood and Christopher Clark Lab

More information

Monocular SLAM for a Small-Size Humanoid Robot

Monocular SLAM for a Small-Size Humanoid Robot Tamkang Journal of Science and Engineering, Vol. 14, No. 2, pp. 123 129 (2011) 123 Monocular SLAM for a Small-Size Humanoid Robot Yin-Tien Wang*, Duen-Yan Hung and Sheng-Hsien Cheng Department of Mechanical

More information

AN INCREMENTAL SLAM ALGORITHM FOR INDOOR AUTONOMOUS NAVIGATION

AN INCREMENTAL SLAM ALGORITHM FOR INDOOR AUTONOMOUS NAVIGATION 20th IMEKO TC4 International Symposium and 18th International Workshop on ADC Modelling and Testing Research on Electric and Electronic Measurement for the Economic Upturn Benevento, Italy, September 15-17,

More information

2005 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media,

2005 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, 25 IEEE Personal use of this material is permitted Permission from IEEE must be obtained for all other uses in any current or future media including reprinting/republishing this material for advertising

More information

Graph-based SLAM (Simultaneous Localization And Mapping) for Bridge Inspection Using UAV (Unmanned Aerial Vehicle)

Graph-based SLAM (Simultaneous Localization And Mapping) for Bridge Inspection Using UAV (Unmanned Aerial Vehicle) Graph-based SLAM (Simultaneous Localization And Mapping) for Bridge Inspection Using UAV (Unmanned Aerial Vehicle) Taekjun Oh 1), Sungwook Jung 2), Seungwon Song 3), and Hyun Myung 4) 1), 2), 3), 4) Urban

More information

Accurate Motion Estimation and High-Precision 3D Reconstruction by Sensor Fusion

Accurate Motion Estimation and High-Precision 3D Reconstruction by Sensor Fusion 007 IEEE International Conference on Robotics and Automation Roma, Italy, 0-4 April 007 FrE5. Accurate Motion Estimation and High-Precision D Reconstruction by Sensor Fusion Yunsu Bok, Youngbae Hwang,

More information

Robotics. Lecture 7: Simultaneous Localisation and Mapping (SLAM)

Robotics. Lecture 7: Simultaneous Localisation and Mapping (SLAM) Robotics Lecture 7: Simultaneous Localisation and Mapping (SLAM) See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information. Andrew Davison Department of Computing Imperial College

More information

Data Association for SLAM

Data Association for SLAM CALIFORNIA INSTITUTE OF TECHNOLOGY ME/CS 132a, Winter 2011 Lab #2 Due: Mar 10th, 2011 Part I Data Association for SLAM 1 Introduction For this part, you will experiment with a simulation of an EKF SLAM

More information

Laser sensors. Transmitter. Receiver. Basilio Bona ROBOTICA 03CFIOR

Laser sensors. Transmitter. Receiver. Basilio Bona ROBOTICA 03CFIOR Mobile & Service Robotics Sensors for Robotics 3 Laser sensors Rays are transmitted and received coaxially The target is illuminated by collimated rays The receiver measures the time of flight (back and

More information

Survey: Simultaneous Localisation and Mapping (SLAM) Ronja Güldenring Master Informatics Project Intellgient Robotics University of Hamburg

Survey: Simultaneous Localisation and Mapping (SLAM) Ronja Güldenring Master Informatics Project Intellgient Robotics University of Hamburg Survey: Simultaneous Localisation and Mapping (SLAM) Ronja Güldenring Master Informatics Project Intellgient Robotics University of Hamburg Introduction EKF-SLAM FastSLAM Loop Closure 01.06.17 Ronja Güldenring

More information

Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview

Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview Arun Das 05/09/2017 Arun Das Waterloo Autonomous Vehicles Lab Introduction What s in a name? Arun Das Waterloo Autonomous

More information

Quaternion-Based Tracking of Multiple Objects in Synchronized Videos

Quaternion-Based Tracking of Multiple Objects in Synchronized Videos Quaternion-Based Tracking of Multiple Objects in Synchronized Videos Quming Zhou 1, Jihun Park 2, and J.K. Aggarwal 1 1 Department of Electrical and Computer Engineering The University of Texas at Austin

More information

Simuntaneous Localisation and Mapping with a Single Camera. Abhishek Aneja and Zhichao Chen

Simuntaneous Localisation and Mapping with a Single Camera. Abhishek Aneja and Zhichao Chen Simuntaneous Localisation and Mapping with a Single Camera Abhishek Aneja and Zhichao Chen 3 December, Simuntaneous Localisation and Mapping with asinglecamera 1 Abstract Image reconstruction is common

More information

Real Time Data Association for FastSLAM

Real Time Data Association for FastSLAM Real Time Data Association for FastSLAM Juan Nieto, Jose Guivant, Eduardo Nebot Australian Centre for Field Robotics, The University of Sydney, Australia fj.nieto,jguivant,nebotg@acfr.usyd.edu.au Sebastian

More information

Simulation of a mobile robot with a LRF in a 2D environment and map building

Simulation of a mobile robot with a LRF in a 2D environment and map building Simulation of a mobile robot with a LRF in a 2D environment and map building Teslić L. 1, Klančar G. 2, and Škrjanc I. 3 1 Faculty of Electrical Engineering, University of Ljubljana, Tržaška 25, 1000 Ljubljana,

More information

arxiv: v1 [cs.ro] 18 Jul 2016

arxiv: v1 [cs.ro] 18 Jul 2016 GENERATIVE SIMULTANEOUS LOCALIZATION AND MAPPING (G-SLAM) NIKOS ZIKOS AND VASSILIOS PETRIDIS arxiv:1607.05217v1 [cs.ro] 18 Jul 2016 Abstract. Environment perception is a crucial ability for robot s interaction

More information

Hartley - Zisserman reading club. Part I: Hartley and Zisserman Appendix 6: Part II: Zhengyou Zhang: Presented by Daniel Fontijne

Hartley - Zisserman reading club. Part I: Hartley and Zisserman Appendix 6: Part II: Zhengyou Zhang: Presented by Daniel Fontijne Hartley - Zisserman reading club Part I: Hartley and Zisserman Appendix 6: Iterative estimation methods Part II: Zhengyou Zhang: A Flexible New Technique for Camera Calibration Presented by Daniel Fontijne

More information

Localization and Map Building

Localization and Map Building Localization and Map Building Noise and aliasing; odometric position estimation To localize or not to localize Belief representation Map representation Probabilistic map-based localization Other examples

More information

SLAM WITH KLT POINT FEATURES. P.O. Box 22, Thammasat-Rangsit Post Office, Pathumthani Thailand

SLAM WITH KLT POINT FEATURES. P.O. Box 22, Thammasat-Rangsit Post Office, Pathumthani Thailand SLAM WITH KLT POINT FEATURES Yoichi Nakaguro, 1 Matthew Dailey, Stanislav Makhanov 1 1 Sirindhorn International Institute of Technology, Thammasat University P.O. Box, Thammasat-Rangsit Post Office, Pathumthani

More information

ctbuh.org/papers Structural Member Monitoring of High-Rise Buildings Using a 2D Laser Scanner Title:

ctbuh.org/papers Structural Member Monitoring of High-Rise Buildings Using a 2D Laser Scanner Title: ctbuh.org/papers Title: Authors: Subject: Keyword: Structural Member Monitoring of High-Rise Buildings Using a 2D Laser Scanner Bub-Ryur Kim, Masters Degree Candidate, Yonsei University Dong Chel Lee,

More information

L15. POSE-GRAPH SLAM. NA568 Mobile Robotics: Methods & Algorithms

L15. POSE-GRAPH SLAM. NA568 Mobile Robotics: Methods & Algorithms L15. POSE-GRAPH SLAM NA568 Mobile Robotics: Methods & Algorithms Today s Topic Nonlinear Least Squares Pose-Graph SLAM Incremental Smoothing and Mapping Feature-Based SLAM Filtering Problem: Motion Prediction

More information

Robot Localization based on Geo-referenced Images and G raphic Methods

Robot Localization based on Geo-referenced Images and G raphic Methods Robot Localization based on Geo-referenced Images and G raphic Methods Sid Ahmed Berrabah Mechanical Department, Royal Military School, Belgium, sidahmed.berrabah@rma.ac.be Janusz Bedkowski, Łukasz Lubasiński,

More information

Zürich. Roland Siegwart Margarita Chli Martin Rufli Davide Scaramuzza. ETH Master Course: L Autonomous Mobile Robots Summary

Zürich. Roland Siegwart Margarita Chli Martin Rufli Davide Scaramuzza. ETH Master Course: L Autonomous Mobile Robots Summary Roland Siegwart Margarita Chli Martin Rufli Davide Scaramuzza ETH Master Course: 151-0854-00L Autonomous Mobile Robots Summary 2 Lecture Overview Mobile Robot Control Scheme knowledge, data base mission

More information

A Genetic Algorithm for Simultaneous Localization and Mapping

A Genetic Algorithm for Simultaneous Localization and Mapping A Genetic Algorithm for Simultaneous Localization and Mapping Tom Duckett Centre for Applied Autonomous Sensor Systems Dept. of Technology, Örebro University SE-70182 Örebro, Sweden http://www.aass.oru.se

More information

Path and Viewpoint Planning of Mobile Robots with Multiple Observation Strategies

Path and Viewpoint Planning of Mobile Robots with Multiple Observation Strategies Path and Viewpoint Planning of Mobile s with Multiple Observation Strategies Atsushi Yamashita Kazutoshi Fujita Toru Kaneko Department of Mechanical Engineering Shizuoka University 3 5 1 Johoku, Hamamatsu-shi,

More information

Simultaneous Localization and Mapping

Simultaneous Localization and Mapping Sebastian Lembcke SLAM 1 / 29 MIN Faculty Department of Informatics Simultaneous Localization and Mapping Visual Loop-Closure Detection University of Hamburg Faculty of Mathematics, Informatics and Natural

More information

The vslam Algorithm for Robust Localization and Mapping

The vslam Algorithm for Robust Localization and Mapping Published in Proc. of Int. Conf. on Robotics and Automation (ICRA) 2005 The vslam Algorithm for Robust Localization and Mapping Niklas Karlsson, Enrico Di Bernardo, Jim Ostrowski, Luis Goncalves, Paolo

More information

Gaussian Multi-Robot SLAM

Gaussian Multi-Robot SLAM Submitted to NIPS*2003 Gaussian Multi-Robot SLAM Yufeng Liu and Sebastian Thrun School of Computer Science Carnegie Mellon University Abstract We present an algorithm for the multi-robot simultaneous localization

More information

Robotics. Lecture 5: Monte Carlo Localisation. See course website for up to date information.

Robotics. Lecture 5: Monte Carlo Localisation. See course website  for up to date information. Robotics Lecture 5: Monte Carlo Localisation See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information. Andrew Davison Department of Computing Imperial College London Review:

More information

On-line Convex Optimization based Solution for Mapping in VSLAM

On-line Convex Optimization based Solution for Mapping in VSLAM On-line Convex Optimization based Solution for Mapping in VSLAM by Abdul Hafeez, Shivudu Bhuvanagiri, Madhava Krishna, C.V.Jawahar in IROS-2008 (Intelligent Robots and Systems) Report No: IIIT/TR/2008/178

More information

L10. PARTICLE FILTERING CONTINUED. NA568 Mobile Robotics: Methods & Algorithms

L10. PARTICLE FILTERING CONTINUED. NA568 Mobile Robotics: Methods & Algorithms L10. PARTICLE FILTERING CONTINUED NA568 Mobile Robotics: Methods & Algorithms Gaussian Filters The Kalman filter and its variants can only model (unimodal) Gaussian distributions Courtesy: K. Arras Motivation

More information

Camera Model and Calibration

Camera Model and Calibration Camera Model and Calibration Lecture-10 Camera Calibration Determine extrinsic and intrinsic parameters of camera Extrinsic 3D location and orientation of camera Intrinsic Focal length The size of the

More information

Efficient particle filter algorithm for ultrasonic sensor based 2D range-only SLAM application

Efficient particle filter algorithm for ultrasonic sensor based 2D range-only SLAM application Efficient particle filter algorithm for ultrasonic sensor based 2D range-only SLAM application Po Yang School of Computing, Science & Engineering, University of Salford, Greater Manchester, United Kingdom

More information

Introduction to Mobile Robotics SLAM Grid-based FastSLAM. Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Diego Tipaldi, Luciano Spinello

Introduction to Mobile Robotics SLAM Grid-based FastSLAM. Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Diego Tipaldi, Luciano Spinello Introduction to Mobile Robotics SLAM Grid-based FastSLAM Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Diego Tipaldi, Luciano Spinello 1 The SLAM Problem SLAM stands for simultaneous localization

More information

Improving Simultaneous Mapping and Localization in 3D Using Global Constraints

Improving Simultaneous Mapping and Localization in 3D Using Global Constraints Improving Simultaneous Mapping and Localization in 3D Using Global Constraints Rudolph Triebel and Wolfram Burgard Department of Computer Science, University of Freiburg George-Koehler-Allee 79, 79108

More information

Estimating Geospatial Trajectory of a Moving Camera

Estimating Geospatial Trajectory of a Moving Camera Estimating Geospatial Trajectory of a Moving Camera Asaad Hakeem 1, Roberto Vezzani 2, Mubarak Shah 1, Rita Cucchiara 2 1 School of Electrical Engineering and Computer Science, University of Central Florida,

More information

Appearance-Based Minimalistic Metric SLAM

Appearance-Based Minimalistic Metric SLAM Appearance-Based Minimalistic Metric SLAM Paul E. Rybski, Stergios I. Roumeliotis, Maria Gini, Nikolaos Papanikolopoulos Center for Distributed Robotics Department of Computer Science and Engineering University

More information

Practical Robotics (PRAC)

Practical Robotics (PRAC) Practical Robotics (PRAC) A Mobile Robot Navigation System (1) - Sensor and Kinematic Modelling Nick Pears University of York, Department of Computer Science December 17, 2014 nep (UoY CS) PRAC Practical

More information

Final project: 45% of the grade, 10% presentation, 35% write-up. Presentations: in lecture Dec 1 and schedule:

Final project: 45% of the grade, 10% presentation, 35% write-up. Presentations: in lecture Dec 1 and schedule: Announcements PS2: due Friday 23:59pm. Final project: 45% of the grade, 10% presentation, 35% write-up Presentations: in lecture Dec 1 and 3 --- schedule: CS 287: Advanced Robotics Fall 2009 Lecture 24:

More information

Probabilistic Robotics

Probabilistic Robotics Probabilistic Robotics Sebastian Thrun Wolfram Burgard Dieter Fox The MIT Press Cambridge, Massachusetts London, England Preface xvii Acknowledgments xix I Basics 1 1 Introduction 3 1.1 Uncertainty in

More information

Robotics. Lecture 8: Simultaneous Localisation and Mapping (SLAM)

Robotics. Lecture 8: Simultaneous Localisation and Mapping (SLAM) Robotics Lecture 8: Simultaneous Localisation and Mapping (SLAM) See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information. Andrew Davison Department of Computing Imperial College

More information

Announcements. Recap Landmark based SLAM. Types of SLAM-Problems. Occupancy Grid Maps. Grid-based SLAM. Page 1. CS 287: Advanced Robotics Fall 2009

Announcements. Recap Landmark based SLAM. Types of SLAM-Problems. Occupancy Grid Maps. Grid-based SLAM. Page 1. CS 287: Advanced Robotics Fall 2009 Announcements PS2: due Friday 23:59pm. Final project: 45% of the grade, 10% presentation, 35% write-up Presentations: in lecture Dec 1 and 3 --- schedule: CS 287: Advanced Robotics Fall 2009 Lecture 24:

More information

Matching Evaluation of 2D Laser Scan Points using Observed Probability in Unstable Measurement Environment

Matching Evaluation of 2D Laser Scan Points using Observed Probability in Unstable Measurement Environment Matching Evaluation of D Laser Scan Points using Observed Probability in Unstable Measurement Environment Taichi Yamada, and Akihisa Ohya Abstract In the real environment such as urban areas sidewalk,

More information

MAPPING ALGORITHM FOR AUTONOMOUS NAVIGATION OF LAWN MOWER USING SICK LASER

MAPPING ALGORITHM FOR AUTONOMOUS NAVIGATION OF LAWN MOWER USING SICK LASER MAPPING ALGORITHM FOR AUTONOMOUS NAVIGATION OF LAWN MOWER USING SICK LASER A thesis submitted in partial fulfillment of the requirements for the degree of Master of Science in Engineering By SHASHIDHAR

More information

(W: 12:05-1:50, 50-N202)

(W: 12:05-1:50, 50-N202) 2016 School of Information Technology and Electrical Engineering at the University of Queensland Schedule of Events Week Date Lecture (W: 12:05-1:50, 50-N202) 1 27-Jul Introduction 2 Representing Position

More information

Towards Gaussian Multi-Robot SLAM for Underwater Robotics

Towards Gaussian Multi-Robot SLAM for Underwater Robotics Towards Gaussian Multi-Robot SLAM for Underwater Robotics Dave Kroetsch davek@alumni.uwaterloo.ca Christoper Clark cclark@mecheng1.uwaterloo.ca Lab for Autonomous and Intelligent Robotics University of

More information

Environment Identification by Comparing Maps of Landmarks

Environment Identification by Comparing Maps of Landmarks Environment Identification by Comparing Maps of Landmarks Jens-Steffen Gutmann Masaki Fukuchi Kohtaro Sabe Digital Creatures Laboratory Sony Corporation -- Kitashinagawa, Shinagawa-ku Tokyo, 4- Japan Email:

More information

Introduction to Mobile Robotics SLAM Landmark-based FastSLAM

Introduction to Mobile Robotics SLAM Landmark-based FastSLAM Introduction to Mobile Robotics SLAM Landmark-based FastSLAM Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Diego Tipaldi, Luciano Spinello Partial slide courtesy of Mike Montemerlo 1 The SLAM Problem

More information

Machine learning based automatic extrinsic calibration of an onboard monocular camera for driving assistance applications on smart mobile devices

Machine learning based automatic extrinsic calibration of an onboard monocular camera for driving assistance applications on smart mobile devices Technical University of Cluj-Napoca Image Processing and Pattern Recognition Research Center www.cv.utcluj.ro Machine learning based automatic extrinsic calibration of an onboard monocular camera for driving

More information

Illumination-Robust Face Recognition based on Gabor Feature Face Intrinsic Identity PCA Model

Illumination-Robust Face Recognition based on Gabor Feature Face Intrinsic Identity PCA Model Illumination-Robust Face Recognition based on Gabor Feature Face Intrinsic Identity PCA Model TAE IN SEOL*, SUN-TAE CHUNG*, SUNHO KI**, SEONGWON CHO**, YUN-KWANG HONG*** *School of Electronic Engineering

More information

Computer Vision cmput 428/615

Computer Vision cmput 428/615 Computer Vision cmput 428/615 Basic 2D and 3D geometry and Camera models Martin Jagersand The equation of projection Intuitively: How do we develop a consistent mathematical framework for projection calculations?

More information

ME 597/747 Autonomous Mobile Robots. Mid Term Exam. Duration: 2 hour Total Marks: 100

ME 597/747 Autonomous Mobile Robots. Mid Term Exam. Duration: 2 hour Total Marks: 100 ME 597/747 Autonomous Mobile Robots Mid Term Exam Duration: 2 hour Total Marks: 100 Instructions: Read the exam carefully before starting. Equations are at the back, but they are NOT necessarily valid

More information

Unit 2: Locomotion Kinematics of Wheeled Robots: Part 3

Unit 2: Locomotion Kinematics of Wheeled Robots: Part 3 Unit 2: Locomotion Kinematics of Wheeled Robots: Part 3 Computer Science 4766/6778 Department of Computer Science Memorial University of Newfoundland January 28, 2014 COMP 4766/6778 (MUN) Kinematics of

More information

Improving Vision-Based Distance Measurements using Reference Objects

Improving Vision-Based Distance Measurements using Reference Objects Improving Vision-Based Distance Measurements using Reference Objects Matthias Jüngel, Heinrich Mellmann, and Michael Spranger Humboldt-Universität zu Berlin, Künstliche Intelligenz Unter den Linden 6,

More information

Structured Light. Tobias Nöll Thanks to Marc Pollefeys, David Nister and David Lowe

Structured Light. Tobias Nöll Thanks to Marc Pollefeys, David Nister and David Lowe Structured Light Tobias Nöll tobias.noell@dfki.de Thanks to Marc Pollefeys, David Nister and David Lowe Introduction Previous lecture: Dense reconstruction Dense matching of non-feature pixels Patch-based

More information

Modeling, Parameter Estimation, and Navigation of Indoor Quadrotor Robots

Modeling, Parameter Estimation, and Navigation of Indoor Quadrotor Robots Brigham Young University BYU ScholarsArchive All Theses and Dissertations 2013-04-29 Modeling, Parameter Estimation, and Navigation of Indoor Quadrotor Robots Stephen C. Quebe Brigham Young University

More information

3D-2D Laser Range Finder calibration using a conic based geometry shape

3D-2D Laser Range Finder calibration using a conic based geometry shape 3D-2D Laser Range Finder calibration using a conic based geometry shape Miguel Almeida 1, Paulo Dias 1, Miguel Oliveira 2, Vítor Santos 2 1 Dept. of Electronics, Telecom. and Informatics, IEETA, University

More information

Navigation methods and systems

Navigation methods and systems Navigation methods and systems Navigare necesse est Content: Navigation of mobile robots a short overview Maps Motion Planning SLAM (Simultaneous Localization and Mapping) Navigation of mobile robots a

More information

Collaborative Multi-Vehicle Localization and Mapping in Marine Environments

Collaborative Multi-Vehicle Localization and Mapping in Marine Environments Collaborative Multi-Vehicle Localization and Mapping in Marine Environments Moratuwage M.D.P., Wijesoma W.S., Kalyan B., Dong J.F., Namal Senarathne P.G.C., Franz S. Hover, Nicholas M. Patrikalakis School

More information

What is the SLAM problem?

What is the SLAM problem? SLAM Tutorial Slides by Marios Xanthidis, C. Stachniss, P. Allen, C. Fermuller Paul Furgale, Margarita Chli, Marco Hutter, Martin Rufli, Davide Scaramuzza, Roland Siegwart What is the SLAM problem? The

More information

Introduction to Mobile Robotics. SLAM: Simultaneous Localization and Mapping

Introduction to Mobile Robotics. SLAM: Simultaneous Localization and Mapping Introduction to Mobile Robotics SLAM: Simultaneous Localization and Mapping The SLAM Problem SLAM is the process by which a robot builds a map of the environment and, at the same time, uses this map to

More information