Kalman Filter Based. Localization

Size: px
Start display at page:

Download "Kalman Filter Based. Localization"

Transcription

1 Autonomous Mobile Robots Localization "Position" Global Map Cognition Environment Model Local Map Path Perception Real World Environment Motion Control Kalman Filter Based Localization & SLAM Zürich Autonomous Systems Lab

2 2 Kalman Filter Localization

3 3 Introduction to Kalman Filter () Two measurements Weighted least-square Finding minimum error After some calculation and rearrangements

4 4 Introduction to Kalman Filter (2) In Kalman Filter notation

5 5 Introduction to Kalman Filter (3) Dynamic Prediction (robot moving) st hour, Motion u = velocity w = noise Combining fusion and dynamic prediction

6 6 The Five Steps for Map-Based Localization Encoder Prediction Measurement and Position position estimate Estimation (odometry) (fusion) Map data base ures ns dicted feat bservation Pred o YES Matching matched predictions and observations raw sensor data or. Prediction based on previous estimate and odometry extracted features 2. Observation with on-board sensors 3. Measurement prediction based on prediction and map Observation on-board sensors 4. Matching of observation and map 5. Estimation -> position update (posteriori position) perce eption

7 5 7 Kalman Filter for Mobile Robot Localization: Robot Position Prediction In a first step, the robots position at time step + is predicted based on its old location (time step ) and its movement due to the control input u(): p ˆ( ) = f ( pˆ( ), u( )) + f: Odometry function T Σ p ( + ) = p f Σ p ( ) p f + u f Σ u ( ) u f T

8 5 8 Kalman Filter for Mobile Robot Localization: Robot Position Prediction: Example s xˆ( ) + s pˆ ( + ) = pˆ( ) + u( ) = yˆ( ) θ ˆ( ) p p p p T + sl sr sl cos( θ+ 2 2b + s l sr s l sin( θ+ 2 2b sr sl b r ) r ) Σ ( + ) = f Σ ( ) f + f Σ ( ) u u u f T Odometry Σ u ( ) = r s 0 r l 0 s l

9 9 Kalman Filter for Mobile Robot Localization: Observation The second step it to obtain the observation Z(+) ( )(measurements) from the robot s sensors at the new location at time + The observation usually consists of a set n 0 of single observations z j (+) extracted from the different sensors signals. It can represent raw data scans as well as features lie lines, doors or any ind of landmars. The parameters of the targets are usually observed in the sensor frame {S}. Therefore the observations have to be transformed to the world frame {W} or the measurement prediction have to be transformed to the sensor frame {S}. This transformation is specified in the function h i (see later).

10 0 Kalman Filter for Mobile Robot Localization: Observation: Example Raw Date of Laser Scanner Extracted Lines Extracted Lines in Model Space line j αj r j j R Sensor j ( + = (robot) frame z ) α rj σαα σα r ΣR, j( + ) = σrα σrr j

11 5 Kalman Filter for Mobile Robot Localization: Measurement Prediction ( + ) pˆ = In the next step we use the predicted robot position and the map M() to generate multiple predicted observations z t. They have to be transformed into the sensor frame ẑ ( + ) = h ( z,pˆ p ( + ) i i t We can now define the measurement prediction as the set containing all n i predicted observations Ẑ ( + ) = ẑ ( + )( i ) { } i n i The function h i is mainly the coordinate transformation between the world frame and the sensor frame

12 5 2 Kalman Filter for Mobile Robot Localization: Measurement Prediction: Example For prediction, only the walls that are in the field of view of the robot are selected.

13 3 Kalman Filter for Mobile Robot Localization: Matching: Example

14 5 4 Kalman Filter for Mobile Robot Localization: Estimation: Applying the Kalman Filter Kalman filter gain: Update of robot s position estimate: The associate variance

15 5 Kalman Filter for Mobile Robot Localization: Estimation: D Case For the one-dimensional case with we can show that the estimation corresponds to the Kalman filter for one-dimension presented earlier.

16 6 Kalman Filter for Mobile Robot Localization: Estimation: Example Kalman filter estimation of the new robot position : pˆ ( ) ) By fusing the prediction of robot position (magenta) with the innovation gained by the measurements (green) we get the updated estimate of the robot position (red)

17 7 Autonomous Indoor Navigation (Pygmalion EPFL) very robust on-the-fly localization one of the first systems with probabilistic sensor fusion 47 steps,78 meter length, realistic office environment, conducted 6 times > m overall distance partially difficult surfaces (laser), partially few vertical edges (vision)

18 8 Videos videos (Pygmalion and Recsys)

19 9 Summary of the EKF Localization EKF = Extended Kalman Filter The ingredients of the EKF Localization problem are: Prediction of the next robot position X based on the previous position X- and the odometry input u- X f X, u ) [ X = N X, P ) ] = ( ( Statistical error model of the odometry is Gaussian with: Statistical observation model is Gaussian: The map is given! [ u = N( u, Q )] [ z = N h( x ), R )] ( EKF Equations: PREDICTION X = f ( X, u ) P = F x PF T x ESTIMATION T T X = X + P H [ H P H + R ][ z h ( X ) ] + F QF u T u P + + = P P H T [ ] T T H P H + R H P

20 5 20 Summary of the EKF Localization: Prediction Phase What is the state X? X = (X,Y,θ) What s the odometry input u and its covariance matrix Q? = L R s s Q α α 0 0 What s the odometry input u and its covariance matrix Q? u = ), ( L R S S α s 0 What is f? ), ( = u X f X s s s s b s s s s Y X Y X L R L R L R L R sin 2 cos 2 θ θ + + = b s s b Y Y L R 2 sin 2 θ θ θ

21 5 2 Summary of the EKF Localization: Observation Phase Let s assume that the map is made of points with world coordinates m i = ( xi, yi ) Let s assume that the robot uses a laser range finder which measures the range r and the bearing θ of each point in the robot frame. i i r The observation z is then z = i ϕ m 2 = ( x2, y2) ) And the predicted observation z = h(x ) m = x, ) m i = ( xi, yi ) ) i ) i r z = = h ( X ) i ϕ = 2 2 ( x X ) + ( y Y ) i y arctan xi i i Y X ( y ϕ ϕ i θ

22 22 Summary of the EKF Localization: Innovation The observation (magenta): z ) z = h(x) The predicted observation (green) = what the robot would observe if its position was the one estimated in the prediction phase ) z z The difference is called Innovation

23 23 Summary of the EKF Localization: Estimation During he Estimation phase the robot position and uncertainty are updated by fusing the information of the prediction phase with that from the observation (red)

24 24 Autonomous Map Building (SLAM) Starting from an arbitrary initial point, a mobile robot should be able to autonomously explore the environment with its on board sensors, gain nowledge about it, interpret the scene, build an appropriate map and localize itself relative to this map. SLAM The Simultaneous Localization and Mapping Problem

25 25 References for interested people Simultaneous Localization and Mapping: Part I by HUGH DURRANT- WHYTE and TIM BAILEY, Robotics and Automation Magazine, June, 2006 Simultaneous Localization and Mapping: Part II by TIM BAILEY and HUGH DURRANT-WHYTE, Robotics and Automation Magazine, September 2006

26 26 Map Building: How to Establish a Map. By Hand Automatically: Map Building The robot learns its environment Motivation: - by hand: hard and costly - dynamically changing environment - different loo due to different perception

27 27 Map Building: The Problems. Map Maintaining: Keeping trac of changes in the environment 2. Representation and Reduction of Uncertainty e.g. disappearing cupboard position of robot -> position of wall? position of wall -> position of robot - i.e. measure of belief of each environment feature probability densities for feature positions inconsistent map due to motion drift

28 28 Overview of the SLAM problem Overview Simultaneous localization and mapping (SLAM) is a technique used by robots and autonomous vehicles to build up a map within an unnown environment while at the same time eeping trac of their current position. This is not as straightforward as it might sound due to inherent uncertainties ti in discerning i the robot's relative movement from its various sensors. If at the next iteration of map building the measured distance and direction traveled has a slight inaccuracy, then any features being added to the map will contain corresponding errors. If uncheced, these positional errors build cumulatively, grossly distorting the map and therefore the robot's ability to now its precise location. There are various techniques to compensate for this such as recognizing i features that t it has come across previously and re-sewing recent parts of the map to mae sure the two instances of that feature become one. Some of the statistical techniques used in SLAM include Kalman filters, particle filters and matching of laser scan data or visual landmars. Sensors characteristics A sensor is characterized principally by:. Noise 2. Dimensionality of input a. Planar laser range finder (2D points) b. 3D laser range finder (3D point cloud) c. Camera features (2D coordinates (pixels) if single camera or 3D coordinates if stereo camera) 3. Frame of reference a. Laser/camera in robot frame b. GPS in earth coord. Frame c. Accelerometer/Gyros in inertial coord. frame

29 29 SLAM overview

30 30

31 3

32 32

33 33

34 34

35 EKF based SLAM 5 35 The framewor is the same we have seen in the localization problem and can be solved by introducing an extended state X! LOCALIZATION The state X is the ONLY robot SLAM The state X is the robot configuration The state X is the ONLY robot configuration, that is: X = (X,Y,θ) g (X,Y,θ) PLUS the features m i =(x i,y i ) in the map, that is: X = (X,Y,θ,x,y,,x 2,y 2,,x i,y i ) And the prediction function was And the Prediction Phase function is X f X ) ( + + b s s s s L R L R 2 cos 2 θ ), ( = u X f X = y x u X f y x X M M ), ( = b s s b s s s s b Y X Y X L R L R L R 2 sin θ θ θ i i i i y x y x M M

36 EKF based SLAM: Estimation Phase 5 36 The observation model is the same as the EKF localization: ( ) ( ) + = = = i i i i i i X x Y y Y y X x X h r z arctan ) ( 2 2 ϕ ) ) ), ( y x m = ) ( i x i X In principle, the estimation can now proceed in the same manner as a ), ( y x m = ), ( i i y xi m = proceed in the same manner as a conventional EKF. ϕ X θ ϕ i = = i y x m X X M M = O M M M L L L m m m m X m m m m m Xm Xm Xm XX P P P P P P P P P P i i i y x m O M M M

37 37 Summary of the EKF SLAM EKF = Extended Kalman Filter The ingredients of the EKF Localization problem are: Prediction of the next robot position X based on the previous position X- and the odometry input u- X = ( X, u ) X = N X P f [ (, )] [ u = N( u, Q )] [ z = N h( x ), R )] Statistical error model of the odometry is Gaussian with: Statistical observation model is Gaussian: The map is now given and is part of the state to X! ( EKF Equations: PREDICTION X = f ( X, u ) P = F x PF T x ESTIMATION T T X = X + P H [ H P H + R ][ z h ( X ) ] + F QF u T u P + + = P P H T [ ] T T H P H + R H P

38 4d 38 EKF based Based SLAM (ETH - ASL) 4d - Perception - Features corner features extracted from lines connecting corners to structures

39 5 39 Particle Filter SLAM (FastSLAM, Montemerlo 2002) (/2) FastSLAM approach It solves the SLAM problem using particle filters. Particle filters are mathematical models that represent probability distribution as a set of discrete particles which occupy the state space. Particle filter update In the update step, a new particle distribution is generated given the motion model and the controls applied. a) For each particle:. Compare particle s prediction of measurements with actual measurements 2. Particles whose predictions match the measurements are given a high weight b) Filter resample: Resample particles based on weight Filter resample Assign each particle a weight depending on how well its estimate of the state agrees with the measurements and randomly draw particles from previous distribution based on weights creating a new distribution.

40 5 40 Particle Filter SLAM (FastSLAM, Montemerlo 2002) (2/2) FastSLAM approach (continuation) Particle set The robot posterior is solved using a Rao Blacwellized particle filtering using landmars. Each landmar estimation is represented by a 2x2 EKF (Extended Kalman Filter). Each particle is independent (due the factorization) from the others and maintains the estimate of M landmar positions.

41 4 Cyclic Environments Small local error accumulate to arbitrary large global errors! This is usually irrelevant for navigation However, when closing loops, global error does matter Courtesy of Sebastian Thrun

42 42 Probabilistic Feature Based 3D Maps (ETH-ASL) raw 3D scan of the same scene raw data photo of the scene decompose space into grid cells fill cells with data find a plane for every cell using RANSAC one plane per grid cell fuse similar neighboring planes together segmented planar segments final segmentation

43 4d - Perception - Features 5 43 Example Result photograph of corridor at ASL raw 3D scan plane segmentation result extracted planes for every cube

44 44 Probabilistic Plane Based 3D SLAM (EKF based) close-up of a reconstructed hallway close-up of reconstructed booshelves The same experiment as before but this time planar segments are visualized and integrated into the estimation process

45 45 Dynamic Environments Dynamical changes require continuous mapping If extraction of high-level features would be possible, the mapping in dynamic environments would become significantly more straightforward. e.g. difference between human and wall Environment modeling is a ey factor for robustness?

46 46 Key Problems in SLAM One of the first ey problem of SLAM is the recognition of a place already visited In order to reduce the drift the robot should detect loops in the enviroments This tass can be solved by detecting distinctive features from the environments. Laser scan features can lead however to ambiguous situations (T junctions in corridors, etc.) In the last 5 years place recognition (useful for loop detection) algorithms using cameras have been developed, d which h use distinctive visual features (e.g. SIFT features, Harris, features, etc.). Indeed, images offer a much larger quantity of information compared to 2D and 3D laser. Also SLAM algorithms have been developed d which h only use a single camera! See next slides

47 Autonomous Mobile Robots Real Time 3D Visual SLAM with a Single Camera This presentation is based on the following papers: Andrew J. Davison, Real-Time Simultaneous Localization and Mapping with a Single Camera, ICCV 2003 Nicholas Molton, Andrew J. Davison and Ian Reid, Locally Planar Patch Features for Real-Time Structure from Motion, BMVC 2004 Zürich Autonomous Systems Lab

48 48 Structure From Motion (SFM) Structure from Motion: Tae some images of the object to reconstruct t Features (points, lines, ) are extracted from all frames and matched among them All images are processed simultaneously Both camera motion and 3D structure can be recovered by optimally fusing the overall information, up to a scale factor Robust solution but far from being Real- Time!

49 49 Visual SLAM (Video From Davison 03) It is Real-Time!

50 50 State of the System The State vector and the first-order uncertainty are: Where: Where r is the camera position r = (X,Y,Z) and q is the camera orientation which is represented using quatternion (4-element vector). v and w are the translational and rotational velocity! In the SLAM the Covariance matrix is not diagonal, that is, the uncertainty of any feature affects the position estimate of all other features and the camera

51 5 EKF based Visual SLAM The state vector X and the Covariance matrix P are updated during camera motion using an EKF: Prediction: a motion model is used to predict where the camera will be in the next time step (P increases) X = f ( X, u ) P = F PF + x T x F u QF Observation: a new feature is measured (P decreases) X + = X + P H T T T u [ T H P H + R][ z h( X )] [ + ] T T H P H R H P P = P P H + +?

52 52 Prediction Phase (/2): A Motion Model for Smoothly Moving Camera Prediction Phase Attempts to predict where the camera will be in the next time step In the case the camera is attached to a person, the unnown intentions of the person can be statistically modeled Davison uses a Constant Velocity Model!

53 53 Prediction Phase (2/2): Constant Velocity Model The unnown intentions, and so unnown accelerations, are taen into account by modeling the acceleration as a process of zero mean and Gaussian distribution: By setting the covariance matrix of n to small or large values, we define the smoothness or rapidity of the motion we expect. In practice these values were used: 4 m / s 6 rad / s

54 54 Localization: Predicted Observation (/2) Predicted Observation By predicting the next camera pose, we can predict where each features is going g liely to appear: z ) i u = = h( X ) v = where And the uncertainty of the predicted observation is:

55 55 Localization: Predicted Observation (2/2) At each frame, the features occurring at previous step are searched in the elliptic region where they are expected to be according to the motion model (Normalized Sum of Square Differences is used for matching) Once the features are matched, the entire state of the system is updated according to EKF

56 56 SLAM: Adding new features in the map The feature is firstly initialized as a 3D line Along this line, 00 possible feature positions are set uniformly in the range m A each time, each hypothesis is tested by projecting it into the image Each matching produces a lielihood for each hypothesis and their probabilities are recomputed

57 57 Improved Feature Matching Up to now, traced features were treated as 2D templates in image space Long-term tracing can be improved by approximating the feature as a locally planar region on 3D world surfaces

58 58 Locally Planar 3D Patch Features

59 59 References A. J. Davison, Real-Time Simultaneous Localization and Mapping with a Single Camera, ICCV 2003 N. Molton, A. J. Davison and I. Reid, Locally Planar Patch Features for Real-Time Structure from Motion, BMVC 2004 A. J. Davison, Y. Gonzalez Cid, N. Kita, Real-Time 3D SLAM with wide-angle vision, IFAC Symposium on Intelligent Autonomous Vehicles, 2004 J. Shi and C. Tomasi, Good features to trac, CVPR, 994.

60 Autonomous Mobile Robots Applications: i Current Wor on Visual 3D SLAM done at the ETH-ASL Zürich Autonomous Systems Lab

61 Omnidirectional Visual Odometry Assumptions Method 2D Planar Motion Calibrated omnidirectional camera SIFT and homography for the translation [Triggs 98] Appearance based method for rotation. [Scaramuzza, 2008] IEEE Trans. On Robotics, October [Scaramuzza, 2008] ECCV - OMNIVIS 08

62 Ground Tracing View View 2 h n Image coordinates of two different views of a plane are related by a Homography H λ x2 = H x T n h T For a calibrated camera, motion can be computed from H using the = Triggs algorithm [Triggs 98] H R +

63 Visual Odometry Results Error at the loop closure: 6.5 m Error of orientation: 5 deg Trajectory length: 400 m

64 Loop Detection and Closing

65 Visual Odometry Results Before Loop Closure Error at the loop closure: 6.5 m Error of orientation: 5 deg Trajectory length: 400 m After Loop Closure

66 SMART trajectory on Google Earth drawn manually from the images (NOT GPS)

67

68

69 Our 3D Models built from Point Clouds and Images Lowenstrasse, Zurich Scaramuzza, 2009

70 70 Current open student projects The previous slides show some of our current research in visual odometry, SLAM, and 3D modeling of city environments. If you are a motivated student looing for a Master, Bachelor, or Semester project contact Davide Scaramuzza

Localization and Map Building

Localization and Map Building Localization and Map Building Noise and aliasing; odometric position estimation To localize or not to localize Belief representation Map representation Probabilistic map-based localization Other examples

More information

Localization and Map Building

Localization and Map Building Localization and Map Building Noise and aliasing; odometric position estimation To localize or not to localize Belief representation Map representation Probabilistic map-based localization Other examples

More information

Uncertainties: Representation and Propagation & Line Extraction from Range data

Uncertainties: Representation and Propagation & Line Extraction from Range data 41 Uncertainties: Representation and Propagation & Line Extraction from Range data 42 Uncertainty Representation Section 4.1.3 of the book Sensing in the real world is always uncertain How can uncertainty

More information

Zürich. Roland Siegwart Margarita Chli Martin Rufli Davide Scaramuzza. ETH Master Course: L Autonomous Mobile Robots Summary

Zürich. Roland Siegwart Margarita Chli Martin Rufli Davide Scaramuzza. ETH Master Course: L Autonomous Mobile Robots Summary Roland Siegwart Margarita Chli Martin Rufli Davide Scaramuzza ETH Master Course: 151-0854-00L Autonomous Mobile Robots Summary 2 Lecture Overview Mobile Robot Control Scheme knowledge, data base mission

More information

Autonomous Mobile Robot Design

Autonomous Mobile Robot Design Autonomous Mobile Robot Design Topic: EKF-based SLAM Dr. Kostas Alexis (CSE) These slides have partially relied on the course of C. Stachniss, Robot Mapping - WS 2013/14 Autonomous Robot Challenges Where

More information

Data Association for SLAM

Data Association for SLAM CALIFORNIA INSTITUTE OF TECHNOLOGY ME/CS 132a, Winter 2011 Lab #2 Due: Mar 10th, 2011 Part I Data Association for SLAM 1 Introduction For this part, you will experiment with a simulation of an EKF SLAM

More information

Perception. Autonomous Mobile Robots. Sensors Vision Uncertainties, Line extraction from laser scans. Autonomous Systems Lab. Zürich.

Perception. Autonomous Mobile Robots. Sensors Vision Uncertainties, Line extraction from laser scans. Autonomous Systems Lab. Zürich. Autonomous Mobile Robots Localization "Position" Global Map Cognition Environment Model Local Map Path Perception Real World Environment Motion Control Perception Sensors Vision Uncertainties, Line extraction

More information

Particle Filters. CSE-571 Probabilistic Robotics. Dependencies. Particle Filter Algorithm. Fast-SLAM Mapping

Particle Filters. CSE-571 Probabilistic Robotics. Dependencies. Particle Filter Algorithm. Fast-SLAM Mapping CSE-571 Probabilistic Robotics Fast-SLAM Mapping Particle Filters Represent belief by random samples Estimation of non-gaussian, nonlinear processes Sampling Importance Resampling (SIR) principle Draw

More information

RS-SLAM: RANSAC Sampling for Visual FastSLAM

RS-SLAM: RANSAC Sampling for Visual FastSLAM 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems September 25-30, 2011. San Francisco, CA, USA RS-SLAM: RANSAC Sampling for Visual FastSLAM Gim Hee Lee, Friedrich Fraundorfer, and

More information

Spring Localization II. Roland Siegwart, Margarita Chli, Martin Rufli. ASL Autonomous Systems Lab. Autonomous Mobile Robots

Spring Localization II. Roland Siegwart, Margarita Chli, Martin Rufli. ASL Autonomous Systems Lab. Autonomous Mobile Robots Spring 2016 Localization II Localization I 25.04.2016 1 knowledge, data base mission commands Localization Map Building environment model local map position global map Cognition Path Planning path Perception

More information

What is the SLAM problem?

What is the SLAM problem? SLAM Tutorial Slides by Marios Xanthidis, C. Stachniss, P. Allen, C. Fermuller Paul Furgale, Margarita Chli, Marco Hutter, Martin Rufli, Davide Scaramuzza, Roland Siegwart What is the SLAM problem? The

More information

Robot Simultaneous Localization and Mapping Based on Self-Detected Waypoint

Robot Simultaneous Localization and Mapping Based on Self-Detected Waypoint BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 16, No 2 Sofia 2016 Print ISSN: 1311-9702; Online ISSN: 1314-4081 DOI: 10.1515/cait-2016-0031 Robot Simultaneous Localization

More information

COS Lecture 13 Autonomous Robot Navigation

COS Lecture 13 Autonomous Robot Navigation COS 495 - Lecture 13 Autonomous Robot Navigation Instructor: Chris Clark Semester: Fall 2011 1 Figures courtesy of Siegwart & Nourbakhsh Control Structure Prior Knowledge Operator Commands Localization

More information

Probabilistic Robotics

Probabilistic Robotics Probabilistic Robotics FastSLAM Sebastian Thrun (abridged and adapted by Rodrigo Ventura in Oct-2008) The SLAM Problem SLAM stands for simultaneous localization and mapping The task of building a map while

More information

Robotics. Lecture 7: Simultaneous Localisation and Mapping (SLAM)

Robotics. Lecture 7: Simultaneous Localisation and Mapping (SLAM) Robotics Lecture 7: Simultaneous Localisation and Mapping (SLAM) See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information. Andrew Davison Department of Computing Imperial College

More information

Survey: Simultaneous Localisation and Mapping (SLAM) Ronja Güldenring Master Informatics Project Intellgient Robotics University of Hamburg

Survey: Simultaneous Localisation and Mapping (SLAM) Ronja Güldenring Master Informatics Project Intellgient Robotics University of Hamburg Survey: Simultaneous Localisation and Mapping (SLAM) Ronja Güldenring Master Informatics Project Intellgient Robotics University of Hamburg Introduction EKF-SLAM FastSLAM Loop Closure 01.06.17 Ronja Güldenring

More information

Probabilistic Robotics

Probabilistic Robotics Probabilistic Robotics Bayes Filter Implementations Discrete filters, Particle filters Piecewise Constant Representation of belief 2 Discrete Bayes Filter Algorithm 1. Algorithm Discrete_Bayes_filter(

More information

Accurate Motion Estimation and High-Precision 3D Reconstruction by Sensor Fusion

Accurate Motion Estimation and High-Precision 3D Reconstruction by Sensor Fusion 007 IEEE International Conference on Robotics and Automation Roma, Italy, 0-4 April 007 FrE5. Accurate Motion Estimation and High-Precision D Reconstruction by Sensor Fusion Yunsu Bok, Youngbae Hwang,

More information

Humanoid Robotics. Monte Carlo Localization. Maren Bennewitz

Humanoid Robotics. Monte Carlo Localization. Maren Bennewitz Humanoid Robotics Monte Carlo Localization Maren Bennewitz 1 Basis Probability Rules (1) If x and y are independent: Bayes rule: Often written as: The denominator is a normalizing constant that ensures

More information

Particle Filter in Brief. Robot Mapping. FastSLAM Feature-based SLAM with Particle Filters. Particle Representation. Particle Filter Algorithm

Particle Filter in Brief. Robot Mapping. FastSLAM Feature-based SLAM with Particle Filters. Particle Representation. Particle Filter Algorithm Robot Mapping FastSLAM Feature-based SLAM with Particle Filters Cyrill Stachniss Particle Filter in Brief! Non-parametric, recursive Bayes filter! Posterior is represented by a set of weighted samples!

More information

Revising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History

Revising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History Revising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History Simon Thompson and Satoshi Kagami Digital Human Research Center National Institute of Advanced

More information

Spring Localization II. Roland Siegwart, Margarita Chli, Juan Nieto, Nick Lawrance. ASL Autonomous Systems Lab. Autonomous Mobile Robots

Spring Localization II. Roland Siegwart, Margarita Chli, Juan Nieto, Nick Lawrance. ASL Autonomous Systems Lab. Autonomous Mobile Robots Spring 2018 Localization II Localization I 16.04.2018 1 knowledge, data base mission commands Localization Map Building environment model local map position global map Cognition Path Planning path Perception

More information

Practical Course WS12/13 Introduction to Monte Carlo Localization

Practical Course WS12/13 Introduction to Monte Carlo Localization Practical Course WS12/13 Introduction to Monte Carlo Localization Cyrill Stachniss and Luciano Spinello 1 State Estimation Estimate the state of a system given observations and controls Goal: 2 Bayes Filter

More information

Visual Bearing-Only Simultaneous Localization and Mapping with Improved Feature Matching

Visual Bearing-Only Simultaneous Localization and Mapping with Improved Feature Matching Visual Bearing-Only Simultaneous Localization and Mapping with Improved Feature Matching Hauke Strasdat, Cyrill Stachniss, Maren Bennewitz, and Wolfram Burgard Computer Science Institute, University of

More information

Probabilistic Robotics. FastSLAM

Probabilistic Robotics. FastSLAM Probabilistic Robotics FastSLAM The SLAM Problem SLAM stands for simultaneous localization and mapping The task of building a map while estimating the pose of the robot relative to this map Why is SLAM

More information

Overview. EECS 124, UC Berkeley, Spring 2008 Lecture 23: Localization and Mapping. Statistical Models

Overview. EECS 124, UC Berkeley, Spring 2008 Lecture 23: Localization and Mapping. Statistical Models Introduction ti to Embedded dsystems EECS 124, UC Berkeley, Spring 2008 Lecture 23: Localization and Mapping Gabe Hoffmann Ph.D. Candidate, Aero/Astro Engineering Stanford University Statistical Models

More information

Robotics. Lecture 8: Simultaneous Localisation and Mapping (SLAM)

Robotics. Lecture 8: Simultaneous Localisation and Mapping (SLAM) Robotics Lecture 8: Simultaneous Localisation and Mapping (SLAM) See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information. Andrew Davison Department of Computing Imperial College

More information

Augmented Reality, Advanced SLAM, Applications

Augmented Reality, Advanced SLAM, Applications Augmented Reality, Advanced SLAM, Applications Prof. Didier Stricker & Dr. Alain Pagani alain.pagani@dfki.de Lecture 3D Computer Vision AR, SLAM, Applications 1 Introduction Previous lectures: Basics (camera,

More information

Probabilistic Robotics

Probabilistic Robotics Probabilistic Robotics Sebastian Thrun Wolfram Burgard Dieter Fox The MIT Press Cambridge, Massachusetts London, England Preface xvii Acknowledgments xix I Basics 1 1 Introduction 3 1.1 Uncertainty in

More information

Robot Localization based on Geo-referenced Images and G raphic Methods

Robot Localization based on Geo-referenced Images and G raphic Methods Robot Localization based on Geo-referenced Images and G raphic Methods Sid Ahmed Berrabah Mechanical Department, Royal Military School, Belgium, sidahmed.berrabah@rma.ac.be Janusz Bedkowski, Łukasz Lubasiński,

More information

Simultaneous Localization and Mapping

Simultaneous Localization and Mapping Sebastian Lembcke SLAM 1 / 29 MIN Faculty Department of Informatics Simultaneous Localization and Mapping Visual Loop-Closure Detection University of Hamburg Faculty of Mathematics, Informatics and Natural

More information

Offline Simultaneous Localization and Mapping (SLAM) using Miniature Robots

Offline Simultaneous Localization and Mapping (SLAM) using Miniature Robots Offline Simultaneous Localization and Mapping (SLAM) using Miniature Robots Objectives SLAM approaches SLAM for ALICE EKF for Navigation Mapping and Network Modeling Test results Philipp Schaer and Adrian

More information

7630 Autonomous Robotics Probabilities for Robotics

7630 Autonomous Robotics Probabilities for Robotics 7630 Autonomous Robotics Probabilities for Robotics Basics of probability theory The Bayes-Filter Introduction to localization problems Monte-Carlo-Localization Based on material from R. Triebel, R. Kästner

More information

Image Augmented Laser Scan Matching for Indoor Localization

Image Augmented Laser Scan Matching for Indoor Localization Image Augmented Laser Scan Matching for Indoor Localization Nikhil Naikal Avideh Zakhor John Kua Electrical Engineering and Computer Sciences University of California at Berkeley Technical Report No. UCB/EECS-2009-35

More information

IROS 05 Tutorial. MCL: Global Localization (Sonar) Monte-Carlo Localization. Particle Filters. Rao-Blackwellized Particle Filters and Loop Closing

IROS 05 Tutorial. MCL: Global Localization (Sonar) Monte-Carlo Localization. Particle Filters. Rao-Blackwellized Particle Filters and Loop Closing IROS 05 Tutorial SLAM - Getting it Working in Real World Applications Rao-Blackwellized Particle Filters and Loop Closing Cyrill Stachniss and Wolfram Burgard University of Freiburg, Dept. of Computer

More information

Sensor Modalities. Sensor modality: Different modalities:

Sensor Modalities. Sensor modality: Different modalities: Sensor Modalities Sensor modality: Sensors which measure same form of energy and process it in similar ways Modality refers to the raw input used by the sensors Different modalities: Sound Pressure Temperature

More information

Implementation of Odometry with EKF for Localization of Hector SLAM Method

Implementation of Odometry with EKF for Localization of Hector SLAM Method Implementation of Odometry with EKF for Localization of Hector SLAM Method Kao-Shing Hwang 1 Wei-Cheng Jiang 2 Zuo-Syuan Wang 3 Department of Electrical Engineering, National Sun Yat-sen University, Kaohsiung,

More information

Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview

Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview Arun Das 05/09/2017 Arun Das Waterloo Autonomous Vehicles Lab Introduction What s in a name? Arun Das Waterloo Autonomous

More information

EE565:Mobile Robotics Lecture 3

EE565:Mobile Robotics Lecture 3 EE565:Mobile Robotics Lecture 3 Welcome Dr. Ahmad Kamal Nasir Today s Objectives Motion Models Velocity based model (Dead-Reckoning) Odometry based model (Wheel Encoders) Sensor Models Beam model of range

More information

CSE 527: Introduction to Computer Vision

CSE 527: Introduction to Computer Vision CSE 527: Introduction to Computer Vision Week 10 Class 2: Visual Odometry November 2nd, 2017 Today Visual Odometry Intro Algorithm SLAM Visual Odometry Input Output Images, Video Camera trajectory, motion

More information

ME 597/747 Autonomous Mobile Robots. Mid Term Exam. Duration: 2 hour Total Marks: 100

ME 597/747 Autonomous Mobile Robots. Mid Term Exam. Duration: 2 hour Total Marks: 100 ME 597/747 Autonomous Mobile Robots Mid Term Exam Duration: 2 hour Total Marks: 100 Instructions: Read the exam carefully before starting. Equations are at the back, but they are NOT necessarily valid

More information

Robot Mapping. A Short Introduction to the Bayes Filter and Related Models. Gian Diego Tipaldi, Wolfram Burgard

Robot Mapping. A Short Introduction to the Bayes Filter and Related Models. Gian Diego Tipaldi, Wolfram Burgard Robot Mapping A Short Introduction to the Bayes Filter and Related Models Gian Diego Tipaldi, Wolfram Burgard 1 State Estimation Estimate the state of a system given observations and controls Goal: 2 Recursive

More information

Introduction to Mobile Robotics SLAM Grid-based FastSLAM. Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Diego Tipaldi, Luciano Spinello

Introduction to Mobile Robotics SLAM Grid-based FastSLAM. Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Diego Tipaldi, Luciano Spinello Introduction to Mobile Robotics SLAM Grid-based FastSLAM Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Diego Tipaldi, Luciano Spinello 1 The SLAM Problem SLAM stands for simultaneous localization

More information

Robotics. Lecture 5: Monte Carlo Localisation. See course website for up to date information.

Robotics. Lecture 5: Monte Carlo Localisation. See course website  for up to date information. Robotics Lecture 5: Monte Carlo Localisation See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information. Andrew Davison Department of Computing Imperial College London Review:

More information

Basics of Localization, Mapping and SLAM. Jari Saarinen Aalto University Department of Automation and systems Technology

Basics of Localization, Mapping and SLAM. Jari Saarinen Aalto University Department of Automation and systems Technology Basics of Localization, Mapping and SLAM Jari Saarinen Aalto University Department of Automation and systems Technology Content Introduction to Problem (s) Localization A few basic equations Dead Reckoning

More information

SLAM WITH KLT POINT FEATURES. P.O. Box 22, Thammasat-Rangsit Post Office, Pathumthani Thailand

SLAM WITH KLT POINT FEATURES. P.O. Box 22, Thammasat-Rangsit Post Office, Pathumthani Thailand SLAM WITH KLT POINT FEATURES Yoichi Nakaguro, 1 Matthew Dailey, Stanislav Makhanov 1 1 Sirindhorn International Institute of Technology, Thammasat University P.O. Box, Thammasat-Rangsit Post Office, Pathumthani

More information

L17. OCCUPANCY MAPS. NA568 Mobile Robotics: Methods & Algorithms

L17. OCCUPANCY MAPS. NA568 Mobile Robotics: Methods & Algorithms L17. OCCUPANCY MAPS NA568 Mobile Robotics: Methods & Algorithms Today s Topic Why Occupancy Maps? Bayes Binary Filters Log-odds Occupancy Maps Inverse sensor model Learning inverse sensor model ML map

More information

Segmentation and Tracking of Partial Planar Templates

Segmentation and Tracking of Partial Planar Templates Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract

More information

USING 3D DATA FOR MONTE CARLO LOCALIZATION IN COMPLEX INDOOR ENVIRONMENTS. Oliver Wulf, Bernardo Wagner

USING 3D DATA FOR MONTE CARLO LOCALIZATION IN COMPLEX INDOOR ENVIRONMENTS. Oliver Wulf, Bernardo Wagner USING 3D DATA FOR MONTE CARLO LOCALIZATION IN COMPLEX INDOOR ENVIRONMENTS Oliver Wulf, Bernardo Wagner Institute for Systems Engineering (RTS/ISE), University of Hannover, Germany Mohamed Khalaf-Allah

More information

Simuntaneous Localisation and Mapping with a Single Camera. Abhishek Aneja and Zhichao Chen

Simuntaneous Localisation and Mapping with a Single Camera. Abhishek Aneja and Zhichao Chen Simuntaneous Localisation and Mapping with a Single Camera Abhishek Aneja and Zhichao Chen 3 December, Simuntaneous Localisation and Mapping with asinglecamera 1 Abstract Image reconstruction is common

More information

Visual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

Visual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech Visual Odometry Features, Tracking, Essential Matrix, and RANSAC Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline The

More information

Pedestrian Detection Using Multi-layer LIDAR

Pedestrian Detection Using Multi-layer LIDAR 1 st International Conference on Transportation Infrastructure and Materials (ICTIM 2016) ISBN: 978-1-60595-367-0 Pedestrian Detection Using Multi-layer LIDAR Mingfang Zhang 1, Yuping Lu 2 and Tong Liu

More information

Robotic Perception and Action: Vehicle SLAM Assignment

Robotic Perception and Action: Vehicle SLAM Assignment Robotic Perception and Action: Vehicle SLAM Assignment Mariolino De Cecco Mariolino De Cecco, Mattia Tavernini 1 CONTENTS Vehicle SLAM Assignment Contents Assignment Scenario 3 Odometry Localization...........................................

More information

Vision-based Mobile Robot Localization and Mapping using Scale-Invariant Features

Vision-based Mobile Robot Localization and Mapping using Scale-Invariant Features Vision-based Mobile Robot Localization and Mapping using Scale-Invariant Features Stephen Se, David Lowe, Jim Little Department of Computer Science University of British Columbia Presented by Adam Bickett

More information

Introduction to Mobile Robotics SLAM Landmark-based FastSLAM

Introduction to Mobile Robotics SLAM Landmark-based FastSLAM Introduction to Mobile Robotics SLAM Landmark-based FastSLAM Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Diego Tipaldi, Luciano Spinello Partial slide courtesy of Mike Montemerlo 1 The SLAM Problem

More information

Final Exam Practice Fall Semester, 2012

Final Exam Practice Fall Semester, 2012 COS 495 - Autonomous Robot Navigation Final Exam Practice Fall Semester, 2012 Duration: Total Marks: 70 Closed Book 2 hours Start Time: End Time: By signing this exam, I agree to the honor code Name: Signature:

More information

Introduction to Autonomous Mobile Robots

Introduction to Autonomous Mobile Robots Introduction to Autonomous Mobile Robots second edition Roland Siegwart, Illah R. Nourbakhsh, and Davide Scaramuzza The MIT Press Cambridge, Massachusetts London, England Contents Acknowledgments xiii

More information

3D Environment Reconstruction

3D Environment Reconstruction 3D Environment Reconstruction Using Modified Color ICP Algorithm by Fusion of a Camera and a 3D Laser Range Finder The 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems October 11-15,

More information

ECE276A: Sensing & Estimation in Robotics Lecture 11: Simultaneous Localization and Mapping using a Particle Filter

ECE276A: Sensing & Estimation in Robotics Lecture 11: Simultaneous Localization and Mapping using a Particle Filter ECE276A: Sensing & Estimation in Robotics Lecture 11: Simultaneous Localization and Mapping using a Particle Filter Lecturer: Nikolay Atanasov: natanasov@ucsd.edu Teaching Assistants: Siwei Guo: s9guo@eng.ucsd.edu

More information

An Application of Omnidirectional Vision to Grid-based SLAM in Indoor Environments

An Application of Omnidirectional Vision to Grid-based SLAM in Indoor Environments An Application of Omnidirectional Vision to Grid-based SLAM in Indoor Environments Vito Macchia, Stefano Rosa, Luca Carlone, Basilio Bona Abstract In this wor we study the use of an omnidirectional camera

More information

Monte Carlo Localization using Dynamically Expanding Occupancy Grids. Karan M. Gupta

Monte Carlo Localization using Dynamically Expanding Occupancy Grids. Karan M. Gupta 1 Monte Carlo Localization using Dynamically Expanding Occupancy Grids Karan M. Gupta Agenda Introduction Occupancy Grids Sonar Sensor Model Dynamically Expanding Occupancy Grids Monte Carlo Localization

More information

Vehicle Localization. Hannah Rae Kerner 21 April 2015

Vehicle Localization. Hannah Rae Kerner 21 April 2015 Vehicle Localization Hannah Rae Kerner 21 April 2015 Spotted in Mtn View: Google Car Why precision localization? in order for a robot to follow a road, it needs to know where the road is to stay in a particular

More information

Odometry-based Online Extrinsic Sensor Calibration

Odometry-based Online Extrinsic Sensor Calibration 213 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) November 3-7, 213. Toyo, Japan Odometry-based Online Extrinsic Sensor Calibration Sebastian Schneider, Thorsten Luettel and

More information

Robot Mapping. SLAM Front-Ends. Cyrill Stachniss. Partial image courtesy: Edwin Olson 1

Robot Mapping. SLAM Front-Ends. Cyrill Stachniss. Partial image courtesy: Edwin Olson 1 Robot Mapping SLAM Front-Ends Cyrill Stachniss Partial image courtesy: Edwin Olson 1 Graph-Based SLAM Constraints connect the nodes through odometry and observations Robot pose Constraint 2 Graph-Based

More information

Monocular SLAM for a Small-Size Humanoid Robot

Monocular SLAM for a Small-Size Humanoid Robot Tamkang Journal of Science and Engineering, Vol. 14, No. 2, pp. 123 129 (2011) 123 Monocular SLAM for a Small-Size Humanoid Robot Yin-Tien Wang*, Duen-Yan Hung and Sheng-Hsien Cheng Department of Mechanical

More information

PML-SLAM: a solution for localization in large-scale urban environments

PML-SLAM: a solution for localization in large-scale urban environments PML-SLAM: a solution for localization in large-scale urban environments Zayed Alsayed, Guillaume Bresson, Fawzi Nashashibi and Anne Verroust-Blondet Institut VEDECOM Versailles, France firstname.name@vedecom.fr

More information

Introduction to SLAM Part II. Paul Robertson

Introduction to SLAM Part II. Paul Robertson Introduction to SLAM Part II Paul Robertson Localization Review Tracking, Global Localization, Kidnapping Problem. Kalman Filter Quadratic Linear (unless EKF) SLAM Loop closing Scaling: Partition space

More information

LOCAL AND GLOBAL DESCRIPTORS FOR PLACE RECOGNITION IN ROBOTICS

LOCAL AND GLOBAL DESCRIPTORS FOR PLACE RECOGNITION IN ROBOTICS 8th International DAAAM Baltic Conference "INDUSTRIAL ENGINEERING - 19-21 April 2012, Tallinn, Estonia LOCAL AND GLOBAL DESCRIPTORS FOR PLACE RECOGNITION IN ROBOTICS Shvarts, D. & Tamre, M. Abstract: The

More information

Lecture 13 Visual Inertial Fusion

Lecture 13 Visual Inertial Fusion Lecture 13 Visual Inertial Fusion Davide Scaramuzza Course Evaluation Please fill the evaluation form you received by email! Provide feedback on Exercises: good and bad Course: good and bad How to improve

More information

NAVIGATION SYSTEM OF AN OUTDOOR SERVICE ROBOT WITH HYBRID LOCOMOTION SYSTEM

NAVIGATION SYSTEM OF AN OUTDOOR SERVICE ROBOT WITH HYBRID LOCOMOTION SYSTEM NAVIGATION SYSTEM OF AN OUTDOOR SERVICE ROBOT WITH HYBRID LOCOMOTION SYSTEM Jorma Selkäinaho, Aarne Halme and Janne Paanajärvi Automation Technology Laboratory, Helsinki University of Technology, Espoo,

More information

Integrated Laser-Camera Sensor for the Detection and Localization of Landmarks for Robotic Applications

Integrated Laser-Camera Sensor for the Detection and Localization of Landmarks for Robotic Applications 28 IEEE International Conference on Robotics and Automation Pasadena, CA, USA, May 19-23, 28 Integrated Laser-Camera Sensor for the Detection and Localization of Landmarks for Robotic Applications Dilan

More information

Localization, Where am I?

Localization, Where am I? 5.1 Localization, Where am I?? position Position Update (Estimation?) Encoder Prediction of Position (e.g. odometry) YES matched observations Map data base predicted position Matching Odometry, Dead Reckoning

More information

Generalizing Random-Vector SLAM with Random Finite Sets

Generalizing Random-Vector SLAM with Random Finite Sets Generalizing Random-Vector SLAM with Random Finite Sets Keith Y. K. Leung, Felipe Inostroza, Martin Adams Advanced Mining Technology Center AMTC, Universidad de Chile, Santiago, Chile eith.leung@amtc.uchile.cl,

More information

Mobile Robots Summery. Autonomous Mobile Robots

Mobile Robots Summery. Autonomous Mobile Robots Mobile Robots Summery Roland Siegwart Mike Bosse, Marco Hutter, Martin Rufli, Davide Scaramuzza, (Margarita Chli, Paul Furgale) Mobile Robots Summery 1 Introduction probabilistic map-based localization

More information

On the Use of Inverse Scaling in Monocular SLAM

On the Use of Inverse Scaling in Monocular SLAM 29 IEEE International Conference on Robotics and Automation Kobe International Conference Center Kobe, Japan, May 12-17, 29 On the Use of Inverse Scaling in Monocular SLAM Daniele Marzorati, Matteo Matteucci,

More information

Dealing with Scale. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

Dealing with Scale. Stephan Weiss Computer Vision Group NASA-JPL / CalTech Dealing with Scale Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline Why care about size? The IMU as scale provider: The

More information

Stable Vision-Aided Navigation for Large-Area Augmented Reality

Stable Vision-Aided Navigation for Large-Area Augmented Reality Stable Vision-Aided Navigation for Large-Area Augmented Reality Taragay Oskiper, Han-Pang Chiu, Zhiwei Zhu Supun Samarasekera, Rakesh Teddy Kumar Vision and Robotics Laboratory SRI-International Sarnoff,

More information

Introduction to Mobile Robotics. SLAM: Simultaneous Localization and Mapping

Introduction to Mobile Robotics. SLAM: Simultaneous Localization and Mapping Introduction to Mobile Robotics SLAM: Simultaneous Localization and Mapping The SLAM Problem SLAM is the process by which a robot builds a map of the environment and, at the same time, uses this map to

More information

Davide Scaramuzza. University of Zurich

Davide Scaramuzza. University of Zurich Davide Scaramuzza University of Zurich Robotics and Perception Group http://rpg.ifi.uzh.ch/ Scaramuzza, D., Fraundorfer, F., Visual Odometry: Part I - The First 30 Years and Fundamentals, IEEE Robotics

More information

Introduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization. Wolfram Burgard

Introduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization. Wolfram Burgard Introduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization Wolfram Burgard 1 Motivation Recall: Discrete filter Discretize the continuous state space High memory complexity

More information

F1/10 th Autonomous Racing. Localization. Nischal K N

F1/10 th Autonomous Racing. Localization. Nischal K N F1/10 th Autonomous Racing Localization Nischal K N System Overview Mapping Hector Mapping Localization Path Planning Control System Overview Mapping Hector Mapping Localization Adaptive Monte Carlo Localization

More information

CSE 490R P1 - Localization using Particle Filters Due date: Sun, Jan 28-11:59 PM

CSE 490R P1 - Localization using Particle Filters Due date: Sun, Jan 28-11:59 PM CSE 490R P1 - Localization using Particle Filters Due date: Sun, Jan 28-11:59 PM 1 Introduction In this assignment you will implement a particle filter to localize your car within a known map. This will

More information

Efficient SLAM Scheme Based ICP Matching Algorithm Using Image and Laser Scan Information

Efficient SLAM Scheme Based ICP Matching Algorithm Using Image and Laser Scan Information Proceedings of the World Congress on Electrical Engineering and Computer Systems and Science (EECSS 2015) Barcelona, Spain July 13-14, 2015 Paper No. 335 Efficient SLAM Scheme Based ICP Matching Algorithm

More information

Probabilistic Robotics

Probabilistic Robotics Probabilistic Robotics Discrete Filters and Particle Filters Models Some slides adopted from: Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Kai Arras and Probabilistic Robotics Book SA-1 Probabilistic

More information

Full Scaled 3D Visual Odometry from a Single Wearable Omnidirectional Camera

Full Scaled 3D Visual Odometry from a Single Wearable Omnidirectional Camera Full Scaled 3D Visual Odometry from a Single Wearable Omnidirectional Camera Daniel Gutiérrez-Gómez, Luis Puig, J.J. Guerrero Aragon Institute of Engineering Research (I3A) University of Zaragoza, Mariano

More information

Mobile Robotics. Mathematics, Models, and Methods. HI Cambridge. Alonzo Kelly. Carnegie Mellon University UNIVERSITY PRESS

Mobile Robotics. Mathematics, Models, and Methods. HI Cambridge. Alonzo Kelly. Carnegie Mellon University UNIVERSITY PRESS Mobile Robotics Mathematics, Models, and Methods Alonzo Kelly Carnegie Mellon University HI Cambridge UNIVERSITY PRESS Contents Preface page xiii 1 Introduction 1 1.1 Applications of Mobile Robots 2 1.2

More information

Particle Filter for Robot Localization ECE 478 Homework #1

Particle Filter for Robot Localization ECE 478 Homework #1 Particle Filter for Robot Localization ECE 478 Homework #1 Phil Lamb pjl@pdx.edu November 15, 2012 1 Contents 1 Introduction 3 2 Implementation 3 2.1 Assumptions and Simplifications.............................

More information

Building Reliable 2D Maps from 3D Features

Building Reliable 2D Maps from 3D Features Building Reliable 2D Maps from 3D Features Dipl. Technoinform. Jens Wettach, Prof. Dr. rer. nat. Karsten Berns TU Kaiserslautern; Robotics Research Lab 1, Geb. 48; Gottlieb-Daimler- Str.1; 67663 Kaiserslautern;

More information

High-precision, consistent EKF-based visual-inertial odometry

High-precision, consistent EKF-based visual-inertial odometry High-precision, consistent EKF-based visual-inertial odometry Mingyang Li and Anastasios I. Mourikis, IJRR 2013 Ao Li Introduction What is visual-inertial odometry (VIO)? The problem of motion tracking

More information

Indoor Localization Algorithms for a Human-Operated Backpack System

Indoor Localization Algorithms for a Human-Operated Backpack System Indoor Localization Algorithms for a Human-Operated Backpack System George Chen, John Kua, Stephen Shum, Nikhil Naikal, Matthew Carlberg, Avideh Zakhor Video and Image Processing Lab, University of California,

More information

Artificial Intelligence for Robotics: A Brief Summary

Artificial Intelligence for Robotics: A Brief Summary Artificial Intelligence for Robotics: A Brief Summary This document provides a summary of the course, Artificial Intelligence for Robotics, and highlights main concepts. Lesson 1: Localization (using Histogram

More information

Proceedings of ESDA2006 8th Biennial ASME Conference on Engineering Systems Design and Analysis July 4-7, 2006, Torino, Italy

Proceedings of ESDA2006 8th Biennial ASME Conference on Engineering Systems Design and Analysis July 4-7, 2006, Torino, Italy Proceedings of ESDA2006 8th Biennial ASME Conference on Engineering Systems Design and Analysis July 4-7, 2006, Torino, Italy ESDA2006-95412 LOCALIZATION OF A MOBILE ROBOT WITH OMNIDIRECTIONAL WHEELS USING

More information

CAMERA POSE ESTIMATION OF RGB-D SENSORS USING PARTICLE FILTERING

CAMERA POSE ESTIMATION OF RGB-D SENSORS USING PARTICLE FILTERING CAMERA POSE ESTIMATION OF RGB-D SENSORS USING PARTICLE FILTERING By Michael Lowney Senior Thesis in Electrical Engineering University of Illinois at Urbana-Champaign Advisor: Professor Minh Do May 2015

More information

5. Tests and results Scan Matching Optimization Parameters Influence

5. Tests and results Scan Matching Optimization Parameters Influence 126 5. Tests and results This chapter presents results obtained using the proposed method on simulated and real data. First, it is analyzed the scan matching optimization; after that, the Scan Matching

More information

Geometrical Feature Extraction Using 2D Range Scanner

Geometrical Feature Extraction Using 2D Range Scanner Geometrical Feature Extraction Using 2D Range Scanner Sen Zhang Lihua Xie Martin Adams Fan Tang BLK S2, School of Electrical and Electronic Engineering Nanyang Technological University, Singapore 639798

More information

This chapter explains two techniques which are frequently used throughout

This chapter explains two techniques which are frequently used throughout Chapter 2 Basic Techniques This chapter explains two techniques which are frequently used throughout this thesis. First, we will introduce the concept of particle filters. A particle filter is a recursive

More information

BBR Progress Report 006: Autonomous 2-D Mapping of a Building Floor

BBR Progress Report 006: Autonomous 2-D Mapping of a Building Floor BBR Progress Report 006: Autonomous 2-D Mapping of a Building Floor Andy Sayler & Constantin Berzan November 30, 2010 Abstract In the past two weeks, we implemented and tested landmark extraction based

More information

Tutorial on 3D Surface Reconstruction in Laparoscopic Surgery. Simultaneous Localization and Mapping for Minimally Invasive Surgery

Tutorial on 3D Surface Reconstruction in Laparoscopic Surgery. Simultaneous Localization and Mapping for Minimally Invasive Surgery Tutorial on 3D Surface Reconstruction in Laparoscopic Surgery Simultaneous Localization and Mapping for Minimally Invasive Surgery Introduction University of Bristol using particle filters to track football

More information

Localization Assistance for Multiple Mobile Robots using Target Tracking Techniques

Localization Assistance for Multiple Mobile Robots using Target Tracking Techniques Localization Assistance for Multiple Mobile Robots using Target Tracing Techniques Peshala Jayaseara a, Hidei Hashimoto b and Taashi Kubota c a Dept. of Electrical Engineering, University of Toyo, Toyo,

More information

Localization of Multiple Robots with Simple Sensors

Localization of Multiple Robots with Simple Sensors Proceedings of the IEEE International Conference on Mechatronics & Automation Niagara Falls, Canada July 2005 Localization of Multiple Robots with Simple Sensors Mike Peasgood and Christopher Clark Lab

More information