Kalman Filter Based. Localization
|
|
- Clifford Lynch
- 5 years ago
- Views:
Transcription
1 Autonomous Mobile Robots Localization "Position" Global Map Cognition Environment Model Local Map Path Perception Real World Environment Motion Control Kalman Filter Based Localization & SLAM Zürich Autonomous Systems Lab
2 2 Kalman Filter Localization
3 3 Introduction to Kalman Filter () Two measurements Weighted least-square Finding minimum error After some calculation and rearrangements
4 4 Introduction to Kalman Filter (2) In Kalman Filter notation
5 5 Introduction to Kalman Filter (3) Dynamic Prediction (robot moving) st hour, Motion u = velocity w = noise Combining fusion and dynamic prediction
6 6 The Five Steps for Map-Based Localization Encoder Prediction Measurement and Position position estimate Estimation (odometry) (fusion) Map data base ures ns dicted feat bservation Pred o YES Matching matched predictions and observations raw sensor data or. Prediction based on previous estimate and odometry extracted features 2. Observation with on-board sensors 3. Measurement prediction based on prediction and map Observation on-board sensors 4. Matching of observation and map 5. Estimation -> position update (posteriori position) perce eption
7 5 7 Kalman Filter for Mobile Robot Localization: Robot Position Prediction In a first step, the robots position at time step + is predicted based on its old location (time step ) and its movement due to the control input u(): p ˆ( ) = f ( pˆ( ), u( )) + f: Odometry function T Σ p ( + ) = p f Σ p ( ) p f + u f Σ u ( ) u f T
8 5 8 Kalman Filter for Mobile Robot Localization: Robot Position Prediction: Example s xˆ( ) + s pˆ ( + ) = pˆ( ) + u( ) = yˆ( ) θ ˆ( ) p p p p T + sl sr sl cos( θ+ 2 2b + s l sr s l sin( θ+ 2 2b sr sl b r ) r ) Σ ( + ) = f Σ ( ) f + f Σ ( ) u u u f T Odometry Σ u ( ) = r s 0 r l 0 s l
9 9 Kalman Filter for Mobile Robot Localization: Observation The second step it to obtain the observation Z(+) ( )(measurements) from the robot s sensors at the new location at time + The observation usually consists of a set n 0 of single observations z j (+) extracted from the different sensors signals. It can represent raw data scans as well as features lie lines, doors or any ind of landmars. The parameters of the targets are usually observed in the sensor frame {S}. Therefore the observations have to be transformed to the world frame {W} or the measurement prediction have to be transformed to the sensor frame {S}. This transformation is specified in the function h i (see later).
10 0 Kalman Filter for Mobile Robot Localization: Observation: Example Raw Date of Laser Scanner Extracted Lines Extracted Lines in Model Space line j αj r j j R Sensor j ( + = (robot) frame z ) α rj σαα σα r ΣR, j( + ) = σrα σrr j
11 5 Kalman Filter for Mobile Robot Localization: Measurement Prediction ( + ) pˆ = In the next step we use the predicted robot position and the map M() to generate multiple predicted observations z t. They have to be transformed into the sensor frame ẑ ( + ) = h ( z,pˆ p ( + ) i i t We can now define the measurement prediction as the set containing all n i predicted observations Ẑ ( + ) = ẑ ( + )( i ) { } i n i The function h i is mainly the coordinate transformation between the world frame and the sensor frame
12 5 2 Kalman Filter for Mobile Robot Localization: Measurement Prediction: Example For prediction, only the walls that are in the field of view of the robot are selected.
13 3 Kalman Filter for Mobile Robot Localization: Matching: Example
14 5 4 Kalman Filter for Mobile Robot Localization: Estimation: Applying the Kalman Filter Kalman filter gain: Update of robot s position estimate: The associate variance
15 5 Kalman Filter for Mobile Robot Localization: Estimation: D Case For the one-dimensional case with we can show that the estimation corresponds to the Kalman filter for one-dimension presented earlier.
16 6 Kalman Filter for Mobile Robot Localization: Estimation: Example Kalman filter estimation of the new robot position : pˆ ( ) ) By fusing the prediction of robot position (magenta) with the innovation gained by the measurements (green) we get the updated estimate of the robot position (red)
17 7 Autonomous Indoor Navigation (Pygmalion EPFL) very robust on-the-fly localization one of the first systems with probabilistic sensor fusion 47 steps,78 meter length, realistic office environment, conducted 6 times > m overall distance partially difficult surfaces (laser), partially few vertical edges (vision)
18 8 Videos videos (Pygmalion and Recsys)
19 9 Summary of the EKF Localization EKF = Extended Kalman Filter The ingredients of the EKF Localization problem are: Prediction of the next robot position X based on the previous position X- and the odometry input u- X f X, u ) [ X = N X, P ) ] = ( ( Statistical error model of the odometry is Gaussian with: Statistical observation model is Gaussian: The map is given! [ u = N( u, Q )] [ z = N h( x ), R )] ( EKF Equations: PREDICTION X = f ( X, u ) P = F x PF T x ESTIMATION T T X = X + P H [ H P H + R ][ z h ( X ) ] + F QF u T u P + + = P P H T [ ] T T H P H + R H P
20 5 20 Summary of the EKF Localization: Prediction Phase What is the state X? X = (X,Y,θ) What s the odometry input u and its covariance matrix Q? = L R s s Q α α 0 0 What s the odometry input u and its covariance matrix Q? u = ), ( L R S S α s 0 What is f? ), ( = u X f X s s s s b s s s s Y X Y X L R L R L R L R sin 2 cos 2 θ θ + + = b s s b Y Y L R 2 sin 2 θ θ θ
21 5 2 Summary of the EKF Localization: Observation Phase Let s assume that the map is made of points with world coordinates m i = ( xi, yi ) Let s assume that the robot uses a laser range finder which measures the range r and the bearing θ of each point in the robot frame. i i r The observation z is then z = i ϕ m 2 = ( x2, y2) ) And the predicted observation z = h(x ) m = x, ) m i = ( xi, yi ) ) i ) i r z = = h ( X ) i ϕ = 2 2 ( x X ) + ( y Y ) i y arctan xi i i Y X ( y ϕ ϕ i θ
22 22 Summary of the EKF Localization: Innovation The observation (magenta): z ) z = h(x) The predicted observation (green) = what the robot would observe if its position was the one estimated in the prediction phase ) z z The difference is called Innovation
23 23 Summary of the EKF Localization: Estimation During he Estimation phase the robot position and uncertainty are updated by fusing the information of the prediction phase with that from the observation (red)
24 24 Autonomous Map Building (SLAM) Starting from an arbitrary initial point, a mobile robot should be able to autonomously explore the environment with its on board sensors, gain nowledge about it, interpret the scene, build an appropriate map and localize itself relative to this map. SLAM The Simultaneous Localization and Mapping Problem
25 25 References for interested people Simultaneous Localization and Mapping: Part I by HUGH DURRANT- WHYTE and TIM BAILEY, Robotics and Automation Magazine, June, 2006 Simultaneous Localization and Mapping: Part II by TIM BAILEY and HUGH DURRANT-WHYTE, Robotics and Automation Magazine, September 2006
26 26 Map Building: How to Establish a Map. By Hand Automatically: Map Building The robot learns its environment Motivation: - by hand: hard and costly - dynamically changing environment - different loo due to different perception
27 27 Map Building: The Problems. Map Maintaining: Keeping trac of changes in the environment 2. Representation and Reduction of Uncertainty e.g. disappearing cupboard position of robot -> position of wall? position of wall -> position of robot - i.e. measure of belief of each environment feature probability densities for feature positions inconsistent map due to motion drift
28 28 Overview of the SLAM problem Overview Simultaneous localization and mapping (SLAM) is a technique used by robots and autonomous vehicles to build up a map within an unnown environment while at the same time eeping trac of their current position. This is not as straightforward as it might sound due to inherent uncertainties ti in discerning i the robot's relative movement from its various sensors. If at the next iteration of map building the measured distance and direction traveled has a slight inaccuracy, then any features being added to the map will contain corresponding errors. If uncheced, these positional errors build cumulatively, grossly distorting the map and therefore the robot's ability to now its precise location. There are various techniques to compensate for this such as recognizing i features that t it has come across previously and re-sewing recent parts of the map to mae sure the two instances of that feature become one. Some of the statistical techniques used in SLAM include Kalman filters, particle filters and matching of laser scan data or visual landmars. Sensors characteristics A sensor is characterized principally by:. Noise 2. Dimensionality of input a. Planar laser range finder (2D points) b. 3D laser range finder (3D point cloud) c. Camera features (2D coordinates (pixels) if single camera or 3D coordinates if stereo camera) 3. Frame of reference a. Laser/camera in robot frame b. GPS in earth coord. Frame c. Accelerometer/Gyros in inertial coord. frame
29 29 SLAM overview
30 30
31 3
32 32
33 33
34 34
35 EKF based SLAM 5 35 The framewor is the same we have seen in the localization problem and can be solved by introducing an extended state X! LOCALIZATION The state X is the ONLY robot SLAM The state X is the robot configuration The state X is the ONLY robot configuration, that is: X = (X,Y,θ) g (X,Y,θ) PLUS the features m i =(x i,y i ) in the map, that is: X = (X,Y,θ,x,y,,x 2,y 2,,x i,y i ) And the prediction function was And the Prediction Phase function is X f X ) ( + + b s s s s L R L R 2 cos 2 θ ), ( = u X f X = y x u X f y x X M M ), ( = b s s b s s s s b Y X Y X L R L R L R 2 sin θ θ θ i i i i y x y x M M
36 EKF based SLAM: Estimation Phase 5 36 The observation model is the same as the EKF localization: ( ) ( ) + = = = i i i i i i X x Y y Y y X x X h r z arctan ) ( 2 2 ϕ ) ) ), ( y x m = ) ( i x i X In principle, the estimation can now proceed in the same manner as a ), ( y x m = ), ( i i y xi m = proceed in the same manner as a conventional EKF. ϕ X θ ϕ i = = i y x m X X M M = O M M M L L L m m m m X m m m m m Xm Xm Xm XX P P P P P P P P P P i i i y x m O M M M
37 37 Summary of the EKF SLAM EKF = Extended Kalman Filter The ingredients of the EKF Localization problem are: Prediction of the next robot position X based on the previous position X- and the odometry input u- X = ( X, u ) X = N X P f [ (, )] [ u = N( u, Q )] [ z = N h( x ), R )] Statistical error model of the odometry is Gaussian with: Statistical observation model is Gaussian: The map is now given and is part of the state to X! ( EKF Equations: PREDICTION X = f ( X, u ) P = F x PF T x ESTIMATION T T X = X + P H [ H P H + R ][ z h ( X ) ] + F QF u T u P + + = P P H T [ ] T T H P H + R H P
38 4d 38 EKF based Based SLAM (ETH - ASL) 4d - Perception - Features corner features extracted from lines connecting corners to structures
39 5 39 Particle Filter SLAM (FastSLAM, Montemerlo 2002) (/2) FastSLAM approach It solves the SLAM problem using particle filters. Particle filters are mathematical models that represent probability distribution as a set of discrete particles which occupy the state space. Particle filter update In the update step, a new particle distribution is generated given the motion model and the controls applied. a) For each particle:. Compare particle s prediction of measurements with actual measurements 2. Particles whose predictions match the measurements are given a high weight b) Filter resample: Resample particles based on weight Filter resample Assign each particle a weight depending on how well its estimate of the state agrees with the measurements and randomly draw particles from previous distribution based on weights creating a new distribution.
40 5 40 Particle Filter SLAM (FastSLAM, Montemerlo 2002) (2/2) FastSLAM approach (continuation) Particle set The robot posterior is solved using a Rao Blacwellized particle filtering using landmars. Each landmar estimation is represented by a 2x2 EKF (Extended Kalman Filter). Each particle is independent (due the factorization) from the others and maintains the estimate of M landmar positions.
41 4 Cyclic Environments Small local error accumulate to arbitrary large global errors! This is usually irrelevant for navigation However, when closing loops, global error does matter Courtesy of Sebastian Thrun
42 42 Probabilistic Feature Based 3D Maps (ETH-ASL) raw 3D scan of the same scene raw data photo of the scene decompose space into grid cells fill cells with data find a plane for every cell using RANSAC one plane per grid cell fuse similar neighboring planes together segmented planar segments final segmentation
43 4d - Perception - Features 5 43 Example Result photograph of corridor at ASL raw 3D scan plane segmentation result extracted planes for every cube
44 44 Probabilistic Plane Based 3D SLAM (EKF based) close-up of a reconstructed hallway close-up of reconstructed booshelves The same experiment as before but this time planar segments are visualized and integrated into the estimation process
45 45 Dynamic Environments Dynamical changes require continuous mapping If extraction of high-level features would be possible, the mapping in dynamic environments would become significantly more straightforward. e.g. difference between human and wall Environment modeling is a ey factor for robustness?
46 46 Key Problems in SLAM One of the first ey problem of SLAM is the recognition of a place already visited In order to reduce the drift the robot should detect loops in the enviroments This tass can be solved by detecting distinctive features from the environments. Laser scan features can lead however to ambiguous situations (T junctions in corridors, etc.) In the last 5 years place recognition (useful for loop detection) algorithms using cameras have been developed, d which h use distinctive visual features (e.g. SIFT features, Harris, features, etc.). Indeed, images offer a much larger quantity of information compared to 2D and 3D laser. Also SLAM algorithms have been developed d which h only use a single camera! See next slides
47 Autonomous Mobile Robots Real Time 3D Visual SLAM with a Single Camera This presentation is based on the following papers: Andrew J. Davison, Real-Time Simultaneous Localization and Mapping with a Single Camera, ICCV 2003 Nicholas Molton, Andrew J. Davison and Ian Reid, Locally Planar Patch Features for Real-Time Structure from Motion, BMVC 2004 Zürich Autonomous Systems Lab
48 48 Structure From Motion (SFM) Structure from Motion: Tae some images of the object to reconstruct t Features (points, lines, ) are extracted from all frames and matched among them All images are processed simultaneously Both camera motion and 3D structure can be recovered by optimally fusing the overall information, up to a scale factor Robust solution but far from being Real- Time!
49 49 Visual SLAM (Video From Davison 03) It is Real-Time!
50 50 State of the System The State vector and the first-order uncertainty are: Where: Where r is the camera position r = (X,Y,Z) and q is the camera orientation which is represented using quatternion (4-element vector). v and w are the translational and rotational velocity! In the SLAM the Covariance matrix is not diagonal, that is, the uncertainty of any feature affects the position estimate of all other features and the camera
51 5 EKF based Visual SLAM The state vector X and the Covariance matrix P are updated during camera motion using an EKF: Prediction: a motion model is used to predict where the camera will be in the next time step (P increases) X = f ( X, u ) P = F PF + x T x F u QF Observation: a new feature is measured (P decreases) X + = X + P H T T T u [ T H P H + R][ z h( X )] [ + ] T T H P H R H P P = P P H + +?
52 52 Prediction Phase (/2): A Motion Model for Smoothly Moving Camera Prediction Phase Attempts to predict where the camera will be in the next time step In the case the camera is attached to a person, the unnown intentions of the person can be statistically modeled Davison uses a Constant Velocity Model!
53 53 Prediction Phase (2/2): Constant Velocity Model The unnown intentions, and so unnown accelerations, are taen into account by modeling the acceleration as a process of zero mean and Gaussian distribution: By setting the covariance matrix of n to small or large values, we define the smoothness or rapidity of the motion we expect. In practice these values were used: 4 m / s 6 rad / s
54 54 Localization: Predicted Observation (/2) Predicted Observation By predicting the next camera pose, we can predict where each features is going g liely to appear: z ) i u = = h( X ) v = where And the uncertainty of the predicted observation is:
55 55 Localization: Predicted Observation (2/2) At each frame, the features occurring at previous step are searched in the elliptic region where they are expected to be according to the motion model (Normalized Sum of Square Differences is used for matching) Once the features are matched, the entire state of the system is updated according to EKF
56 56 SLAM: Adding new features in the map The feature is firstly initialized as a 3D line Along this line, 00 possible feature positions are set uniformly in the range m A each time, each hypothesis is tested by projecting it into the image Each matching produces a lielihood for each hypothesis and their probabilities are recomputed
57 57 Improved Feature Matching Up to now, traced features were treated as 2D templates in image space Long-term tracing can be improved by approximating the feature as a locally planar region on 3D world surfaces
58 58 Locally Planar 3D Patch Features
59 59 References A. J. Davison, Real-Time Simultaneous Localization and Mapping with a Single Camera, ICCV 2003 N. Molton, A. J. Davison and I. Reid, Locally Planar Patch Features for Real-Time Structure from Motion, BMVC 2004 A. J. Davison, Y. Gonzalez Cid, N. Kita, Real-Time 3D SLAM with wide-angle vision, IFAC Symposium on Intelligent Autonomous Vehicles, 2004 J. Shi and C. Tomasi, Good features to trac, CVPR, 994.
60 Autonomous Mobile Robots Applications: i Current Wor on Visual 3D SLAM done at the ETH-ASL Zürich Autonomous Systems Lab
61 Omnidirectional Visual Odometry Assumptions Method 2D Planar Motion Calibrated omnidirectional camera SIFT and homography for the translation [Triggs 98] Appearance based method for rotation. [Scaramuzza, 2008] IEEE Trans. On Robotics, October [Scaramuzza, 2008] ECCV - OMNIVIS 08
62 Ground Tracing View View 2 h n Image coordinates of two different views of a plane are related by a Homography H λ x2 = H x T n h T For a calibrated camera, motion can be computed from H using the = Triggs algorithm [Triggs 98] H R +
63 Visual Odometry Results Error at the loop closure: 6.5 m Error of orientation: 5 deg Trajectory length: 400 m
64 Loop Detection and Closing
65 Visual Odometry Results Before Loop Closure Error at the loop closure: 6.5 m Error of orientation: 5 deg Trajectory length: 400 m After Loop Closure
66 SMART trajectory on Google Earth drawn manually from the images (NOT GPS)
67
68
69 Our 3D Models built from Point Clouds and Images Lowenstrasse, Zurich Scaramuzza, 2009
70 70 Current open student projects The previous slides show some of our current research in visual odometry, SLAM, and 3D modeling of city environments. If you are a motivated student looing for a Master, Bachelor, or Semester project contact Davide Scaramuzza
Localization and Map Building
Localization and Map Building Noise and aliasing; odometric position estimation To localize or not to localize Belief representation Map representation Probabilistic map-based localization Other examples
More informationLocalization and Map Building
Localization and Map Building Noise and aliasing; odometric position estimation To localize or not to localize Belief representation Map representation Probabilistic map-based localization Other examples
More informationUncertainties: Representation and Propagation & Line Extraction from Range data
41 Uncertainties: Representation and Propagation & Line Extraction from Range data 42 Uncertainty Representation Section 4.1.3 of the book Sensing in the real world is always uncertain How can uncertainty
More informationZürich. Roland Siegwart Margarita Chli Martin Rufli Davide Scaramuzza. ETH Master Course: L Autonomous Mobile Robots Summary
Roland Siegwart Margarita Chli Martin Rufli Davide Scaramuzza ETH Master Course: 151-0854-00L Autonomous Mobile Robots Summary 2 Lecture Overview Mobile Robot Control Scheme knowledge, data base mission
More informationAutonomous Mobile Robot Design
Autonomous Mobile Robot Design Topic: EKF-based SLAM Dr. Kostas Alexis (CSE) These slides have partially relied on the course of C. Stachniss, Robot Mapping - WS 2013/14 Autonomous Robot Challenges Where
More informationData Association for SLAM
CALIFORNIA INSTITUTE OF TECHNOLOGY ME/CS 132a, Winter 2011 Lab #2 Due: Mar 10th, 2011 Part I Data Association for SLAM 1 Introduction For this part, you will experiment with a simulation of an EKF SLAM
More informationPerception. Autonomous Mobile Robots. Sensors Vision Uncertainties, Line extraction from laser scans. Autonomous Systems Lab. Zürich.
Autonomous Mobile Robots Localization "Position" Global Map Cognition Environment Model Local Map Path Perception Real World Environment Motion Control Perception Sensors Vision Uncertainties, Line extraction
More informationParticle Filters. CSE-571 Probabilistic Robotics. Dependencies. Particle Filter Algorithm. Fast-SLAM Mapping
CSE-571 Probabilistic Robotics Fast-SLAM Mapping Particle Filters Represent belief by random samples Estimation of non-gaussian, nonlinear processes Sampling Importance Resampling (SIR) principle Draw
More informationRS-SLAM: RANSAC Sampling for Visual FastSLAM
2011 IEEE/RSJ International Conference on Intelligent Robots and Systems September 25-30, 2011. San Francisco, CA, USA RS-SLAM: RANSAC Sampling for Visual FastSLAM Gim Hee Lee, Friedrich Fraundorfer, and
More informationSpring Localization II. Roland Siegwart, Margarita Chli, Martin Rufli. ASL Autonomous Systems Lab. Autonomous Mobile Robots
Spring 2016 Localization II Localization I 25.04.2016 1 knowledge, data base mission commands Localization Map Building environment model local map position global map Cognition Path Planning path Perception
More informationWhat is the SLAM problem?
SLAM Tutorial Slides by Marios Xanthidis, C. Stachniss, P. Allen, C. Fermuller Paul Furgale, Margarita Chli, Marco Hutter, Martin Rufli, Davide Scaramuzza, Roland Siegwart What is the SLAM problem? The
More informationRobot Simultaneous Localization and Mapping Based on Self-Detected Waypoint
BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 16, No 2 Sofia 2016 Print ISSN: 1311-9702; Online ISSN: 1314-4081 DOI: 10.1515/cait-2016-0031 Robot Simultaneous Localization
More informationCOS Lecture 13 Autonomous Robot Navigation
COS 495 - Lecture 13 Autonomous Robot Navigation Instructor: Chris Clark Semester: Fall 2011 1 Figures courtesy of Siegwart & Nourbakhsh Control Structure Prior Knowledge Operator Commands Localization
More informationProbabilistic Robotics
Probabilistic Robotics FastSLAM Sebastian Thrun (abridged and adapted by Rodrigo Ventura in Oct-2008) The SLAM Problem SLAM stands for simultaneous localization and mapping The task of building a map while
More informationRobotics. Lecture 7: Simultaneous Localisation and Mapping (SLAM)
Robotics Lecture 7: Simultaneous Localisation and Mapping (SLAM) See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information. Andrew Davison Department of Computing Imperial College
More informationSurvey: Simultaneous Localisation and Mapping (SLAM) Ronja Güldenring Master Informatics Project Intellgient Robotics University of Hamburg
Survey: Simultaneous Localisation and Mapping (SLAM) Ronja Güldenring Master Informatics Project Intellgient Robotics University of Hamburg Introduction EKF-SLAM FastSLAM Loop Closure 01.06.17 Ronja Güldenring
More informationProbabilistic Robotics
Probabilistic Robotics Bayes Filter Implementations Discrete filters, Particle filters Piecewise Constant Representation of belief 2 Discrete Bayes Filter Algorithm 1. Algorithm Discrete_Bayes_filter(
More informationAccurate Motion Estimation and High-Precision 3D Reconstruction by Sensor Fusion
007 IEEE International Conference on Robotics and Automation Roma, Italy, 0-4 April 007 FrE5. Accurate Motion Estimation and High-Precision D Reconstruction by Sensor Fusion Yunsu Bok, Youngbae Hwang,
More informationHumanoid Robotics. Monte Carlo Localization. Maren Bennewitz
Humanoid Robotics Monte Carlo Localization Maren Bennewitz 1 Basis Probability Rules (1) If x and y are independent: Bayes rule: Often written as: The denominator is a normalizing constant that ensures
More informationParticle Filter in Brief. Robot Mapping. FastSLAM Feature-based SLAM with Particle Filters. Particle Representation. Particle Filter Algorithm
Robot Mapping FastSLAM Feature-based SLAM with Particle Filters Cyrill Stachniss Particle Filter in Brief! Non-parametric, recursive Bayes filter! Posterior is represented by a set of weighted samples!
More informationRevising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History
Revising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History Simon Thompson and Satoshi Kagami Digital Human Research Center National Institute of Advanced
More informationSpring Localization II. Roland Siegwart, Margarita Chli, Juan Nieto, Nick Lawrance. ASL Autonomous Systems Lab. Autonomous Mobile Robots
Spring 2018 Localization II Localization I 16.04.2018 1 knowledge, data base mission commands Localization Map Building environment model local map position global map Cognition Path Planning path Perception
More informationPractical Course WS12/13 Introduction to Monte Carlo Localization
Practical Course WS12/13 Introduction to Monte Carlo Localization Cyrill Stachniss and Luciano Spinello 1 State Estimation Estimate the state of a system given observations and controls Goal: 2 Bayes Filter
More informationVisual Bearing-Only Simultaneous Localization and Mapping with Improved Feature Matching
Visual Bearing-Only Simultaneous Localization and Mapping with Improved Feature Matching Hauke Strasdat, Cyrill Stachniss, Maren Bennewitz, and Wolfram Burgard Computer Science Institute, University of
More informationProbabilistic Robotics. FastSLAM
Probabilistic Robotics FastSLAM The SLAM Problem SLAM stands for simultaneous localization and mapping The task of building a map while estimating the pose of the robot relative to this map Why is SLAM
More informationOverview. EECS 124, UC Berkeley, Spring 2008 Lecture 23: Localization and Mapping. Statistical Models
Introduction ti to Embedded dsystems EECS 124, UC Berkeley, Spring 2008 Lecture 23: Localization and Mapping Gabe Hoffmann Ph.D. Candidate, Aero/Astro Engineering Stanford University Statistical Models
More informationRobotics. Lecture 8: Simultaneous Localisation and Mapping (SLAM)
Robotics Lecture 8: Simultaneous Localisation and Mapping (SLAM) See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information. Andrew Davison Department of Computing Imperial College
More informationAugmented Reality, Advanced SLAM, Applications
Augmented Reality, Advanced SLAM, Applications Prof. Didier Stricker & Dr. Alain Pagani alain.pagani@dfki.de Lecture 3D Computer Vision AR, SLAM, Applications 1 Introduction Previous lectures: Basics (camera,
More informationProbabilistic Robotics
Probabilistic Robotics Sebastian Thrun Wolfram Burgard Dieter Fox The MIT Press Cambridge, Massachusetts London, England Preface xvii Acknowledgments xix I Basics 1 1 Introduction 3 1.1 Uncertainty in
More informationRobot Localization based on Geo-referenced Images and G raphic Methods
Robot Localization based on Geo-referenced Images and G raphic Methods Sid Ahmed Berrabah Mechanical Department, Royal Military School, Belgium, sidahmed.berrabah@rma.ac.be Janusz Bedkowski, Łukasz Lubasiński,
More informationSimultaneous Localization and Mapping
Sebastian Lembcke SLAM 1 / 29 MIN Faculty Department of Informatics Simultaneous Localization and Mapping Visual Loop-Closure Detection University of Hamburg Faculty of Mathematics, Informatics and Natural
More informationOffline Simultaneous Localization and Mapping (SLAM) using Miniature Robots
Offline Simultaneous Localization and Mapping (SLAM) using Miniature Robots Objectives SLAM approaches SLAM for ALICE EKF for Navigation Mapping and Network Modeling Test results Philipp Schaer and Adrian
More information7630 Autonomous Robotics Probabilities for Robotics
7630 Autonomous Robotics Probabilities for Robotics Basics of probability theory The Bayes-Filter Introduction to localization problems Monte-Carlo-Localization Based on material from R. Triebel, R. Kästner
More informationImage Augmented Laser Scan Matching for Indoor Localization
Image Augmented Laser Scan Matching for Indoor Localization Nikhil Naikal Avideh Zakhor John Kua Electrical Engineering and Computer Sciences University of California at Berkeley Technical Report No. UCB/EECS-2009-35
More informationIROS 05 Tutorial. MCL: Global Localization (Sonar) Monte-Carlo Localization. Particle Filters. Rao-Blackwellized Particle Filters and Loop Closing
IROS 05 Tutorial SLAM - Getting it Working in Real World Applications Rao-Blackwellized Particle Filters and Loop Closing Cyrill Stachniss and Wolfram Burgard University of Freiburg, Dept. of Computer
More informationSensor Modalities. Sensor modality: Different modalities:
Sensor Modalities Sensor modality: Sensors which measure same form of energy and process it in similar ways Modality refers to the raw input used by the sensors Different modalities: Sound Pressure Temperature
More informationImplementation of Odometry with EKF for Localization of Hector SLAM Method
Implementation of Odometry with EKF for Localization of Hector SLAM Method Kao-Shing Hwang 1 Wei-Cheng Jiang 2 Zuo-Syuan Wang 3 Department of Electrical Engineering, National Sun Yat-sen University, Kaohsiung,
More informationNonlinear State Estimation for Robotics and Computer Vision Applications: An Overview
Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview Arun Das 05/09/2017 Arun Das Waterloo Autonomous Vehicles Lab Introduction What s in a name? Arun Das Waterloo Autonomous
More informationEE565:Mobile Robotics Lecture 3
EE565:Mobile Robotics Lecture 3 Welcome Dr. Ahmad Kamal Nasir Today s Objectives Motion Models Velocity based model (Dead-Reckoning) Odometry based model (Wheel Encoders) Sensor Models Beam model of range
More informationCSE 527: Introduction to Computer Vision
CSE 527: Introduction to Computer Vision Week 10 Class 2: Visual Odometry November 2nd, 2017 Today Visual Odometry Intro Algorithm SLAM Visual Odometry Input Output Images, Video Camera trajectory, motion
More informationME 597/747 Autonomous Mobile Robots. Mid Term Exam. Duration: 2 hour Total Marks: 100
ME 597/747 Autonomous Mobile Robots Mid Term Exam Duration: 2 hour Total Marks: 100 Instructions: Read the exam carefully before starting. Equations are at the back, but they are NOT necessarily valid
More informationRobot Mapping. A Short Introduction to the Bayes Filter and Related Models. Gian Diego Tipaldi, Wolfram Burgard
Robot Mapping A Short Introduction to the Bayes Filter and Related Models Gian Diego Tipaldi, Wolfram Burgard 1 State Estimation Estimate the state of a system given observations and controls Goal: 2 Recursive
More informationIntroduction to Mobile Robotics SLAM Grid-based FastSLAM. Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Diego Tipaldi, Luciano Spinello
Introduction to Mobile Robotics SLAM Grid-based FastSLAM Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Diego Tipaldi, Luciano Spinello 1 The SLAM Problem SLAM stands for simultaneous localization
More informationRobotics. Lecture 5: Monte Carlo Localisation. See course website for up to date information.
Robotics Lecture 5: Monte Carlo Localisation See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information. Andrew Davison Department of Computing Imperial College London Review:
More informationBasics of Localization, Mapping and SLAM. Jari Saarinen Aalto University Department of Automation and systems Technology
Basics of Localization, Mapping and SLAM Jari Saarinen Aalto University Department of Automation and systems Technology Content Introduction to Problem (s) Localization A few basic equations Dead Reckoning
More informationSLAM WITH KLT POINT FEATURES. P.O. Box 22, Thammasat-Rangsit Post Office, Pathumthani Thailand
SLAM WITH KLT POINT FEATURES Yoichi Nakaguro, 1 Matthew Dailey, Stanislav Makhanov 1 1 Sirindhorn International Institute of Technology, Thammasat University P.O. Box, Thammasat-Rangsit Post Office, Pathumthani
More informationL17. OCCUPANCY MAPS. NA568 Mobile Robotics: Methods & Algorithms
L17. OCCUPANCY MAPS NA568 Mobile Robotics: Methods & Algorithms Today s Topic Why Occupancy Maps? Bayes Binary Filters Log-odds Occupancy Maps Inverse sensor model Learning inverse sensor model ML map
More informationSegmentation and Tracking of Partial Planar Templates
Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract
More informationUSING 3D DATA FOR MONTE CARLO LOCALIZATION IN COMPLEX INDOOR ENVIRONMENTS. Oliver Wulf, Bernardo Wagner
USING 3D DATA FOR MONTE CARLO LOCALIZATION IN COMPLEX INDOOR ENVIRONMENTS Oliver Wulf, Bernardo Wagner Institute for Systems Engineering (RTS/ISE), University of Hannover, Germany Mohamed Khalaf-Allah
More informationSimuntaneous Localisation and Mapping with a Single Camera. Abhishek Aneja and Zhichao Chen
Simuntaneous Localisation and Mapping with a Single Camera Abhishek Aneja and Zhichao Chen 3 December, Simuntaneous Localisation and Mapping with asinglecamera 1 Abstract Image reconstruction is common
More informationVisual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech
Visual Odometry Features, Tracking, Essential Matrix, and RANSAC Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline The
More informationPedestrian Detection Using Multi-layer LIDAR
1 st International Conference on Transportation Infrastructure and Materials (ICTIM 2016) ISBN: 978-1-60595-367-0 Pedestrian Detection Using Multi-layer LIDAR Mingfang Zhang 1, Yuping Lu 2 and Tong Liu
More informationRobotic Perception and Action: Vehicle SLAM Assignment
Robotic Perception and Action: Vehicle SLAM Assignment Mariolino De Cecco Mariolino De Cecco, Mattia Tavernini 1 CONTENTS Vehicle SLAM Assignment Contents Assignment Scenario 3 Odometry Localization...........................................
More informationVision-based Mobile Robot Localization and Mapping using Scale-Invariant Features
Vision-based Mobile Robot Localization and Mapping using Scale-Invariant Features Stephen Se, David Lowe, Jim Little Department of Computer Science University of British Columbia Presented by Adam Bickett
More informationIntroduction to Mobile Robotics SLAM Landmark-based FastSLAM
Introduction to Mobile Robotics SLAM Landmark-based FastSLAM Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Diego Tipaldi, Luciano Spinello Partial slide courtesy of Mike Montemerlo 1 The SLAM Problem
More informationFinal Exam Practice Fall Semester, 2012
COS 495 - Autonomous Robot Navigation Final Exam Practice Fall Semester, 2012 Duration: Total Marks: 70 Closed Book 2 hours Start Time: End Time: By signing this exam, I agree to the honor code Name: Signature:
More informationIntroduction to Autonomous Mobile Robots
Introduction to Autonomous Mobile Robots second edition Roland Siegwart, Illah R. Nourbakhsh, and Davide Scaramuzza The MIT Press Cambridge, Massachusetts London, England Contents Acknowledgments xiii
More information3D Environment Reconstruction
3D Environment Reconstruction Using Modified Color ICP Algorithm by Fusion of a Camera and a 3D Laser Range Finder The 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems October 11-15,
More informationECE276A: Sensing & Estimation in Robotics Lecture 11: Simultaneous Localization and Mapping using a Particle Filter
ECE276A: Sensing & Estimation in Robotics Lecture 11: Simultaneous Localization and Mapping using a Particle Filter Lecturer: Nikolay Atanasov: natanasov@ucsd.edu Teaching Assistants: Siwei Guo: s9guo@eng.ucsd.edu
More informationAn Application of Omnidirectional Vision to Grid-based SLAM in Indoor Environments
An Application of Omnidirectional Vision to Grid-based SLAM in Indoor Environments Vito Macchia, Stefano Rosa, Luca Carlone, Basilio Bona Abstract In this wor we study the use of an omnidirectional camera
More informationMonte Carlo Localization using Dynamically Expanding Occupancy Grids. Karan M. Gupta
1 Monte Carlo Localization using Dynamically Expanding Occupancy Grids Karan M. Gupta Agenda Introduction Occupancy Grids Sonar Sensor Model Dynamically Expanding Occupancy Grids Monte Carlo Localization
More informationVehicle Localization. Hannah Rae Kerner 21 April 2015
Vehicle Localization Hannah Rae Kerner 21 April 2015 Spotted in Mtn View: Google Car Why precision localization? in order for a robot to follow a road, it needs to know where the road is to stay in a particular
More informationOdometry-based Online Extrinsic Sensor Calibration
213 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) November 3-7, 213. Toyo, Japan Odometry-based Online Extrinsic Sensor Calibration Sebastian Schneider, Thorsten Luettel and
More informationRobot Mapping. SLAM Front-Ends. Cyrill Stachniss. Partial image courtesy: Edwin Olson 1
Robot Mapping SLAM Front-Ends Cyrill Stachniss Partial image courtesy: Edwin Olson 1 Graph-Based SLAM Constraints connect the nodes through odometry and observations Robot pose Constraint 2 Graph-Based
More informationMonocular SLAM for a Small-Size Humanoid Robot
Tamkang Journal of Science and Engineering, Vol. 14, No. 2, pp. 123 129 (2011) 123 Monocular SLAM for a Small-Size Humanoid Robot Yin-Tien Wang*, Duen-Yan Hung and Sheng-Hsien Cheng Department of Mechanical
More informationPML-SLAM: a solution for localization in large-scale urban environments
PML-SLAM: a solution for localization in large-scale urban environments Zayed Alsayed, Guillaume Bresson, Fawzi Nashashibi and Anne Verroust-Blondet Institut VEDECOM Versailles, France firstname.name@vedecom.fr
More informationIntroduction to SLAM Part II. Paul Robertson
Introduction to SLAM Part II Paul Robertson Localization Review Tracking, Global Localization, Kidnapping Problem. Kalman Filter Quadratic Linear (unless EKF) SLAM Loop closing Scaling: Partition space
More informationLOCAL AND GLOBAL DESCRIPTORS FOR PLACE RECOGNITION IN ROBOTICS
8th International DAAAM Baltic Conference "INDUSTRIAL ENGINEERING - 19-21 April 2012, Tallinn, Estonia LOCAL AND GLOBAL DESCRIPTORS FOR PLACE RECOGNITION IN ROBOTICS Shvarts, D. & Tamre, M. Abstract: The
More informationLecture 13 Visual Inertial Fusion
Lecture 13 Visual Inertial Fusion Davide Scaramuzza Course Evaluation Please fill the evaluation form you received by email! Provide feedback on Exercises: good and bad Course: good and bad How to improve
More informationNAVIGATION SYSTEM OF AN OUTDOOR SERVICE ROBOT WITH HYBRID LOCOMOTION SYSTEM
NAVIGATION SYSTEM OF AN OUTDOOR SERVICE ROBOT WITH HYBRID LOCOMOTION SYSTEM Jorma Selkäinaho, Aarne Halme and Janne Paanajärvi Automation Technology Laboratory, Helsinki University of Technology, Espoo,
More informationIntegrated Laser-Camera Sensor for the Detection and Localization of Landmarks for Robotic Applications
28 IEEE International Conference on Robotics and Automation Pasadena, CA, USA, May 19-23, 28 Integrated Laser-Camera Sensor for the Detection and Localization of Landmarks for Robotic Applications Dilan
More informationLocalization, Where am I?
5.1 Localization, Where am I?? position Position Update (Estimation?) Encoder Prediction of Position (e.g. odometry) YES matched observations Map data base predicted position Matching Odometry, Dead Reckoning
More informationGeneralizing Random-Vector SLAM with Random Finite Sets
Generalizing Random-Vector SLAM with Random Finite Sets Keith Y. K. Leung, Felipe Inostroza, Martin Adams Advanced Mining Technology Center AMTC, Universidad de Chile, Santiago, Chile eith.leung@amtc.uchile.cl,
More informationMobile Robots Summery. Autonomous Mobile Robots
Mobile Robots Summery Roland Siegwart Mike Bosse, Marco Hutter, Martin Rufli, Davide Scaramuzza, (Margarita Chli, Paul Furgale) Mobile Robots Summery 1 Introduction probabilistic map-based localization
More informationOn the Use of Inverse Scaling in Monocular SLAM
29 IEEE International Conference on Robotics and Automation Kobe International Conference Center Kobe, Japan, May 12-17, 29 On the Use of Inverse Scaling in Monocular SLAM Daniele Marzorati, Matteo Matteucci,
More informationDealing with Scale. Stephan Weiss Computer Vision Group NASA-JPL / CalTech
Dealing with Scale Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline Why care about size? The IMU as scale provider: The
More informationStable Vision-Aided Navigation for Large-Area Augmented Reality
Stable Vision-Aided Navigation for Large-Area Augmented Reality Taragay Oskiper, Han-Pang Chiu, Zhiwei Zhu Supun Samarasekera, Rakesh Teddy Kumar Vision and Robotics Laboratory SRI-International Sarnoff,
More informationIntroduction to Mobile Robotics. SLAM: Simultaneous Localization and Mapping
Introduction to Mobile Robotics SLAM: Simultaneous Localization and Mapping The SLAM Problem SLAM is the process by which a robot builds a map of the environment and, at the same time, uses this map to
More informationDavide Scaramuzza. University of Zurich
Davide Scaramuzza University of Zurich Robotics and Perception Group http://rpg.ifi.uzh.ch/ Scaramuzza, D., Fraundorfer, F., Visual Odometry: Part I - The First 30 Years and Fundamentals, IEEE Robotics
More informationIntroduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization. Wolfram Burgard
Introduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization Wolfram Burgard 1 Motivation Recall: Discrete filter Discretize the continuous state space High memory complexity
More informationF1/10 th Autonomous Racing. Localization. Nischal K N
F1/10 th Autonomous Racing Localization Nischal K N System Overview Mapping Hector Mapping Localization Path Planning Control System Overview Mapping Hector Mapping Localization Adaptive Monte Carlo Localization
More informationCSE 490R P1 - Localization using Particle Filters Due date: Sun, Jan 28-11:59 PM
CSE 490R P1 - Localization using Particle Filters Due date: Sun, Jan 28-11:59 PM 1 Introduction In this assignment you will implement a particle filter to localize your car within a known map. This will
More informationEfficient SLAM Scheme Based ICP Matching Algorithm Using Image and Laser Scan Information
Proceedings of the World Congress on Electrical Engineering and Computer Systems and Science (EECSS 2015) Barcelona, Spain July 13-14, 2015 Paper No. 335 Efficient SLAM Scheme Based ICP Matching Algorithm
More informationProbabilistic Robotics
Probabilistic Robotics Discrete Filters and Particle Filters Models Some slides adopted from: Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Kai Arras and Probabilistic Robotics Book SA-1 Probabilistic
More informationFull Scaled 3D Visual Odometry from a Single Wearable Omnidirectional Camera
Full Scaled 3D Visual Odometry from a Single Wearable Omnidirectional Camera Daniel Gutiérrez-Gómez, Luis Puig, J.J. Guerrero Aragon Institute of Engineering Research (I3A) University of Zaragoza, Mariano
More informationMobile Robotics. Mathematics, Models, and Methods. HI Cambridge. Alonzo Kelly. Carnegie Mellon University UNIVERSITY PRESS
Mobile Robotics Mathematics, Models, and Methods Alonzo Kelly Carnegie Mellon University HI Cambridge UNIVERSITY PRESS Contents Preface page xiii 1 Introduction 1 1.1 Applications of Mobile Robots 2 1.2
More informationParticle Filter for Robot Localization ECE 478 Homework #1
Particle Filter for Robot Localization ECE 478 Homework #1 Phil Lamb pjl@pdx.edu November 15, 2012 1 Contents 1 Introduction 3 2 Implementation 3 2.1 Assumptions and Simplifications.............................
More informationBuilding Reliable 2D Maps from 3D Features
Building Reliable 2D Maps from 3D Features Dipl. Technoinform. Jens Wettach, Prof. Dr. rer. nat. Karsten Berns TU Kaiserslautern; Robotics Research Lab 1, Geb. 48; Gottlieb-Daimler- Str.1; 67663 Kaiserslautern;
More informationHigh-precision, consistent EKF-based visual-inertial odometry
High-precision, consistent EKF-based visual-inertial odometry Mingyang Li and Anastasios I. Mourikis, IJRR 2013 Ao Li Introduction What is visual-inertial odometry (VIO)? The problem of motion tracking
More informationIndoor Localization Algorithms for a Human-Operated Backpack System
Indoor Localization Algorithms for a Human-Operated Backpack System George Chen, John Kua, Stephen Shum, Nikhil Naikal, Matthew Carlberg, Avideh Zakhor Video and Image Processing Lab, University of California,
More informationArtificial Intelligence for Robotics: A Brief Summary
Artificial Intelligence for Robotics: A Brief Summary This document provides a summary of the course, Artificial Intelligence for Robotics, and highlights main concepts. Lesson 1: Localization (using Histogram
More informationProceedings of ESDA2006 8th Biennial ASME Conference on Engineering Systems Design and Analysis July 4-7, 2006, Torino, Italy
Proceedings of ESDA2006 8th Biennial ASME Conference on Engineering Systems Design and Analysis July 4-7, 2006, Torino, Italy ESDA2006-95412 LOCALIZATION OF A MOBILE ROBOT WITH OMNIDIRECTIONAL WHEELS USING
More informationCAMERA POSE ESTIMATION OF RGB-D SENSORS USING PARTICLE FILTERING
CAMERA POSE ESTIMATION OF RGB-D SENSORS USING PARTICLE FILTERING By Michael Lowney Senior Thesis in Electrical Engineering University of Illinois at Urbana-Champaign Advisor: Professor Minh Do May 2015
More information5. Tests and results Scan Matching Optimization Parameters Influence
126 5. Tests and results This chapter presents results obtained using the proposed method on simulated and real data. First, it is analyzed the scan matching optimization; after that, the Scan Matching
More informationGeometrical Feature Extraction Using 2D Range Scanner
Geometrical Feature Extraction Using 2D Range Scanner Sen Zhang Lihua Xie Martin Adams Fan Tang BLK S2, School of Electrical and Electronic Engineering Nanyang Technological University, Singapore 639798
More informationThis chapter explains two techniques which are frequently used throughout
Chapter 2 Basic Techniques This chapter explains two techniques which are frequently used throughout this thesis. First, we will introduce the concept of particle filters. A particle filter is a recursive
More informationBBR Progress Report 006: Autonomous 2-D Mapping of a Building Floor
BBR Progress Report 006: Autonomous 2-D Mapping of a Building Floor Andy Sayler & Constantin Berzan November 30, 2010 Abstract In the past two weeks, we implemented and tested landmark extraction based
More informationTutorial on 3D Surface Reconstruction in Laparoscopic Surgery. Simultaneous Localization and Mapping for Minimally Invasive Surgery
Tutorial on 3D Surface Reconstruction in Laparoscopic Surgery Simultaneous Localization and Mapping for Minimally Invasive Surgery Introduction University of Bristol using particle filters to track football
More informationLocalization Assistance for Multiple Mobile Robots using Target Tracking Techniques
Localization Assistance for Multiple Mobile Robots using Target Tracing Techniques Peshala Jayaseara a, Hidei Hashimoto b and Taashi Kubota c a Dept. of Electrical Engineering, University of Toyo, Toyo,
More informationLocalization of Multiple Robots with Simple Sensors
Proceedings of the IEEE International Conference on Mechatronics & Automation Niagara Falls, Canada July 2005 Localization of Multiple Robots with Simple Sensors Mike Peasgood and Christopher Clark Lab
More information