Motion estimation of unmanned marine vehicles Massimo Caccia
|
|
- Adelia Dalton
- 5 years ago
- Views:
Transcription
1 Motion estimation of unmanned marine vehicles Massimo Caccia Consiglio Nazionale delle Ricerche Istituto di Studi sui Sistemi Intelligenti per l Automazione Via Amendola 122 D/O, 70126, Bari, Italy massimo.caccia@ge.issia.cnr.it
2 Unmanned Surface Vehicles Onboard sensors position Configuration 1 GPS Ashtech GG24C heading GPS, DGPS Charlie USV DGPS position & 1 Hz Compass compass KVH Azimuth Gyrotrack True North, pitch & 2 Hz speed GPS, DVL
3 USV motion equations USV longitudinal asimmetry effects sea current
4 USV motion estimation scheme
5 USV motion estimation: basic issues and tricks (1) model-based yaw motion estimation using extended Kalman filter model-based prediction of angular acceleration is used as a kind of virtual sensor the delay of the estimate is reduced improving the stability of closedloop heading system GPS signal could present local discontinuities practical approach modeling a piecewise constant measurement disturbance the Guidance & Control module needs to deal with a continuous estimate of the vehicle motion
6 USV motion estimation: basic issues and tricks (2 yaw motion estimation) angles are measured in radians or degrees in a 2p or 360 deg interval, typically (-p, p] rad or (-180, 180] deg REMEMBER this fact when executing operations with angles
7 USV motion estimation: basic issues and tricks (3 GPS discontinuity) piecewise constant GPS measurement disturbance S(t) : step function x, y : disturbance step amplitudes measurement equations
8 USV motion estimation: basic issues and tricks (4 GPS discontinuity)
9 USV motion estimation: results (yaw motion)
10 USV motion estimation: results (x-y motion)
11 ROV motion estimation Vertical motion estimation depth estimation altitude estimation Horizontal motion estimation dead-reckoning vision-based motion estimation, including SLAM (basics)
12 Optical vision ROV downward-looking camera
13 Camera sensor model f : focal length [m n] : image point in the image plane [ m, n ] : image motion field in the image plane [X Y Z] : coordinates of the generic point in the 3-D space (referred to the camera frame) Camera perspective model [] [ ] m=f X n Z Y
14 Motion field of a stationary scene point Speed of a generic 3-D point in the camera frame [ ] [ ] [ ][ ] 0 r q X X u 0 p Y Y = v r q p 0 Z w Z Motion field [] [] [ ] [ ] [ ] [] m = f u w m r n f q pn qm m Z v Z n f n m p n neglecting terms of higher order [] [] [ ] [ ] [ ] m = f u w m r n f q Z v Z n n m p
15 Motion from token displacements Motion field neglecting pitch and roll [] [] [ ] [ ] m = f u w m r n Z v Z n n m Normalising with respect to the scene depth Z, assumed constant u v w u = ; v = ; w= Z Z Z when N tokens are tracked [ f 0 f 0 0 f 0 f [] ] [ ] m1 n1 m 1 n1 m1 u n 1 v = w mn nn m N r n N n N m N and then... Least Squares
16 Open questions How to define and extract tokens from images of the seabed? How to track tokens? How to compute the image depth?
17 Image token definition Definition of an image token in an unstructured environment region with high local variance
18 Image token tracking Cross-correlation Measure of similarity between two waveforms as a function of a time-lag applied to one of them m= corr f, g [ n ] = m= f [ m ] g [ n m ] where f is the complex conjugate of f When the functions f and g match, the value of their correlation is maximized. The reason for this is that when lumps (positives areas) are aligned, they contribute to making the sum larger. Also, when the troughs (negative areas) align, they also make a positive contribution to the sum because the product of two negative numbers is positive. Normalised cross-correlation N: number of pixels in template t 1 x, y f x, y f t x, y t corr f x, y,t x, y = n 1 f t
19 Image depth estimation Stereo vision Acoustic echo-sounder time-of-flight sensor Laser spot tracking
20 Laser spot tracking instrument
21 Laser spot tracking instrument
22 Laser spot position estimation ll a lx Z ll Z ll X ll = X Laser ray equations (l=1..nl) Y ll =Y ll a Yl Z ll Z ll Laser ray parameters are estimated during the device calibration Set of measurements at different measured ranges and Least Squares Once detected the image laser spots, combining laser ray equations and camera perspective models, laser spot depths are estimated with Least Squares Z = L l L m ll n ll l a Yl Y ll a lx X f f 2 L m ll n a lx l a Yl f f 2,l =1..N L
23 Seabed model Z = tan cos X tan sin Y h : seabed slope : seabed maximum slope orientation Small tracked image: seabed is assumed locally flat Z=h
24 Vision-based motion estimation architecture
25 Experimental results: token detection & tracking
26 Experimental results: estimated surge & pitch rate (from inclinometers)
27 SLAM Simultaneous Localization And Mapping Problem formulation The ability for a robotic vehicle to move in an unknown environment building a map, using only relative observations of the environment, and simultaneously using this map to navigate Classical extended Kalman filter solution Dissanayake et al. (2001) A solution to the simultaneous localization and map building (SLAM) problem. IEEE Transactions on Robotics and Automation (2001)
28 SLAM state equations vehicle model [ [ ][ ] ] x k 1 = x k R k u k u k t y k 1 y k v k v k landmark model [ ][ ] X i k 1 X k = i Y i k 1 Y i k Classical SLAM kinematic model Vehicle state: position Control input: horizontal speed and heading The encapsulation of the uncertainty of the vehicle speed in the control input allows its projection on the vehicle motion direction avoiding a generic, overestimated, system noise on the vehicle Cartesian coordinates
29 SLAM observation model z i =hi x, y,, X i, Y i The estimation-theoretical formulation of SLAM relies on the assumption that the vector of control input and observations are independent [ ][ ] [ ] x k 1 = x k R k u k u k t y k 1 y k v k v k
30 SLAM estimation & data association Estimation Kalman filter according to the theoretic formulation of the SLAM problem (Dissanayake et al., 2001) Data association Associating an existing landmark or a new landmark or nothing to a measurement MAIN ISSUE in the case of error in data association the vehicle is lost
31 Vision-based SLAM architecture
32 Data association & estimation Estimation Kalman filter according to the theoretic formulation of the SLAM problem (Dissanayake et al., 2001) Data association Searching in the actual image any landmark template that could be observed by the camera Template search, based on correlation, is performed only if the landmark is expected to be in the camera field of view
33 Processing steps Laser spots extraction/tracking and image depth computing Inhibition of laser spot neighbourhoods for template extraction and tracking SLAM template tracking. Searching new landmark if no one is tracked and the vehicle is far from the existing ones Inhibition of SLAM template neighbourhood for speed template extraction and tracking Speed template tracking and speed computing SLAM prediction, observation and update SLAM state augmentation, if needed Speed template extraction
34 Observation model [ ] [ ][ ] mi = f RT X i x m Z ni Y i y n SLAM landmark position measurement [mi, ni] is provided by the SLAM Optical Template Detector & Tracker SLAM control input [ui, vi] is provided by the Speed Estimate Optical Template Detector & Tracker The estimation-theoretical formulation of SLAM relies on the assumption that the vector of control input and observations are independent No overlapping between landmark and speed estimate image templates
35 Experimental results Romeo ROV & SLAM
36 Experimental results predicted & SLAM estimated position
37 Experimental results estimated position, covariance & tracked templates
38 Experimental results tracked landmarks
39 Experimental results landmark 0 tracking
40 Ground truth ROV configuration does not include high precision DVL and Gyros or high frequency LBL Ground truth verification procedure Select a couple of images Tracking a set of 16 x 16 pixels templates in the two images Compute the corresponding displacement Compare the result with SLAM-computed displacement Results: more than 1 cm accuracy when a visual landmark is tracked
41 What can happen with acoustics... Exploration of the underwater part of the Campbell glacier ice tongue with the Romeo ROV (1998)
42 Romeo ROV configuration upward-looking sonar forward-looking sonar
43 Romeo ROV telemetry data
44 Acoustics is not random, it's physics (M. Cardew)
45 Acoustic multi-path approaching the ice tongue
SLAM with SIFT (aka Mobile Robot Localization and Mapping with Uncertainty using Scale-Invariant Visual Landmarks ) Se, Lowe, and Little
SLAM with SIFT (aka Mobile Robot Localization and Mapping with Uncertainty using Scale-Invariant Visual Landmarks ) Se, Lowe, and Little + Presented by Matt Loper CS296-3: Robot Learning and Autonomy Brown
More informationMobile Robotics. Mathematics, Models, and Methods. HI Cambridge. Alonzo Kelly. Carnegie Mellon University UNIVERSITY PRESS
Mobile Robotics Mathematics, Models, and Methods Alonzo Kelly Carnegie Mellon University HI Cambridge UNIVERSITY PRESS Contents Preface page xiii 1 Introduction 1 1.1 Applications of Mobile Robots 2 1.2
More informationData Association for SLAM
CALIFORNIA INSTITUTE OF TECHNOLOGY ME/CS 132a, Winter 2011 Lab #2 Due: Mar 10th, 2011 Part I Data Association for SLAM 1 Introduction For this part, you will experiment with a simulation of an EKF SLAM
More informationDYNAMIC POSITIONING CONFERENCE September 16-17, Sensors
DYNAMIC POSITIONING CONFERENCE September 16-17, 2003 Sensors An Integrated acoustic positioning and inertial navigation system Jan Erik Faugstadmo, Hans Petter Jacobsen Kongsberg Simrad, Norway Revisions
More informationNavigational Aids 1 st Semester/2007/TF 7:30 PM -9:00 PM
Glossary of Navigation Terms accelerometer. A device that senses inertial reaction to measure linear or angular acceleration. In its simplest form, it consists of a case-mounted spring and mass arrangement
More informationCamera Parameters Estimation from Hand-labelled Sun Sositions in Image Sequences
Camera Parameters Estimation from Hand-labelled Sun Sositions in Image Sequences Jean-François Lalonde, Srinivasa G. Narasimhan and Alexei A. Efros {jlalonde,srinivas,efros}@cs.cmu.edu CMU-RI-TR-8-32 July
More informationTEST RESULTS OF A GPS/INERTIAL NAVIGATION SYSTEM USING A LOW COST MEMS IMU
TEST RESULTS OF A GPS/INERTIAL NAVIGATION SYSTEM USING A LOW COST MEMS IMU Alison K. Brown, Ph.D.* NAVSYS Corporation, 1496 Woodcarver Road, Colorado Springs, CO 891 USA, e-mail: abrown@navsys.com Abstract
More informationExam in DD2426 Robotics and Autonomous Systems
Exam in DD2426 Robotics and Autonomous Systems Lecturer: Patric Jensfelt KTH, March 16, 2010, 9-12 No aids are allowed on the exam, i.e. no notes, no books, no calculators, etc. You need a minimum of 20
More informationExterior Orientation Parameters
Exterior Orientation Parameters PERS 12/2001 pp 1321-1332 Karsten Jacobsen, Institute for Photogrammetry and GeoInformation, University of Hannover, Germany The georeference of any photogrammetric product
More informationChapter 13. Vision Based Guidance. Beard & McLain, Small Unmanned Aircraft, Princeton University Press, 2012,
Chapter 3 Vision Based Guidance Beard & McLain, Small Unmanned Aircraft, Princeton University Press, 22, Chapter 3: Slide Architecture w/ Camera targets to track/avoid vision-based guidance waypoints status
More informationTopology Estimation and Global Alignment
Topology Estimation and Global Alignment Ricard Campos Nuno Gracias Rafael Garcia VICOROB http://vicorob.udg.edu/ COMPUTER VISION AND ROBOTICS GROUP UNDERWATER VISION LABORATORY Introduction Photomosaic
More informationUnmanned Aerial Vehicles
Unmanned Aerial Vehicles Embedded Control Edited by Rogelio Lozano WILEY Table of Contents Chapter 1. Aerodynamic Configurations and Dynamic Models 1 Pedro CASTILLO and Alejandro DZUL 1.1. Aerodynamic
More informationInertial Systems. Ekinox Series TACTICAL GRADE MEMS. Motion Sensing & Navigation IMU AHRS MRU INS VG
Ekinox Series TACTICAL GRADE MEMS Inertial Systems IMU AHRS MRU INS VG ITAR Free 0.05 RMS Motion Sensing & Navigation AEROSPACE GROUND MARINE Ekinox Series R&D specialists usually compromise between high
More informationMULTI-MODAL MAPPING. Robotics Day, 31 Mar Frank Mascarich, Shehryar Khattak, Tung Dang
MULTI-MODAL MAPPING Robotics Day, 31 Mar 2017 Frank Mascarich, Shehryar Khattak, Tung Dang Application-Specific Sensors Cameras TOF Cameras PERCEPTION LiDAR IMU Localization Mapping Autonomy Robotic Perception
More informationAutonomous Mobile Robot Design
Autonomous Mobile Robot Design Topic: EKF-based SLAM Dr. Kostas Alexis (CSE) These slides have partially relied on the course of C. Stachniss, Robot Mapping - WS 2013/14 Autonomous Robot Challenges Where
More informationStereo SLAM. Davide Migliore, PhD Department of Electronics and Information, Politecnico di Milano, Italy
Stereo SLAM, PhD migliore@elet.polimi.it Department of Electronics and Information, Politecnico di Milano, Italy What is a Stereo Camera? Slide n 2 Do you remember the pin-hole camera? What is a Stereo
More informationChecking the values using backscatter data
A Technique for using Backscatter Imagery to Calibrate your Multibeam sonar Harold Orlinsky Harold@Hypack.com Checking the values using backscatter data The collection of Backscatter is co located with
More informationUsing Side Scan Sonar to Relative Navigation
Using Side Scan Sonar to Relative Navigation Miguel Pinto, Bruno Ferreira, Aníbal Matos, Nuno Cruz FEUP-DEEC, Rua Dr. Roberto Frias, 4200-465 Porto, Portugal ee04134@fe.up.pt, ee04018@fe.up.pt, anibal@fe.up.pt,
More informationOutline. ETN-FPI Training School on Plenoptic Sensing
Outline Introduction Part I: Basics of Mathematical Optimization Linear Least Squares Nonlinear Optimization Part II: Basics of Computer Vision Camera Model Multi-Camera Model Multi-Camera Calibration
More informationImplementation of underwater precise navigation system for a remotely operated mine disposal vehicle
International Journal of Ocean System Engineering 1(2) (211) 12-19 DOI 1.74/IJOSE.211.1.2.12 International Journal of Ocean System Engineering Implementation of underwater precise navigation system for
More informationRobot Localization based on Geo-referenced Images and G raphic Methods
Robot Localization based on Geo-referenced Images and G raphic Methods Sid Ahmed Berrabah Mechanical Department, Royal Military School, Belgium, sidahmed.berrabah@rma.ac.be Janusz Bedkowski, Łukasz Lubasiński,
More informationRange Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation
Obviously, this is a very slow process and not suitable for dynamic scenes. To speed things up, we can use a laser that projects a vertical line of light onto the scene. This laser rotates around its vertical
More informationLAIR. UNDERWATER ROBOTICS Field Explorations in Marine Biology, Oceanography, and Archeology
UNDERWATER ROBOTICS Field Explorations in Marine Biology, Oceanography, and Archeology COS 402: Artificial Intelligence - Sept. 2011 Christopher M. Clark Outline! Past Projects! Maltese Cistern Mapping!
More informationUAV Autonomous Navigation in a GPS-limited Urban Environment
UAV Autonomous Navigation in a GPS-limited Urban Environment Yoko Watanabe DCSD/CDIN JSO-Aerial Robotics 2014/10/02-03 Introduction 2 Global objective Development of a UAV onboard system to maintain flight
More informationME 597/747 Autonomous Mobile Robots. Mid Term Exam. Duration: 2 hour Total Marks: 100
ME 597/747 Autonomous Mobile Robots Mid Term Exam Duration: 2 hour Total Marks: 100 Instructions: Read the exam carefully before starting. Equations are at the back, but they are NOT necessarily valid
More informationVision-based Localization of an Underwater Robot in a Structured Environment
Vision-based Localization of an Underwater Robot in a Structured Environment M. Carreras, P. Ridao, R. Garcia and T. Nicosevici Institute of Informatics and Applications University of Girona Campus Montilivi,
More informationChapters 1 9: Overview
Chapters 1 9: Overview Chapter 1: Introduction Chapters 2 4: Data acquisition Chapters 5 9: Data manipulation Chapter 5: Vertical imagery Chapter 6: Image coordinate measurements and refinements Chapters
More informationNonlinear State Estimation for Robotics and Computer Vision Applications: An Overview
Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview Arun Das 05/09/2017 Arun Das Waterloo Autonomous Vehicles Lab Introduction What s in a name? Arun Das Waterloo Autonomous
More informationUsing 3D Laser Range Data for SLAM in Outdoor Environments
Using 3D Laser Range Data for SLAM in Outdoor Environments Christian Brenneke, Oliver Wulf, Bernardo Wagner Institute for Systems Engineering, University of Hannover, Germany [brenneke, wulf, wagner]@rts.uni-hannover.de
More informationSensor technology for mobile robots
Laser application, vision application, sonar application and sensor fusion (6wasserf@informatik.uni-hamburg.de) Outline Introduction Mobile robots perception Definitions Sensor classification Sensor Performance
More informationMEMS technology quality requirements as applied to multibeam echosounder. Jerzy DEMKOWICZ, Krzysztof BIKONIS
MEMS technology quality requirements as applied to multibeam echosounder Jerzy DEMKOWICZ, Krzysztof BIKONIS Gdansk University of Technology Gdansk, Narutowicza str. 11/12, Poland demjot@eti.pg.gda.pl Small,
More informationZürich. Roland Siegwart Margarita Chli Martin Rufli Davide Scaramuzza. ETH Master Course: L Autonomous Mobile Robots Summary
Roland Siegwart Margarita Chli Martin Rufli Davide Scaramuzza ETH Master Course: 151-0854-00L Autonomous Mobile Robots Summary 2 Lecture Overview Mobile Robot Control Scheme knowledge, data base mission
More informationROBOT TEAMS CH 12. Experiments with Cooperative Aerial-Ground Robots
ROBOT TEAMS CH 12 Experiments with Cooperative Aerial-Ground Robots Gaurav S. Sukhatme, James F. Montgomery, and Richard T. Vaughan Speaker: Jeff Barnett Paper Focus Heterogeneous Teams for Surveillance
More informationDevelopment of a Ground Based Cooperating Spacecraft Testbed for Research and Education
DIPARTIMENTO DI INGEGNERIA INDUSTRIALE Development of a Ground Based Cooperating Spacecraft Testbed for Research and Education Mattia Mazzucato, Sergio Tronco, Andrea Valmorbida, Fabio Scibona and Enrico
More informationSelection and Integration of Sensors Alex Spitzer 11/23/14
Selection and Integration of Sensors Alex Spitzer aes368@cornell.edu 11/23/14 Sensors Perception of the outside world Cameras, DVL, Sonar, Pressure Accelerometers, Gyroscopes, Magnetometers Position vs
More informationLUMS Mine Detector Project
LUMS Mine Detector Project Using visual information to control a robot (Hutchinson et al. 1996). Vision may or may not be used in the feedback loop. Visual (image based) features such as points, lines
More informationHigh Accuracy Navigation Using Laser Range Sensors in Outdoor Applications
Proceedings of the 2000 IEEE International Conference on Robotics & Automation San Francisco, CA April 2000 High Accuracy Navigation Using Laser Range Sensors in Outdoor Applications Jose Guivant, Eduardo
More informationEE565:Mobile Robotics Lecture 3
EE565:Mobile Robotics Lecture 3 Welcome Dr. Ahmad Kamal Nasir Today s Objectives Motion Models Velocity based model (Dead-Reckoning) Odometry based model (Wheel Encoders) Sensor Models Beam model of range
More informationState Estimation for Continuous-Time Systems with Perspective Outputs from Discrete Noisy Time-Delayed Measurements
State Estimation for Continuous-Time Systems with Perspective Outputs from Discrete Noisy Time-Delayed Measurements António Pedro Aguiar aguiar@ece.ucsb.edu João Pedro Hespanha hespanha@ece.ucsb.edu Dept.
More informationOptimization of the Simultaneous Localization and Map-Building Algorithm for Real-Time Implementation
242 IEEE TRANSACTIONS ON ROBOTICS AND AUTOMATION, VOL. 17, NO. 3, JUNE 2001 Optimization of the Simultaneous Localization and Map-Building Algorithm for Real-Time Implementation José E. Guivant and Eduardo
More informationEE631 Cooperating Autonomous Mobile Robots
EE631 Cooperating Autonomous Mobile Robots Lecture: Multi-Robot Motion Planning Prof. Yi Guo ECE Department Plan Introduction Premises and Problem Statement A Multi-Robot Motion Planning Algorithm Implementation
More informationMarker Based Localization of a Quadrotor. Akshat Agarwal & Siddharth Tanwar
Marker Based Localization of a Quadrotor Akshat Agarwal & Siddharth Tanwar Objective Introduction Objective: To implement a high level control pipeline on a quadrotor which could autonomously take-off,
More informationTowards Gaussian Multi-Robot SLAM for Underwater Robotics
Towards Gaussian Multi-Robot SLAM for Underwater Robotics Dave Kroetsch davek@alumni.uwaterloo.ca Christoper Clark cclark@mecheng1.uwaterloo.ca Lab for Autonomous and Intelligent Robotics University of
More informationSmartphone Video Guidance Sensor for Small Satellites
SSC13-I-7 Smartphone Video Guidance Sensor for Small Satellites Christopher Becker, Richard Howard, John Rakoczy NASA Marshall Space Flight Center Mail Stop EV42, Huntsville, AL 35812; 256-544-0114 christophermbecker@nasagov
More informationAdvanced Vision Guided Robotics. David Bruce Engineering Manager FANUC America Corporation
Advanced Vision Guided Robotics David Bruce Engineering Manager FANUC America Corporation Traditional Vision vs. Vision based Robot Guidance Traditional Machine Vision Determine if a product passes or
More informationDeepwater Spoolpiece Metrology and INS
Deepwater Spoolpiece Metrology and INS ir. Wilbert Brink AVANS Hogeschool - 16 June 2009 Introduction What is a deepwater spoolpiece metrology? What is the classical way of doing a metrology? How can we
More informationAutonomous Navigation for Flying Robots
Computer Vision Group Prof. Daniel Cremers Autonomous Navigation for Flying Robots Lecture 7.1: 2D Motion Estimation in Images Jürgen Sturm Technische Universität München 3D to 2D Perspective Projections
More informationSensor Fusion: Potential, Challenges and Applications. Presented by KVH Industries and Geodetics, Inc. December 2016
Sensor Fusion: Potential, Challenges and Applications Presented by KVH Industries and Geodetics, Inc. December 2016 1 KVH Industries Overview Innovative technology company 600 employees worldwide Focused
More informationLecture: Autonomous micro aerial vehicles
Lecture: Autonomous micro aerial vehicles Friedrich Fraundorfer Remote Sensing Technology TU München 1/41 Autonomous operation@eth Zürich Start 2/41 Autonomous operation@eth Zürich 3/41 Outline MAV system
More informationHumanoid Robotics. Monte Carlo Localization. Maren Bennewitz
Humanoid Robotics Monte Carlo Localization Maren Bennewitz 1 Basis Probability Rules (1) If x and y are independent: Bayes rule: Often written as: The denominator is a normalizing constant that ensures
More informationTargetless Calibration of a Lidar - Perspective Camera Pair. Levente Tamás, Zoltán Kató
Targetless Calibration of a Lidar - Perspective Camera Pair Levente Tamás, Zoltán Kató Content Introduction Region-based calibration framework Evaluation on synthetic data Real data experiments Conclusions
More informationOutline Sensors. EE Sensors. H.I. Bozma. Electric Electronic Engineering Bogazici University. December 13, 2017
Electric Electronic Engineering Bogazici University December 13, 2017 Absolute position measurement Outline Motion Odometry Inertial systems Environmental Tactile Proximity Sensing Ground-Based RF Beacons
More informationA non-contact optical technique for vehicle tracking along bounded trajectories
Home Search Collections Journals About Contact us My IOPscience A non-contact optical technique for vehicle tracking along bounded trajectories This content has been downloaded from IOPscience. Please
More informationCalibration of a rotating multi-beam Lidar
The 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems October 18-22, 2010, Taipei, Taiwan Calibration of a rotating multi-beam Lidar Naveed Muhammad 1,2 and Simon Lacroix 1,2 Abstract
More informationnavigation Isaac Skog
Foot-mounted zerovelocity aided inertial navigation Isaac Skog skog@kth.se Course Outline 1. Foot-mounted inertial navigation a. Basic idea b. Pros and cons 2. Inertial navigation a. The inertial sensors
More informationInertial Navigation Systems
Inertial Navigation Systems Kiril Alexiev University of Pavia March 2017 1 /89 Navigation Estimate the position and orientation. Inertial navigation one of possible instruments. Newton law is used: F =
More informationSpatio-Temporal Stereo Disparity Integration
Spatio-Temporal Stereo Disparity Integration Sandino Morales and Reinhard Klette The.enpeda.. Project, The University of Auckland Tamaki Innovation Campus, Auckland, New Zealand pmor085@aucklanduni.ac.nz
More informationStructure from Motion. Prof. Marco Marcon
Structure from Motion Prof. Marco Marcon Summing-up 2 Stereo is the most powerful clue for determining the structure of a scene Another important clue is the relative motion between the scene and (mono)
More informationMachine learning based automatic extrinsic calibration of an onboard monocular camera for driving assistance applications on smart mobile devices
Technical University of Cluj-Napoca Image Processing and Pattern Recognition Research Center www.cv.utcluj.ro Machine learning based automatic extrinsic calibration of an onboard monocular camera for driving
More informationGround Plane Motion Parameter Estimation For Non Circular Paths
Ground Plane Motion Parameter Estimation For Non Circular Paths G.J.Ellwood Y.Zheng S.A.Billings Department of Automatic Control and Systems Engineering University of Sheffield, Sheffield, UK J.E.W.Mayhew
More informationPatch Test & Stability Check Report
Patch Test & Stability Check Report Storebælt, 2009 SB Cable Project CT Offshore Final Report November, 2009 SB Cable Project November 2009 8-10 Teglbaekvej DK-8361 Hasselager Aarhus, Denmark Tel: +45
More information(W: 12:05-1:50, 50-N202)
2016 School of Information Technology and Electrical Engineering at the University of Queensland Schedule of Events Week Date Lecture (W: 12:05-1:50, 50-N202) 1 27-Jul Introduction 2 Representing Position
More informationStrapdown inertial navigation technology
Strapdown inertial navigation technology D. H. Titterton and J. L. Weston Peter Peregrinus Ltd. on behalf of the Institution of Electrical Engineers Contents Preface Page xiii 1 Introduction 1 1.1 Navigation
More informationGI-Eye II GPS/Inertial System For Target Geo-Location and Image Geo-Referencing
GI-Eye II GPS/Inertial System For Target Geo-Location and Image Geo-Referencing David Boid, Alison Brown, Ph. D., Mark Nylund, Dan Sullivan NAVSYS Corporation 14960 Woodcarver Road, Colorado Springs, CO
More informationCHAPTER 2 SENSOR DATA SIMULATION: A KINEMATIC APPROACH
27 CHAPTER 2 SENSOR DATA SIMULATION: A KINEMATIC APPROACH 2.1 INTRODUCTION The standard technique of generating sensor data for navigation is the dynamic approach. As revealed in the literature (John Blakelock
More informationDealing with Scale. Stephan Weiss Computer Vision Group NASA-JPL / CalTech
Dealing with Scale Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline Why care about size? The IMU as scale provider: The
More informationTightly-Integrated Visual and Inertial Navigation for Pinpoint Landing on Rugged Terrains
Tightly-Integrated Visual and Inertial Navigation for Pinpoint Landing on Rugged Terrains PhD student: Jeff DELAUNE ONERA Director: Guy LE BESNERAIS ONERA Advisors: Jean-Loup FARGES Clément BOURDARIAS
More informationStereo Observation Models
Stereo Observation Models Gabe Sibley June 16, 2003 Abstract This technical report describes general stereo vision triangulation and linearized error modeling. 0.1 Standard Model Equations If the relative
More informationMTRX 4700 Experimental Robotics
Course Outline Mtrx 4700: Experimental Robotics Dr. Stefan B. Williams Slide 1 Wk. Date Content Labs Due Dates 1 4 Mar Introduction, history & philosophy of robotics 2 11 Mar Robot kinematics & dynamics
More informationVisual Bearing-Only Simultaneous Localization and Mapping with Improved Feature Matching
Visual Bearing-Only Simultaneous Localization and Mapping with Improved Feature Matching Hauke Strasdat, Cyrill Stachniss, Maren Bennewitz, and Wolfram Burgard Computer Science Institute, University of
More informationStereo camera de-calibration detection based on observing kinematic attributes of detected objects and the camera rig
Technical University of Dortmund Stereo camera de-calibration detection based on observing kinematic attributes of detected objects and the camera rig by Venkata Rama Prasad Donda A thesis submitted in
More informationPerspective Sensing for Inertial Stabilization
Perspective Sensing for Inertial Stabilization Dr. Bernard A. Schnaufer Jeremy Nadke Advanced Technology Center Rockwell Collins, Inc. Cedar Rapids, IA Agenda Rockwell Collins & the Advanced Technology
More informationDEVICE CONFIGURATION INSTRUCTIONS. WinFrog Device Group:
WinFrog Device Group: Device Name/Model: Device Manufacturer: Device Data String(s) Output to WinFrog: WinFrog Data String(s) Output to Device: WinFrog Data Item(s) and their RAW record: ROV Olympian T2
More informationVehicle Localization. Hannah Rae Kerner 21 April 2015
Vehicle Localization Hannah Rae Kerner 21 April 2015 Spotted in Mtn View: Google Car Why precision localization? in order for a robot to follow a road, it needs to know where the road is to stay in a particular
More informationLaser sensors. Transmitter. Receiver. Basilio Bona ROBOTICA 03CFIOR
Mobile & Service Robotics Sensors for Robotics 3 Laser sensors Rays are transmitted and received coaxially The target is illuminated by collimated rays The receiver measures the time of flight (back and
More informationTracking a Mobile Robot Position Using Vision and Inertial Sensor
Tracking a Mobile Robot Position Using Vision and Inertial Sensor Francisco Coito, António Eleutério, Stanimir Valtchev, and Fernando Coito Faculdade de Ciências e Tecnologia Universidade Nova de Lisboa,
More informationCamera Model and Calibration
Camera Model and Calibration Lecture-10 Camera Calibration Determine extrinsic and intrinsic parameters of camera Extrinsic 3D location and orientation of camera Intrinsic Focal length The size of the
More informationVision-based Mobile Robot Localization and Mapping using Scale-Invariant Features
Vision-based Mobile Robot Localization and Mapping using Scale-Invariant Features Stephen Se, David Lowe, Jim Little Department of Computer Science University of British Columbia Presented by Adam Bickett
More informationGPS denied Navigation Solutions
GPS denied Navigation Solutions Krishnraj Singh Gaur and Mangal Kothari ksgaur@iitk.ac.in, mangal@iitk.ac.in https://www.iitk.ac.in/aero/mangal/ Intelligent Guidance and Control Laboratory Indian Institute
More informationRobotics. Lecture 7: Simultaneous Localisation and Mapping (SLAM)
Robotics Lecture 7: Simultaneous Localisation and Mapping (SLAM) See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information. Andrew Davison Department of Computing Imperial College
More informationTHE problem of building a functional autonomous mobile
IEEE TRANSACTIONS ON ROBOTICS, VOL. 20, NO. 5, OCTOBER 2004 851 Localization Methods for a Mobile Robot in Urban Environments Atanas Georgiev, Member, IEEE, and Peter K. Allen, Member, IEEE Abstract This
More informationCAMERA POSE ESTIMATION OF RGB-D SENSORS USING PARTICLE FILTERING
CAMERA POSE ESTIMATION OF RGB-D SENSORS USING PARTICLE FILTERING By Michael Lowney Senior Thesis in Electrical Engineering University of Illinois at Urbana-Champaign Advisor: Professor Minh Do May 2015
More informationRoadmaps. Vertex Visibility Graph. Reduced visibility graph, i.e., not including segments that extend into obstacles on either side.
Roadmaps Vertex Visibility Graph Full visibility graph Reduced visibility graph, i.e., not including segments that extend into obstacles on either side. (but keeping endpoints roads) what else might we
More informationSimuntaneous Localisation and Mapping with a Single Camera. Abhishek Aneja and Zhichao Chen
Simuntaneous Localisation and Mapping with a Single Camera Abhishek Aneja and Zhichao Chen 3 December, Simuntaneous Localisation and Mapping with asinglecamera 1 Abstract Image reconstruction is common
More informationifp Universität Stuttgart Performance of IGI AEROcontrol-IId GPS/Inertial System Final Report
Universität Stuttgart Performance of IGI AEROcontrol-IId GPS/Inertial System Final Report Institute for Photogrammetry (ifp) University of Stuttgart ifp Geschwister-Scholl-Str. 24 D M. Cramer: Final report
More informationUnscented Kalman Filter for Vision Based Target Localisation with a Quadrotor
Unscented Kalman Filter for Vision Based Target Localisation with a Quadrotor Jos Alejandro Dena Ruiz, Nabil Aouf Centre of Electronic Warfare, Defence Academy of the United Kingdom Cranfield University,
More informationCamera and Inertial Sensor Fusion
January 6, 2018 For First Robotics 2018 Camera and Inertial Sensor Fusion David Zhang david.chao.zhang@gmail.com Version 4.1 1 My Background Ph.D. of Physics - Penn State Univ. Research scientist at SRI
More informationAlignment of Centimeter Scale Bathymetry using Six Degrees of Freedom
Alignment of Centimeter Scale Bathymetry using Six Degrees of Freedom Ethan Slattery, University of California Santa Cruz Mentors: David Caress Summer 2018 Keywords: point-clouds, iterative closest point,
More informationMapping of Translating, Rotating Icebergs With an Autonomous Underwater Vehicle
196 IEEE JOURNAL OF OCEANIC ENGINEERING, VOL 40, NO 1, JANUARY 2015 Mapping of Translating, Rotating Icebergs With an Autonomous Underwater Vehicle Peter W Kimball and Stephen M Rock Abstract This paper
More informationModelling and Simulation of the Autonomous Underwater Vehicle (AUV) Robot
21st International Congress on Modelling and Simulation, Gold Coast, Australia, 29 Nov to 4 Dec 2015 www.mssanz.org.au/modsim2015 Modelling and Simulation of the Autonomous Underwater Vehicle (AUV) Robot
More informationInternational Journal of Advance Engineering and Research Development
Scientific Journal of Impact Factor (SJIF): 4.14 International Journal of Advance Engineering and Research Development Volume 3, Issue 3, March -2016 e-issn (O): 2348-4470 p-issn (P): 2348-6406 Research
More informationGPU-Based Real-Time SAS Processing On-Board Autonomous Underwater Vehicles
GPU Technology Conference 2013 GPU-Based Real-Time SAS Processing On-Board Autonomous Underwater Vehicles Jesús Ortiz Advanced Robotics Department (ADVR) Istituto Italiano di tecnologia (IIT) Francesco
More informationGraph based INS-camera calibration
Graph based INS-camera calibration D. Bender a,b, M. Schikora a,b, J. Sturm b, D. Cremers b a Dept. Sensor Data and Information Fusion, Fraunhofer FKIE, Wachtberg, Germany b Computer Vision Group, Technical
More informationHomography Based Kalman Filter for Mosaic Building. Applications to UAV position estimation
2007 IEEE International Conference on Robotics and Automation Roma, Italy, 10-14 April 2007 ThA12.1 Homography Based Kalman Filter for Mosaic Building. Applications to UAV position estimation Fernando
More informationToward Real-Time Visually Augmented Navigation for Autonomous Search and Inspection of Ship Hulls and Port Facilities
Toward Real-Time Visually Augmented Navigation for Autonomous Search and Inspection of Ship Hulls and Port Facilities Ryan M. Eustice Dept. of Naval Architecture & Marine Engineering University of Michigan
More informationTurning an Automated System into an Autonomous system using Model-Based Design Autonomous Tech Conference 2018
Turning an Automated System into an Autonomous system using Model-Based Design Autonomous Tech Conference 2018 Asaf Moses Systematics Ltd., Technical Product Manager aviasafm@systematics.co.il 1 Autonomous
More informationFeature Based Navigation for a Platform Inspection AUV
Feature Based Navigation for a latform Inspection AUV ehar Tangirala* Chris Debrunner Walter Feldman Alan Fettinger Locheed Martin 00 E.7 th treet Riviera Beach, FL 33404 (56) 47-4336 *sehar.tangirala@lmco.com
More informationGeometric Correction
CEE 6150: Digital Image Processing Geometric Correction 1 Sources of Distortion Sensor Characteristics optical distortion aspect ratio non-linear mirror velocity detector geometry & scanning sequence Viewing
More informationCorrecting INS Drift in Terrain Surface Measurements. Heather Chemistruck Ph.D. Student Mechanical Engineering Vehicle Terrain Performance Lab
Correcting INS Drift in Terrain Surface Measurements Ph.D. Student Mechanical Engineering Vehicle Terrain Performance Lab October 25, 2010 Outline Laboratory Overview Vehicle Terrain Measurement System
More informationLow Cost solution for Pose Estimation of Quadrotor
Low Cost solution for Pose Estimation of Quadrotor mangal@iitk.ac.in https://www.iitk.ac.in/aero/mangal/ Intelligent Guidance and Control Laboratory Indian Institute of Technology, Kanpur Mangal Kothari
More information