H2020 Space Robotic SRC- OG4

Size: px
Start display at page:

Download "H2020 Space Robotic SRC- OG4"

Transcription

1 H2020 Space Robotic SRC- OG4 CCT/COMET ORB Workshop on Space Rendezvous 05/12/2017 «Smart Sensors for Smart Missions» Contacts: Sabrina Andiappane, Vincent Dubanchet, This project has received funding from the European Union s Horizon 2020 research and innovation programme under grant agreement No

2 Outline 1. General Presentation of I3DS Context What is I3DS? I3DS Consortium 2. Development plan 3. Design of I3DS suite 4. Validation usecase 5. GNC elements 6. Conclusion & Q&A 2

3 1.0 - Context I3DS (Integrated 3D Sensors) is one the six consortiums (Operational Grant 4) involved in the frame of the Strategic Research Cluster (SRC) on Space Robotics technologies Horizon 2020 Space call 2016 to enhance the EU industry competitiveness in space robotics. I3DS project develops a generic modular Inspector Sensors Suite (INSES) which is a smart collection of building blocks and a common set of various sensors answering the needs of near-future space exploration missions in terms of remote and contact sensors with integrated pre-processing and data concentration functions. 3

4 1.1 - I3DS: Integrated 3D Sensors 4

5 1.2 - I3DS Objectives Modular approach: Inter-changeable items Scalable design Standard I/F vs satellite/rover platform Realise a suite of perception sensors for both orbital and planetary applications that will allow localisation and map-making for robotic inspection of orbital assets and for planetary surface exploration Subsets to be defined according to the application: Planetary mission: rover Non-cooperative target capture: debris removal Cooperative rendezvous: servicing, spacetug 5

6 2. Development Plan Development on-going for I3DS building blocks: ICU, sensors, software, pre-processing algorithms and mechanical housing To have a top down approach for design: System requirements Functional requirements Sensors requirements To have a bottom up approach for validation Test each sensor individually Test at system level for one selected scenario to validate the coherence of the sensor suite 6

7 3. I3DS Suite - Elements Credit SINTEF Photo: Werner Juvik 7

8 3. I3DS Suite - Pre-processing Goal: Embedded algorithms on the ICU For all the cameras: Stereo Camera, HR Camera and TIR Camera Vignetting Correction Optical Distortion Correction Histogram Equalisation For the Stereo Camera Stereo Rectification For the HR camera + Pattern Projector Structured light pattern description Depth map/point cloud production from structured light 8

9 4. Validation Usecases Two main usecases Orbital : on-orbit rendezvous & capture for servicing Planetary : autonomous science while traversing (mapping, sample characterization) 9

10 4. Validation Usecases - Orbital Missions Phases Step 1: Inertial Navigation / Drifting orbit Step2: Far-range acquisition of the target Step3: Inspection of the target Step4: Final Approach + Tracking Step5: Capture Step5-bis: Emergency manoeuver Medium-range Approach > 200m Inspection / Observation ~ 20m Rendezvous < 20m Berthing / Docking ~ 3m Servicing 0m 10

11 4. Validation Usecases - Orbital Robotic test bench used to reproduce free-floating dynamics Sensors integrated in one mechanical housing for testing 11

12 4. Validation Usecases - Orbital Closed loop validation (foreseen in summer 2018) Need for extra pre-processing and data fusion algorithms to be on the EGSE (limit with other consortium specialized in data fusion: OG3) o Dense or sparse stereo matching o Dense or sparse point cloud production o Point cloud classification and outlier rejection o Data fusion: for instance Stereo + Lidar Development of a simplified GNC loop for testing Hardware-in-the-loop validation mixing sensors HW and simulator SW 12

13 5. GNC Elements - Guidance Cooperative/Collaborative target No tumbling motion Dedicated markers and grapple fixture for capture Two main sub-phases of the rendezvous performed for validation Ellipse of inspection Straight line approach 13

14 5. GNC Elements - Guidance Simulator outputs for Navigation algorithms Open loop guidance : results obtained from simulations (at the moment) Ellipse of inspection (20-60 m) LIDAR (8 key points) STEREO (Relative pos/att) 14

15 5. GNC Elements - Navigation Different processing algorithms by sensor Stereo Camera: pose estimation based on SLAM algorithm Monocular Camera : pose estimation based on visual odometry LIDAR: Pose estimation based on raw ICP algorithm Camera+Projector: point cloud generation using sine pattern Data fusion foreseen between LIDAR and Stereo Camera pose estimation refinement using Kalman filter 15

16 5. GNC Elements - Navigation Geometric Calibration Checkerboard for calibrating camera on a given set of frames Gray Code patterns for calibrating projector (depending on projector resolution) Relative pose estimation between camera and projector 16

17 5. GNC Elements 3D mapping Innovative 3D mapping technology Monocular camera + Pattern projector Short-range sensor < 5m (end-effector sensor) Able to overcome the sunlight 17

18 5. GNC Elements 3D mapping Innovative 3D mapping technology Preliminary testing with different panels 18

19 5. GNC Elements Image Generation Software-in-the-Loop validation of navigation algorithms Using internal tool «SPICAM» o Camera electronics o Camera optics o Lighting conditions o Thermal imaging Planck CAD model displayed by SpiCam Ganymede and Jupiter displayed by SpiCam I3DS target spacecraft derived from ELITE Platform 19

20 Conclusion I3DS is a system on its own with its own hardware and software with embedded processing for on-board sensing capabilities Enables Robotic payload coupling with state-of-the-art platforms Key building block for future robotic missions Tests foreseen in summer 2018 Hope to present them in the next COMET ORB Worskshop! Credit Thales Alenia Space 20

21 Thank you for your attention! 21

H2020 Space Robotic SRC- OG4

H2020 Space Robotic SRC- OG4 H2020 Space Robotic SRC- OG4 2 nd PERASPERA workshop Presentation by Sabrina Andiappane Thales Alenia Space France This project has received funding from the European Union s Horizon 2020 research and

More information

Rendezvous sensors and navigation

Rendezvous sensors and navigation Clean Space Industrial Days Rendezvous sensors and navigation 23-27 th May 2016 Nicolas Deslaef (TAS-F) Julien Christy (TAS-F) 83230918-DOC-TAS-FR-001 OPEN Agenda 2 Collection of R&T studies enabling vision

More information

I3DS SENSOR SUITE FOR SPACE ROBOTICS

I3DS SENSOR SUITE FOR SPACE ROBOTICS I3DS SENSOR SUITE FOR SPACE ROBOTICS Kristoffer Nyborg Gregertsen, PhD Ada-Europe 2018, Lisbon ESA/Hubble About me SINTEF Digital Department of Mathematics and Cybernetics Research manager for automation

More information

InFuse: A Comprehensive Framework for Data Fusion in Space Robotics

InFuse: A Comprehensive Framework for Data Fusion in Space Robotics InFuse InFuse: A Comprehensive Framework for Data Fusion in Space Robotics June 20 th, 2017 Shashank Govindaraj (Space Applications Services, Belgium) Overview 1. Motivations & Objectives 2. InFuse within

More information

t Bench for Robotics and Autonomy Andrea Merlo

t Bench for Robotics and Autonomy Andrea Merlo t Bench for Robotics and Autonomy Andrea Merlo Agenda Introduction TBRA Overview Objectives Architecture / Technical Description Status Test Results Roadmap he context of a Rover, dance, Navigation and

More information

W4. Perception & Situation Awareness & Decision making

W4. Perception & Situation Awareness & Decision making W4. Perception & Situation Awareness & Decision making Robot Perception for Dynamic environments: Outline & DP-Grids concept Dynamic Probabilistic Grids Bayesian Occupancy Filter concept Dynamic Probabilistic

More information

MULTI-MODAL MAPPING. Robotics Day, 31 Mar Frank Mascarich, Shehryar Khattak, Tung Dang

MULTI-MODAL MAPPING. Robotics Day, 31 Mar Frank Mascarich, Shehryar Khattak, Tung Dang MULTI-MODAL MAPPING Robotics Day, 31 Mar 2017 Frank Mascarich, Shehryar Khattak, Tung Dang Application-Specific Sensors Cameras TOF Cameras PERCEPTION LiDAR IMU Localization Mapping Autonomy Robotic Perception

More information

CNES robotics activities : Towards long distance on-board decision-making navigation

CNES robotics activities : Towards long distance on-board decision-making navigation CNES robotics activities : Towards long distance on-board decision-making navigation S.MORENO sabine.moreno@cnes.fr Contents I. Introduction 1.Context 2.Definition II. CNES activities 1.Perception 2.Localisation

More information

Visual Pose Estimation System for Autonomous Rendezvous of Spacecraft

Visual Pose Estimation System for Autonomous Rendezvous of Spacecraft Visual Pose Estimation System for Autonomous Rendezvous of Spacecraft Mark A. Post1, Junquan Li2, and Craig Clark2 Space Mechatronic Systems Technology Laboratory Dept. of Design, Manufacture & Engineering

More information

Application questions. Theoretical questions

Application questions. Theoretical questions The oral exam will last 30 minutes and will consist of one application question followed by two theoretical questions. Please find below a non exhaustive list of possible application questions. The list

More information

LAUROPE Six Legged Walking Robot for Planetary Exploration participating in the SpaceBot Cup

LAUROPE Six Legged Walking Robot for Planetary Exploration participating in the SpaceBot Cup FZI RESEARCH CENTER FOR INFORMATION TECHNOLOGY LAUROPE Six Legged Walking Robot for Planetary Exploration participating in the SpaceBot Cup ASTRA 2015, ESA/ESTEC, Noordwijk 12. Mai 2015 Outline SpaceBot

More information

3D Time-of-Flight Image Sensor Solutions for Mobile Devices

3D Time-of-Flight Image Sensor Solutions for Mobile Devices 3D Time-of-Flight Image Sensor Solutions for Mobile Devices SEMICON Europa 2015 Imaging Conference Bernd Buxbaum 2015 pmdtechnologies gmbh c o n f i d e n t i a l Content Introduction Motivation for 3D

More information

Planetary Rover Absolute Localization by Combining Visual Odometry with Orbital Image Measurements

Planetary Rover Absolute Localization by Combining Visual Odometry with Orbital Image Measurements Planetary Rover Absolute Localization by Combining Visual Odometry with Orbital Image Measurements M. Lourakis and E. Hourdakis Institute of Computer Science Foundation for Research and Technology Hellas

More information

VALIDATION OF 3D ENVIRONMENT PERCEPTION FOR LANDING ON SMALL BODIES USING UAV PLATFORMS

VALIDATION OF 3D ENVIRONMENT PERCEPTION FOR LANDING ON SMALL BODIES USING UAV PLATFORMS ASTRA 2015 VALIDATION OF 3D ENVIRONMENT PERCEPTION FOR LANDING ON SMALL BODIES USING UAV PLATFORMS Property of GMV All rights reserved PERIGEO PROJECT The work presented here is part of the PERIGEO project

More information

Autonomous navigation in industrial cluttered environments using embedded stereo-vision

Autonomous navigation in industrial cluttered environments using embedded stereo-vision Autonomous navigation in industrial cluttered environments using embedded stereo-vision Julien Marzat ONERA Palaiseau Aerial Robotics workshop, Paris, 8-9 March 2017 1 Copernic Lab (ONERA Palaiseau) Research

More information

Introduction to Autonomous Mobile Robots

Introduction to Autonomous Mobile Robots Introduction to Autonomous Mobile Robots second edition Roland Siegwart, Illah R. Nourbakhsh, and Davide Scaramuzza The MIT Press Cambridge, Massachusetts London, England Contents Acknowledgments xiii

More information

Dealing with Scale. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

Dealing with Scale. Stephan Weiss Computer Vision Group NASA-JPL / CalTech Dealing with Scale Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline Why care about size? The IMU as scale provider: The

More information

Development of a Ground Based Cooperating Spacecraft Testbed for Research and Education

Development of a Ground Based Cooperating Spacecraft Testbed for Research and Education DIPARTIMENTO DI INGEGNERIA INDUSTRIALE Development of a Ground Based Cooperating Spacecraft Testbed for Research and Education Mattia Mazzucato, Sergio Tronco, Andrea Valmorbida, Fabio Scibona and Enrico

More information

On-Orbit Testing of Target-less TriDAR 3D Rendezvous and Docking Sensor

On-Orbit Testing of Target-less TriDAR 3D Rendezvous and Docking Sensor On-Orbit Testing of Target-less TriDAR 3D Rendezvous and Docking Sensor Stephane Ruel, Tim Luu, Andrew Berube* *Neptec Design Group, Canada e-mail: sruel@neptec.com Abstract TriDAR is a vision system developed

More information

Stereo and Monocular Vision Applied to Computer Aided Navigation for Aerial and Ground Vehicles

Stereo and Monocular Vision Applied to Computer Aided Navigation for Aerial and Ground Vehicles Riccardo Giubilato Centro di Ateneo di Studi e Attivita Spaziali Giuseppe Colombo CISAS University of Padova 1 Introduction Previous Work Research Objectives Research Methodology Robot vision: Process

More information

Deep Incremental Scene Understanding. Federico Tombari & Christian Rupprecht Technical University of Munich, Germany

Deep Incremental Scene Understanding. Federico Tombari & Christian Rupprecht Technical University of Munich, Germany Deep Incremental Scene Understanding Federico Tombari & Christian Rupprecht Technical University of Munich, Germany C. Couprie et al. "Toward Real-time Indoor Semantic Segmentation Using Depth Information"

More information

VISION FOR AUTOMOTIVE DRIVING

VISION FOR AUTOMOTIVE DRIVING VISION FOR AUTOMOTIVE DRIVING French Japanese Workshop on Deep Learning & AI, Paris, October 25th, 2017 Quoc Cuong PHAM, PhD Vision and Content Engineering Lab AI & MACHINE LEARNING FOR ADAS AND SELF-DRIVING

More information

Low Cost solution for Pose Estimation of Quadrotor

Low Cost solution for Pose Estimation of Quadrotor Low Cost solution for Pose Estimation of Quadrotor mangal@iitk.ac.in https://www.iitk.ac.in/aero/mangal/ Intelligent Guidance and Control Laboratory Indian Institute of Technology, Kanpur Mangal Kothari

More information

UAV Autonomous Navigation in a GPS-limited Urban Environment

UAV Autonomous Navigation in a GPS-limited Urban Environment UAV Autonomous Navigation in a GPS-limited Urban Environment Yoko Watanabe DCSD/CDIN JSO-Aerial Robotics 2014/10/02-03 Introduction 2 Global objective Development of a UAV onboard system to maintain flight

More information

3D object recognition used by team robotto

3D object recognition used by team robotto 3D object recognition used by team robotto Workshop Juliane Hoebel February 1, 2016 Faculty of Computer Science, Otto-von-Guericke University Magdeburg Content 1. Introduction 2. Depth sensor 3. 3D object

More information

Safe Long-Range Travel for Planetary Rovers through Forward Sensing

Safe Long-Range Travel for Planetary Rovers through Forward Sensing Safe Long-Range Travel for Planetary Rovers through Forward Sensing 12 th ESA Workshop on Advanced Space Technologies for Robotics & Automation Yashodhan Nevatia Space Applications Services May 16 th,

More information

Innovative Test-beds for Validation of Motion Recognition Systems Intended for Orbital Capture Manoeuvres

Innovative Test-beds for Validation of Motion Recognition Systems Intended for Orbital Capture Manoeuvres Innovative Test-beds for Validation of Motion Recognition Systems Intended for Orbital Capture Manoeuvres Tomasz Rybus 1, Karol Seweryn 1, Janusz Grzybowski 1, Janusz Nicolau-Kukliński 1, Rafał Przybyła

More information

Mixed-Reality for Intuitive Photo-Realistic 3D-Model Generation

Mixed-Reality for Intuitive Photo-Realistic 3D-Model Generation Mixed-Reality for Intuitive Photo-Realistic 3D-Model Generation Wolfgang Sepp, Tim Bodenmueller, Michael Suppa, and Gerd Hirzinger DLR, Institut für Robotik und Mechatronik @ GI-Workshop VR/AR 2009 Folie

More information

Localization, Where am I?

Localization, Where am I? 5.1 Localization, Where am I?? position Position Update (Estimation?) Encoder Prediction of Position (e.g. odometry) YES matched observations Map data base predicted position Matching Odometry, Dead Reckoning

More information

Model-based Visual Tracking:

Model-based Visual Tracking: Technische Universität München Model-based Visual Tracking: the OpenTL framework Giorgio Panin Technische Universität München Institut für Informatik Lehrstuhl für Echtzeitsysteme und Robotik (Prof. Alois

More information

Hybrids Mixed Approaches

Hybrids Mixed Approaches Hybrids Mixed Approaches Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline Why mixing? Parallel Tracking and Mapping Benefits

More information

Visual Pose Estimation and Identification for Satellite Rendezvous Operations

Visual Pose Estimation and Identification for Satellite Rendezvous Operations Post, Mark A. and Yan, Xiu T. (2015) Visual pose estimation and identification for satellite rendezvous operations. In: Sixth China- Scotland SIPRA workshop "Recent Advances in Signal and Image Processing",

More information

Visual SLAM. An Overview. L. Freda. ALCOR Lab DIAG University of Rome La Sapienza. May 3, 2016

Visual SLAM. An Overview. L. Freda. ALCOR Lab DIAG University of Rome La Sapienza. May 3, 2016 An Overview L. Freda ALCOR Lab DIAG University of Rome La Sapienza May 3, 2016 L. Freda (University of Rome La Sapienza ) Visual SLAM May 3, 2016 1 / 39 Outline 1 Introduction What is SLAM Motivations

More information

OUTDOOR AND INDOOR NAVIGATION WITH MICROSOFT KINECT

OUTDOOR AND INDOOR NAVIGATION WITH MICROSOFT KINECT DICA-Dept. of Civil and Environmental Engineering Geodesy and Geomatics Section OUTDOOR AND INDOOR NAVIGATION WITH MICROSOFT KINECT Diana Pagliari Livio Pinto OUTLINE 2 The Microsoft Kinect sensor The

More information

ME/CS 132: Introduction to Vision-based Robot Navigation! Low-level Image Processing" Larry Matthies"

ME/CS 132: Introduction to Vision-based Robot Navigation! Low-level Image Processing Larry Matthies ME/CS 132: Introduction to Vision-based Robot Navigation! Low-level Image Processing" Larry Matthies" lhm@jpl.nasa.gov, 818-354-3722" Announcements" First homework grading is done! Second homework is due

More information

Sensor Fusion: Potential, Challenges and Applications. Presented by KVH Industries and Geodetics, Inc. December 2016

Sensor Fusion: Potential, Challenges and Applications. Presented by KVH Industries and Geodetics, Inc. December 2016 Sensor Fusion: Potential, Challenges and Applications Presented by KVH Industries and Geodetics, Inc. December 2016 1 KVH Industries Overview Innovative technology company 600 employees worldwide Focused

More information

Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview

Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview Arun Das 05/09/2017 Arun Das Waterloo Autonomous Vehicles Lab Introduction What s in a name? Arun Das Waterloo Autonomous

More information

CSE 527: Introduction to Computer Vision

CSE 527: Introduction to Computer Vision CSE 527: Introduction to Computer Vision Week 10 Class 2: Visual Odometry November 2nd, 2017 Today Visual Odometry Intro Algorithm SLAM Visual Odometry Input Output Images, Video Camera trajectory, motion

More information

4D Crop Analysis for Plant Geometry Estimation in Precision Agriculture

4D Crop Analysis for Plant Geometry Estimation in Precision Agriculture 4D Crop Analysis for Plant Geometry Estimation in Precision Agriculture MIT Laboratory for Information & Decision Systems IEEE RAS TC on Agricultural Robotics and Automation Webinar #37 Acknowledgements

More information

Flexible Visual Inspection. IAS-13 Industrial Forum Horizon 2020 Dr. Eng. Stefano Tonello - CEO

Flexible Visual Inspection. IAS-13 Industrial Forum Horizon 2020 Dr. Eng. Stefano Tonello - CEO Flexible Visual Inspection IAS-13 Industrial Forum Horizon 2020 Dr. Eng. Stefano Tonello - CEO IT+Robotics Spin-off of University of Padua founded in 2005 Strong relationship with IAS-LAB (Intelligent

More information

More processing power - Driving requirements

More processing power - Driving requirements Software More processing power - Driving requirements Autonomy Task based commanding / resource management rather than timeline remote control E.g. planetary rovers drive from A B, look out for interesting

More information

Presentation Outline

Presentation Outline Presentation Outline Phd Activities during the three years mainly concentrated on the development and testing of the SPARTANS cooperating spacecraft hardware testbed Development of the Translation Module

More information

Tightly-Integrated Visual and Inertial Navigation for Pinpoint Landing on Rugged Terrains

Tightly-Integrated Visual and Inertial Navigation for Pinpoint Landing on Rugged Terrains Tightly-Integrated Visual and Inertial Navigation for Pinpoint Landing on Rugged Terrains PhD student: Jeff DELAUNE ONERA Director: Guy LE BESNERAIS ONERA Advisors: Jean-Loup FARGES Clément BOURDARIAS

More information

Jena-Optronik Relative Navigation Sensor Activities

Jena-Optronik Relative Navigation Sensor Activities Jena-Optronik Relative Navigation Sensor Activities ESA AIM Industry Day Jena-Optronik Relative Navigation Sensor Activities ESA AIM Industry Day. Page 1 Space developments from Jena. 1976 1980 1991 1994

More information

Camera Drones Lecture 3 3D data generation

Camera Drones Lecture 3 3D data generation Camera Drones Lecture 3 3D data generation Ass.Prof. Friedrich Fraundorfer WS 2017 Outline SfM introduction SfM concept Feature matching Camera pose estimation Bundle adjustment Dense matching Data products

More information

Space Robotics. Ioannis Rekleitis

Space Robotics. Ioannis Rekleitis Space Robotics Ioannis Rekleitis On-Orbit Servicing of Satellites Work done at the Canadian Space Agency Guy Rouleau, Ioannis Rekleitis, Régent L'Archevêque, Eric Martin, Kourosh Parsa, and Erick Dupuis

More information

GPS denied Navigation Solutions

GPS denied Navigation Solutions GPS denied Navigation Solutions Krishnraj Singh Gaur and Mangal Kothari ksgaur@iitk.ac.in, mangal@iitk.ac.in https://www.iitk.ac.in/aero/mangal/ Intelligent Guidance and Control Laboratory Indian Institute

More information

Vision-based navigation

Vision-based navigation Vision-based navigation Simon Lacroix To cite this version: Simon Lacroix. Vision-based navigation. Doctoral. Spring School on Location-Based Services, École Nationale d Aviation Civile, Toulouse (France),

More information

VALIDATION AND TUNNING OF DENSE STEREO-VISION SYSTEMS USING HI-RESOLUTION 3D REFERENCE MODELS

VALIDATION AND TUNNING OF DENSE STEREO-VISION SYSTEMS USING HI-RESOLUTION 3D REFERENCE MODELS VALIDATION AND TUNNING OF DENSE STEREOVISION SYSTEMS USING HIRESOLUTION 3D REFERENCE MODELS E. REMETEAN (CNES) S. MAS (MAGELLIUM) JB GINESTET (MAGELLIUM) L.RASTEL (CNES) ASTRA 14.04.2011 CONTENT Validation

More information

Dense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm

Dense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm Computer Vision Group Prof. Daniel Cremers Dense Tracking and Mapping for Autonomous Quadrocopters Jürgen Sturm Joint work with Frank Steinbrücker, Jakob Engel, Christian Kerl, Erik Bylow, and Daniel Cremers

More information

Training models for road scene understanding with automated ground truth Dan Levi

Training models for road scene understanding with automated ground truth Dan Levi Training models for road scene understanding with automated ground truth Dan Levi With: Noa Garnett, Ethan Fetaya, Shai Silberstein, Rafi Cohen, Shaul Oron, Uri Verner, Ariel Ayash, Kobi Horn, Vlad Golder,

More information

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Structured Light II Johannes Köhler Johannes.koehler@dfki.de Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Introduction Previous lecture: Structured Light I Active Scanning Camera/emitter

More information

Towards Fully-automated Driving. tue-mps.org. Challenges and Potential Solutions. Dr. Gijs Dubbelman Mobile Perception Systems EE-SPS/VCA

Towards Fully-automated Driving. tue-mps.org. Challenges and Potential Solutions. Dr. Gijs Dubbelman Mobile Perception Systems EE-SPS/VCA Towards Fully-automated Driving Challenges and Potential Solutions Dr. Gijs Dubbelman Mobile Perception Systems EE-SPS/VCA Mobile Perception Systems 6 PhDs, 1 postdoc, 1 project manager, 2 software engineers

More information

Localization and Map Building

Localization and Map Building Localization and Map Building Noise and aliasing; odometric position estimation To localize or not to localize Belief representation Map representation Probabilistic map-based localization Other examples

More information

5 member consortium o University of Strathclyde o Wideblue o National Nuclear Laboratory o Sellafield Ltd o Inspectahire

5 member consortium o University of Strathclyde o Wideblue o National Nuclear Laboratory o Sellafield Ltd o Inspectahire 3 year, 1.24M Innovate UK funded Collaborative Research and Development Project (Nuclear Call) o Commenced April 2015 o Follows on from a successful 6 month Innovate UK funded feasibility study 2013-2014

More information

Visual SLAM for small Unmanned Aerial Vehicles

Visual SLAM for small Unmanned Aerial Vehicles Visual SLAM for small Unmanned Aerial Vehicles Margarita Chli Autonomous Systems Lab, ETH Zurich Simultaneous Localization And Mapping How can a body navigate in a previously unknown environment while

More information

The WiMUST H2020 project: Widely scalable Mobile Underwater Sonar Technology

The WiMUST H2020 project: Widely scalable Mobile Underwater Sonar Technology The WiMUST H2020 project: Widely scalable Mobile Underwater Sonar Technology Giovanni Indiveri Coordinator - ISME WiMUST - Widely scalable Mobile Underwater Sonar Technology Grant agreement no: 645141

More information

3D Fusion of Infrared Images with Dense RGB Reconstruction from Multiple Views - with Application to Fire-fighting Robots

3D Fusion of Infrared Images with Dense RGB Reconstruction from Multiple Views - with Application to Fire-fighting Robots 3D Fusion of Infrared Images with Dense RGB Reconstruction from Multiple Views - with Application to Fire-fighting Robots Yuncong Chen 1 and Will Warren 2 1 Department of Computer Science and Engineering,

More information

COMPUTER VISION FOR REAL-TIME RELATIVE NAVIGATION WITH A NON-COOPERATIVE AND SPINNING TARGET SPACECRAFT

COMPUTER VISION FOR REAL-TIME RELATIVE NAVIGATION WITH A NON-COOPERATIVE AND SPINNING TARGET SPACECRAFT COMPUTER VISION FOR REAL-TIME RELATIVE NAVIGATION WITH A NON-COOPERATIVE AND SPINNING TARGET SPACECRAFT Setareh Yazdkhasti, Steve Ulrich, and Jurek Z. Sasiadek Carleton University, Ottawa, Ontario, K1S

More information

Monitoring the environment with a Cloud Robotics Service

Monitoring the environment with a Cloud Robotics Service Monitoring the environment with a Cloud Robotics Service Ludovico O. Russo ludus.russo@gmail.com Outline Introduction Service Robotics Autonomous Navigation Cloud Robotics Robot Based Environmental Monitoring

More information

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University. 3D Computer Vision Structured Light II Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de 1 Introduction

More information

A Modular Sensor Fusion Approach for Agricultural Machines

A Modular Sensor Fusion Approach for Agricultural Machines A Modular Sensor Fusion Approach for Agricultural Machines Sebastian Blank (1), Georg Kormann (2), Karsten Berns (1) (1) Robotics Research Lab University of Kaiserslautern, Germany (2) John Deere European

More information

Beate Klinger and Torsten Mayer-Gürr Institute of Geodesy NAWI Graz, Graz University of Technology

Beate Klinger and Torsten Mayer-Gürr Institute of Geodesy NAWI Graz, Graz University of Technology and Torsten Mayer-Gürr Institute of Geodesy NAWI Graz, Graz University of Technology Outline Motivation GRACE Preprocessing GRACE sensor fusion Accelerometer simulation & calibration Impact on monthly

More information

Direct Methods in Visual Odometry

Direct Methods in Visual Odometry Direct Methods in Visual Odometry July 24, 2017 Direct Methods in Visual Odometry July 24, 2017 1 / 47 Motivation for using Visual Odometry Wheel odometry is affected by wheel slip More accurate compared

More information

V-Sentinel: A Novel Framework for Situational Awareness and Surveillance

V-Sentinel: A Novel Framework for Situational Awareness and Surveillance V-Sentinel: A Novel Framework for Situational Awareness and Surveillance Suya You Integrated Media Systems Center Computer Science Department University of Southern California March 2005 1 Objective Developing

More information

WP6 D6.2 Project website

WP6 D6.2 Project website WP6 D6.2 Project website Project title: Project Acronym: Promoting Youth Scientific Career Awareness and it Attractiveness through Multistakeholder Cooperation MultiCO Project ID: 665100 Prepared by: University

More information

USE OF COTS ROBOTICS FOR ON-GROUND VALIDATION OF SPACE GNC SYSTEMS: PLATFORM DYNAMIC TEST BENCH I-SAIRAS 2012 TURIN, ITALY 4-6 SEPTEMBER 2012

USE OF COTS ROBOTICS FOR ON-GROUND VALIDATION OF SPACE GNC SYSTEMS: PLATFORM DYNAMIC TEST BENCH I-SAIRAS 2012 TURIN, ITALY 4-6 SEPTEMBER 2012 USE OF COTS ROBOTICS FOR ON-GROUND VALIDATION OF SPACE GNC SYSTEMS: PLATFORM DYNAMIC TEST BENCH I-SAIRAS 212 TURIN, ITALY 4-6 SEPTEMBER 212 Matteo Suatoni (1), Luis Mollinedo (1), Valentín Barrena (1),

More information

AIRBUS DS VISION BASED NAVIGATION SOLUTIONS TESTED ON LIRIS EXPERIMENT DATA

AIRBUS DS VISION BASED NAVIGATION SOLUTIONS TESTED ON LIRIS EXPERIMENT DATA AIRBUS DS VISION BASED NAVIGATION SOLUTIONS TESTED ON LIRIS EXPERIMENT DATA A. Masson(1), C. Haskamp (2), I. Ahrns (2), R. Brochard(1), P. Duteis (1), K. Kanani(1), R. Delage(1) (1) Airbus Defence and

More information

MONOCULAR OBJECT LOCALIZATION BY SUPERQUADRICS CURVATURE REPROJECTION AND MATCHING

MONOCULAR OBJECT LOCALIZATION BY SUPERQUADRICS CURVATURE REPROJECTION AND MATCHING GraphiCon 2012 (Moscow, Russia) University of Trento, Italy MONOCULAR OBJECT LOCALIZATION BY SUPERQUADRICS CURVATURE REPROJECTION AND MATCHING Enrico Zappia, Ilya Afanasyev, Nicolo Biasi, Mattia Tavernini,

More information

Concept and Performance Simulation with ASTOS

Concept and Performance Simulation with ASTOS Concept and Performance Simulation with ASTOS Andreas Wiegand (1), Sven Weikert (1) Astos Solutions GmbH (1) Meitnerstraße 8, 70563 Stuttgart, Germany andreas.wiegand@astos.de ABSTRACT Advanced space missions,

More information

On Satellite Vision-aided Robotics Experiment

On Satellite Vision-aided Robotics Experiment On Satellite Vision-aided Robotics Experiment Maarten Vergauwen, Marc Pollefeys, Tinne Tuytelaars and Luc Van Gool ESAT PSI, K.U.Leuven, Kard. Mercierlaan 94, B-3001 Heverlee, Belgium Phone: +32-16-32.10.64,

More information

Leica Systems Overview

Leica Systems Overview RC30 AERIAL CAMERA SYSTEM Leica Systems Overview The Leica RC30 aerial film camera is the culmination of decades of development, started with Wild's first aerial camera in the 1920s. Beautifully engineered

More information

On-line and Off-line 3D Reconstruction for Crisis Management Applications

On-line and Off-line 3D Reconstruction for Crisis Management Applications On-line and Off-line 3D Reconstruction for Crisis Management Applications Geert De Cubber Royal Military Academy, Department of Mechanical Engineering (MSTA) Av. de la Renaissance 30, 1000 Brussels geert.de.cubber@rma.ac.be

More information

AUTONOMOUS LONG-RANGE ROVER NAVIGATION EXPERIMENTAL RESULTS

AUTONOMOUS LONG-RANGE ROVER NAVIGATION EXPERIMENTAL RESULTS AUTONOMOUS LONG-RANGE ROVER NAVIGATION EXPERIMENTAL RESULTS Erick Dupuis, Joseph Nsasi Bakambu, Ioannis Rekleitis, Jean-Luc Bedwani, Sébastien Gemme, Jean-Patrice Rivest-Caissy Canadian Space Agency 6767

More information

SpatialFuser. Offers support for GNSS and INS navigation systems. Supports multiple remote sensing equipment configurations.

SpatialFuser. Offers support for GNSS and INS navigation systems. Supports multiple remote sensing equipment configurations. SPATIALFUSER SpatialFuser SpatialFuser is designed to georeference data acquired by our LiDAR mapping systems into common mapping formats such as LAS/LAZ or GPX. The main purpose of SpatialFuser is to

More information

GNC for a Rendezvous in Space with an Uncooperative Target.

GNC for a Rendezvous in Space with an Uncooperative Target. GNC for a Rendezvous in Space with an Uncooperative Target Josef Sommer (1), Ingo Ahrns (2) (1) Astrium GmbH, Airbusallee 1, 28199 Bremen, Germany, +49 421 539 4440, josef.sommer@astrium.eads.net (2) Astrium

More information

Basics of Localization, Mapping and SLAM. Jari Saarinen Aalto University Department of Automation and systems Technology

Basics of Localization, Mapping and SLAM. Jari Saarinen Aalto University Department of Automation and systems Technology Basics of Localization, Mapping and SLAM Jari Saarinen Aalto University Department of Automation and systems Technology Content Introduction to Problem (s) Localization A few basic equations Dead Reckoning

More information

CONTENT ENGINEERING & VISION LABORATORY. Régis Vinciguerra

CONTENT ENGINEERING & VISION LABORATORY. Régis Vinciguerra CONTENT ENGINEERING & VISION LABORATORY Régis Vinciguerra regis.vinciguerra@cea.fr ALTERNATIVE ENERGIES AND ATOMIC ENERGY COMMISSION Military Applications Division (DAM) Nuclear Energy Division (DEN) Technological

More information

Autonomous Navigation in Complex Indoor and Outdoor Environments with Micro Aerial Vehicles

Autonomous Navigation in Complex Indoor and Outdoor Environments with Micro Aerial Vehicles Autonomous Navigation in Complex Indoor and Outdoor Environments with Micro Aerial Vehicles Shaojie Shen Dept. of Electrical and Systems Engineering & GRASP Lab, University of Pennsylvania Committee: Daniel

More information

Incremental Real-time Bundle Adjustment for Multi-camera Systems with Points at Infinity

Incremental Real-time Bundle Adjustment for Multi-camera Systems with Points at Infinity Incremental Real-time Bundle Adjustment for Multi-camera Systems with Points at Infinity Johannes Schneider, Thomas Läbe, Wolfgang Förstner 1 Department of Photogrammetry Institute of Geodesy and Geoinformation

More information

Dense 3D Reconstruction from Autonomous Quadrocopters

Dense 3D Reconstruction from Autonomous Quadrocopters Dense 3D Reconstruction from Autonomous Quadrocopters Computer Science & Mathematics TU Munich Martin Oswald, Jakob Engel, Christian Kerl, Frank Steinbrücker, Jan Stühmer & Jürgen Sturm Autonomous Quadrocopters

More information

PROJECT HIGHLIGHTS, EXPECTED IMPACT & FUTURE DIRECTIONS

PROJECT HIGHLIGHTS, EXPECTED IMPACT & FUTURE DIRECTIONS PROJECT HIGHLIGHTS, EXPECTED IMPACT & FUTURE DIRECTIONS Dr. Angelos Amditis (project coordinator) Institute of Communication and Computer Systems (ICCS), Research Director Final Event AstaZero, Sandhult,

More information

How can the Future Internet

How can the Future Internet How can the Future Internet enable Smart Energy? Presented by Werner Mohr (Coordinator), Nokia Siemens Networks on behalf of the FINSENY project Smart Energy enabled by Future Internet Workshop FINSENY

More information

INTELLIGENT AUTONOMOUS SYSTEMS LAB

INTELLIGENT AUTONOMOUS SYSTEMS LAB Matteo Munaro 1,3, Alex Horn 2, Randy Illum 2, Jeff Burke 2, and Radu Bogdan Rusu 3 1 IAS-Lab at Department of Information Engineering, University of Padova 2 Center for Research in Engineering, Media

More information

GNSS/INS for High Accuracy Mobile Mapping. Olaf Gross 11 th Terrasolid European User Event Kittilä, Finland

GNSS/INS for High Accuracy Mobile Mapping. Olaf Gross 11 th Terrasolid European User Event Kittilä, Finland GNSS/INS for High Accuracy Mobile Mapping Olaf Gross 11 th Terrasolid European User Event 15.02.2012 Kittilä, Finland IGI mbh Founded 1978 25 staff about half in R&D More than 380 customers in 60 different

More information

Photoneo's brand new PhoXi 3D Camera is the highest resolution and highest accuracy area based 3D

Photoneo's brand new PhoXi 3D Camera is the highest resolution and highest accuracy area based 3D Company: Photoneo s.r.o. Germany Contact: Veronika Pulisova E-mail: pulisova@photoneo.com PhoXi 3D Camera Author: Tomas Kovacovsky & Jan Zizka Description of the innovation: General description Photoneo's

More information

Miniature faking. In close-up photo, the depth of field is limited.

Miniature faking. In close-up photo, the depth of field is limited. Miniature faking In close-up photo, the depth of field is limited. http://en.wikipedia.org/wiki/file:jodhpur_tilt_shift.jpg Miniature faking Miniature faking http://en.wikipedia.org/wiki/file:oregon_state_beavers_tilt-shift_miniature_greg_keene.jpg

More information

SGEO: Overview and Product Offering. Marco R. Fuchs. Marco R. R. Fuchs. DLR-ESA Workshop on ARTES 11. Marco R. Fuchs OHB Technology AG

SGEO: Overview and Product Offering. Marco R. Fuchs. Marco R. R. Fuchs. DLR-ESA Workshop on ARTES 11. Marco R. Fuchs OHB Technology AG DLR-ESA Workshop on ARTES 11 SGEO: Overview and Product Offering Marco R. R. Fuchs June June29, 29, 2006 2006 Tegernsee, Tegernsee, Germany Germany Marco R. Fuchs Marco R. Fuchs OHB Technology AG OHB Technology

More information

3D Point Cloud Processing

3D Point Cloud Processing 3D Point Cloud Processing The image depicts how our robot Irma3D sees itself in a mirror. The laser looking into itself creates distortions as well as changes in intensity that give the robot a single

More information

DARPA Perspective on Space

DARPA Perspective on Space DARPA Perspective on Space Ms. Pamela A. Melroy, Deputy Director DARPA Tactical Technology Office Briefing prepared for International Symposium for Personal and Commercial Spaceflight (ISPCS) 2015 October

More information

Outline. ETN-FPI Training School on Plenoptic Sensing

Outline. ETN-FPI Training School on Plenoptic Sensing Outline Introduction Part I: Basics of Mathematical Optimization Linear Least Squares Nonlinear Optimization Part II: Basics of Computer Vision Camera Model Multi-Camera Model Multi-Camera Calibration

More information

Advanced Driver Assistance: Modular Image Sensor Concept

Advanced Driver Assistance: Modular Image Sensor Concept Vision Advanced Driver Assistance: Modular Image Sensor Concept Supplying value. Integrated Passive and Active Safety Systems Active Safety Passive Safety Scope Reduction of accident probability Get ready

More information

New Sony DepthSense TM ToF Technology

New Sony DepthSense TM ToF Technology ADVANCED MATERIAL HANDLING WITH New Sony DepthSense TM ToF Technology Jenson Chang Product Marketing November 7, 2018 1 3D SENSING APPLICATIONS Pick and Place Drones Collision Detection People Counting

More information

Solid State LiDAR for Ubiquitous 3D Sensing

Solid State LiDAR for Ubiquitous 3D Sensing April 6, 2016 Solid State LiDAR for Ubiquitous 3D Sensing Louay Eldada, Ph.D. CEO, Co-founder Quanergy Systems New Paradigm in 3D Sensing Disruptive Technologies: Solid State 3D LiDAR sensors Embedded

More information

Managing and mining (streaming) sensor data

Managing and mining (streaming) sensor data Petr Čížek Artificial Intelligence Center Czech Technical University in Prague November 3, 2016 Petr Čížek VPD 1 / 1 Stream data mining / stream data querying Problem definition Data can not be stored

More information

Augmented Reality, Advanced SLAM, Applications

Augmented Reality, Advanced SLAM, Applications Augmented Reality, Advanced SLAM, Applications Prof. Didier Stricker & Dr. Alain Pagani alain.pagani@dfki.de Lecture 3D Computer Vision AR, SLAM, Applications 1 Introduction Previous lectures: Basics (camera,

More information

Why the Self-Driving Revolution Hinges on one Enabling Technology: LiDAR

Why the Self-Driving Revolution Hinges on one Enabling Technology: LiDAR Why the Self-Driving Revolution Hinges on one Enabling Technology: LiDAR Markus Prison Director Business Development Europe Quanergy ID: 23328 Who We Are The leader in LiDAR (laser-based 3D spatial sensor)

More information

Overview. EECS 124, UC Berkeley, Spring 2008 Lecture 23: Localization and Mapping. Statistical Models

Overview. EECS 124, UC Berkeley, Spring 2008 Lecture 23: Localization and Mapping. Statistical Models Introduction ti to Embedded dsystems EECS 124, UC Berkeley, Spring 2008 Lecture 23: Localization and Mapping Gabe Hoffmann Ph.D. Candidate, Aero/Astro Engineering Stanford University Statistical Models

More information

Visual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

Visual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech Visual Odometry Features, Tracking, Essential Matrix, and RANSAC Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline The

More information

Personal Navigation and Indoor Mapping: Performance Characterization of Kinect Sensor-based Trajectory Recovery

Personal Navigation and Indoor Mapping: Performance Characterization of Kinect Sensor-based Trajectory Recovery Personal Navigation and Indoor Mapping: Performance Characterization of Kinect Sensor-based Trajectory Recovery 1 Charles TOTH, 1 Dorota BRZEZINSKA, USA 2 Allison KEALY, Australia, 3 Guenther RETSCHER,

More information