H2020 Space Robotic SRC- OG4
|
|
- Bennett Collins
- 5 years ago
- Views:
Transcription
1 H2020 Space Robotic SRC- OG4 CCT/COMET ORB Workshop on Space Rendezvous 05/12/2017 «Smart Sensors for Smart Missions» Contacts: Sabrina Andiappane, Vincent Dubanchet, This project has received funding from the European Union s Horizon 2020 research and innovation programme under grant agreement No
2 Outline 1. General Presentation of I3DS Context What is I3DS? I3DS Consortium 2. Development plan 3. Design of I3DS suite 4. Validation usecase 5. GNC elements 6. Conclusion & Q&A 2
3 1.0 - Context I3DS (Integrated 3D Sensors) is one the six consortiums (Operational Grant 4) involved in the frame of the Strategic Research Cluster (SRC) on Space Robotics technologies Horizon 2020 Space call 2016 to enhance the EU industry competitiveness in space robotics. I3DS project develops a generic modular Inspector Sensors Suite (INSES) which is a smart collection of building blocks and a common set of various sensors answering the needs of near-future space exploration missions in terms of remote and contact sensors with integrated pre-processing and data concentration functions. 3
4 1.1 - I3DS: Integrated 3D Sensors 4
5 1.2 - I3DS Objectives Modular approach: Inter-changeable items Scalable design Standard I/F vs satellite/rover platform Realise a suite of perception sensors for both orbital and planetary applications that will allow localisation and map-making for robotic inspection of orbital assets and for planetary surface exploration Subsets to be defined according to the application: Planetary mission: rover Non-cooperative target capture: debris removal Cooperative rendezvous: servicing, spacetug 5
6 2. Development Plan Development on-going for I3DS building blocks: ICU, sensors, software, pre-processing algorithms and mechanical housing To have a top down approach for design: System requirements Functional requirements Sensors requirements To have a bottom up approach for validation Test each sensor individually Test at system level for one selected scenario to validate the coherence of the sensor suite 6
7 3. I3DS Suite - Elements Credit SINTEF Photo: Werner Juvik 7
8 3. I3DS Suite - Pre-processing Goal: Embedded algorithms on the ICU For all the cameras: Stereo Camera, HR Camera and TIR Camera Vignetting Correction Optical Distortion Correction Histogram Equalisation For the Stereo Camera Stereo Rectification For the HR camera + Pattern Projector Structured light pattern description Depth map/point cloud production from structured light 8
9 4. Validation Usecases Two main usecases Orbital : on-orbit rendezvous & capture for servicing Planetary : autonomous science while traversing (mapping, sample characterization) 9
10 4. Validation Usecases - Orbital Missions Phases Step 1: Inertial Navigation / Drifting orbit Step2: Far-range acquisition of the target Step3: Inspection of the target Step4: Final Approach + Tracking Step5: Capture Step5-bis: Emergency manoeuver Medium-range Approach > 200m Inspection / Observation ~ 20m Rendezvous < 20m Berthing / Docking ~ 3m Servicing 0m 10
11 4. Validation Usecases - Orbital Robotic test bench used to reproduce free-floating dynamics Sensors integrated in one mechanical housing for testing 11
12 4. Validation Usecases - Orbital Closed loop validation (foreseen in summer 2018) Need for extra pre-processing and data fusion algorithms to be on the EGSE (limit with other consortium specialized in data fusion: OG3) o Dense or sparse stereo matching o Dense or sparse point cloud production o Point cloud classification and outlier rejection o Data fusion: for instance Stereo + Lidar Development of a simplified GNC loop for testing Hardware-in-the-loop validation mixing sensors HW and simulator SW 12
13 5. GNC Elements - Guidance Cooperative/Collaborative target No tumbling motion Dedicated markers and grapple fixture for capture Two main sub-phases of the rendezvous performed for validation Ellipse of inspection Straight line approach 13
14 5. GNC Elements - Guidance Simulator outputs for Navigation algorithms Open loop guidance : results obtained from simulations (at the moment) Ellipse of inspection (20-60 m) LIDAR (8 key points) STEREO (Relative pos/att) 14
15 5. GNC Elements - Navigation Different processing algorithms by sensor Stereo Camera: pose estimation based on SLAM algorithm Monocular Camera : pose estimation based on visual odometry LIDAR: Pose estimation based on raw ICP algorithm Camera+Projector: point cloud generation using sine pattern Data fusion foreseen between LIDAR and Stereo Camera pose estimation refinement using Kalman filter 15
16 5. GNC Elements - Navigation Geometric Calibration Checkerboard for calibrating camera on a given set of frames Gray Code patterns for calibrating projector (depending on projector resolution) Relative pose estimation between camera and projector 16
17 5. GNC Elements 3D mapping Innovative 3D mapping technology Monocular camera + Pattern projector Short-range sensor < 5m (end-effector sensor) Able to overcome the sunlight 17
18 5. GNC Elements 3D mapping Innovative 3D mapping technology Preliminary testing with different panels 18
19 5. GNC Elements Image Generation Software-in-the-Loop validation of navigation algorithms Using internal tool «SPICAM» o Camera electronics o Camera optics o Lighting conditions o Thermal imaging Planck CAD model displayed by SpiCam Ganymede and Jupiter displayed by SpiCam I3DS target spacecraft derived from ELITE Platform 19
20 Conclusion I3DS is a system on its own with its own hardware and software with embedded processing for on-board sensing capabilities Enables Robotic payload coupling with state-of-the-art platforms Key building block for future robotic missions Tests foreseen in summer 2018 Hope to present them in the next COMET ORB Worskshop! Credit Thales Alenia Space 20
21 Thank you for your attention! 21
H2020 Space Robotic SRC- OG4
H2020 Space Robotic SRC- OG4 2 nd PERASPERA workshop Presentation by Sabrina Andiappane Thales Alenia Space France This project has received funding from the European Union s Horizon 2020 research and
More informationRendezvous sensors and navigation
Clean Space Industrial Days Rendezvous sensors and navigation 23-27 th May 2016 Nicolas Deslaef (TAS-F) Julien Christy (TAS-F) 83230918-DOC-TAS-FR-001 OPEN Agenda 2 Collection of R&T studies enabling vision
More informationI3DS SENSOR SUITE FOR SPACE ROBOTICS
I3DS SENSOR SUITE FOR SPACE ROBOTICS Kristoffer Nyborg Gregertsen, PhD Ada-Europe 2018, Lisbon ESA/Hubble About me SINTEF Digital Department of Mathematics and Cybernetics Research manager for automation
More informationInFuse: A Comprehensive Framework for Data Fusion in Space Robotics
InFuse InFuse: A Comprehensive Framework for Data Fusion in Space Robotics June 20 th, 2017 Shashank Govindaraj (Space Applications Services, Belgium) Overview 1. Motivations & Objectives 2. InFuse within
More informationt Bench for Robotics and Autonomy Andrea Merlo
t Bench for Robotics and Autonomy Andrea Merlo Agenda Introduction TBRA Overview Objectives Architecture / Technical Description Status Test Results Roadmap he context of a Rover, dance, Navigation and
More informationW4. Perception & Situation Awareness & Decision making
W4. Perception & Situation Awareness & Decision making Robot Perception for Dynamic environments: Outline & DP-Grids concept Dynamic Probabilistic Grids Bayesian Occupancy Filter concept Dynamic Probabilistic
More informationMULTI-MODAL MAPPING. Robotics Day, 31 Mar Frank Mascarich, Shehryar Khattak, Tung Dang
MULTI-MODAL MAPPING Robotics Day, 31 Mar 2017 Frank Mascarich, Shehryar Khattak, Tung Dang Application-Specific Sensors Cameras TOF Cameras PERCEPTION LiDAR IMU Localization Mapping Autonomy Robotic Perception
More informationCNES robotics activities : Towards long distance on-board decision-making navigation
CNES robotics activities : Towards long distance on-board decision-making navigation S.MORENO sabine.moreno@cnes.fr Contents I. Introduction 1.Context 2.Definition II. CNES activities 1.Perception 2.Localisation
More informationVisual Pose Estimation System for Autonomous Rendezvous of Spacecraft
Visual Pose Estimation System for Autonomous Rendezvous of Spacecraft Mark A. Post1, Junquan Li2, and Craig Clark2 Space Mechatronic Systems Technology Laboratory Dept. of Design, Manufacture & Engineering
More informationApplication questions. Theoretical questions
The oral exam will last 30 minutes and will consist of one application question followed by two theoretical questions. Please find below a non exhaustive list of possible application questions. The list
More informationLAUROPE Six Legged Walking Robot for Planetary Exploration participating in the SpaceBot Cup
FZI RESEARCH CENTER FOR INFORMATION TECHNOLOGY LAUROPE Six Legged Walking Robot for Planetary Exploration participating in the SpaceBot Cup ASTRA 2015, ESA/ESTEC, Noordwijk 12. Mai 2015 Outline SpaceBot
More information3D Time-of-Flight Image Sensor Solutions for Mobile Devices
3D Time-of-Flight Image Sensor Solutions for Mobile Devices SEMICON Europa 2015 Imaging Conference Bernd Buxbaum 2015 pmdtechnologies gmbh c o n f i d e n t i a l Content Introduction Motivation for 3D
More informationPlanetary Rover Absolute Localization by Combining Visual Odometry with Orbital Image Measurements
Planetary Rover Absolute Localization by Combining Visual Odometry with Orbital Image Measurements M. Lourakis and E. Hourdakis Institute of Computer Science Foundation for Research and Technology Hellas
More informationVALIDATION OF 3D ENVIRONMENT PERCEPTION FOR LANDING ON SMALL BODIES USING UAV PLATFORMS
ASTRA 2015 VALIDATION OF 3D ENVIRONMENT PERCEPTION FOR LANDING ON SMALL BODIES USING UAV PLATFORMS Property of GMV All rights reserved PERIGEO PROJECT The work presented here is part of the PERIGEO project
More informationAutonomous navigation in industrial cluttered environments using embedded stereo-vision
Autonomous navigation in industrial cluttered environments using embedded stereo-vision Julien Marzat ONERA Palaiseau Aerial Robotics workshop, Paris, 8-9 March 2017 1 Copernic Lab (ONERA Palaiseau) Research
More informationIntroduction to Autonomous Mobile Robots
Introduction to Autonomous Mobile Robots second edition Roland Siegwart, Illah R. Nourbakhsh, and Davide Scaramuzza The MIT Press Cambridge, Massachusetts London, England Contents Acknowledgments xiii
More informationDealing with Scale. Stephan Weiss Computer Vision Group NASA-JPL / CalTech
Dealing with Scale Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline Why care about size? The IMU as scale provider: The
More informationDevelopment of a Ground Based Cooperating Spacecraft Testbed for Research and Education
DIPARTIMENTO DI INGEGNERIA INDUSTRIALE Development of a Ground Based Cooperating Spacecraft Testbed for Research and Education Mattia Mazzucato, Sergio Tronco, Andrea Valmorbida, Fabio Scibona and Enrico
More informationOn-Orbit Testing of Target-less TriDAR 3D Rendezvous and Docking Sensor
On-Orbit Testing of Target-less TriDAR 3D Rendezvous and Docking Sensor Stephane Ruel, Tim Luu, Andrew Berube* *Neptec Design Group, Canada e-mail: sruel@neptec.com Abstract TriDAR is a vision system developed
More informationStereo and Monocular Vision Applied to Computer Aided Navigation for Aerial and Ground Vehicles
Riccardo Giubilato Centro di Ateneo di Studi e Attivita Spaziali Giuseppe Colombo CISAS University of Padova 1 Introduction Previous Work Research Objectives Research Methodology Robot vision: Process
More informationDeep Incremental Scene Understanding. Federico Tombari & Christian Rupprecht Technical University of Munich, Germany
Deep Incremental Scene Understanding Federico Tombari & Christian Rupprecht Technical University of Munich, Germany C. Couprie et al. "Toward Real-time Indoor Semantic Segmentation Using Depth Information"
More informationVISION FOR AUTOMOTIVE DRIVING
VISION FOR AUTOMOTIVE DRIVING French Japanese Workshop on Deep Learning & AI, Paris, October 25th, 2017 Quoc Cuong PHAM, PhD Vision and Content Engineering Lab AI & MACHINE LEARNING FOR ADAS AND SELF-DRIVING
More informationLow Cost solution for Pose Estimation of Quadrotor
Low Cost solution for Pose Estimation of Quadrotor mangal@iitk.ac.in https://www.iitk.ac.in/aero/mangal/ Intelligent Guidance and Control Laboratory Indian Institute of Technology, Kanpur Mangal Kothari
More informationUAV Autonomous Navigation in a GPS-limited Urban Environment
UAV Autonomous Navigation in a GPS-limited Urban Environment Yoko Watanabe DCSD/CDIN JSO-Aerial Robotics 2014/10/02-03 Introduction 2 Global objective Development of a UAV onboard system to maintain flight
More information3D object recognition used by team robotto
3D object recognition used by team robotto Workshop Juliane Hoebel February 1, 2016 Faculty of Computer Science, Otto-von-Guericke University Magdeburg Content 1. Introduction 2. Depth sensor 3. 3D object
More informationSafe Long-Range Travel for Planetary Rovers through Forward Sensing
Safe Long-Range Travel for Planetary Rovers through Forward Sensing 12 th ESA Workshop on Advanced Space Technologies for Robotics & Automation Yashodhan Nevatia Space Applications Services May 16 th,
More informationInnovative Test-beds for Validation of Motion Recognition Systems Intended for Orbital Capture Manoeuvres
Innovative Test-beds for Validation of Motion Recognition Systems Intended for Orbital Capture Manoeuvres Tomasz Rybus 1, Karol Seweryn 1, Janusz Grzybowski 1, Janusz Nicolau-Kukliński 1, Rafał Przybyła
More informationMixed-Reality for Intuitive Photo-Realistic 3D-Model Generation
Mixed-Reality for Intuitive Photo-Realistic 3D-Model Generation Wolfgang Sepp, Tim Bodenmueller, Michael Suppa, and Gerd Hirzinger DLR, Institut für Robotik und Mechatronik @ GI-Workshop VR/AR 2009 Folie
More informationLocalization, Where am I?
5.1 Localization, Where am I?? position Position Update (Estimation?) Encoder Prediction of Position (e.g. odometry) YES matched observations Map data base predicted position Matching Odometry, Dead Reckoning
More informationModel-based Visual Tracking:
Technische Universität München Model-based Visual Tracking: the OpenTL framework Giorgio Panin Technische Universität München Institut für Informatik Lehrstuhl für Echtzeitsysteme und Robotik (Prof. Alois
More informationHybrids Mixed Approaches
Hybrids Mixed Approaches Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline Why mixing? Parallel Tracking and Mapping Benefits
More informationVisual Pose Estimation and Identification for Satellite Rendezvous Operations
Post, Mark A. and Yan, Xiu T. (2015) Visual pose estimation and identification for satellite rendezvous operations. In: Sixth China- Scotland SIPRA workshop "Recent Advances in Signal and Image Processing",
More informationVisual SLAM. An Overview. L. Freda. ALCOR Lab DIAG University of Rome La Sapienza. May 3, 2016
An Overview L. Freda ALCOR Lab DIAG University of Rome La Sapienza May 3, 2016 L. Freda (University of Rome La Sapienza ) Visual SLAM May 3, 2016 1 / 39 Outline 1 Introduction What is SLAM Motivations
More informationOUTDOOR AND INDOOR NAVIGATION WITH MICROSOFT KINECT
DICA-Dept. of Civil and Environmental Engineering Geodesy and Geomatics Section OUTDOOR AND INDOOR NAVIGATION WITH MICROSOFT KINECT Diana Pagliari Livio Pinto OUTLINE 2 The Microsoft Kinect sensor The
More informationME/CS 132: Introduction to Vision-based Robot Navigation! Low-level Image Processing" Larry Matthies"
ME/CS 132: Introduction to Vision-based Robot Navigation! Low-level Image Processing" Larry Matthies" lhm@jpl.nasa.gov, 818-354-3722" Announcements" First homework grading is done! Second homework is due
More informationSensor Fusion: Potential, Challenges and Applications. Presented by KVH Industries and Geodetics, Inc. December 2016
Sensor Fusion: Potential, Challenges and Applications Presented by KVH Industries and Geodetics, Inc. December 2016 1 KVH Industries Overview Innovative technology company 600 employees worldwide Focused
More informationNonlinear State Estimation for Robotics and Computer Vision Applications: An Overview
Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview Arun Das 05/09/2017 Arun Das Waterloo Autonomous Vehicles Lab Introduction What s in a name? Arun Das Waterloo Autonomous
More informationCSE 527: Introduction to Computer Vision
CSE 527: Introduction to Computer Vision Week 10 Class 2: Visual Odometry November 2nd, 2017 Today Visual Odometry Intro Algorithm SLAM Visual Odometry Input Output Images, Video Camera trajectory, motion
More information4D Crop Analysis for Plant Geometry Estimation in Precision Agriculture
4D Crop Analysis for Plant Geometry Estimation in Precision Agriculture MIT Laboratory for Information & Decision Systems IEEE RAS TC on Agricultural Robotics and Automation Webinar #37 Acknowledgements
More informationFlexible Visual Inspection. IAS-13 Industrial Forum Horizon 2020 Dr. Eng. Stefano Tonello - CEO
Flexible Visual Inspection IAS-13 Industrial Forum Horizon 2020 Dr. Eng. Stefano Tonello - CEO IT+Robotics Spin-off of University of Padua founded in 2005 Strong relationship with IAS-LAB (Intelligent
More informationMore processing power - Driving requirements
Software More processing power - Driving requirements Autonomy Task based commanding / resource management rather than timeline remote control E.g. planetary rovers drive from A B, look out for interesting
More informationPresentation Outline
Presentation Outline Phd Activities during the three years mainly concentrated on the development and testing of the SPARTANS cooperating spacecraft hardware testbed Development of the Translation Module
More informationTightly-Integrated Visual and Inertial Navigation for Pinpoint Landing on Rugged Terrains
Tightly-Integrated Visual and Inertial Navigation for Pinpoint Landing on Rugged Terrains PhD student: Jeff DELAUNE ONERA Director: Guy LE BESNERAIS ONERA Advisors: Jean-Loup FARGES Clément BOURDARIAS
More informationJena-Optronik Relative Navigation Sensor Activities
Jena-Optronik Relative Navigation Sensor Activities ESA AIM Industry Day Jena-Optronik Relative Navigation Sensor Activities ESA AIM Industry Day. Page 1 Space developments from Jena. 1976 1980 1991 1994
More informationCamera Drones Lecture 3 3D data generation
Camera Drones Lecture 3 3D data generation Ass.Prof. Friedrich Fraundorfer WS 2017 Outline SfM introduction SfM concept Feature matching Camera pose estimation Bundle adjustment Dense matching Data products
More informationSpace Robotics. Ioannis Rekleitis
Space Robotics Ioannis Rekleitis On-Orbit Servicing of Satellites Work done at the Canadian Space Agency Guy Rouleau, Ioannis Rekleitis, Régent L'Archevêque, Eric Martin, Kourosh Parsa, and Erick Dupuis
More informationGPS denied Navigation Solutions
GPS denied Navigation Solutions Krishnraj Singh Gaur and Mangal Kothari ksgaur@iitk.ac.in, mangal@iitk.ac.in https://www.iitk.ac.in/aero/mangal/ Intelligent Guidance and Control Laboratory Indian Institute
More informationVision-based navigation
Vision-based navigation Simon Lacroix To cite this version: Simon Lacroix. Vision-based navigation. Doctoral. Spring School on Location-Based Services, École Nationale d Aviation Civile, Toulouse (France),
More informationVALIDATION AND TUNNING OF DENSE STEREO-VISION SYSTEMS USING HI-RESOLUTION 3D REFERENCE MODELS
VALIDATION AND TUNNING OF DENSE STEREOVISION SYSTEMS USING HIRESOLUTION 3D REFERENCE MODELS E. REMETEAN (CNES) S. MAS (MAGELLIUM) JB GINESTET (MAGELLIUM) L.RASTEL (CNES) ASTRA 14.04.2011 CONTENT Validation
More informationDense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm
Computer Vision Group Prof. Daniel Cremers Dense Tracking and Mapping for Autonomous Quadrocopters Jürgen Sturm Joint work with Frank Steinbrücker, Jakob Engel, Christian Kerl, Erik Bylow, and Daniel Cremers
More informationTraining models for road scene understanding with automated ground truth Dan Levi
Training models for road scene understanding with automated ground truth Dan Levi With: Noa Garnett, Ethan Fetaya, Shai Silberstein, Rafi Cohen, Shaul Oron, Uri Verner, Ariel Ayash, Kobi Horn, Vlad Golder,
More informationStructured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov
Structured Light II Johannes Köhler Johannes.koehler@dfki.de Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Introduction Previous lecture: Structured Light I Active Scanning Camera/emitter
More informationTowards Fully-automated Driving. tue-mps.org. Challenges and Potential Solutions. Dr. Gijs Dubbelman Mobile Perception Systems EE-SPS/VCA
Towards Fully-automated Driving Challenges and Potential Solutions Dr. Gijs Dubbelman Mobile Perception Systems EE-SPS/VCA Mobile Perception Systems 6 PhDs, 1 postdoc, 1 project manager, 2 software engineers
More informationLocalization and Map Building
Localization and Map Building Noise and aliasing; odometric position estimation To localize or not to localize Belief representation Map representation Probabilistic map-based localization Other examples
More information5 member consortium o University of Strathclyde o Wideblue o National Nuclear Laboratory o Sellafield Ltd o Inspectahire
3 year, 1.24M Innovate UK funded Collaborative Research and Development Project (Nuclear Call) o Commenced April 2015 o Follows on from a successful 6 month Innovate UK funded feasibility study 2013-2014
More informationVisual SLAM for small Unmanned Aerial Vehicles
Visual SLAM for small Unmanned Aerial Vehicles Margarita Chli Autonomous Systems Lab, ETH Zurich Simultaneous Localization And Mapping How can a body navigate in a previously unknown environment while
More informationThe WiMUST H2020 project: Widely scalable Mobile Underwater Sonar Technology
The WiMUST H2020 project: Widely scalable Mobile Underwater Sonar Technology Giovanni Indiveri Coordinator - ISME WiMUST - Widely scalable Mobile Underwater Sonar Technology Grant agreement no: 645141
More information3D Fusion of Infrared Images with Dense RGB Reconstruction from Multiple Views - with Application to Fire-fighting Robots
3D Fusion of Infrared Images with Dense RGB Reconstruction from Multiple Views - with Application to Fire-fighting Robots Yuncong Chen 1 and Will Warren 2 1 Department of Computer Science and Engineering,
More informationCOMPUTER VISION FOR REAL-TIME RELATIVE NAVIGATION WITH A NON-COOPERATIVE AND SPINNING TARGET SPACECRAFT
COMPUTER VISION FOR REAL-TIME RELATIVE NAVIGATION WITH A NON-COOPERATIVE AND SPINNING TARGET SPACECRAFT Setareh Yazdkhasti, Steve Ulrich, and Jurek Z. Sasiadek Carleton University, Ottawa, Ontario, K1S
More informationMonitoring the environment with a Cloud Robotics Service
Monitoring the environment with a Cloud Robotics Service Ludovico O. Russo ludus.russo@gmail.com Outline Introduction Service Robotics Autonomous Navigation Cloud Robotics Robot Based Environmental Monitoring
More information3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.
3D Computer Vision Structured Light II Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de 1 Introduction
More informationA Modular Sensor Fusion Approach for Agricultural Machines
A Modular Sensor Fusion Approach for Agricultural Machines Sebastian Blank (1), Georg Kormann (2), Karsten Berns (1) (1) Robotics Research Lab University of Kaiserslautern, Germany (2) John Deere European
More informationBeate Klinger and Torsten Mayer-Gürr Institute of Geodesy NAWI Graz, Graz University of Technology
and Torsten Mayer-Gürr Institute of Geodesy NAWI Graz, Graz University of Technology Outline Motivation GRACE Preprocessing GRACE sensor fusion Accelerometer simulation & calibration Impact on monthly
More informationDirect Methods in Visual Odometry
Direct Methods in Visual Odometry July 24, 2017 Direct Methods in Visual Odometry July 24, 2017 1 / 47 Motivation for using Visual Odometry Wheel odometry is affected by wheel slip More accurate compared
More informationV-Sentinel: A Novel Framework for Situational Awareness and Surveillance
V-Sentinel: A Novel Framework for Situational Awareness and Surveillance Suya You Integrated Media Systems Center Computer Science Department University of Southern California March 2005 1 Objective Developing
More informationWP6 D6.2 Project website
WP6 D6.2 Project website Project title: Project Acronym: Promoting Youth Scientific Career Awareness and it Attractiveness through Multistakeholder Cooperation MultiCO Project ID: 665100 Prepared by: University
More informationUSE OF COTS ROBOTICS FOR ON-GROUND VALIDATION OF SPACE GNC SYSTEMS: PLATFORM DYNAMIC TEST BENCH I-SAIRAS 2012 TURIN, ITALY 4-6 SEPTEMBER 2012
USE OF COTS ROBOTICS FOR ON-GROUND VALIDATION OF SPACE GNC SYSTEMS: PLATFORM DYNAMIC TEST BENCH I-SAIRAS 212 TURIN, ITALY 4-6 SEPTEMBER 212 Matteo Suatoni (1), Luis Mollinedo (1), Valentín Barrena (1),
More informationAIRBUS DS VISION BASED NAVIGATION SOLUTIONS TESTED ON LIRIS EXPERIMENT DATA
AIRBUS DS VISION BASED NAVIGATION SOLUTIONS TESTED ON LIRIS EXPERIMENT DATA A. Masson(1), C. Haskamp (2), I. Ahrns (2), R. Brochard(1), P. Duteis (1), K. Kanani(1), R. Delage(1) (1) Airbus Defence and
More informationMONOCULAR OBJECT LOCALIZATION BY SUPERQUADRICS CURVATURE REPROJECTION AND MATCHING
GraphiCon 2012 (Moscow, Russia) University of Trento, Italy MONOCULAR OBJECT LOCALIZATION BY SUPERQUADRICS CURVATURE REPROJECTION AND MATCHING Enrico Zappia, Ilya Afanasyev, Nicolo Biasi, Mattia Tavernini,
More informationConcept and Performance Simulation with ASTOS
Concept and Performance Simulation with ASTOS Andreas Wiegand (1), Sven Weikert (1) Astos Solutions GmbH (1) Meitnerstraße 8, 70563 Stuttgart, Germany andreas.wiegand@astos.de ABSTRACT Advanced space missions,
More informationOn Satellite Vision-aided Robotics Experiment
On Satellite Vision-aided Robotics Experiment Maarten Vergauwen, Marc Pollefeys, Tinne Tuytelaars and Luc Van Gool ESAT PSI, K.U.Leuven, Kard. Mercierlaan 94, B-3001 Heverlee, Belgium Phone: +32-16-32.10.64,
More informationLeica Systems Overview
RC30 AERIAL CAMERA SYSTEM Leica Systems Overview The Leica RC30 aerial film camera is the culmination of decades of development, started with Wild's first aerial camera in the 1920s. Beautifully engineered
More informationOn-line and Off-line 3D Reconstruction for Crisis Management Applications
On-line and Off-line 3D Reconstruction for Crisis Management Applications Geert De Cubber Royal Military Academy, Department of Mechanical Engineering (MSTA) Av. de la Renaissance 30, 1000 Brussels geert.de.cubber@rma.ac.be
More informationAUTONOMOUS LONG-RANGE ROVER NAVIGATION EXPERIMENTAL RESULTS
AUTONOMOUS LONG-RANGE ROVER NAVIGATION EXPERIMENTAL RESULTS Erick Dupuis, Joseph Nsasi Bakambu, Ioannis Rekleitis, Jean-Luc Bedwani, Sébastien Gemme, Jean-Patrice Rivest-Caissy Canadian Space Agency 6767
More informationSpatialFuser. Offers support for GNSS and INS navigation systems. Supports multiple remote sensing equipment configurations.
SPATIALFUSER SpatialFuser SpatialFuser is designed to georeference data acquired by our LiDAR mapping systems into common mapping formats such as LAS/LAZ or GPX. The main purpose of SpatialFuser is to
More informationGNC for a Rendezvous in Space with an Uncooperative Target.
GNC for a Rendezvous in Space with an Uncooperative Target Josef Sommer (1), Ingo Ahrns (2) (1) Astrium GmbH, Airbusallee 1, 28199 Bremen, Germany, +49 421 539 4440, josef.sommer@astrium.eads.net (2) Astrium
More informationBasics of Localization, Mapping and SLAM. Jari Saarinen Aalto University Department of Automation and systems Technology
Basics of Localization, Mapping and SLAM Jari Saarinen Aalto University Department of Automation and systems Technology Content Introduction to Problem (s) Localization A few basic equations Dead Reckoning
More informationCONTENT ENGINEERING & VISION LABORATORY. Régis Vinciguerra
CONTENT ENGINEERING & VISION LABORATORY Régis Vinciguerra regis.vinciguerra@cea.fr ALTERNATIVE ENERGIES AND ATOMIC ENERGY COMMISSION Military Applications Division (DAM) Nuclear Energy Division (DEN) Technological
More informationAutonomous Navigation in Complex Indoor and Outdoor Environments with Micro Aerial Vehicles
Autonomous Navigation in Complex Indoor and Outdoor Environments with Micro Aerial Vehicles Shaojie Shen Dept. of Electrical and Systems Engineering & GRASP Lab, University of Pennsylvania Committee: Daniel
More informationIncremental Real-time Bundle Adjustment for Multi-camera Systems with Points at Infinity
Incremental Real-time Bundle Adjustment for Multi-camera Systems with Points at Infinity Johannes Schneider, Thomas Läbe, Wolfgang Förstner 1 Department of Photogrammetry Institute of Geodesy and Geoinformation
More informationDense 3D Reconstruction from Autonomous Quadrocopters
Dense 3D Reconstruction from Autonomous Quadrocopters Computer Science & Mathematics TU Munich Martin Oswald, Jakob Engel, Christian Kerl, Frank Steinbrücker, Jan Stühmer & Jürgen Sturm Autonomous Quadrocopters
More informationPROJECT HIGHLIGHTS, EXPECTED IMPACT & FUTURE DIRECTIONS
PROJECT HIGHLIGHTS, EXPECTED IMPACT & FUTURE DIRECTIONS Dr. Angelos Amditis (project coordinator) Institute of Communication and Computer Systems (ICCS), Research Director Final Event AstaZero, Sandhult,
More informationHow can the Future Internet
How can the Future Internet enable Smart Energy? Presented by Werner Mohr (Coordinator), Nokia Siemens Networks on behalf of the FINSENY project Smart Energy enabled by Future Internet Workshop FINSENY
More informationINTELLIGENT AUTONOMOUS SYSTEMS LAB
Matteo Munaro 1,3, Alex Horn 2, Randy Illum 2, Jeff Burke 2, and Radu Bogdan Rusu 3 1 IAS-Lab at Department of Information Engineering, University of Padova 2 Center for Research in Engineering, Media
More informationGNSS/INS for High Accuracy Mobile Mapping. Olaf Gross 11 th Terrasolid European User Event Kittilä, Finland
GNSS/INS for High Accuracy Mobile Mapping Olaf Gross 11 th Terrasolid European User Event 15.02.2012 Kittilä, Finland IGI mbh Founded 1978 25 staff about half in R&D More than 380 customers in 60 different
More informationPhotoneo's brand new PhoXi 3D Camera is the highest resolution and highest accuracy area based 3D
Company: Photoneo s.r.o. Germany Contact: Veronika Pulisova E-mail: pulisova@photoneo.com PhoXi 3D Camera Author: Tomas Kovacovsky & Jan Zizka Description of the innovation: General description Photoneo's
More informationMiniature faking. In close-up photo, the depth of field is limited.
Miniature faking In close-up photo, the depth of field is limited. http://en.wikipedia.org/wiki/file:jodhpur_tilt_shift.jpg Miniature faking Miniature faking http://en.wikipedia.org/wiki/file:oregon_state_beavers_tilt-shift_miniature_greg_keene.jpg
More informationSGEO: Overview and Product Offering. Marco R. Fuchs. Marco R. R. Fuchs. DLR-ESA Workshop on ARTES 11. Marco R. Fuchs OHB Technology AG
DLR-ESA Workshop on ARTES 11 SGEO: Overview and Product Offering Marco R. R. Fuchs June June29, 29, 2006 2006 Tegernsee, Tegernsee, Germany Germany Marco R. Fuchs Marco R. Fuchs OHB Technology AG OHB Technology
More information3D Point Cloud Processing
3D Point Cloud Processing The image depicts how our robot Irma3D sees itself in a mirror. The laser looking into itself creates distortions as well as changes in intensity that give the robot a single
More informationDARPA Perspective on Space
DARPA Perspective on Space Ms. Pamela A. Melroy, Deputy Director DARPA Tactical Technology Office Briefing prepared for International Symposium for Personal and Commercial Spaceflight (ISPCS) 2015 October
More informationOutline. ETN-FPI Training School on Plenoptic Sensing
Outline Introduction Part I: Basics of Mathematical Optimization Linear Least Squares Nonlinear Optimization Part II: Basics of Computer Vision Camera Model Multi-Camera Model Multi-Camera Calibration
More informationAdvanced Driver Assistance: Modular Image Sensor Concept
Vision Advanced Driver Assistance: Modular Image Sensor Concept Supplying value. Integrated Passive and Active Safety Systems Active Safety Passive Safety Scope Reduction of accident probability Get ready
More informationNew Sony DepthSense TM ToF Technology
ADVANCED MATERIAL HANDLING WITH New Sony DepthSense TM ToF Technology Jenson Chang Product Marketing November 7, 2018 1 3D SENSING APPLICATIONS Pick and Place Drones Collision Detection People Counting
More informationSolid State LiDAR for Ubiquitous 3D Sensing
April 6, 2016 Solid State LiDAR for Ubiquitous 3D Sensing Louay Eldada, Ph.D. CEO, Co-founder Quanergy Systems New Paradigm in 3D Sensing Disruptive Technologies: Solid State 3D LiDAR sensors Embedded
More informationManaging and mining (streaming) sensor data
Petr Čížek Artificial Intelligence Center Czech Technical University in Prague November 3, 2016 Petr Čížek VPD 1 / 1 Stream data mining / stream data querying Problem definition Data can not be stored
More informationAugmented Reality, Advanced SLAM, Applications
Augmented Reality, Advanced SLAM, Applications Prof. Didier Stricker & Dr. Alain Pagani alain.pagani@dfki.de Lecture 3D Computer Vision AR, SLAM, Applications 1 Introduction Previous lectures: Basics (camera,
More informationWhy the Self-Driving Revolution Hinges on one Enabling Technology: LiDAR
Why the Self-Driving Revolution Hinges on one Enabling Technology: LiDAR Markus Prison Director Business Development Europe Quanergy ID: 23328 Who We Are The leader in LiDAR (laser-based 3D spatial sensor)
More informationOverview. EECS 124, UC Berkeley, Spring 2008 Lecture 23: Localization and Mapping. Statistical Models
Introduction ti to Embedded dsystems EECS 124, UC Berkeley, Spring 2008 Lecture 23: Localization and Mapping Gabe Hoffmann Ph.D. Candidate, Aero/Astro Engineering Stanford University Statistical Models
More informationVisual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech
Visual Odometry Features, Tracking, Essential Matrix, and RANSAC Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline The
More informationPersonal Navigation and Indoor Mapping: Performance Characterization of Kinect Sensor-based Trajectory Recovery
Personal Navigation and Indoor Mapping: Performance Characterization of Kinect Sensor-based Trajectory Recovery 1 Charles TOTH, 1 Dorota BRZEZINSKA, USA 2 Allison KEALY, Australia, 3 Guenther RETSCHER,
More information