MASTER: Mobile Autonomous Scientist for Terrestrial and Extra-terrestrial Research
|
|
- Dorothy Hodge
- 5 years ago
- Views:
Transcription
1 MASTER: Mobile Autonomous Scientist for Terrestrial and Extra-terrestrial Research Dr. Iain Wallace, Dr. Mark Woods Autonomous & Intelligent Systems Group Wednesday 13 th May 2015
2 What is MASTER? Why do we need it? 2
3 Missing the Obvious? NASA/JPL Retracing its steps 3
4 4 Why are SCISYS doing this work?
5 Science Autonomy Heritage 5
6 6
7 Neuro-Fuzzy, Computer Vision Based Science Classification Used FL flight code 7
8 Fusing Technologies Scene Understanding + Tactical Planning = Closed loop science Autonomously detect what s interesting to scientists Autonomously and safely work out how to gather more data We have a basis for an autonomous scientific robotic apprentice 8
9 9 STOP-PRESS: ProviScout win in FP 7 CREST Autonomous Robotic Scientist Showed that autonomous, science data acquisition/investigation was feasible
10 Full Autonomy Trials in Etna, Iceland, or Tenerife Unusual Material WP4 Interesting Stop WP3 Out Crop WP1 WP2 Unusual Material WP0 10
11 Step Back Need for fundamental approach and review 11
12 Objectives Conceive a generic novelty detection system architecture Test with Planetary Science and EO examples But this is an open research problem Currently several algorithmic choices for detection pipeline steps but which ones work best? MASTER is GSP Early TRL so we: built the prototype architecture implementation and a methodology to evaluate the algorithmic choices System allows you to use alternate algs. for different steps and therefore compare 12
13 MASTER Team Prime Contractor, Req s Definition, System Design Development and Test SCISYS State of the Art Literature Review MRG Uni. of Oxford Planetary Science Dataset Uni. of Leicester Earth Observation Dataset Pixalytics 13
14 Robotic ARM Autonomy and Robotics Group Scope Remote Operations Operators Autonomy Framework Mission Planning Payload Insts. Mission Planning Scene Analysis 3D Data Understanding Data Autonomous Navigation Robotic Platform + low-level software 14
15 The MASTER Problem Can you spot what s unusual in this image? 1 image ~10cm traverse ~2.5s operations
16 The MASTER Problem Can you spot it now? 9 images ~1m traverse ~25s operations and you know what you re looking for.
17 The MASTER Problem Can you spot it now? 25 images ~2.5m traverse ~1 min operations
18 The MASTER Problem Can you spot it now? 100 images ~10m traverse ~5 mins operations
19 The MASTER Problem Miss anything else interesting? 250 images, ~25m traverse, ~15 mins operations How many can be saved/returned? How do you choose?
20 MASTER Objectives A domain independent software architecture for detecting scientific events in images.» This was narrowed to detecting (and defining) novelty. Capturing expert scientist s knowledge. Two test domains» Earth Observation» Planetary Science System output is location and extent of regions of interest. Project output is a detailed analysis of the problem and properties of the presented solution.
21 Classes of Novelty Novel Not Novel Expected This case corresponds to easily classified phenomena seen in training. Unusual Phenomena that can be classified, but an expert has somehow indicated would be unusual. Unexpected The phenomena primarily applied to different contexts in training. Classified Novelty Reasonably probable the classifier has identified the phenomena, but it is a statistical outlier. Unclassified Novelty Where the output is above the threshold to create a new class, this is unclassified novelty.
22 Test Datasets - PS NASA/JPL Photojournal Planetary Image archive
23 Test Datasets PS SAFER Collected during ESA SAFER trial.
24 Test Datasets - Seeker Rover NAVCAM from ESA Seeker trial
25 Test Datasets - EO Processed MERIS Level 1 data
26 System Overview
27 Saliency What s outstanding in this image?» Looking for local interest. Many algorithms evaluated» Complete system can select from a library. Example Saliency Maps Left-to-right: Original image, scale-space saliency map, image-signature saliency map.
28 Saliency Evaluation Conclusions AUC SK Pssafer PS EO Performance varies by domain» Varies by phenomena of interest too. SAFER Dataset performance appears poor, but this is a relative measure not absolute.» Reflects the relatively sparse labelling. Seeker data also exhibited very tight error bounds due to very homogenous data. Operating point selection important, as saliency represents a bound on performance.
29 Classifiers classifier Feature Vector 0.8 Well-proven HOG/Visual Words/Spatial histogram based features» Others also tested, ICF, ACF SVM Classifiers» libsvm implementation» One classifier per class» Trained one versus the rest Parameter selection, then training on complete training set.
30 Summary of Classifier Performance On average, they work most of the time
31 Classification Conclusions Performance on the whole good.» Generality exceeded our expectations. Best performance with homogeneity between test and training.» As seen in SAFER and Seeker data. More training examples improve performance.» 25 samples a sensible lower bound in homogenous data. Classes should be well defined.» Poor relative performance of roi and flare classes.
32 Novelty Detection Simple Novelty» If the most confident classifier is above a threshold, assign a not-novel class according to training frequency. Otherwise Unclassified Novelty. Group Novelty Classification» Similar to the above, but considers also many confident classes (confusion) to be unclassified novelty, and defines classified novelty as a large margin to an uncertain class or many unsure classes.
33 System Evaluation What is novelty, in relation to the training data? How important is the detection of novelty in an image? How important is identifying the location and extent of any novelty? How desirable is it to classify regions of interest that are not novel? What is the appropriate trade-off between detecting novelty, and the cost of false positives? Is the problem one of selection is this image novel? Is the problem one of ranking which is the most novel image?
34 Recall Varying Thresholds Recall Varying Novelty Varying Saliency False-Discovery Rate False-Discovery Rate
35 Labelling Future Possibilities
36 Carnot Discovery - Saliency
37 UKSA Chameleon - Classifiers
38 UKSA Chameleon - Performance average time from command to capture (seconds) CPU consumption (watts) sensor power consumption (hardware) memory usage (component specified, Mb) energy cost over idle overseer (joules, rough) Data product or Test std. dev data product size (Mb) notes idle system na na n/a running overseer, no sensors or data capture, average over 10 minutes na na Mb in 253s - generating map xtion m imu Hz SSC Mode 1 rectified images PAIR Hz hokuyo to PC Hz XB3 rectified images Hz commanded map to occupancy grid (includes dtm) n/a SSC 1 pose 5Hz, PC 0.5hz ROR, Mapp SSC Mode 7 rectified image Hz Xtion QVGA@60Hz PC mode 2, 60hz, SOR filtered XB3 PC Hz Xtion VGA@30Hz PC hz, mode 0, SOR filtered SSC Mode 1 VO estimate Hz still camera on tripod teval SSC mode1, 15Hz XB3 VO estimate Hz commanded SSC Mode 1 PC no filtering Hz (commanded) SSC Mode 7 VO estimate Hz mapping eval eval 1Hz, SSC mode1 for PC 1Hz no f generating map SSC SSC 1Hz PC no filter, Mapping 1Hz generating map SSC SSC 1Hz PC no filter, Mapping 1Hz XB3 PC SOR filter Hz XB3 PC ROR filter Hz SSC Mode 7 PC Hz SSC Mode 1 PC SOR Hz SSC Mode 1 PC ROR Hz SSC Mode 7 PC SOR filter Hz SSC Mode 7 PC ROR filter Hz
39 General Conclusions General classifiers show good performance Visual saliency shows good utility. Evaluation of novelty detection is not simple. Labelling and annotation is a big deal.» Several good datasets created. Future application performance baselines have implications:» Where all images are inspected e.g. labelling» Selecting a few images for downlink» No images are inspected e.g. navcams. Evaluation is slow, execution is fast.
40 We are recruiting for autonomy R&D engineers. Thank you See our rovers in the exhibition area this morning!
Presented by Dr. Mark Woods. 15 th May 2013 ESTEC
Presented by Dr. Mark Woods 15 th May 2013 ESTEC Seeker Long Range Autonomous Navigation in the Atacama Desert - Chile Field Trial Team - Andy Shaw, Brian Maddison, Aron Kisdi, Wayne Tubbey, Estelle Tidey,
More informationQuasi-thematic Features Detection & Tracking. Future Rover Long-Distance Autonomous Navigation
Quasi-thematic Feature Detection And Tracking For Future Rover Long-Distance Autonomous Navigation Authors: Affan Shaukat, Conrad Spiteri, Yang Gao, Said Al-Milli, and Abhinav Bajpai Surrey Space Centre,
More informationTightly-Integrated Visual and Inertial Navigation for Pinpoint Landing on Rugged Terrains
Tightly-Integrated Visual and Inertial Navigation for Pinpoint Landing on Rugged Terrains PhD student: Jeff DELAUNE ONERA Director: Guy LE BESNERAIS ONERA Advisors: Jean-Loup FARGES Clément BOURDARIAS
More informationRendezvous sensors and navigation
Clean Space Industrial Days Rendezvous sensors and navigation 23-27 th May 2016 Nicolas Deslaef (TAS-F) Julien Christy (TAS-F) 83230918-DOC-TAS-FR-001 OPEN Agenda 2 Collection of R&T studies enabling vision
More informationVisual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech
Visual Odometry Features, Tracking, Essential Matrix, and RANSAC Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline The
More informationPlanetary Rover Absolute Localization by Combining Visual Odometry with Orbital Image Measurements
Planetary Rover Absolute Localization by Combining Visual Odometry with Orbital Image Measurements M. Lourakis and E. Hourdakis Institute of Computer Science Foundation for Research and Technology Hellas
More informationROBUST VISUAL ODOMETRY FOR SPACE EXPLORATION: 12 TH SYMPOSIUM ON ADVANCED SPACE TECHNOLOGIES IN ROBOTICS AND AUTOMATION
ROBUST VISUAL ODOMETRY FOR SPACE EXPLORATION: 12 TH SYMPOSIUM ON ADVANCED SPACE TECHNOLOGIES IN ROBOTICS AND AUTOMATION Dr Andrew Shaw (1), Dr Mark Woods (1), Dr Winston Churchill (2), Prof. Paul Newman
More information4D Crop Analysis for Plant Geometry Estimation in Precision Agriculture
4D Crop Analysis for Plant Geometry Estimation in Precision Agriculture MIT Laboratory for Information & Decision Systems IEEE RAS TC on Agricultural Robotics and Automation Webinar #37 Acknowledgements
More informationVALIDATION OF 3D ENVIRONMENT PERCEPTION FOR LANDING ON SMALL BODIES USING UAV PLATFORMS
ASTRA 2015 VALIDATION OF 3D ENVIRONMENT PERCEPTION FOR LANDING ON SMALL BODIES USING UAV PLATFORMS Property of GMV All rights reserved PERIGEO PROJECT The work presented here is part of the PERIGEO project
More informationEvaluation Metrics. (Classifiers) CS229 Section Anand Avati
Evaluation Metrics (Classifiers) CS Section Anand Avati Topics Why? Binary classifiers Metrics Rank view Thresholding Confusion Matrix Point metrics: Accuracy, Precision, Recall / Sensitivity, Specificity,
More informationSafe Long-Range Travel for Planetary Rovers through Forward Sensing
Safe Long-Range Travel for Planetary Rovers through Forward Sensing 12 th ESA Workshop on Advanced Space Technologies for Robotics & Automation Yashodhan Nevatia Space Applications Services May 16 th,
More informationExploiting scene constraints to improve object detection algorithms for industrial applications
Exploiting scene constraints to improve object detection algorithms for industrial applications PhD Public Defense Steven Puttemans Promotor: Toon Goedemé 2 A general introduction Object detection? Help
More informationDense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm
Computer Vision Group Prof. Daniel Cremers Dense Tracking and Mapping for Autonomous Quadrocopters Jürgen Sturm Joint work with Frank Steinbrücker, Jakob Engel, Christian Kerl, Erik Bylow, and Daniel Cremers
More informationLAUROPE Six Legged Walking Robot for Planetary Exploration participating in the SpaceBot Cup
FZI RESEARCH CENTER FOR INFORMATION TECHNOLOGY LAUROPE Six Legged Walking Robot for Planetary Exploration participating in the SpaceBot Cup ASTRA 2015, ESA/ESTEC, Noordwijk 12. Mai 2015 Outline SpaceBot
More informationApplying Model Predictive Control Architecture for Efficient Autonomous Data Collection and Operations on Planetary Missions
Applying Model Predictive Control Architecture for Efficient Autonomous Data Collection and Operations on Planetary Missions M. Lieber, E. Schindhelm, R. Rohrschneider, C. Weimer, S. Roark Cahill Center
More informationCAP 6412 Advanced Computer Vision
CAP 6412 Advanced Computer Vision http://www.cs.ucf.edu/~bgong/cap6412.html Boqing Gong April 21st, 2016 Today Administrivia Free parameters in an approach, model, or algorithm? Egocentric videos by Aisha
More informationSmartHMA Introduction of SmartHMA project objectives
SmartHMA Introduction of SmartHMA project objectives Daniel Zinkiewicz (daniel.zinkiewicz@wasat.pl) Wasat Sp. z o.o. Slide 1 SmartHMA SmartHMA SmartHMA mobile platform for deployment of HMA standardised
More information3D object recognition used by team robotto
3D object recognition used by team robotto Workshop Juliane Hoebel February 1, 2016 Faculty of Computer Science, Otto-von-Guericke University Magdeburg Content 1. Introduction 2. Depth sensor 3. 3D object
More informationSpace Robotics. Ioannis Rekleitis
Space Robotics Ioannis Rekleitis On-Orbit Servicing of Satellites Work done at the Canadian Space Agency Guy Rouleau, Ioannis Rekleitis, Régent L'Archevêque, Eric Martin, Kourosh Parsa, and Erick Dupuis
More informationIntensity-Difference Based 3D Video Stabilization for Planetary Robots
Intensity-Difference Based 3D Video Stabilization for Planetary Robots Geovanni Martinez Image Processing and Computer Vision Research Laboratory (IPCV-LAB) Escuela de Ingeniería Eléctrica Universidad
More informationCOSC160: Detection and Classification. Jeremy Bolton, PhD Assistant Teaching Professor
COSC160: Detection and Classification Jeremy Bolton, PhD Assistant Teaching Professor Outline I. Problem I. Strategies II. Features for training III. Using spatial information? IV. Reducing dimensionality
More informationBAYESIAN GLOBAL OPTIMIZATION
BAYESIAN GLOBAL OPTIMIZATION Using Optimal Learning to Tune Deep Learning Pipelines Scott Clark scott@sigopt.com OUTLINE 1. Why is Tuning AI Models Hard? 2. Comparison of Tuning Methods 3. Bayesian Global
More informationROBOT TEAMS CH 12. Experiments with Cooperative Aerial-Ground Robots
ROBOT TEAMS CH 12 Experiments with Cooperative Aerial-Ground Robots Gaurav S. Sukhatme, James F. Montgomery, and Richard T. Vaughan Speaker: Jeff Barnett Paper Focus Heterogeneous Teams for Surveillance
More informationt Bench for Robotics and Autonomy Andrea Merlo
t Bench for Robotics and Autonomy Andrea Merlo Agenda Introduction TBRA Overview Objectives Architecture / Technical Description Status Test Results Roadmap he context of a Rover, dance, Navigation and
More informationTowards Fully-automated Driving. tue-mps.org. Challenges and Potential Solutions. Dr. Gijs Dubbelman Mobile Perception Systems EE-SPS/VCA
Towards Fully-automated Driving Challenges and Potential Solutions Dr. Gijs Dubbelman Mobile Perception Systems EE-SPS/VCA Mobile Perception Systems 6 PhDs, 1 postdoc, 1 project manager, 2 software engineers
More informationAn Automatic Method for Adjustment of a Camera Calibration Room
An Automatic Method for Adjustment of a Camera Calibration Room Presented at the FIG Working Week 2017, May 29 - June 2, 2017 in Helsinki, Finland Theory, algorithms, implementation, and two advanced applications.
More informationSpacecraft Communications Payload (SCP) for Swampworks
Spacecraft Communications Payload (SCP) for Swampworks Joseph A. Hauser Praxis, Inc. 2550 Huntington Avenue Suite 300 Alexandria, VA 22303 phone: (703) 682-2059 fax: (703) 837-8500 email: hauserj@pxi.com
More informationLearning to Grasp Objects: A Novel Approach for Localizing Objects Using Depth Based Segmentation
Learning to Grasp Objects: A Novel Approach for Localizing Objects Using Depth Based Segmentation Deepak Rao, Arda Kara, Serena Yeung (Under the guidance of Quoc V. Le) Stanford University Abstract We
More informationToward Retail Product Recognition on Grocery Shelves
Toward Retail Product Recognition on Grocery Shelves Gül Varol gul.varol@boun.edu.tr Boğaziçi University, İstanbul, Turkey İdea Teknoloji Çözümleri, İstanbul, Turkey Rıdvan S. Kuzu ridvan.salih@boun.edu.tr
More informationRecognizing Apples by Piecing Together the Segmentation Puzzle
Recognizing Apples by Piecing Together the Segmentation Puzzle Kyle Wilshusen 1 and Stephen Nuske 2 Abstract This paper presents a system that can provide yield estimates in apple orchards. This is done
More informationToday MAPS AND MAPPING. Features. process of creating maps. More likely features are things that can be extracted from images:
MAPS AND MAPPING Features In the first class we said that navigation begins with what the robot can see. There are several forms this might take, but it will depend on: What sensors the robot has What
More informationCS 4758: Automated Semantic Mapping of Environment
CS 4758: Automated Semantic Mapping of Environment Dongsu Lee, ECE, M.Eng., dl624@cornell.edu Aperahama Parangi, CS, 2013, alp75@cornell.edu Abstract The purpose of this project is to program an Erratic
More informationH2020 Space Robotic SRC- OG4
H2020 Space Robotic SRC- OG4 CCT/COMET ORB Workshop on Space Rendezvous 05/12/2017 «Smart Sensors for Smart Missions» Contacts: Sabrina Andiappane, sabrina.andiappane@thalesaleniaspace.com Vincent Dubanchet,
More informationRecognition of Animal Skin Texture Attributes in the Wild. Amey Dharwadker (aap2174) Kai Zhang (kz2213)
Recognition of Animal Skin Texture Attributes in the Wild Amey Dharwadker (aap2174) Kai Zhang (kz2213) Motivation Patterns and textures are have an important role in object description and understanding
More informationAUTOMATIC PARKING OF SELF-DRIVING CAR BASED ON LIDAR
AUTOMATIC PARKING OF SELF-DRIVING CAR BASED ON LIDAR Bijun Lee a, Yang Wei a, I. Yuan Guo a a State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University,
More informationCameramanVis: where the camera should look? CPSC 547 Project proposal
CameramanVis: where the camera should look? CPSC 547 Project proposal Jianhui Chen Computer Science Department University of British Columbia jhchen14@cs.ubc.ca 1. Domain, task and dataset reconstruction
More informationMachine Learning in Data Quality Monitoring
CERN openlab workshop on Machine Learning and Data Analytics April 27 th, 2017 Machine Learning in Data Quality Monitoring a point of view Goal Maximize the best Quality Data for physics analysis Data
More informationCOMPUTER VISION. Dr. Sukhendu Das Deptt. of Computer Science and Engg., IIT Madras, Chennai
COMPUTER VISION Dr. Sukhendu Das Deptt. of Computer Science and Engg., IIT Madras, Chennai 600036. Email: sdas@iitm.ac.in URL: //www.cs.iitm.ernet.in/~sdas 1 INTRODUCTION 2 Human Vision System (HVS) Vs.
More informationHuman Detection and Tracking for Video Surveillance: A Cognitive Science Approach
Human Detection and Tracking for Video Surveillance: A Cognitive Science Approach Vandit Gajjar gajjar.vandit.381@ldce.ac.in Ayesha Gurnani gurnani.ayesha.52@ldce.ac.in Yash Khandhediya khandhediya.yash.364@ldce.ac.in
More informationDrywall state detection in image data for automatic indoor progress monitoring C. Kropp, C. Koch and M. König
Drywall state detection in image data for automatic indoor progress monitoring C. Kropp, C. Koch and M. König Chair for Computing in Engineering, Department of Civil and Environmental Engineering, Ruhr-Universität
More informationYu Tao, Panos Sidiropoulos, Jan-Peter Muller Imaging group, Mullard Space Science Laboratory, University College London BiDS 17 March 2016
Big Data Applications from the EU FP-7 imars and PRoViDE projects: automated DTM and Supoer-resolution restoration from NASA MRO Cameras and in future from TGO16 CaSSiS Yu Tao, Panos Sidiropoulos, Jan-Peter
More informationThird Rock from the Sun
Geodesy 101 AHD LiDAR Best Practice The Mystery of LiDAR Best Practice Glenn Jones SSSi GIS in the Coastal Environment Batemans Bay November 9, 2010 Light Detection and Ranging (LiDAR) Basic principles
More informationCHAPTER 4 DETECTION OF DISEASES IN PLANT LEAF USING IMAGE SEGMENTATION
CHAPTER 4 DETECTION OF DISEASES IN PLANT LEAF USING IMAGE SEGMENTATION 4.1. Introduction Indian economy is highly dependent of agricultural productivity. Therefore, in field of agriculture, detection of
More informationContents. Preface to the Second Edition
Preface to the Second Edition v 1 Introduction 1 1.1 What Is Data Mining?....................... 4 1.2 Motivating Challenges....................... 5 1.3 The Origins of Data Mining....................
More informationUnmanned Vehicle Technology Researches for Outdoor Environments. *Ju-Jang Lee 1)
Keynote Paper Unmanned Vehicle Technology Researches for Outdoor Environments *Ju-Jang Lee 1) 1) Department of Electrical Engineering, KAIST, Daejeon 305-701, Korea 1) jjlee@ee.kaist.ac.kr ABSTRACT The
More informationSimultaneous Localization and Mapping (SLAM)
Simultaneous Localization and Mapping (SLAM) RSS Lecture 16 April 8, 2013 Prof. Teller Text: Siegwart and Nourbakhsh S. 5.8 SLAM Problem Statement Inputs: No external coordinate reference Time series of
More informationShort Survey on Static Hand Gesture Recognition
Short Survey on Static Hand Gesture Recognition Huu-Hung Huynh University of Science and Technology The University of Danang, Vietnam Duc-Hoang Vo University of Science and Technology The University of
More informationThe YellowScan Surveyor: 5cm Accuracy Demonstrated
The YellowScan Surveyor: 5cm Accuracy Demonstrated Pierre Chaponnière1 and Tristan Allouis2 1 Application Engineer, YellowScan 2 CTO, YellowScan Introduction YellowScan Surveyor, the very latest lightweight
More informationArticulated Pose Estimation with Flexible Mixtures-of-Parts
Articulated Pose Estimation with Flexible Mixtures-of-Parts PRESENTATION: JESSE DAVIS CS 3710 VISUAL RECOGNITION Outline Modeling Special Cases Inferences Learning Experiments Problem and Relevance Problem:
More informationMotivation: why study sound? Motivation (2) Related Work. Our Study. Related Work (2)
Motivation: why study sound? Interactive Learning of the Acoustic Properties of Objects by a Robot Sound Producing Event Jivko Sinapov Mark Wiemer Alexander Stoytchev {jsinapov banff alexs}@iastate.edu
More informationMore processing power - Driving requirements
Software More processing power - Driving requirements Autonomy Task based commanding / resource management rather than timeline remote control E.g. planetary rovers drive from A B, look out for interesting
More informationOpenStreetSLAM: Global Vehicle Localization using OpenStreetMaps
OpenStreetSLAM: Global Vehicle Localization using OpenStreetMaps Georgios Floros, Benito van der Zander and Bastian Leibe RWTH Aachen University, Germany http://www.vision.rwth-aachen.de floros@vision.rwth-aachen.de
More informationIntern Presentation:
: Gripper Stereo and Assisted Teleoperation Stanford University December 13, 2010 Outline 1. Introduction 2. Hardware 3. Research 4. Packages 5. Conclusion Introduction Hardware Research Packages Conclusion
More informationAerial and Mobile LiDAR Data Fusion
Creating Value Delivering Solutions Aerial and Mobile LiDAR Data Fusion Dr. Srini Dharmapuri, CP, PMP What You Will Learn About LiDAR Fusion Mobile and Aerial LiDAR Technology Components & Parameters Project
More informationH2020 Space Robotic SRC- OG4
H2020 Space Robotic SRC- OG4 2 nd PERASPERA workshop Presentation by Sabrina Andiappane Thales Alenia Space France This project has received funding from the European Union s Horizon 2020 research and
More informationMULTI-MODAL MAPPING. Robotics Day, 31 Mar Frank Mascarich, Shehryar Khattak, Tung Dang
MULTI-MODAL MAPPING Robotics Day, 31 Mar 2017 Frank Mascarich, Shehryar Khattak, Tung Dang Application-Specific Sensors Cameras TOF Cameras PERCEPTION LiDAR IMU Localization Mapping Autonomy Robotic Perception
More informationLearning video saliency from human gaze using candidate selection
Learning video saliency from human gaze using candidate selection Rudoy, Goldman, Shechtman, Zelnik-Manor CVPR 2013 Paper presentation by Ashish Bora Outline What is saliency? Image vs video Candidates
More informationPedestrian Detection Using Correlated Lidar and Image Data EECS442 Final Project Fall 2016
edestrian Detection Using Correlated Lidar and Image Data EECS442 Final roject Fall 2016 Samuel Rohrer University of Michigan rohrer@umich.edu Ian Lin University of Michigan tiannis@umich.edu Abstract
More informationSeminar Heidelberg University
Seminar Heidelberg University Mobile Human Detection Systems Pedestrian Detection by Stereo Vision on Mobile Robots Philip Mayer Matrikelnummer: 3300646 Motivation Fig.1: Pedestrians Within Bounding Box
More informationTEGRA K1 AND THE AUTOMOTIVE INDUSTRY. Gernot Ziegler, Timo Stich
TEGRA K1 AND THE AUTOMOTIVE INDUSTRY Gernot Ziegler, Timo Stich Previously: Tegra in Automotive Infotainment / Navigation Digital Instrument Cluster Passenger Entertainment TEGRA K1 with Kepler GPU GPU:
More informationLecture 19: Depth Cameras. Visual Computing Systems CMU , Fall 2013
Lecture 19: Depth Cameras Visual Computing Systems Continuing theme: computational photography Cameras capture light, then extensive processing produces the desired image Today: - Capturing scene depth
More informationPROBABILITY OF FAILURE ANALYSIS STANDARDS AND GUIDELINES FOR ELVS
PROBABILITY OF FAILURE ANALYSIS STANDARDS AND GUIDELINES FOR ELVS Federal Aviation Administration 6th IAASS Conference Session 10: Launch Safety Part 1 By:, Elisabeth Morse (Valador Inc.), Paul Rosati
More informationA Reconfigurable Multiclass Support Vector Machine Architecture for Real-Time Embedded Systems Classification
A Reconfigurable Multiclass Support Vector Machine Architecture for Real-Time Embedded Systems Classification Jason Kane, Robert Hernandez, and Qing Yang University of Rhode Island 1 Overview Background
More informationLive Metric 3D Reconstruction on Mobile Phones ICCV 2013
Live Metric 3D Reconstruction on Mobile Phones ICCV 2013 Main Contents 1. Target & Related Work 2. Main Features of This System 3. System Overview & Workflow 4. Detail of This System 5. Experiments 6.
More informationPROBLEM FORMULATION AND RESEARCH METHODOLOGY
PROBLEM FORMULATION AND RESEARCH METHODOLOGY ON THE SOFT COMPUTING BASED APPROACHES FOR OBJECT DETECTION AND TRACKING IN VIDEOS CHAPTER 3 PROBLEM FORMULATION AND RESEARCH METHODOLOGY The foregoing chapter
More informationQadeer Baig, Mathias Perrollaz, Jander Botelho Do Nascimento, Christian Laugier
Using Fast Classification of Static and Dynamic Environment for Improving Bayesian Occupancy Filter (BOF) and Tracking Qadeer Baig, Mathias Perrollaz, Jander Botelho Do Nascimento, Christian Laugier e-motion
More informationSpatial Localization and Detection. Lecture 8-1
Lecture 8: Spatial Localization and Detection Lecture 8-1 Administrative - Project Proposals were due on Saturday Homework 2 due Friday 2/5 Homework 1 grades out this week Midterm will be in-class on Wednesday
More informationSensor Fusion: Potential, Challenges and Applications. Presented by KVH Industries and Geodetics, Inc. December 2016
Sensor Fusion: Potential, Challenges and Applications Presented by KVH Industries and Geodetics, Inc. December 2016 1 KVH Industries Overview Innovative technology company 600 employees worldwide Focused
More informationVisual-Inertial Localization and Mapping for Robot Navigation
Visual-Inertial Localization and Mapping for Robot Navigation Dr. Guillermo Gallego Robotics & Perception Group University of Zurich Davide Scaramuzza University of Zurich - http://rpg.ifi.uzh.ch Mocular,
More informationScalable Sensor Data Processor Development Status DSP Day - September 2014
Scalable Sensor Data Processor Development Status DSP Day - September 2014 83230347-DOC-TAS-EN-003 Presentation of the SSDP ASIC Program & Context SSDP will be the first Space Qualified European multi-core
More informationLecture: Autonomous micro aerial vehicles
Lecture: Autonomous micro aerial vehicles Friedrich Fraundorfer Remote Sensing Technology TU München 1/41 Autonomous operation@eth Zürich Start 2/41 Autonomous operation@eth Zürich 3/41 Outline MAV system
More informationHOG-based Pedestriant Detector Training
HOG-based Pedestriant Detector Training evs embedded Vision Systems Srl c/o Computer Science Park, Strada Le Grazie, 15 Verona- Italy http: // www. embeddedvisionsystems. it Abstract This paper describes
More informationHow to get realistic C-states latency and residency? Vincent Guittot
How to get realistic C-states latency and residency? Vincent Guittot Agenda Overview Exit latency Enter latency Residency Conclusion Overview Overview PMWG uses hikey960 for testing our dev on b/l system
More informationThe CEDA Archive: Data, Services and Infrastructure
The CEDA Archive: Data, Services and Infrastructure Kevin Marsh Centre for Environmental Data Archival (CEDA) www.ceda.ac.uk with thanks to V. Bennett, P. Kershaw, S. Donegan and the rest of the CEDA Team
More informationAutonomous Robot Navigation: Using Multiple Semi-supervised Models for Obstacle Detection
Autonomous Robot Navigation: Using Multiple Semi-supervised Models for Obstacle Detection Adam Bates University of Colorado at Boulder Abstract: This paper proposes a novel approach to efficiently creating
More informationTri-modal Human Body Segmentation
Tri-modal Human Body Segmentation Master of Science Thesis Cristina Palmero Cantariño Advisor: Sergio Escalera Guerrero February 6, 2014 Outline 1 Introduction 2 Tri-modal dataset 3 Proposed baseline 4
More informationCS 229 Final Project Report Learning to Decode Cognitive States of Rat using Functional Magnetic Resonance Imaging Time Series
CS 229 Final Project Report Learning to Decode Cognitive States of Rat using Functional Magnetic Resonance Imaging Time Series Jingyuan Chen //Department of Electrical Engineering, cjy2010@stanford.edu//
More informationarxiv: v1 [cs.lg] 31 Oct 2018
UNDERSTANDING DEEP NEURAL NETWORKS USING TOPOLOGICAL DATA ANALYSIS DANIEL GOLDFARB arxiv:1811.00852v1 [cs.lg] 31 Oct 2018 Abstract. Deep neural networks (DNN) are black box algorithms. They are trained
More informationVISION-BASED HUMANOID NAVIGATION USING SELF-SUPERVISED OBSTACLE DETECTION
International Journal of Humanoid Robotics c World Scientific Publishing Company VISION-BASED HUMANOID NAVIGATION USING SELF-SUPERVISED OBSTACLE DETECTION DANIEL MAIER CYRILL STACHNISS MAREN BENNEWITZ
More informationAUTONOMOUS PLANETARY ROVER CONTROL USING INVERSE SIMULATION
AUTONOMOUS PLANETARY ROVER CONTROL USING INVERSE SIMULATION Kevin Worrall (1), Douglas Thomson (1), Euan McGookin (1), Thaleia Flessa (1) (1)University of Glasgow, Glasgow, G12 8QQ, UK, Email: kevin.worrall@glasgow.ac.uk
More informationLearning Semantic Environment Perception for Cognitive Robots
Learning Semantic Environment Perception for Cognitive Robots Sven Behnke University of Bonn, Germany Computer Science Institute VI Autonomous Intelligent Systems Some of Our Cognitive Robots Equipped
More informationRobotics Programming Laboratory
Chair of Software Engineering Robotics Programming Laboratory Bertrand Meyer Jiwon Shin Lecture 8: Robot Perception Perception http://pascallin.ecs.soton.ac.uk/challenges/voc/databases.html#caltech car
More information(Deep) Learning for Robot Perception and Navigation. Wolfram Burgard
(Deep) Learning for Robot Perception and Navigation Wolfram Burgard Deep Learning for Robot Perception (and Navigation) Lifeng Bo, Claas Bollen, Thomas Brox, Andreas Eitel, Dieter Fox, Gabriel L. Oliveira,
More informationEarth Observation, Climate and Space for Smarter Government
Earth Observation, Climate and Space for Smarter Government Beth Greenaway, Head of Earth Observation 30 March 2015 http://www.bis.gov.uk/ukspaceagency Overview UK Space Agency EO Importance and priorities
More information2. POINT CLOUD DATA PROCESSING
Point Cloud Generation from suas-mounted iphone Imagery: Performance Analysis A. D. Ladai, J. Miller Towill, Inc., 2300 Clayton Road, Suite 1200, Concord, CA 94520-2176, USA - (andras.ladai, jeffrey.miller)@towill.com
More informationAUTONOMOUS DRONE NAVIGATION WITH DEEP LEARNING
AUTONOMOUS DRONE NAVIGATION WITH DEEP LEARNING Nikolai Smolyanskiy, Alexey Kamenev, Jeffrey Smith Project Redtail May 8, 2017 100% AUTONOMOUS FLIGHT OVER 1 KM FOREST TRAIL AT 3 M/S 2 Why autonomous path
More informationProject 3 Q&A. Jonathan Krause
Project 3 Q&A Jonathan Krause 1 Outline R-CNN Review Error metrics Code Overview Project 3 Report Project 3 Presentations 2 Outline R-CNN Review Error metrics Code Overview Project 3 Report Project 3 Presentations
More informationTHE SPEED-LIMIT SIGN DETECTION AND RECOGNITION SYSTEM
THE SPEED-LIMIT SIGN DETECTION AND RECOGNITION SYSTEM Kuo-Hsin Tu ( 塗國星 ), Chiou-Shann Fuh ( 傅楸善 ) Dept. of Computer Science and Information Engineering, National Taiwan University, Taiwan E-mail: p04922004@csie.ntu.edu.tw,
More informationLearning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009
Learning and Inferring Depth from Monocular Images Jiyan Pan April 1, 2009 Traditional ways of inferring depth Binocular disparity Structure from motion Defocus Given a single monocular image, how to infer
More informationTowards smart sensing based on field-programmable technology
Towards smart sensing based on field-programmable technology Wayne Luk Imperial College 28 January 2017 Outline 1. Smart sensing + field-programmable technology 2. Example: remote sensing 3. Acceleration:
More informationC. Premsai 1, Prof. A. Kavya 2 School of Computer Science, School of Computer Science Engineering, Engineering VIT Chennai, VIT Chennai
Traffic Sign Detection Via Graph-Based Ranking and Segmentation Algorithm C. Premsai 1, Prof. A. Kavya 2 School of Computer Science, School of Computer Science Engineering, Engineering VIT Chennai, VIT
More informationABSTRACT SERVICE ROBOT FOR THE VISUALLY IMPAIRED: PROVIDING NAVIGATIONAL ASSISTANCE USING DEEP LEARNING. by Amlaan Shakeel
ABSTRACT SERVICE ROBOT FOR THE VISUALLY IMPAIRED: PROVIDING NAVIGATIONAL ASSISTANCE USING DEEP LEARNING by Amlaan Shakeel Assistive technology helps improve the day to day activities for people with disabilities.
More informationMotion Planning Technologies for Planetary Rovers and Manipulators
Motion Planning Technologies for Planetary Rovers and Manipulators Eric T. Baumgartner NASA s Jet Propulsion Laboratory International Workshop on Motion Planning in Virtual Environments LAAS-CNRS Toulouse,
More informationMeasuring the World: Designing Robust Vehicle Localization for Autonomous Driving. Frank Schuster, Dr. Martin Haueis
Measuring the World: Designing Robust Vehicle Localization for Autonomous Driving Frank Schuster, Dr. Martin Haueis Agenda Motivation: Why measure the world for autonomous driving? Map Content: What do
More informationSpring 2016 :: :: Robot Autonomy :: Team 7 Motion Planning for Autonomous All-Terrain Vehicle
Spring 2016 :: 16662 :: Robot Autonomy :: Team 7 Motion Planning for Autonomous All-Terrain Vehicle Guan-Horng Liu, Samuel Wang, Shu-Kai Lin, Chris Wang, Tiffany May Advisor : Mr. George Kantor OUTLINE
More informationBus Detection and recognition for visually impaired people
Bus Detection and recognition for visually impaired people Hangrong Pan, Chucai Yi, and Yingli Tian The City College of New York The Graduate Center The City University of New York MAP4VIP Outline Motivation
More informationSGEO: Overview and Product Offering. Marco R. Fuchs. Marco R. R. Fuchs. DLR-ESA Workshop on ARTES 11. Marco R. Fuchs OHB Technology AG
DLR-ESA Workshop on ARTES 11 SGEO: Overview and Product Offering Marco R. R. Fuchs June June29, 29, 2006 2006 Tegernsee, Tegernsee, Germany Germany Marco R. Fuchs Marco R. Fuchs OHB Technology AG OHB Technology
More informationProcess of Interaction Design and Design Languages
Process of Interaction Design and Design Languages Process of Interaction Design This week, we will explore how we can design and build interactive products What is different in interaction design compared
More informationStereo and Monocular Vision Applied to Computer Aided Navigation for Aerial and Ground Vehicles
Riccardo Giubilato Centro di Ateneo di Studi e Attivita Spaziali Giuseppe Colombo CISAS University of Padova 1 Introduction Previous Work Research Objectives Research Methodology Robot vision: Process
More informationSpeaker: Ming-Ming Cheng Nankai University 15-Sep-17 Towards Weakly Supervised Image Understanding
Towards Weakly Supervised Image Understanding (WSIU) Speaker: Ming-Ming Cheng Nankai University http://mmcheng.net/ 1/50 Understanding Visual Information Image by kirkh.deviantart.com 2/50 Dataset Annotation
More information