Deep Neural Network Enhanced VSLAM Landmark Selection
|
|
- Eugene Baldwin
- 5 years ago
- Views:
Transcription
1 Deep Neural Network Enhanced VSLAM Landmark Selection Dr. Patrick Benavidez
2 Overview 1 Introduction 2 Background on methods used in VSLAM 3 Proposed Method 4 Testbed 5 Preliminary Results
3 What is VSLAM? Visual Simultaneous Localization and Mapping Use of vision and depth sensors to acquire features from an environment, map them and to navigate with the map
4 Motivation to use VSLAM Similar to methods used by humans GPS-denied and contested environments Spoofing attacks on GPS Cloud-based robotics Data simplification and organization
5 What processes are involved in VSLAM? Sensors capture properties of the surrounding environment Operations to transform captured environmental data with robot pose data Algorithms to place transformed environmental data into existing map Methods to determine whether the robot has already visited a particular location Operations to update existing data in the map Loop closure operations to constrain the bounds of a map
6 Typical Scenes for VSLAM Indoors Outdoors
7 VSLAM Mapping Process feature detectors find useful feature rich points in an image feature descriptors describe sets of features feature matching match features into map
8 Feature Detection Corner detectors - Harris, Shi Tomasi Scale Invariant Feature Transform (SIFT) better than Harris Corner detector Speeded-Up Robust Features (SURF) faster version of SIFT Features from Accelerated Segment Test (FAST) fast enough for SLAM [ beta/doc/py tutorials/py feature2d/py sift intro/py sift intro.html #sift-intro]
9 Feature Descriptors Feature descriptors describe sets of features Feature descriptors are saved in a database or similar structure Binary Robust Independent Elementary Features (BRIEF) Oriented FAST and Rotated BRIEF (ORB) SIFT & SURF are patented, ORB is free Fusion of FAST Keypoint Detector and BRIEF descriptor methods with increased performance [ beta/doc/py tutorials/py feature2d/py orb/py orb.html#orb]
10 Feature Matching Commonly used methods are brute force and FLANN matching algorithms These methods match feature descriptors of a newly acquired image to those saved in the database
11 Bag of Words A bag of words in natural language processing is the decomposition of a sentence into its constituent components (words) and storing them in a container (bag) Example [ model]: Sentence 1 John likes to watch movies. Mary likes movies too. Sentence 2 John also likes to watch football games. Bag of words used to describe sentence 1 and sentence 2 [ John, likes, to, watch, movies, also, football, games, Mary, too ]
12 Visual Bag of Words A visual bag of words is where an image is broken down into its component regions of interest (ROI) in an image (words) and stored in a collection (bag) Labels can be applied to the ROI in a bag Example: [
13 Application of Visual Bag of Words Scene Identification A visual bag of words (collection of known images) is created to describe particular components for each scene Features taken from the latest camera image are compared to those in the visual bag of words for each scene A collection of the most relevant words (images) in each bag matching the input image are generated The bag most closely matching the current image identifies the scene
14 Application of Visual Bag of Words Mapping A visual bag of words (collection of unknown images) is created at runtime to describe particular components discovered by a robot Locations where the words have been discovered are input into the map Features taken from the latest camera image are compared to those in the visual bag of words to determine if the object has been seen before New objects are added to the bag Objects already in the bag are used to identify where the agent is in a map if it has been to that location before
15 Problems with existing methods Features are made too general by design Recall that features are simple components of an image: corners, edges, intersections, etc. Almost every type of object can contain these features Problem: Both static and dynamic objects in the environment are registered in the map in the same context
16 Problems with existing methods (continued) Objects with freedom to move around the environment Examples: people, animals, robots, mobile carts, etc. Problem: Traditional SLAM/VSLAM will fail to produce meaningful maps with multiple agents working in the same environment
17 Problems with existing methods (continued) Time varying objects that do not travel around the environment Examples: trees, plants, tracking solar panels, windmills, flags, banners, televisions, digital billboards Problem: Features should not be taken off of the dynamic portions of these items Features can be acquired from their static components (planted/grounded base, static frame) [
18 Problems with existing methods (continued) Loop closure revisiting the same place twice produces multiple paths due to odometry errors
19 Problems with existing methods (continued) 2D maps from laser scanners contain low levels of information about the environment without use of a vision sensor
20 Method Overview Deep learning for object classification Association of classified objects to known properties Map classified objects by their properties - static/dynamic Localize on the event of identifying clusters of adjoining objects (preferably static object clusters)
21 Deep Neural Network Use deep neural networks to classify objects into known objects classes Known object classes can be any of the following examples: desk, chair, wall, door frame, door, UGV, UAV, cup, trash can, wheels, etc. ImageNet - currently has 14,197,122 images, synsets (synonym sets) indexed [ Convolution Neural Networks (CNNs) will be used for this work
22 Deep Neural Network
23 Deep Neural Network *source:
24 Association of classified objects to known properties Dynamic properties of an object can be either referenced from a database or measured from the environment Measurement of an object s dynamic properties from observation would entail detection of movement with or without perturbation
25 Mapping Process Identify static/dynamic properties of classified objects Map dynamic objects as temporary obstacles Map static objects as landmark components Identify clusters of landmark components as a landmark Perform loop closure (re-adjustment/alignment of map) with knowledge of landmark locations
26 Computers for training models Custom-built machine learning optimized desktop computers Relevant Specifications EVGA NVIDIA GeForce GTX-1080 Video Card 2560 CUDA Cores, 8 GB GDDR5X Intel Core i GHz Quad Core 32GB DDR4 RAM 240GB SSD for operating system 2TB HDD for storage Acquired these today on 4/7/2017 will setup after this talk
27 Computer for processing input NVIDIA Jetson TX-1 Relevant Specifications Mobile Supercomputer on a Chip NVIDIA Maxwell architecture 256 CUDA Cores 64-bit CPU Quad ARM R A57 4GB LPDDR4 RAM 16GB emmc for operating system SD card slot for storage Multiple high speed camera connections (USB3, CSI)
28 Robot Hardware A variety of systems capable of mapping the environment
29 Preliminary Results Use of TensorFlow, Inception V3 and GoogLeNet for detecting various objects and recording their locations in a radial map Convolutional Neural Network to classify people from a car s perspective 97% Accuracy Can be modified to work from the MAV s perspective
30 Thank You Any Questions???
Hardware Acceleration of Feature Detection and Description Algorithms on Low Power Embedded Platforms
Hardware Acceleration of Feature Detection and Description Algorithms on LowPower Embedded Platforms Onur Ulusel, Christopher Picardo, Christopher Harris, Sherief Reda, R. Iris Bahar, School of Engineering,
More informationApplication questions. Theoretical questions
The oral exam will last 30 minutes and will consist of one application question followed by two theoretical questions. Please find below a non exhaustive list of possible application questions. The list
More informationDeep Learning Based Real-time Object Recognition System with Image Web Crawler
, pp.103-110 http://dx.doi.org/10.14257/astl.2016.142.19 Deep Learning Based Real-time Object Recognition System with Image Web Crawler Myung-jae Lee 1, Hyeok-june Jeong 1, Young-guk Ha 2 1 Department
More informationImproving Visual SLAM Algorithms for use in Realtime Robotic Applications
Improving Visual SLAM Algorithms for use in Realtime Robotic Applications Patrick Benavidez, Mohan Kumar Muppidi, and Mo Jamshidi, Ph.D.Lutcher Brown Endowed Chair Professor Autonomous Control Engineering
More informationEnsemble of Bayesian Filters for Loop Closure Detection
Ensemble of Bayesian Filters for Loop Closure Detection Mohammad Omar Salameh, Azizi Abdullah, Shahnorbanun Sahran Pattern Recognition Research Group Center for Artificial Intelligence Faculty of Information
More informationNvidia Jetson TX2 and its Software Toolset. João Fernandes 2017/2018
Nvidia Jetson TX2 and its Software Toolset João Fernandes 2017/2018 In this presentation Nvidia Jetson TX2: Hardware Nvidia Jetson TX2: Software Machine Learning: Neural Networks Convolutional Neural Networks
More informationRobot localization method based on visual features and their geometric relationship
, pp.46-50 http://dx.doi.org/10.14257/astl.2015.85.11 Robot localization method based on visual features and their geometric relationship Sangyun Lee 1, Changkyung Eem 2, and Hyunki Hong 3 1 Department
More informationServosila Robotic Heads
Servosila Robotic Heads www.servosila.com TABLE OF CONTENTS SERVOSILA ROBOTIC HEADS 2 SOFTWARE-DEFINED FUNCTIONS OF THE ROBOTIC HEADS 2 SPECIFICATIONS: ROBOTIC HEADS 4 DIMENSIONS OF ROBOTIC HEAD 5 DIMENSIONS
More informationSmall is the New Big: Data Analytics on the Edge
Small is the New Big: Data Analytics on the Edge An overview of processors and algorithms for deep learning techniques on the edge Dr. Abhay Samant VP Engineering, Hiller Measurements Adjunct Faculty,
More informationTEGRA K1 AND THE AUTOMOTIVE INDUSTRY. Gernot Ziegler, Timo Stich
TEGRA K1 AND THE AUTOMOTIVE INDUSTRY Gernot Ziegler, Timo Stich Previously: Tegra in Automotive Infotainment / Navigation Digital Instrument Cluster Passenger Entertainment TEGRA K1 with Kepler GPU GPU:
More informationNonlinear State Estimation for Robotics and Computer Vision Applications: An Overview
Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview Arun Das 05/09/2017 Arun Das Waterloo Autonomous Vehicles Lab Introduction What s in a name? Arun Das Waterloo Autonomous
More informationVISION FOR AUTOMOTIVE DRIVING
VISION FOR AUTOMOTIVE DRIVING French Japanese Workshop on Deep Learning & AI, Paris, October 25th, 2017 Quoc Cuong PHAM, PhD Vision and Content Engineering Lab AI & MACHINE LEARNING FOR ADAS AND SELF-DRIVING
More informationOffloading UAV Navigation Computation to the Cloud Sriramya Ramya Bhamidipati and Grace Xingxin Gao
Offloading UAV Navigation Computation to the Cloud Sriramya Ramya Bhamidipati and Grace Xingxin Gao SCPNT 2016 November 01 2016 UAV Navigation in Urban Areas Challenges Limited battery life Restricted
More informationDesigning a Pick and Place Robotics Application Using MATLAB and Simulink
Designing a Pick and Place Robotics Application Using MATLAB and Simulink Carlos Santacruz-Rosero, PhD Sr Application Engineer Robotics Pulkit Kapur Sr Industry Marketing Manager Robotics 2017 The MathWorks,
More informationIndoor Simultaneous Localization and Mapping for Lego Ev3
2018 International Conference on Computer, Communication and Network Technology (CCNT 2018) ISBN: 978-1-60595-561-2 Indoor Simultaneous Localization and Mapping for Lego Ev3 Qiang CHEN 1,2, Yi-nong CHEN
More informationSimultaneous Localization and Mapping
Sebastian Lembcke SLAM 1 / 29 MIN Faculty Department of Informatics Simultaneous Localization and Mapping Visual Loop-Closure Detection University of Hamburg Faculty of Mathematics, Informatics and Natural
More informationDeep learning in MATLAB From Concept to CUDA Code
Deep learning in MATLAB From Concept to CUDA Code Roy Fahn Applications Engineer Systematics royf@systematics.co.il 03-7660111 Ram Kokku Principal Engineer MathWorks ram.kokku@mathworks.com 2017 The MathWorks,
More informationEmerging Vision Technologies: Enabling a New Era of Intelligent Devices
Emerging Vision Technologies: Enabling a New Era of Intelligent Devices Computer vision overview Computer vision is being integrated in our daily lives Acquiring, processing, and understanding visual data
More informationImage Features: Detection, Description, and Matching and their Applications
Image Features: Detection, Description, and Matching and their Applications Image Representation: Global Versus Local Features Features/ keypoints/ interset points are interesting locations in the image.
More informationTarget Shape Identification for Nanosatellites using Monocular Point Cloud Techniques
Target Shape Identification for Nanosatellites using Monocular Point Cloud Techniques 6th European CubeSat Symposium Oct. 16, 2014 Mark A. Post and Xiu.T. Yan Space Mechatronic Systems Technology (SMeSTech)
More informationP I X E V I A : A I B A S E D, R E A L - T I M E C O M P U T E R V I S I O N S Y S T E M F O R D R O N E S
P I X E V I A : A I B A S E D, R E A L - T I M E C O M P U T E R V I S I O N S Y S T E M F O R D R O N E S Mindaugas Eglinskas, CEO at PIXEVIA www.pixevia.com Origins in R&D projects for Lithuanian MoD.
More informationCharacterization and Benchmarking of Deep Learning. Natalia Vassilieva, PhD Sr. Research Manager
Characterization and Benchmarking of Deep Learning Natalia Vassilieva, PhD Sr. Research Manager Deep learning applications Vision Speech Text Other Search & information extraction Security/Video surveillance
More informationLearning Semantic Environment Perception for Cognitive Robots
Learning Semantic Environment Perception for Cognitive Robots Sven Behnke University of Bonn, Germany Computer Science Institute VI Autonomous Intelligent Systems Some of Our Cognitive Robots Equipped
More informationExploiting Depth Camera for 3D Spatial Relationship Interpretation
Exploiting Depth Camera for 3D Spatial Relationship Interpretation Jun Ye Kien A. Hua Data Systems Group, University of Central Florida Mar 1, 2013 Jun Ye and Kien A. Hua (UCF) 3D directional spatial relationships
More informationArtec Leo. A smart professional 3D scanner for a next-generation user experience
Artec Leo A smart professional 3D scanner for a next-generation user experience Industrial design and manufacturing / Healthcare VR / E-commerce / Science and education Forensics / Art and design Easy
More informationArtec Leo. A smart professional 3D scanner for a next-generation user experience
Artec Leo A smart professional 3D scanner for a next-generation user experience Industrial design and manufacturing / Healthcare VR / E-commerce / Science and education Forensics / Art and design See your
More informationArtec Leo. A smart professional 3D scanner for a next-generation user experience
Artec Leo A smart professional 3D scanner for a next-generation user experience Industrial design and manufacturing / Healthcare VR / E-commerce / Science and education Forensics / Art and design See your
More informationElementary Computing CSC M. Cheng, Computer Science 1
Elementary Computing CSC 100 2014-07-14 M. Cheng, Computer Science 1 CPU and Memory Inside of a computer CPU and RAM Clock Speed & Multi-core Microprocessor How a CPU works Secondary Storage 2014-07-14
More informationVisual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech
Visual Odometry Features, Tracking, Essential Matrix, and RANSAC Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline The
More informationEmbedded Computing without Compromise. Evolution of the Rugged GPGPU Computer Session: SIL7127 Dan Mor PLM -Aitech Systems GTC Israel 2017
Evolution of the Rugged GPGPU Computer Session: SIL7127 Dan Mor PLM - Systems GTC Israel 2017 Agenda Current GPGPU systems NVIDIA Jetson TX1 and TX2 evaluation Conclusions New Products 2 GPGPU Product
More informationINTEGRATING COMPUTER VISION SENSOR INNOVATIONS INTO MOBILE DEVICES. Eli Savransky Principal Architect - CTO Office Mobile BU NVIDIA corp.
INTEGRATING COMPUTER VISION SENSOR INNOVATIONS INTO MOBILE DEVICES Eli Savransky Principal Architect - CTO Office Mobile BU NVIDIA corp. Computer Vision in Mobile Tegra K1 It s time! AGENDA Use cases categories
More informationFinal Project Report: Mobile Pick and Place
Final Project Report: Mobile Pick and Place Xiaoyang Liu (xiaoyan1) Juncheng Zhang (junchen1) Karthik Ramachandran (kramacha) Sumit Saxena (sumits1) Yihao Qian (yihaoq) Adviser: Dr Matthew Travers Carnegie
More informationVisual Recognition and Search April 18, 2008 Joo Hyun Kim
Visual Recognition and Search April 18, 2008 Joo Hyun Kim Introduction Suppose a stranger in downtown with a tour guide book?? Austin, TX 2 Introduction Look at guide What s this? Found Name of place Where
More informationMIOVISION DEEP LEARNING TRAFFIC ANALYTICS SYSTEM FOR REAL-WORLD DEPLOYMENT. Kurtis McBride CEO, Miovision
MIOVISION DEEP LEARNING TRAFFIC ANALYTICS SYSTEM FOR REAL-WORLD DEPLOYMENT Kurtis McBride CEO, Miovision ABOUT MIOVISION COMPANY Founded in 2005 40% growth, year over year Offices in Kitchener, Canada
More information3D object recognition used by team robotto
3D object recognition used by team robotto Workshop Juliane Hoebel February 1, 2016 Faculty of Computer Science, Otto-von-Guericke University Magdeburg Content 1. Introduction 2. Depth sensor 3. 3D object
More informationW4. Perception & Situation Awareness & Decision making
W4. Perception & Situation Awareness & Decision making Robot Perception for Dynamic environments: Outline & DP-Grids concept Dynamic Probabilistic Grids Bayesian Occupancy Filter concept Dynamic Probabilistic
More informationDeveloping Algorithms for Robotics and Autonomous Systems
Developing Algorithms for Robotics and Autonomous Systems Jorik Caljouw 2015 The MathWorks, Inc. 1 Key Takeaway of this Talk Success in developing an autonomous robotics system requires: 1. Multi-domain
More informationSensor Modalities. Sensor modality: Different modalities:
Sensor Modalities Sensor modality: Sensors which measure same form of energy and process it in similar ways Modality refers to the raw input used by the sensors Different modalities: Sound Pressure Temperature
More informationVisual Navigation for Flying Robots. Structure From Motion
Computer Vision Group Prof. Daniel Cremers Visual Navigation for Flying Robots Structure From Motion Dr. Jürgen Sturm VISNAV Oral Team Exam Date and Time Student Name Student Name Student Name Mon, July
More informationA176 Cyclone. GPGPU Fanless Small FF RediBuilt Supercomputer. IT and Instrumentation for industry. Aitech I/O
The A176 Cyclone is the smallest and most powerful Rugged-GPGPU, ideally suited for distributed systems. Its 256 CUDA cores reach 1 TFLOPS, and it consumes less than 17W at full load (8-10W at typical
More informationDeep Learning: Transforming Engineering and Science The MathWorks, Inc.
Deep Learning: Transforming Engineering and Science 1 2015 The MathWorks, Inc. DEEP LEARNING: TRANSFORMING ENGINEERING AND SCIENCE A THE NEW RISE ERA OF OF GPU COMPUTING 3 NVIDIA A IS NEW THE WORLD S ERA
More informationDelivering Deep Learning to Mobile Devices via Offloading
Delivering Deep Learning to Mobile Devices via Offloading Xukan Ran*, Haoliang Chen*, Zhenming Liu 1, Jiasi Chen* *University of California, Riverside 1 College of William and Mary Deep learning on mobile
More informationVisual SLAM for small Unmanned Aerial Vehicles
Visual SLAM for small Unmanned Aerial Vehicles Margarita Chli Autonomous Systems Lab, ETH Zurich Simultaneous Localization And Mapping How can a body navigate in a previously unknown environment while
More informationVehicle Tracking using Optical Features
Vehicle Tracking using Optical Features Songlin Piao Robotics Research Lab University of Kaiserslautern, Germany Outline Motivation State-of-Art Proposed System Experiments Conclusion Motivation Autonomous
More informationA Lightweight YOLOv2:
FPGA2018 @Monterey A Lightweight YOLOv2: A Binarized CNN with a Parallel Support Vector Regression for an FPGA Hiroki Nakahara, Haruyoshi Yonekawa, Tomoya Fujii, Shimpei Sato Tokyo Institute of Technology,
More informationVisual Pose Estimation System for Autonomous Rendezvous of Spacecraft
Visual Pose Estimation System for Autonomous Rendezvous of Spacecraft Mark A. Post1, Junquan Li2, and Craig Clark2 Space Mechatronic Systems Technology Laboratory Dept. of Design, Manufacture & Engineering
More informationMobile Robotics. Mathematics, Models, and Methods. HI Cambridge. Alonzo Kelly. Carnegie Mellon University UNIVERSITY PRESS
Mobile Robotics Mathematics, Models, and Methods Alonzo Kelly Carnegie Mellon University HI Cambridge UNIVERSITY PRESS Contents Preface page xiii 1 Introduction 1 1.1 Applications of Mobile Robots 2 1.2
More informationAppearance-Based Place Recognition Using Whole-Image BRISK for Collaborative MultiRobot Localization
Appearance-Based Place Recognition Using Whole-Image BRISK for Collaborative MultiRobot Localization Jung H. Oh, Gyuho Eoh, and Beom H. Lee Electrical and Computer Engineering, Seoul National University,
More informationCreating Affordable and Reliable Autonomous Vehicle Systems
Creating Affordable and Reliable Autonomous Vehicle Systems Shaoshan Liu shaoshan.liu@perceptin.io Autonomous Driving Localization Most crucial task of autonomous driving Solutions: GNSS but withvariations,
More informationTHE LEADER IN VISUAL COMPUTING
MOBILE EMBEDDED THE LEADER IN VISUAL COMPUTING 2 TAKING OUR VISION TO REALITY HPC DESIGN and VISUALIZATION AUTO GAMING 3 BEST DEVELOPER EXPERIENCE Tools for Fast Development Debug and Performance Tuning
More informationNew approaches to pattern recognition and automated learning
Z Y X New approaches to pattern recognition and automated learning Technology Forum 2015 Johannes Zuegner STEMMER IMAGING GmbH, Puchheim, Germany OUTLINE Introduction Description of the task What does
More informationPERFORMANCE MEASUREMENTS OF FEATURE TRACKING AND HISTOGRAM BASED TRAFFIC CONGESTION ALGORITHMS
PERFORMANCE MEASUREMENTS OF FEATURE TRACKING AND HISTOGRAM BASED TRAFFIC CONGESTION ALGORITHMS Ozgur Altun 1 and Kenan Aksoy 2 Proline Bilisim Sistemleri, Istanbul, Turkey 1 Research and Development Engineer,
More informationPOINT CLOUD DEEP LEARNING
POINT CLOUD DEEP LEARNING Innfarn Yoo, 3/29/28 / 57 Introduction AGENDA Previous Work Method Result Conclusion 2 / 57 INTRODUCTION 3 / 57 2D OBJECT CLASSIFICATION Deep Learning for 2D Object Classification
More informationArtec Leo. A smart professional 3D scanner for a next-generation user experience
001-09/2017-USD-EN Artec Leo A smart professional 3D scanner for a next-generation user experience Industrial design and manufacturing / Healthcare VR / E-commerce / Science and education Forensics / Art
More informationTowards Fully-automated Driving. tue-mps.org. Challenges and Potential Solutions. Dr. Gijs Dubbelman Mobile Perception Systems EE-SPS/VCA
Towards Fully-automated Driving Challenges and Potential Solutions Dr. Gijs Dubbelman Mobile Perception Systems EE-SPS/VCA Mobile Perception Systems 6 PhDs, 1 postdoc, 1 project manager, 2 software engineers
More informationWINDOWS COMPUTER Price List
Last Updated: 10/20/17 Page 1 of 5 WINDOWS COMPUTER Price List Dell Inspiron 13 5000 Series 2-in-1 13-inch 2-in-1 in a sleek and portable design with stunning sound and the latest Intel processors. Optional
More informationFast Hardware For AI
Fast Hardware For AI Karl Freund karl@moorinsightsstrategy.com Sr. Analyst, AI and HPC Moor Insights & Strategy Follow my blogs covering Machine Learning Hardware on Forbes: http://www.forbes.com/sites/moorinsights
More informationIntroduction to Deep Learning in Signal Processing & Communications with MATLAB
Introduction to Deep Learning in Signal Processing & Communications with MATLAB Dr. Amod Anandkumar Pallavi Kar Application Engineering Group, Mathworks India 2019 The MathWorks, Inc. 1 Different Types
More informationIntroduction to Autonomous Mobile Robots
Introduction to Autonomous Mobile Robots second edition Roland Siegwart, Illah R. Nourbakhsh, and Davide Scaramuzza The MIT Press Cambridge, Massachusetts London, England Contents Acknowledgments xiii
More informationRobot Localization based on Geo-referenced Images and G raphic Methods
Robot Localization based on Geo-referenced Images and G raphic Methods Sid Ahmed Berrabah Mechanical Department, Royal Military School, Belgium, sidahmed.berrabah@rma.ac.be Janusz Bedkowski, Łukasz Lubasiński,
More informationPattern recognition (3)
Pattern recognition (3) 1 Things we have discussed until now Statistical pattern recognition Building simple classifiers Supervised classification Minimum distance classifier Bayesian classifier Building
More informationVision-based Mobile Robot Localization and Mapping using Scale-Invariant Features
Vision-based Mobile Robot Localization and Mapping using Scale-Invariant Features Stephen Se, David Lowe, Jim Little Department of Computer Science University of British Columbia Presented by Adam Bickett
More informationLS-ACTS 1.0 USER MANUAL
LS-ACTS 1.0 USER MANUAL VISION GROUP, STATE KEY LAB OF CAD&CG, ZHEJIANG UNIVERSITY HTTP://WWW.ZJUCVG.NET TABLE OF CONTENTS 1 Introduction... 1-3 1.1 Product Specification...1-3 1.2 Feature Functionalities...1-3
More informationAccurate Motion Estimation and High-Precision 3D Reconstruction by Sensor Fusion
007 IEEE International Conference on Robotics and Automation Roma, Italy, 0-4 April 007 FrE5. Accurate Motion Estimation and High-Precision D Reconstruction by Sensor Fusion Yunsu Bok, Youngbae Hwang,
More informationCS 4758: Automated Semantic Mapping of Environment
CS 4758: Automated Semantic Mapping of Environment Dongsu Lee, ECE, M.Eng., dl624@cornell.edu Aperahama Parangi, CS, 2013, alp75@cornell.edu Abstract The purpose of this project is to program an Erratic
More informationSIFT Descriptor Extraction on the GPU for Large-Scale Video Analysis. Hannes Fassold, Jakub Rosner
SIFT Descriptor Extraction on the GPU for Large-Scale Video Analysis Hannes Fassold, Jakub Rosner 2014-03-26 2 Overview GPU-activities @ AVM research group SIFT descriptor extraction Algorithm GPU implementation
More informationAccelerating Reinforcement Learning in Engineering Systems
Accelerating Reinforcement Learning in Engineering Systems Tham Chen Khong with contributions from Zhou Chongyu and Le Van Duc Department of Electrical & Computer Engineering National University of Singapore
More informationCS 378: Autonomous Intelligent Robotics. Instructor: Jivko Sinapov
CS 378: Autonomous Intelligent Robotics Instructor: Jivko Sinapov http://www.cs.utexas.edu/~jsinapov/teaching/cs378/ Visual Registration and Recognition Announcements Homework 6 is out, due 4/5 4/7 Installing
More informationABSTRACT SERVICE ROBOT FOR THE VISUALLY IMPAIRED: PROVIDING NAVIGATIONAL ASSISTANCE USING DEEP LEARNING. by Amlaan Shakeel
ABSTRACT SERVICE ROBOT FOR THE VISUALLY IMPAIRED: PROVIDING NAVIGATIONAL ASSISTANCE USING DEEP LEARNING by Amlaan Shakeel Assistive technology helps improve the day to day activities for people with disabilities.
More informationarxiv: v2 [cs.cv] 31 Aug 2018
Real-time,, and of Moving and Stationary Objects using Multiple Fisheye Images Iljoo Baek, Albert Davies, Geng Yan, and Ragunathan (Raj) Rajkumar arxiv:1803.06077v2 [cs.cv] 31 Aug 2018 Abstract The ability
More information3D Environment Reconstruction
3D Environment Reconstruction Using Modified Color ICP Algorithm by Fusion of a Camera and a 3D Laser Range Finder The 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems October 11-15,
More informationIntegrating multiple representations of spatial knowledge for mapping, navigation, and communication
Integrating multiple representations of spatial knowledge for mapping, navigation, and communication Patrick Beeson Matt MacMahon Joseph Modayil Aniket Murarka Benjamin Kuipers Department of Computer Sciences
More informationMaster Automática y Robótica. Técnicas Avanzadas de Vision: Visual Odometry. by Pascual Campoy Computer Vision Group
Master Automática y Robótica Técnicas Avanzadas de Vision: by Pascual Campoy Computer Vision Group www.vision4uav.eu Centro de Automá
More informationVisual Perception for Robots
Visual Perception for Robots Sven Behnke Computer Science Institute VI Autonomous Intelligent Systems Our Cognitive Robots Complete systems for example scenarios Equipped with rich sensors Flying robot
More informationLecture 12 Recognition
Institute of Informatics Institute of Neuroinformatics Lecture 12 Recognition Davide Scaramuzza 1 Lab exercise today replaced by Deep Learning Tutorial Room ETH HG E 1.1 from 13:15 to 15:00 Optional lab
More informationME132 February 3, 2011
ME132 February 3, 2011 Outline: - active sensors - introduction to lab setup (Player/Stage) - lab assignment - brief overview of OpenCV ME132 February 3, 2011 Outline: - active sensors - introduction to
More informationCluster-based 3D Reconstruction of Aerial Video
Cluster-based 3D Reconstruction of Aerial Video Scott Sawyer (scott.sawyer@ll.mit.edu) MIT Lincoln Laboratory HPEC 12 12 September 2012 This work is sponsored by the Assistant Secretary of Defense for
More informationOpen Source Software in Robotics and Real-Time Control Systems. Gary Crum at OpenWest 2017
Open Source Software in Robotics and Real-Time Control Systems Gary Crum at OpenWest 2017 Background and some videos for context ASI history with some open source and USU academic roots: asirobots.com
More informationAUTONOMOUS DRONE NAVIGATION WITH DEEP LEARNING
AUTONOMOUS DRONE NAVIGATION WITH DEEP LEARNING Nikolai Smolyanskiy, Alexey Kamenev, Jeffrey Smith Project Redtail May 8, 2017 100% AUTONOMOUS FLIGHT OVER 1 KM FOREST TRAIL AT 3 M/S 2 Why autonomous path
More informationFast Natural Feature Tracking for Mobile Augmented Reality Applications
Fast Natural Feature Tracking for Mobile Augmented Reality Applications Jong-Seung Park 1, Byeong-Jo Bae 2, and Ramesh Jain 3 1 Dept. of Computer Science & Eng., University of Incheon, Korea 2 Hyundai
More informationInternational Journal Of Global Innovations -Vol.6, Issue.I Paper Id: SP-V6-I1-P01 ISSN Online:
IMPLEMENTATION OF OBJECT RECOGNITION USING SIFT ALGORITHM ON BEAGLE BOARD XM USING EMBEDDED LINUX #1 T.KRISHNA KUMAR -M. Tech Student, #2 G.SUDHAKAR - Assistant Professor, #3 R. MURALI, HOD - Assistant
More informationIntentionally Blank 0
Intentionally Blank 0 Technology in Action Chapter 2 Looking at Computers: Understanding the Parts 1 Understanding Your Computer: Computers are Data Processing Devices Perform four major functions Input:
More informationUAV Autonomous Navigation in a GPS-limited Urban Environment
UAV Autonomous Navigation in a GPS-limited Urban Environment Yoko Watanabe DCSD/CDIN JSO-Aerial Robotics 2014/10/02-03 Introduction 2 Global objective Development of a UAV onboard system to maintain flight
More informationFPGA-Based Feature Detection
FPGA-Based Feature Detection Wennie Tabib School of Computer Science Carnegie Mellon University Pittsburgh, PA 15213 wtabib@andrew.cmu.edu Abstract Fast, accurate, autonomous robot navigation is essential
More informationArchitectures for Scalable Media Object Search
Architectures for Scalable Media Object Search Dennis Sng Deputy Director & Principal Scientist NVIDIA GPU Technology Workshop 10 July 2014 ROSE LAB OVERVIEW 2 Large Database of Media Objects Next- Generation
More informationRealtime Object Detection and Segmentation for HD Mapping
Realtime Object Detection and Segmentation for HD Mapping William Raveane Lead AI Engineer Bahram Yoosefizonooz Technical Director NavInfo Europe Advanced Research Lab Presented at GTC Europe 2018 AI in
More informationACCELERATED COMPUTING: THE PATH FORWARD. Jen-Hsun Huang, Co-Founder and CEO, NVIDIA SC15 Nov. 16, 2015
ACCELERATED COMPUTING: THE PATH FORWARD Jen-Hsun Huang, Co-Founder and CEO, NVIDIA SC15 Nov. 16, 2015 COMMODITY DISRUPTS CUSTOM SOURCE: Top500 ACCELERATED COMPUTING: THE PATH FORWARD It s time to start
More informationOverview. Augmented reality and applications Marker-based augmented reality. Camera model. Binary markers Textured planar markers
Augmented reality Overview Augmented reality and applications Marker-based augmented reality Binary markers Textured planar markers Camera model Homography Direct Linear Transformation What is augmented
More informationVirtual Testbeds for Planetary Exploration: The Self Localization Aspect
Virtual Testbeds for Planetary Exploration: The Self Localization Aspect, RWTH Aachen University Björn Sondermann Markus Emde Jürgen Roßmann 1 Content Previous Work Self Localization in terrestrial forestry
More informationGPU Accelerating Speeded-Up Robust Features Timothy B. Terriberry, Lindley M. French, and John Helmsen
GPU Accelerating Speeded-Up Robust Features Timothy B. Terriberry, Lindley M. French, and John Helmsen Overview of ArgonST Manufacturer of integrated sensor hardware and sensor analysis systems 2 RF, COMINT,
More informationOpenCV on Zynq: Accelerating 4k60 Dense Optical Flow and Stereo Vision. Kamran Khan, Product Manager, Software Acceleration and Libraries July 2017
OpenCV on Zynq: Accelerating 4k60 Dense Optical Flow and Stereo Vision Kamran Khan, Product Manager, Software Acceleration and Libraries July 2017 Agenda Why Zynq SoCs for Traditional Computer Vision Automated
More informationVisual Pose Estimation and Identification for Satellite Rendezvous Operations
Post, Mark A. and Yan, Xiu T. (2015) Visual pose estimation and identification for satellite rendezvous operations. In: Sixth China- Scotland SIPRA workshop "Recent Advances in Signal and Image Processing",
More informationE27 Computer Vision - Final Project: Creating Panoramas David Nahmias, Dan Spagnolo, Vincent Stigliani Professor Zucker Due 5/10/13
E27 Computer Vision - Final Project: Creating Panoramas David Nahmias, Dan Spagnolo, Vincent Stigliani Professor Zucker Due 5/10/13 Sources Brown, M.; Lowe, D.G., "Recognising panoramas," Computer Vision,
More informationSegmentation and Tracking of Partial Planar Templates
Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract
More informationOutline. Introduction System Overview Camera Calibration Marker Tracking Pose Estimation of Markers Conclusion. Media IC & System Lab Po-Chen Wu 2
Outline Introduction System Overview Camera Calibration Marker Tracking Pose Estimation of Markers Conclusion Media IC & System Lab Po-Chen Wu 2 Outline Introduction System Overview Camera Calibration
More informationDepartment of Electrical and Electronic Engineering, University of Peradeniya, KY 20400, Sri Lanka
WIT: Window Intensity Test Detector and Descriptor T.W.U.Madhushani, D.H.S.Maithripala, J.V.Wijayakulasooriya Postgraduate and Research Unit, Sri Lanka Technological Campus, CO 10500, Sri Lanka. Department
More informationSpecular 3D Object Tracking by View Generative Learning
Specular 3D Object Tracking by View Generative Learning Yukiko Shinozuka, Francois de Sorbier and Hideo Saito Keio University 3-14-1 Hiyoshi, Kohoku-ku 223-8522 Yokohama, Japan shinozuka@hvrl.ics.keio.ac.jp
More informationFace Detection and Tracking Control with Omni Car
Face Detection and Tracking Control with Omni Car Jheng-Hao Chen, Tung-Yu Wu CS 231A Final Report June 31, 2016 Abstract We present a combination of frontal and side face detection approach, using deep
More informationRobotics Programming Laboratory
Chair of Software Engineering Robotics Programming Laboratory Bertrand Meyer Jiwon Shin Lecture 8: Robot Perception Perception http://pascallin.ecs.soton.ac.uk/challenges/voc/databases.html#caltech car
More informationThe Most User-Friendly 3D scanner
The Most User-Friendly 3D scanner The Solutionix C500 is optimized for scanning small- to medium-sized objects. With dual 5.0MP cameras, the C500 provides excellent data quality at a high resolution. In
More information