Using infrared proximity sensors for close 2D localization and object size recognition. Richard Berglind Neonode
|
|
- Cory Lang
- 5 years ago
- Views:
Transcription
1 Using infrared proximity sensors for close 2D localization and object size recognition Richard Berglind Neonode
2 Outline Overview of sensor types IR proximity sensors and their drawbacks Principles of a distributed sensor Robot implementation
3 Sensor types
4 Sensor types Camera LIDAR Time of flight Ultrasound IR proximity
5 Camera 2D image of the surrounding environment Stereo vision or structured light to help determine distance and object shape Image processing required Camera FOV limits the minimum sensor to object distance
6 LIDAR Scans a series of light pulses to generate a point cloud of the surrounding environment Distance based on return time of pulses Suitable for long range applications Expensive
7 Time of flight Phase shift method 2D-chip generates 3D image Requires calibration Temperature sensitive Accuracy ~1 cm Improve accuracy at the expense of frame rate Difficult to measure very short distances
8 Ultrasound Speed of sound ~340 m/s Round trip 50 cm distance ~3ms Sensor array: tradeoff between frame rate and resolution
9 IR proximity
10 Basic proximity components Receiver (PD) Field of view Light blocker Emitter (LED/laser) Emitted light
11 Diffuse object scatters light in all directions
12 A fraction of the scattered light is detected
13 Detector signal decreases with distance
14 What about the reflectivity (bright/dark objects)?
15 Object shape recognition
16 Object shape recognition striped objects
17 Distributed sensor
18 Distributed sensor concept Alternating emitters and receivers Collimating optics Receiver field of view at an angle toward the emitters Each sequentially scanned emitter is used in combination with several receivers Object position is determined based on relative signal levels and geometry Independent of material reflectivity
19 Emitter (laser/led)
20 Receiver field of view
21 Multiple emitters
22 Emitter-receiver maxima for one receiver
23 Emitter-receiver maxima for multiple receivers
24 Two viewing directions for every receiver
25 Full set of emitter-receiver maxima
26 Short distance
27 Long distance
28 Lateral resolution signal ratio between emitters
29 Distance resolution signal ratio between receivers
30 Specifications Accuracy: < 5 mm Resolution: 0.1 mm Maximum scanning frequency: ~900 Hz Typical power consumption: mw
31 Shape recognition Each emitter generates a distance value Distance values for multiple emitters can be translated into an object shape with the same resolution as the distance between the emitters (7.2 mm)
32 Shape recognition different shapes
33 Shape recognition striped objects
34 Robot implementation
35 Roboteam Twente Multidisciplinary team of students from the University of Twente, Netherlands Autonomous robot soccer team with eight robot players Computer controlled during games Cameras above the field track general positions of ball and robots Rapid and precise 2D-localization of the ball relative the robot is required for good ball handling
36 Roboteam Twente soccer playing robot Courtesy of Roboteam Twente
37 Roboteam Twente soccer playing robot Courtesy of Roboteam Twente
38 Roboteam Twente robot response Courtesy of Roboteam Twente
39 Summary Ordinary IR proximity sensor generates ambiguous and possibly erroneous data Distributed sensor overcomes this by using a geometric approach Fast, accurate 2D-localization and shape recognition Well suited for robot applications
40
Sensor Modalities. Sensor modality: Different modalities:
Sensor Modalities Sensor modality: Sensors which measure same form of energy and process it in similar ways Modality refers to the raw input used by the sensors Different modalities: Sound Pressure Temperature
More informationSensor technology for mobile robots
Laser application, vision application, sonar application and sensor fusion (6wasserf@informatik.uni-hamburg.de) Outline Introduction Mobile robots perception Definitions Sensor classification Sensor Performance
More informationSolid-State Hybrid LiDAR for Autonomous Driving Product Description
Solid-State Hybrid LiDAR for Autonomous Driving Product Description What is LiDAR Sensor Who is Using LiDARs How does LiDAR Work Hesai LiDAR Demo Features Terminologies Specifications What is LiDAR A LiDAR
More informationRange Sensors (time of flight) (1)
Range Sensors (time of flight) (1) Large range distance measurement -> called range sensors Range information: key element for localization and environment modeling Ultrasonic sensors, infra-red sensors
More information10/5/09 1. d = 2. Range Sensors (time of flight) (2) Ultrasonic Sensor (time of flight, sound) (1) Ultrasonic Sensor (time of flight, sound) (2) 4.1.
Range Sensors (time of flight) (1) Range Sensors (time of flight) (2) arge range distance measurement -> called range sensors Range information: key element for localization and environment modeling Ultrasonic
More informationDistance Sensors: Sound, Light and Vision DISTANCE SENSORS: SOUND, LIGHT AND VISION - THOMAS MAIER 1
Distance Sensors: Sound, Light and Vision THOMAS MAIER SEMINAR: INTELLIGENT ROBOTICS DISTANCE SENSORS: SOUND, LIGHT AND VISION - THOMAS MAIER 1 Structure Motivation Distance Sensors Sound Light Vision
More informationOutline Sensors. EE Sensors. H.I. Bozma. Electric Electronic Engineering Bogazici University. December 13, 2017
Electric Electronic Engineering Bogazici University December 13, 2017 Absolute position measurement Outline Motion Odometry Inertial systems Environmental Tactile Proximity Sensing Ground-Based RF Beacons
More informationDepth Sensors Kinect V2 A. Fornaser
Depth Sensors Kinect V2 A. Fornaser alberto.fornaser@unitn.it Vision Depth data It is not a 3D data, It is a map of distances Not a 3D, not a 2D it is a 2.5D or Perspective 3D Complete 3D - Tomography
More informationLecture 19: Depth Cameras. Visual Computing Systems CMU , Fall 2013
Lecture 19: Depth Cameras Visual Computing Systems Continuing theme: computational photography Cameras capture light, then extensive processing produces the desired image Today: - Capturing scene depth
More informationMULTI-MODAL MAPPING. Robotics Day, 31 Mar Frank Mascarich, Shehryar Khattak, Tung Dang
MULTI-MODAL MAPPING Robotics Day, 31 Mar 2017 Frank Mascarich, Shehryar Khattak, Tung Dang Application-Specific Sensors Cameras TOF Cameras PERCEPTION LiDAR IMU Localization Mapping Autonomy Robotic Perception
More informationOutline. ETN-FPI Training School on Plenoptic Sensing
Outline Introduction Part I: Basics of Mathematical Optimization Linear Least Squares Nonlinear Optimization Part II: Basics of Computer Vision Camera Model Multi-Camera Model Multi-Camera Calibration
More informationLeica s Pinpoint EDM Technology with Modified Signal Processing and Novel Optomechanical Features
Leica s Pinpoint EDM Technology with Modified Signal Processing and Novel Optomechanical Features Fadi A Bayoud Ph.D. Geomatics Engineering Contents Leica s Total Stations and Telescope ATR PS Leica s
More informationME132 February 3, 2011
ME132 February 3, 2011 Outline: - active sensors - introduction to lab setup (Player/Stage) - lab assignment - brief overview of OpenCV ME132 February 3, 2011 Outline: - active sensors - introduction to
More informationIndoor Mobile Robot Navigation and Obstacle Avoidance Using a 3D Camera and Laser Scanner
AARMS Vol. 15, No. 1 (2016) 51 59. Indoor Mobile Robot Navigation and Obstacle Avoidance Using a 3D Camera and Laser Scanner Peter KUCSERA 1 Thanks to the developing sensor technology in mobile robot navigation
More informationActuated Sensor Networks: a 40-minute rant on convergence of vision, robotics, and sensor networks
1 Actuated Sensor Networks: a 40-minute rant on convergence of vision, robotics, and sensor networks Chad Jenkins Assistant Professor Computer Science Department Brown University Women in Computer Science
More information3D Modeling of Objects Using Laser Scanning
1 3D Modeling of Objects Using Laser Scanning D. Jaya Deepu, LPU University, Punjab, India Email: Jaideepudadi@gmail.com Abstract: In the last few decades, constructing accurate three-dimensional models
More informationCanny Edge Based Self-localization of a RoboCup Middle-sized League Robot
Canny Edge Based Self-localization of a RoboCup Middle-sized League Robot Yoichi Nakaguro Sirindhorn International Institute of Technology, Thammasat University P.O. Box 22, Thammasat-Rangsit Post Office,
More informationRevolutionizing 2D measurement. Maximizing longevity. Challenging expectations. R2100 Multi-Ray LED Scanner
Revolutionizing 2D measurement. Maximizing longevity. Challenging expectations. R2100 Multi-Ray LED Scanner A Distance Ahead A Distance Ahead: Your Crucial Edge in the Market The new generation of distancebased
More informationCS595:Introduction to Computer Vision
CS595:Introduction to Computer Vision Instructor: Qi Li Instructor Course syllabus E-mail: qi.li@cs.wku.edu Office: TCCW 135 Office hours MW: 9:00-10:00, 15:00-16:00 T: 9:00-12:00, 14:00-16:00 F: 9:00-10:00
More informationOverview of Active Vision Techniques
SIGGRAPH 99 Course on 3D Photography Overview of Active Vision Techniques Brian Curless University of Washington Overview Introduction Active vision techniques Imaging radar Triangulation Moire Active
More informationToF Camera for high resolution 3D images with affordable pricing
ToF Camera for high resolution 3D images with affordable pricing Basler AG Jana Bartels, Product Manager 3D Agenda Coming next I. Basler AG II. 3D Purpose and Time-of-Flight - Working Principle III. Advantages
More information3D Time-of-Flight Image Sensor Solutions for Mobile Devices
3D Time-of-Flight Image Sensor Solutions for Mobile Devices SEMICON Europa 2015 Imaging Conference Bernd Buxbaum 2015 pmdtechnologies gmbh c o n f i d e n t i a l Content Introduction Motivation for 3D
More informationSupplier Business Opportunities on ADAS and Autonomous Driving Technologies
AUTOMOTIVE Supplier Business Opportunities on ADAS and Autonomous Driving Technologies 19 October 2016 Tokyo, Japan Masanori Matsubara, Senior Analyst, +81 3 6262 1734, Masanori.Matsubara@ihsmarkit.com
More information3D object recognition used by team robotto
3D object recognition used by team robotto Workshop Juliane Hoebel February 1, 2016 Faculty of Computer Science, Otto-von-Guericke University Magdeburg Content 1. Introduction 2. Depth sensor 3. 3D object
More informationPMD [vision] Day Vol. 3 Munich, November 18, PMD Cameras for Automotive & Outdoor Applications. ifm electronic gmbh, V.Frey. Dr.
R PMD [vision] Day Vol. 3 Munich, November 18, 2010 Dr. Volker Frey ifm electronic gmbh PMD Cameras for Automotive & Outdoor Applications Stand: 27.10.2010 Seite 1 I Working Principle PMD distance measurement
More informationOptical Sensors: Key Technology for the Autonomous Car
Optical Sensors: Key Technology for the Autonomous Car Rajeev Thakur, P.E., Product Marketing Manager, Infrared Business Unit, Osram Opto Semiconductors Autonomously driven cars will combine a variety
More informationVisual Perception Sensors
G. Glaser Visual Perception Sensors 1 / 27 MIN Faculty Department of Informatics Visual Perception Sensors Depth Determination Gerrit Glaser University of Hamburg Faculty of Mathematics, Informatics and
More informationInteraction with the Physical World
Interaction with the Physical World Methods and techniques for sensing and changing the environment Light Sensing and Changing the Environment Motion and acceleration Sound Proximity and touch RFID Sensors
More informationRobotEye RE08 3D LIDAR 3D Laser Scanning System. Product Datasheet
3 D L A S E R S C A N N I N G S Y S T E M S RobotEye RE08 3D LIDAR 3D Laser Scanning System Product Datasheet 2015 Ocular Robotics Ltd All rights reserved RobotEye RE08 3D LIDAR - 3D Laser Scanning System
More informationPhotoneo's brand new PhoXi 3D Camera is the highest resolution and highest accuracy area based 3D
Company: Photoneo s.r.o. Germany Contact: Veronika Pulisova E-mail: pulisova@photoneo.com PhoXi 3D Camera Author: Tomas Kovacovsky & Jan Zizka Description of the innovation: General description Photoneo's
More informationElectronic Travel Aids for Blind Guidance an Industry Landscape Study
Electronic Travel Aids for Blind Guidance an Industry Landscape Study Kun (Linda) Li EECS, UC Berkeley Dec 4 th, 2015 IND ENG 290 Final Presentation Visual Impairment Traditional travel aids are limited,
More informationAre you looking for ultrafast and extremely precise stereovision technology for industrial applications? Learn about
Edition November 2017 Image sensors and vision systems, Smart Industries, imec.engineering Are you looking for ultrafast and extremely precise stereovision technology for industrial applications? Learn
More information3D Photography: Active Ranging, Structured Light, ICP
3D Photography: Active Ranging, Structured Light, ICP Kalin Kolev, Marc Pollefeys Spring 2013 http://cvg.ethz.ch/teaching/2013spring/3dphoto/ Schedule (tentative) Feb 18 Feb 25 Mar 4 Mar 11 Mar 18 Mar
More informationUNIT II SPECIAL SEMICONDUCTOR DEVICES
UNIT II SPECIAL SEMICONDUCTOR DEVICES 1 2 3 4 5 6 7 8 9 10 11 12 13 A photodiode is a semiconductor device that converts light into an electrical current. The current is generated when photons are absorbed
More informationCS4495/6495 Introduction to Computer Vision
CS4495/6495 Introduction to Computer Vision 9C-L1 3D perception Some slides by Kelsey Hawkins Motivation Why do animals, people & robots need vision? To detect and recognize objects/landmarks Is that a
More informationPOME A mobile camera system for accurate indoor pose
POME A mobile camera system for accurate indoor pose Paul Montgomery & Andreas Winter November 2 2016 2010. All rights reserved. 1 ICT Intelligent Construction Tools A 50-50 joint venture between Trimble
More information3D Scanning. Qixing Huang Feb. 9 th Slide Credit: Yasutaka Furukawa
3D Scanning Qixing Huang Feb. 9 th 2017 Slide Credit: Yasutaka Furukawa Geometry Reconstruction Pipeline This Lecture Depth Sensing ICP for Pair-wise Alignment Next Lecture Global Alignment Pairwise Multiple
More informationComputational Imaging for Self-Driving Vehicles
CVPR 2018 Computational Imaging for Self-Driving Vehicles Jan Kautz--------Ramesh Raskar--------Achuta Kadambi--------Guy Satat Computational Imaging for Self-Driving Vehicles Jan Kautz--------Ramesh Raskar--------Achuta
More informationReplacing expensive machine vision systems with low-cost 3D Time of Flight sensors in mobile robotics
White paper PULUWHITE001 Replacing expensive machine vision systems with low-cost 3D Time of Flight sensors in mobile robotics Antti Alhonen, Pulu Robotics Oy, 2018 www.pulurobotics.fi Definitions Autonomous
More informationWhy the Self-Driving Revolution Hinges on one Enabling Technology: LiDAR
Why the Self-Driving Revolution Hinges on one Enabling Technology: LiDAR Markus Prison Director Business Development Europe Quanergy ID: 23328 Who We Are The leader in LiDAR (laser-based 3D spatial sensor)
More informationDedicated Software Algorithms for 3D Clouds of Points
Dedicated Software Algorithms for 3D Clouds of Points Introduction AQSENSE develops and commercializes 3D image acquisition and processing technologies that allow high speed in-line 100% production inspection,
More informationSCAN & PAINT 3D. Product Leaflet NEW PRODUCT. icroflown Technologies Charting sound fields MICROFLOWN // CHARTING SOUND FIELDS
icroflown Technologies Charting sound fields Product Leaflet SCAN & NEW PRODUCT PAINT 3D Microflown Technologies Tivolilaan 205 6824 BV Arnhem The Netherlands Phone : +31 088 0010800 Fax : +31 088 0010810
More informationMulti-View Stereo for Community Photo Collections Michael Goesele, et al, ICCV Venus de Milo
Vision Sensing Multi-View Stereo for Community Photo Collections Michael Goesele, et al, ICCV 2007 Venus de Milo The Digital Michelangelo Project, Stanford How to sense 3D very accurately? How to sense
More information3D scanning. 3D scanning is a family of technologies created as a means of automatic measurement of geometric properties of objects.
Acquiring 3D shape 3D scanning 3D scanning is a family of technologies created as a means of automatic measurement of geometric properties of objects. The produced digital model is formed by geometric
More informationDepth Camera for Mobile Devices
Depth Camera for Mobile Devices Instructor - Simon Lucey 16-423 - Designing Computer Vision Apps Today Stereo Cameras Structured Light Cameras Time of Flight (ToF) Camera Inferring 3D Points Given we have
More informationThermal and Optical Cameras. By Philip Smerkovitz TeleEye South Africa
Thermal and Optical Cameras By Philip Smerkovitz TeleEye South Africa phil@teleeye.co.za OPTICAL CAMERAS OVERVIEW Traditional CCTV Camera s (IP and Analog, many form factors). Colour and Black and White
More informationLecture 8 Active stereo & Volumetric stereo
Lecture 8 Active stereo & Volumetric stereo Active stereo Structured lighting Depth sensing Volumetric stereo: Space carving Shadow carving Voxel coloring Reading: [Szelisky] Chapter 11 Multi-view stereo
More informationRange Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation
Obviously, this is a very slow process and not suitable for dynamic scenes. To speed things up, we can use a laser that projects a vertical line of light onto the scene. This laser rotates around its vertical
More information3D Simultaneous Localization and Mapping and Navigation Planning for Mobile Robots in Complex Environments
3D Simultaneous Localization and Mapping and Navigation Planning for Mobile Robots in Complex Environments Sven Behnke University of Bonn, Germany Computer Science Institute VI Autonomous Intelligent Systems
More informationStructured LED Lighting
Structured LED Lighting Very intense and uniform LED pattern projector Full range of colors: from UV to IR, white Long lifetime and few maintenance Compatible with most objectives Standard version Electronics
More informationSnow cover change detection with laser scanning range and brightness measurements
Snow cover change detection with laser scanning range and brightness measurements Sanna Kaasalainen, Harri Kaartinen, Antero Kukko, Henri Niittymäki Department of Remote Sensing and Photogrammetry 5th
More informationGP2Y0A Series / GP2Y0D Series Application Note
GP2Y0ASeries / GP2Y0DSeries GP2Y0A Series / GP2Y0D Series Application Note Content Page 1. General description 2 2. Measuring principle and features of this sensor 2-1. Principle of triangulation 2 2-2.
More informationAUTOMOTIVE ENVIRONMENT SENSORS
AUTOMOTIVE ENVIRONMENT SENSORS Lecture 3. LIDARs Dr. Szilárd Aradi BME KÖZLEKEDÉSMÉRNÖKI ÉS JÁRMŰMÉRNÖKI KAR 32708-2/2017/INTFIN SZÁMÚ EMMI ÁLTAL TÁMOGATOTT TANANYAG LIDAR intro Light Detection and Ranging
More information3D Computer Vision 1
3D Computer Vision 1 Multiview Stereo Multiview Stereo Multiview Stereo https://www.youtube.com/watch?v=ugkb7itpnae Shape from silhouette Shape from silhouette Shape from silhouette Shape from silhouette
More informationNew Sony DepthSense TM ToF Technology
ADVANCED MATERIAL HANDLING WITH New Sony DepthSense TM ToF Technology Jenson Chang Product Marketing November 7, 2018 1 3D SENSING APPLICATIONS Pick and Place Drones Collision Detection People Counting
More informationCamera Drones Lecture 2 Control and Sensors
Camera Drones Lecture 2 Control and Sensors Ass.Prof. Friedrich Fraundorfer WS 2017 1 Outline Quadrotor control principles Sensors 2 Quadrotor control - Hovering Hovering means quadrotor needs to hold
More information3D Laser Range Finder Topological sensor gathering spatial data from the surrounding environment
Initial Project and Group Identification Document September 19, 2013 3D Laser Range Finder Topological sensor gathering spatial data from the surrounding environment Christian Conrose Jonathan Ulrich Andrew
More information(a) (b) (c) Fig. 1. Omnidirectional camera: (a) principle; (b) physical construction; (c) captured. of a local vision system is more challenging than
An Omnidirectional Vision System that finds and tracks color edges and blobs Felix v. Hundelshausen, Sven Behnke, and Raul Rojas Freie Universität Berlin, Institut für Informatik Takustr. 9, 14195 Berlin,
More informationPhenoBot: A Robot System for Phenotyping in the Greenhouse Using 3D Light-Field Technology
PhenoBot: A Robot System for Phenotyping in the Greenhouse Using 3D Light-Field Technology IEEE TC Agricultural Robotics - AgRA Webinar #20 August 27, 2014. Gerrit Polder, Jan Willem Hofstee, Wageningen
More information1-chip DLP Laser Projector with High Brightness LX-MH502Z
1-chip DLP Laser Projector with High Brightness LX-MH502Z Product Outline The LX-MH502Z is a compact and lightweight 4K UHD DLP laser projector capable of projecting high-resolution 4K images, making it
More informationSet up and Foundation of the Husky
Set up and Foundation of the Husky Marisa Warner, Jeremy Gottlieb, Gabriel Elkaim Worcester Polytechnic Institute University of California, Santa Cruz Abstract Clearpath s Husky A200 is an unmanned ground
More informationAUTONOMOUS IMAGE EXTRACTION AND SEGMENTATION OF IMAGE USING UAV S
AUTONOMOUS IMAGE EXTRACTION AND SEGMENTATION OF IMAGE USING UAV S Radha Krishna Rambola, Associate Professor, NMIMS University, India Akash Agrawal, Student at NMIMS University, India ABSTRACT Due to the
More informationNew Sony DepthSense TM ToF Technology
ADVANCED MATERIAL HANDLING WITH New Sony DepthSense TM ToF Technology Jenson Chang Product Marketing November 7, 2018 1 3D SENSING APPLICATIONS Pick and Place Drones Collision Detection People Counting
More information3D Technologies for Nuclear Safeguards Verification. E. Wolfart, JRC-ITU-Ispra IPNDV WG3 Meeting May 2016, Ispra Italy
3D Technologies for Nuclear Safeguards Verification E. Wolfart, JRC-ITU-Ispra IPNDV WG3 Meeting 12-13 May 2016, Ispra Italy Objectives Nuclear safeguards verification systems based on 3D imaging for: Design
More informationNew Approach in Non- Contact 3D Free Form Scanning
New Approach in Non- Contact 3D Free Form Scanning Contents Abstract Industry Trends The solution A smart laser scanning system Implementation of the laser scanning probe in parts inspection Conclusion
More informationFinal Report. EEL 5666 Intelligent Machines Design Laboratory
Final Report EEL 5666 Intelligent Machines Design Laboratory TAs: Mike Pridgen & Thomas Vermeer Instructors: Dr. A. Antonio Arroyo & Dr. Eric M. Schwartz Hao (Hardy) He Dec 08 th, 2009 Table of Contents
More informationCeilbot vision and mapping system
Ceilbot vision and mapping system Provide depth and camera data from the robot's environment Keep a map of the environment based on the received data Keep track of the robot's location on the map Recognize
More informationRetrieval of 3D-position of a Passive Object Using Infrared LED s and Photodiodes Christensen, Henrik Vie
Aalborg Universitet Retrieval of 3D-position of a Passive Object Using Infrared LED s and Photodiodes Christensen, Henrik Vie Publication date: 24 Document Version Publisher's PDF, also known as Version
More information3D Photography: Stereo
3D Photography: Stereo Marc Pollefeys, Torsten Sattler Spring 2016 http://www.cvg.ethz.ch/teaching/3dvision/ 3D Modeling with Depth Sensors Today s class Obtaining depth maps / range images unstructured
More informationThe world s first and only Windows collaboration display created in partnership with Microsoft
The world s first and only Windows collaboration display created in partnership with Microsoft Accelerating teamwork Collaboration is critical to business success, so choosing the right tools to enhance
More informationCalibration of a rotating multi-beam Lidar
The 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems October 18-22, 2010, Taipei, Taiwan Calibration of a rotating multi-beam Lidar Naveed Muhammad 1,2 and Simon Lacroix 1,2 Abstract
More informationIntroduction to 3D Machine Vision
Introduction to 3D Machine Vision 1 Many methods for 3D machine vision Use Triangulation (Geometry) to Determine the Depth of an Object By Different Methods: Single Line Laser Scan Stereo Triangulation
More informationChaplin, Modern Times, 1936
Chaplin, Modern Times, 1936 [A Bucket of Water and a Glass Matte: Special Effects in Modern Times; bonus feature on The Criterion Collection set] Multi-view geometry problems Structure: Given projections
More informationRobotic Systems ECE 401RB Fall 2006
The following notes are from: Robotic Systems ECE 401RB Fall 2006 Lecture 15: Processors Part 3 Chapter 14, G. McComb, and M. Predko, Robot Builder's Bonanza, Third Edition, Mc- Graw Hill, 2006. I. Peripherals
More information5mm Round With Flange Type Infrared LED Technical Data Sheet. Part No.: LL-503IRT2E-2AC
5mm Round With Flange Type Infrared LED Technical Data Sheet Part No.: LL-503IRT2E-2AC Spec No: B508 X335 Rev No: V.2 Date: Nov/26/2005 Page: 1 OF 8 Features: Standard T-1 3/4 diameter package. Low forward
More informationWorld s first MEMS ultrasonic time-of-flight sensors
Technologies & Products Press Conference 2018 World s first MEMS ultrasonic time-of-flight sensors Miniature high-performance range finding, presence detection and more Dr. David Horsley CTO Chirp Microsystems,
More informationCLASSIFICATION FOR ROADSIDE OBJECTS BASED ON SIMULATED LASER SCANNING
CLASSIFICATION FOR ROADSIDE OBJECTS BASED ON SIMULATED LASER SCANNING Kenta Fukano 1, and Hiroshi Masuda 2 1) Graduate student, Department of Intelligence Mechanical Engineering, The University of Electro-Communications,
More informationMobile Autonomous Robotic Sentry (MARS) with Facial Detection and Recognition
Mobile Autonomous Robotic Sentry (MARS) with Facial Detection and Recognition Tyler M. Lovelly University of Florida, Dept. of Electrical & Computer Engineering 12315 Clarendon Ct Spring Hill, FL 34609
More informationUnderstanding Tracking and StroMotion of Soccer Ball
Understanding Tracking and StroMotion of Soccer Ball Nhat H. Nguyen Master Student 205 Witherspoon Hall Charlotte, NC 28223 704 656 2021 rich.uncc@gmail.com ABSTRACT Soccer requires rapid ball movements.
More informationIR-Signature of the MULDICON Configuration Determined by the IR-Signature Model MIRA
DLR.de Folie 1 IR-Signature of the MULDICON Configuration Determined by the IR-Signature Model MIRA Erwin Lindermeir*, Markus Rütten DLR German Aerospace Center *Remote Sensing Technology Institute Institute
More informationA 3-D Scanner Capturing Range and Color for the Robotics Applications
J.Haverinen & J.Röning, A 3-D Scanner Capturing Range and Color for the Robotics Applications, 24th Workshop of the AAPR - Applications of 3D-Imaging and Graph-based Modeling, May 25-26, Villach, Carinthia,
More information3D Localizers for Surgical Navigation Surgical Navigation Systems
3D Localizers for Surgical Navigation 600.445 Russell H. Taylor Surgical Navigation Systems Workstation Image: dallassinuscenter.com Images Tracking device Images: www.radiologyinfo.org Tool 1 3D localizers
More informationThird Rock from the Sun
Geodesy 101 AHD LiDAR Best Practice The Mystery of LiDAR Best Practice Glenn Jones SSSi GIS in the Coastal Environment Batemans Bay November 9, 2010 Light Detection and Ranging (LiDAR) Basic principles
More informationDD2426 Robotics and Autonomous Systems Lecture 4: Robot Sensors and Perception
DD2426 Robotics and Autonomous Systems Lecture 4: Robot Sensors and Perception Patric Jensfelt Kungliga Tekniska Högskolan patric@kth.se April 8,2008 Example: Robots and sensors B21 (Real world interfaces)
More informationNon-axially-symmetric Lens with extended depth of focus for Machine Vision applications
Non-axially-symmetric Lens with extended depth of focus for Machine Vision applications Category: Sensors & Measuring Techniques Reference: TDI0040 Broker Company Name: D Appolonia Broker Name: Tanya Scalia
More informationDevelopment of 3D Positioning Scheme by Integration of Multiple Wiimote IR Cameras
Proceedings of the 5th IIAE International Conference on Industrial Application Engineering 2017 Development of 3D Positioning Scheme by Integration of Multiple Wiimote IR Cameras Hui-Yuan Chan *, Ting-Hao
More informationDigital Image Processing COSC 6380/4393
Digital Image Processing COSC 6380/4393 Lecture 21 Nov 16 th, 2017 Pranav Mantini Ack: Shah. M Image Processing Geometric Transformation Point Operations Filtering (spatial, Frequency) Input Restoration/
More information7 3-Sep Localization and Navigation (GPS, INS, & SLAM) 8 10-Sep State-space modelling & Controller Design 9 17-Sep Vision-based control
RoboticsCourseWare Contributor 2012 School of Information Technology and Electrical Engineering at the University of Queensland Schedule Week Date Lecture (M: 12-1:30, 43-102) 1 23-Jul Introduction Representing
More informationLAS extrabytes implementation in RIEGL software WHITEPAPER
in RIEGL software WHITEPAPER _ Author: RIEGL Laser Measurement Systems GmbH Date: May 25, 2012 Status: Release Pages: 13 All rights are reserved in the event of the grant or the registration of a utility
More informationRobot-Based Solutions for NDT Inspections: Integration of Laser Ultrasonics and Air Coupled Ultrasounds for Aeronautical Components
19 th World Conference on Non-Destructive Testing 2016 Robot-Based Solutions for NDT Inspections: Integration of Laser Ultrasonics and Air Coupled Ultrasounds for Aeronautical Components Esmeralda CUEVAS
More informationIntroduction to Computer Vision
Introduction to Computer Vision Dr. Gerhard Roth COMP 4102A Winter 2015 Version 2 General Information Instructor: Adjunct Prof. Dr. Gerhard Roth gerhardroth@rogers.com read hourly gerhardroth@cmail.carleton.ca
More informationModeling the Interaction of a Laser Target Detection Device with the Sea
Modeling the Interaction of a Laser Target Detection Device with the Sea Gary Buzzard, Thales Missile Electronics Proximity Fuze Product Technical Manager Content Low Level & Embedded Threats TDD Sensor
More informationEnabling Technologies for Robot Assisted Ultrasound Tomography
Enabling Technologies for Robot Assisted Ultrasound Tomography Seminar Presentation By: Fereshteh Aalamifar Team Members: Rishabh Khurana, Fereshteh Aalamifar Mentors: Emad Boctor, Iulian Iordachita, Russell
More informationComputer and Machine Vision
Computer and Machine Vision Lecture Week 12 Part-2 Additional 3D Scene Considerations March 29, 2014 Sam Siewert Outline of Week 12 Computer Vision APIs and Languages Alternatives to C++ and OpenCV API
More informationActive Stereo Vision. COMP 4900D Winter 2012 Gerhard Roth
Active Stereo Vision COMP 4900D Winter 2012 Gerhard Roth Why active sensors? Project our own texture using light (usually laser) This simplifies correspondence problem (much easier) Pluses Can handle different
More informationWorld s Smallest Time-of-Flight Proximity/Ranging Sensor. ST Imaging Marcus (Xiaoyong) Yang Product Marketing Manager
World s Smallest Time-of-Flight Proximity/Ranging Sensor ST Imaging Marcus (Xiaoyong) Yang Product Marketing Manager FlightSense Breakthrough Technology 2 Measurement at the true speed of light! FlightSense
More information3D Object Representations. COS 526, Fall 2016 Princeton University
3D Object Representations COS 526, Fall 2016 Princeton University 3D Object Representations How do we... Represent 3D objects in a computer? Acquire computer representations of 3D objects? Manipulate computer
More informationPRODUCT INFORMATION Model 1000XP Wide Range Particle Spectrometer (WPS )
PRODUCT INFORMATION Model 1000XP Wide Range Particle Spectrometer (WPS ) With the lower size limit extended from 10nm to 5 nm, the WPS can now count and size aerosol particles automatically from 5nm to
More informationOBSTACLE DETECTION WITH ACTIVE LASER TRIANGULATION
Advances in Production Engineering & Management 2 (2007) 2, 79-90 ISSN 1854-6250 Original scientific paper OBSTACLE DETECTION WITH ACTIVE LASER TRIANGULATION Klančnik, S. * ; Balič, J. * & Planinšič, P.
More informationA Comparison between Active and Passive 3D Vision Sensors: BumblebeeXB3 and Microsoft Kinect
A Comparison between Active and Passive 3D Vision Sensors: BumblebeeXB3 and Microsoft Kinect Diana Beltran and Luis Basañez Technical University of Catalonia, Barcelona, Spain {diana.beltran,luis.basanez}@upc.edu
More information