Jo-Car2 Autonomous Mode. Path Planning (Cost Matrix Algorithm)

Size: px
Start display at page:

Download "Jo-Car2 Autonomous Mode. Path Planning (Cost Matrix Algorithm)"

Transcription

1 Chapter 8.2 Jo-Car2 Autonomous Mode Path Planning (Cost Matrix Algorithm) Introduction: In order to achieve its mission and reach the GPS goal safely; without crashing into obstacles or leaving the lane, Jo-Car2 uses three main sources of information; GPS with magnetometer, Camera and Laser Range Finder. Using the combination between the GPS and magnetometer readings, we can determine the desired direction of the robot. The Camera is used to detect the path lines and also the obstacles, by applying the image processing algorithm described in chapter 9. While the Laser Range Finder is used also to detect obstacles. Camera Image Processing LRF Mapping Cost Matrix Path Selection Motors Driving GPS Direction Detection Magnetometer Path Planning Algorithm Jo-Car2 Autonomous Mode - Cost Matrix Algorithm - Mahmood Shubbak 54

2 The three sensors are then fused into one source of information; The Cost Matrix, that Jo-Car2 depends basically on for its artificial intelligence and path planning. Look at the following diagram. The Cost Matrix: The Cost Matrix is a 7x7 matrix, whose values represent the risk value for each position on the camera image plane; i.e. it has large values where obstacles or path lines exist, and low values in safe positions. Each sequence of matrix elements represents a path; which has its own cost (the sum of elements values), the path of lowest cost is the best path. For thorough understanding of this method, we provide in the following lines a step by step explanation. Step by Step Explanation: 1. Starting from the original image captured by the camera, the first step is to generate a binary image that only includes the obstacles and path lines; this can be achieved by detecting and removing the green grass from the image. Using an image processing algorithm depends basically on the ratio between red, green and blue channels for each pixel; we can get the desired binary image. (For more details about the image processing algorithm refer to Chapter 9). Look at the following example: 2. The second step is to detect and remove noise; this is also achieved by some image processing depends on the size of the white regions; the small ones are considered as noise and removed consequently. (more details can be found in the image processing chapter 9) Look at the following example: Jo-Car2 Autonomous Mode - Cost Matrix Algorithm - Mahmood Shubbak 55

3 Up to this point, we obtain the final binary image that contains the path lines and the obstacles only; without any noise. 3. Now we can generate the cost matrix; which is 7x7 matrix contains the average of pixels intensities as described previously, look at the figure below: Notice that on the obstacles positions (totally white) the matrix value is 255 which is the highest intensity; while on safe positions (totally black) the value is zero which is the lowest intensity. This matrix is the primary framework, to which all other sensors will be fused, and on which the robot will depend in planning its path. 4. The next step is to fuse other sensors into the cost matrix; sensor fusion here means adding the laser range finder LRF, the GPS and the magnetometer information into the path matrix. Firstly, fusing the LRF; by mapping the laser data (which is points on the horizontal ground plane) to the camera image plane. This has been done using some experimental methods for calibration. More details can be found in the LRF Mapping chapter 10. Jo-Car2 Autonomous Mode - Cost Matrix Algorithm - Mahmood Shubbak 56

4 After determining the positions of LRF points on the image plane, and their positions on the matrix consequently, an extra cost of 100 will be added to each position. Look at the following figure. The GPS direction can be then fused into the matrix by detecting wrong way motion; comparing the difference between current position and next GPS with the history data. 5. Now we reach paths evaluation step; on which pre-defined paths will be evaluated in order to determine the best path. Here we will show a sample of five paths; look at the following figure. By adding the cost values of each cell in the forward path we got a sum of 468; which represents the total cost of following the straight forward path. Applying the same procedure, we got results for another four paths as shown in the following figures. Jo-Car2 Autonomous Mode - Cost Matrix Algorithm - Mahmood Shubbak 57

5 Notice that the straight-right path costs 498, while the straight-left path costs 159. The right-straight path costs 713, and the left-straight path costs The last step in our path planning algorithm is choosing the best path; which is the lowest cost path. In our sample image here, it is obvious that the left straight path is the best path with the lowest cost of 45. Jo-Car2 Autonomous Mode - Cost Matrix Algorithm - Mahmood Shubbak 58

6 Summary: After choosing the best path, the controller must convert it into its relevant desired speeds for the both motors according to the differential drive model, explained in the actuators Chapter 6. The PID controller will guarantee that the motor will operate at the desired speed. Jo-Car2 uses a path planning algorithm called cost matrix algorithm; it depends on fusing all sensors into one 7x7 matrix, then evaluating alternative paths on it by their costs. The matrix is generated through six steps; 1 converting the original image from camera to binary image by image processing algorithm, 2 detecting and removing the noise, then 3 generating the matrix by averaging pixels, 4 fusing all sensors, after that 5 evaluating pre-defined paths, and finally 6 selecting the best path with the lowest cost. Jo-Car2 Autonomous Mode - Cost Matrix Algorithm - Mahmood Shubbak 59

W4. Perception & Situation Awareness & Decision making

W4. Perception & Situation Awareness & Decision making W4. Perception & Situation Awareness & Decision making Robot Perception for Dynamic environments: Outline & DP-Grids concept Dynamic Probabilistic Grids Bayesian Occupancy Filter concept Dynamic Probabilistic

More information

HOG-Based Person Following and Autonomous Returning Using Generated Map by Mobile Robot Equipped with Camera and Laser Range Finder

HOG-Based Person Following and Autonomous Returning Using Generated Map by Mobile Robot Equipped with Camera and Laser Range Finder HOG-Based Person Following and Autonomous Returning Using Generated Map by Mobile Robot Equipped with Camera and Laser Range Finder Masashi Awai, Takahito Shimizu and Toru Kaneko Department of Mechanical

More information

One type of these solutions is automatic license plate character recognition (ALPR).

One type of these solutions is automatic license plate character recognition (ALPR). 1.0 Introduction Modelling, Simulation & Computing Laboratory (msclab) A rapid technical growth in the area of computer image processing has increased the need for an efficient and affordable security,

More information

3D Terrain Sensing System using Laser Range Finder with Arm-Type Movable Unit

3D Terrain Sensing System using Laser Range Finder with Arm-Type Movable Unit 3D Terrain Sensing System using Laser Range Finder with Arm-Type Movable Unit 9 Toyomi Fujita and Yuya Kondo Tohoku Institute of Technology Japan 1. Introduction A 3D configuration and terrain sensing

More information

PATENT LIABILITY ANALYSIS. Daniel Barrett Sebastian Hening Sandunmalee Abeyratne Anthony Myers

PATENT LIABILITY ANALYSIS. Daniel Barrett Sebastian Hening Sandunmalee Abeyratne Anthony Myers PATENT LIABILITY ANALYSIS Autonomous Targeting Vehicle (ATV) Daniel Barrett Sebastian Hening Sandunmalee Abeyratne Anthony Myers Autonomous wheeled vehicle with obstacle avoidance Two infrared range finder

More information

Wall-Follower. Xiaodong Fang. EEL5666 Intelligent Machines Design Laboratory University of Florida School of Electrical and Computer Engineering

Wall-Follower. Xiaodong Fang. EEL5666 Intelligent Machines Design Laboratory University of Florida School of Electrical and Computer Engineering Wall-Follower Xiaodong Fang EEL5666 Intelligent Machines Design Laboratory University of Florida School of Electrical and Computer Engineering TAs: Tim Martin Josh Weaver Instructors: Dr. A. Antonio Arroyo

More information

Build and Test Plan: IGV Team

Build and Test Plan: IGV Team Build and Test Plan: IGV Team 2/6/2008 William Burke Donaldson Diego Gonzales David Mustain Ray Laser Range Finder Week 3 Jan 29 The laser range finder will be set-up in the lab and connected to the computer

More information

2002 Intelligent Ground Vehicle Competition Design Report. Grizzly Oakland University

2002 Intelligent Ground Vehicle Competition Design Report. Grizzly Oakland University 2002 Intelligent Ground Vehicle Competition Design Report Grizzly Oakland University June 21, 2002 Submitted By: Matt Rizzo Brian Clark Brian Yurconis Jelena Nikolic I. ABSTRACT Grizzly is the product

More information

Robotics. Chapter 25. Chapter 25 1

Robotics. Chapter 25. Chapter 25 1 Robotics Chapter 25 Chapter 25 1 Outline Robots, Effectors, and Sensors Localization and Mapping Motion Planning Chapter 25 2 Mobile Robots Chapter 25 3 Manipulators P R R R R R Configuration of robot

More information

Exam in DD2426 Robotics and Autonomous Systems

Exam in DD2426 Robotics and Autonomous Systems Exam in DD2426 Robotics and Autonomous Systems Lecturer: Patric Jensfelt KTH, March 16, 2010, 9-12 No aids are allowed on the exam, i.e. no notes, no books, no calculators, etc. You need a minimum of 20

More information

Acoustic/Lidar Sensor Fusion for Car Tracking in City Traffic Scenarios

Acoustic/Lidar Sensor Fusion for Car Tracking in City Traffic Scenarios Sensor Fusion for Car Tracking Acoustic/Lidar Sensor Fusion for Car Tracking in City Traffic Scenarios, Daniel Goehring 1 Motivation Direction to Object-Detection: What is possible with costefficient microphone

More information

Summary of Computing Team s Activities Fall 2007 Siddharth Gauba, Toni Ivanov, Edwin Lai, Gary Soedarsono, Tanya Gupta

Summary of Computing Team s Activities Fall 2007 Siddharth Gauba, Toni Ivanov, Edwin Lai, Gary Soedarsono, Tanya Gupta Summary of Computing Team s Activities Fall 2007 Siddharth Gauba, Toni Ivanov, Edwin Lai, Gary Soedarsono, Tanya Gupta 1 OVERVIEW Input Image Channel Separation Inverse Perspective Mapping The computing

More information

Vehicle Localization. Hannah Rae Kerner 21 April 2015

Vehicle Localization. Hannah Rae Kerner 21 April 2015 Vehicle Localization Hannah Rae Kerner 21 April 2015 Spotted in Mtn View: Google Car Why precision localization? in order for a robot to follow a road, it needs to know where the road is to stay in a particular

More information

Dealing with Scale. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

Dealing with Scale. Stephan Weiss Computer Vision Group NASA-JPL / CalTech Dealing with Scale Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline Why care about size? The IMU as scale provider: The

More information

CS283: Robotics Fall 2016: Software

CS283: Robotics Fall 2016: Software CS283: Robotics Fall 2016: Software Sören Schwertfeger / 师泽仁 ShanghaiTech University Mobile Robotics ShanghaiTech University - SIST - 18.09.2016 2 Review Definition Robot: A machine capable of performing

More information

Turning an Automated System into an Autonomous system using Model-Based Design Autonomous Tech Conference 2018

Turning an Automated System into an Autonomous system using Model-Based Design Autonomous Tech Conference 2018 Turning an Automated System into an Autonomous system using Model-Based Design Autonomous Tech Conference 2018 Asaf Moses Systematics Ltd., Technical Product Manager aviasafm@systematics.co.il 1 Autonomous

More information

Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview

Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview Arun Das 05/09/2017 Arun Das Waterloo Autonomous Vehicles Lab Introduction What s in a name? Arun Das Waterloo Autonomous

More information

St Margaret College Secondary School Verdala. Smart Wheelchair. Form 3 NXT Coursework Project. Marquita Formosa (Class X)

St Margaret College Secondary School Verdala. Smart Wheelchair. Form 3 NXT Coursework Project. Marquita Formosa (Class X) St Margaret College Secondary School Verdala Smart Wheelchair Marquita Formosa (Class X) Marking Scheme (This was given to you during the Computing lesson) Marquita Formosa Page 2 Table of Contents Marking

More information

Fire Bird V Insect - Nex Robotics

Fire Bird V Insect - Nex Robotics Fire Bird V Insect is a small six legged robot. It has three pair of legs driven by one servo each. Robot can navigate itself using Sharp IR range sensors. It can be controlled wirelessly using ZigBee

More information

Sensory Augmentation for Increased Awareness of Driving Environment

Sensory Augmentation for Increased Awareness of Driving Environment Sensory Augmentation for Increased Awareness of Driving Environment Pranay Agrawal John M. Dolan Dec. 12, 2014 Technologies for Safe and Efficient Transportation (T-SET) UTC The Robotics Institute Carnegie

More information

Attack Resilient State Estimation for Vehicular Systems

Attack Resilient State Estimation for Vehicular Systems December 15 th 2013. T-SET Final Report Attack Resilient State Estimation for Vehicular Systems Nicola Bezzo (nicbezzo@seas.upenn.edu) Prof. Insup Lee (lee@cis.upenn.edu) PRECISE Center University of Pennsylvania

More information

Thomas Bräunl EMBEDDED ROBOTICS. Mobile Robot Design and Applications with Embedded Systems. Second Edition. With 233 Figures and 24 Tables.

Thomas Bräunl EMBEDDED ROBOTICS. Mobile Robot Design and Applications with Embedded Systems. Second Edition. With 233 Figures and 24 Tables. Thomas Bräunl EMBEDDED ROBOTICS Mobile Robot Design and Applications with Embedded Systems Second Edition With 233 Figures and 24 Tables Springer CONTENTS PART I: EMBEDDED SYSTEMS 1 Robots and Controllers

More information

NAME :... Signature :... Desk no. :... Question Answer

NAME :... Signature :... Desk no. :... Question Answer Written test Tuesday 19th of December 2000. Aids allowed : All usual aids Weighting : All questions are equally weighted. NAME :................................................... Signature :...................................................

More information

Proc. 14th Int. Conf. on Intelligent Autonomous Systems (IAS-14), 2016

Proc. 14th Int. Conf. on Intelligent Autonomous Systems (IAS-14), 2016 Proc. 14th Int. Conf. on Intelligent Autonomous Systems (IAS-14), 2016 Outdoor Robot Navigation Based on View-based Global Localization and Local Navigation Yohei Inoue, Jun Miura, and Shuji Oishi Department

More information

DTU M.SC. - COURSE EXAM Revised Edition

DTU M.SC. - COURSE EXAM Revised Edition Written test, 16 th of December 1999. Course name : 04250 - Digital Image Analysis Aids allowed : All usual aids Weighting : All questions are equally weighed. Name :...................................................

More information

OCSD-A / AeroCube-7A Status Update

OCSD-A / AeroCube-7A Status Update OCSD-A / AeroCube-7A Status Update Darren Rowen Richard Dolphus Patrick Doyle Addison Faler April 20, 2016 2016 The Aerospace Corporation Agenda Concept of Operations Overview Spacecraft Configuration

More information

EE368 Project: Visual Code Marker Detection

EE368 Project: Visual Code Marker Detection EE368 Project: Visual Code Marker Detection Kahye Song Group Number: 42 Email: kahye@stanford.edu Abstract A visual marker detection algorithm has been implemented and tested with twelve training images.

More information

Localization, Where am I?

Localization, Where am I? 5.1 Localization, Where am I?? position Position Update (Estimation?) Encoder Prediction of Position (e.g. odometry) YES matched observations Map data base predicted position Matching Odometry, Dead Reckoning

More information

THE UNIVERSITY OF AUCKLAND

THE UNIVERSITY OF AUCKLAND THE UNIVERSITY OF AUCKLAND FIRST SEMESTER, 2002 Campus: Tamaki COMPUTER SCIENCE Intelligent Active Vision (Time allowed: TWO hours) NOTE: Attempt questions A, B, C, D, E, and F. This is an open book examination.

More information

Introducing Robotics Vision System to a Manufacturing Robotics Course

Introducing Robotics Vision System to a Manufacturing Robotics Course Paper ID #16241 Introducing Robotics Vision System to a Manufacturing Robotics Course Dr. Yuqiu You, Ohio University c American Society for Engineering Education, 2016 Introducing Robotics Vision System

More information

LaserGuard LG300 area alarm system. 3D laser radar alarm system for motion control and alarm applications. Instruction manual

LaserGuard LG300 area alarm system. 3D laser radar alarm system for motion control and alarm applications. Instruction manual LaserGuard LG300 area alarm system 3D laser radar alarm system for motion control and alarm applications Instruction manual LaserGuard The LaserGuard program is the user interface for the 3D laser scanner

More information

The area processing unit of Caroline

The area processing unit of Caroline 2nd Workshop Robot Vision RobVis '08 February 18-20, 2008 Auckland, New Zealand The area processing unit of Caroline Finding the way through DARPA's urban challenge February 18th, 2008 Kai Berger Christian

More information

Autonomous Vehicle Navigation Using Stereoscopic Imaging

Autonomous Vehicle Navigation Using Stereoscopic Imaging Autonomous Vehicle Navigation Using Stereoscopic Imaging Functional Description and Complete System Block Diagram By: Adam Beach Nick Wlaznik Advisors: Dr. Huggins Dr. Stewart December 14, 2006 I. Introduction

More information

ENGR3390: Robotics Fall 2009

ENGR3390: Robotics Fall 2009 J. Gorasia Vision Lab ENGR339: Robotics ENGR339: Robotics Fall 29 Vision Lab Team Bravo J. Gorasia - 1/4/9 J. Gorasia Vision Lab ENGR339: Robotics Table of Contents 1.Theory and summary of background readings...4

More information

Robot Localization based on Geo-referenced Images and G raphic Methods

Robot Localization based on Geo-referenced Images and G raphic Methods Robot Localization based on Geo-referenced Images and G raphic Methods Sid Ahmed Berrabah Mechanical Department, Royal Military School, Belgium, sidahmed.berrabah@rma.ac.be Janusz Bedkowski, Łukasz Lubasiński,

More information

Localization and Map Building

Localization and Map Building Localization and Map Building Noise and aliasing; odometric position estimation To localize or not to localize Belief representation Map representation Probabilistic map-based localization Other examples

More information

Safe Prediction-Based Local Path Planning using Obstacle Probability Sections

Safe Prediction-Based Local Path Planning using Obstacle Probability Sections Slide 1 Safe Prediction-Based Local Path Planning using Obstacle Probability Sections Tanja Hebecker and Frank Ortmeier Chair of Software Engineering, Otto-von-Guericke University of Magdeburg, Germany

More information

A Reactive Bearing Angle Only Obstacle Avoidance Technique for Unmanned Ground Vehicles

A Reactive Bearing Angle Only Obstacle Avoidance Technique for Unmanned Ground Vehicles Proceedings of the International Conference of Control, Dynamic Systems, and Robotics Ottawa, Ontario, Canada, May 15-16 2014 Paper No. 54 A Reactive Bearing Angle Only Obstacle Avoidance Technique for

More information

Player/Gazebo Simulation Environment John I. Martin 7 January 2005

Player/Gazebo Simulation Environment John I. Martin 7 January 2005 Player/Gazebo Simulation Environment John I. Martin martinj@ece.osu.edu 7 January 2005 Introduction The Player/Gazebo simulation environment provides a virtual world in which high level robot control code

More information

Dominant plane detection using optical flow and Independent Component Analysis

Dominant plane detection using optical flow and Independent Component Analysis Dominant plane detection using optical flow and Independent Component Analysis Naoya OHNISHI 1 and Atsushi IMIYA 2 1 School of Science and Technology, Chiba University, Japan Yayoicho 1-33, Inage-ku, 263-8522,

More information

Laserscanner Based Cooperative Pre-Data-Fusion

Laserscanner Based Cooperative Pre-Data-Fusion Laserscanner Based Cooperative Pre-Data-Fusion 63 Laserscanner Based Cooperative Pre-Data-Fusion F. Ahlers, Ch. Stimming, Ibeo Automobile Sensor GmbH Abstract The Cooperative Pre-Data-Fusion is a novel

More information

Image Processing using LabVIEW. By, Sandip Nair sandipnair.hpage.com

Image Processing using LabVIEW. By, Sandip Nair sandipnair.hpage.com Image Processing using LabVIEW By, Sandip Nair sandipnair06@yahoomail.com sandipnair.hpage.com What is image? An image is two dimensional function, f(x,y), where x and y are spatial coordinates, and the

More information

Dynamic Sensor-based Path Planning and Hostile Target Detection with Mobile Ground Robots. Matt Epperson Dr. Timothy Chung

Dynamic Sensor-based Path Planning and Hostile Target Detection with Mobile Ground Robots. Matt Epperson Dr. Timothy Chung Dynamic Sensor-based Path Planning and Hostile Target Detection with Mobile Ground Robots Matt Epperson Dr. Timothy Chung Brief Bio Matt Epperson Cal Poly, San Luis Obispo Sophmore Computer Engineer NREIP

More information

Handy Board MX. page 1

Handy Board MX. page 1 Handy Board MX The Handy Board MX (Modular extension) was developed as a quick-connect system to help eliminate connection errors, reduce prototyping time, and lower the bar of necessary technical skill.

More information

On-line and Off-line 3D Reconstruction for Crisis Management Applications

On-line and Off-line 3D Reconstruction for Crisis Management Applications On-line and Off-line 3D Reconstruction for Crisis Management Applications Geert De Cubber Royal Military Academy, Department of Mechanical Engineering (MSTA) Av. de la Renaissance 30, 1000 Brussels geert.de.cubber@rma.ac.be

More information

Chapters 1 7: Overview

Chapters 1 7: Overview Chapters 1 7: Overview Photogrammetric mapping: introduction, applications, and tools GNSS/INS-assisted photogrammetric and LiDAR mapping LiDAR mapping: principles, applications, mathematical model, and

More information

Designing a Pick and Place Robotics Application Using MATLAB and Simulink

Designing a Pick and Place Robotics Application Using MATLAB and Simulink Designing a Pick and Place Robotics Application Using MATLAB and Simulink Carlos Santacruz-Rosero, PhD Sr Application Engineer Robotics Pulkit Kapur Sr Industry Marketing Manager Robotics 2017 The MathWorks,

More information

ECE 172A: Introduction to Intelligent Systems: Machine Vision, Fall Midterm Examination

ECE 172A: Introduction to Intelligent Systems: Machine Vision, Fall Midterm Examination ECE 172A: Introduction to Intelligent Systems: Machine Vision, Fall 2008 October 29, 2008 Notes: Midterm Examination This is a closed book and closed notes examination. Please be precise and to the point.

More information

3D-2D Laser Range Finder calibration using a conic based geometry shape

3D-2D Laser Range Finder calibration using a conic based geometry shape 3D-2D Laser Range Finder calibration using a conic based geometry shape Miguel Almeida 1, Paulo Dias 1, Miguel Oliveira 2, Vítor Santos 2 1 Dept. of Electronics, Telecom. and Informatics, IEETA, University

More information

UAV Autonomous Navigation in a GPS-limited Urban Environment

UAV Autonomous Navigation in a GPS-limited Urban Environment UAV Autonomous Navigation in a GPS-limited Urban Environment Yoko Watanabe DCSD/CDIN JSO-Aerial Robotics 2014/10/02-03 Introduction 2 Global objective Development of a UAV onboard system to maintain flight

More information

Image processing techniques for driver assistance. Razvan Itu June 2014, Technical University Cluj-Napoca

Image processing techniques for driver assistance. Razvan Itu June 2014, Technical University Cluj-Napoca Image processing techniques for driver assistance Razvan Itu June 2014, Technical University Cluj-Napoca Introduction Computer vision & image processing from wiki: any form of signal processing for which

More information

Complex Sensors: Cameras, Visual Sensing. The Robotics Primer (Ch. 9) ECE 497: Introduction to Mobile Robotics -Visual Sensors

Complex Sensors: Cameras, Visual Sensing. The Robotics Primer (Ch. 9) ECE 497: Introduction to Mobile Robotics -Visual Sensors Complex Sensors: Cameras, Visual Sensing The Robotics Primer (Ch. 9) Bring your laptop and robot everyday DO NOT unplug the network cables from the desktop computers or the walls Tuesday s Quiz is on Visual

More information

COMPUTER-BASED WORKPIECE DETECTION ON CNC MILLING MACHINE TOOLS USING OPTICAL CAMERA AND NEURAL NETWORKS

COMPUTER-BASED WORKPIECE DETECTION ON CNC MILLING MACHINE TOOLS USING OPTICAL CAMERA AND NEURAL NETWORKS Advances in Production Engineering & Management 5 (2010) 1, 59-68 ISSN 1854-6250 Scientific paper COMPUTER-BASED WORKPIECE DETECTION ON CNC MILLING MACHINE TOOLS USING OPTICAL CAMERA AND NEURAL NETWORKS

More information

Estimating the wavelength composition of scene illumination from image data is an

Estimating the wavelength composition of scene illumination from image data is an Chapter 3 The Principle and Improvement for AWB in DSC 3.1 Introduction Estimating the wavelength composition of scene illumination from image data is an important topics in color engineering. Solutions

More information

3-D MAP GENERATION BY A MOBILE ROBOT EQUIPPED WITH A LASER RANGE FINDER. Takumi Nakamoto, Atsushi Yamashita, and Toru Kaneko

3-D MAP GENERATION BY A MOBILE ROBOT EQUIPPED WITH A LASER RANGE FINDER. Takumi Nakamoto, Atsushi Yamashita, and Toru Kaneko 3-D AP GENERATION BY A OBILE ROBOT EQUIPPED WITH A LAER RANGE FINDER Takumi Nakamoto, Atsushi Yamashita, and Toru Kaneko Department of echanical Engineering, hizuoka Univerty 3-5-1 Johoku, Hamamatsu-shi,

More information

Introduction to Information Science and Technology (IST) Part IV: Intelligent Machines and Robotics Planning

Introduction to Information Science and Technology (IST) Part IV: Intelligent Machines and Robotics Planning Introduction to Information Science and Technology (IST) Part IV: Intelligent Machines and Robotics Planning Sören Schwertfeger / 师泽仁 ShanghaiTech University ShanghaiTech University - SIST - 10.05.2017

More information

Servosila Robotic Heads

Servosila Robotic Heads Servosila Robotic Heads www.servosila.com TABLE OF CONTENTS SERVOSILA ROBOTIC HEADS 2 SOFTWARE-DEFINED FUNCTIONS OF THE ROBOTIC HEADS 2 SPECIFICATIONS: ROBOTIC HEADS 4 DIMENSIONS OF ROBOTIC HEAD 5 DIMENSIONS

More information

Pedestrian Detection Using Correlated Lidar and Image Data EECS442 Final Project Fall 2016

Pedestrian Detection Using Correlated Lidar and Image Data EECS442 Final Project Fall 2016 edestrian Detection Using Correlated Lidar and Image Data EECS442 Final roject Fall 2016 Samuel Rohrer University of Michigan rohrer@umich.edu Ian Lin University of Michigan tiannis@umich.edu Abstract

More information

COS Lecture 10 Autonomous Robot Navigation

COS Lecture 10 Autonomous Robot Navigation COS 495 - Lecture 10 Autonomous Robot Navigation Instructor: Chris Clark Semester: Fall 2011 1 Figures courtesy of Siegwart & Nourbakhsh Control Structure Prior Knowledge Operator Commands Localization

More information

Vol. 21 No. 6, pp ,

Vol. 21 No. 6, pp , Vol. 21 No. 6, pp.69 696, 23 69 3 3 3 Map Generation of a Mobile Robot by Integrating Omnidirectional Stereo and Laser Range Finder Yoshiro Negishi 3, Jun Miura 3 and Yoshiaki Shirai 3 This paper describes

More information

Usability study of 3D Time-of-Flight cameras for automatic plant phenotyping

Usability study of 3D Time-of-Flight cameras for automatic plant phenotyping 93 Usability study of 3D Time-of-Flight cameras for automatic plant phenotyping Ralph Klose, Jaime Penlington, Arno Ruckelshausen University of Applied Sciences Osnabrück/ Faculty of Engineering and Computer

More information

Understanding Tracking and StroMotion of Soccer Ball

Understanding Tracking and StroMotion of Soccer Ball Understanding Tracking and StroMotion of Soccer Ball Nhat H. Nguyen Master Student 205 Witherspoon Hall Charlotte, NC 28223 704 656 2021 rich.uncc@gmail.com ABSTRACT Soccer requires rapid ball movements.

More information

Stereo Vision Based Traversable Region Detection for Mobile Robots Using U-V-Disparity

Stereo Vision Based Traversable Region Detection for Mobile Robots Using U-V-Disparity Stereo Vision Based Traversable Region Detection for Mobile Robots Using U-V-Disparity ZHU Xiaozhou, LU Huimin, Member, IEEE, YANG Xingrui, LI Yubo, ZHANG Hui College of Mechatronics and Automation, National

More information

On Board 6D Visual Sensors for Intersection Driving Assistance Systems

On Board 6D Visual Sensors for Intersection Driving Assistance Systems On Board 6D Visual Sensors for Intersection Driving Assistance Systems S. Nedevschi, T. Marita, R. Danescu, F. Oniga, S. Bota, I. Haller, C. Pantilie, M. Drulea, C. Golban Sergiu.Nedevschi@cs.utcluj.ro

More information

Revising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History

Revising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History Revising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History Simon Thompson and Satoshi Kagami Digital Human Research Center National Institute of Advanced

More information

CS4758: Rovio Augmented Vision Mapping Project

CS4758: Rovio Augmented Vision Mapping Project CS4758: Rovio Augmented Vision Mapping Project Sam Fladung, James Mwaura Abstract The goal of this project is to use the Rovio to create a 2D map of its environment using a camera and a fixed laser pointer

More information

Odyssey. Build and Test Plan

Odyssey. Build and Test Plan Build and Test Plan The construction and programming of this autonomous ground vehicle requires careful planning and project management to ensure that all tasks are completed appropriately in due time.

More information

Color Tracking Robot

Color Tracking Robot Color Tracking Robot 1 Suraksha Bhat, 2 Preeti Kini, 3 Anjana Nair, 4 Neha Athavale Abstract: This Project describes a visual sensor system used in the field of robotics for identification and tracking

More information

Obstacle Classification and 3D Measurement in Unstructured Environments Based on ToF Cameras

Obstacle Classification and 3D Measurement in Unstructured Environments Based on ToF Cameras Sensors 2014, 14, 10753-10782; doi:10.3390/s140610753 Article OPEN ACCESS sensors ISSN 1424-8220 www.mdpi.com/journal/sensors Obstacle Classification and 3D Measurement in Unstructured Environments Based

More information

Basics of Localization, Mapping and SLAM. Jari Saarinen Aalto University Department of Automation and systems Technology

Basics of Localization, Mapping and SLAM. Jari Saarinen Aalto University Department of Automation and systems Technology Basics of Localization, Mapping and SLAM Jari Saarinen Aalto University Department of Automation and systems Technology Content Introduction to Problem (s) Localization A few basic equations Dead Reckoning

More information

Automatic Fatigue Detection System

Automatic Fatigue Detection System Automatic Fatigue Detection System T. Tinoco De Rubira, Stanford University December 11, 2009 1 Introduction Fatigue is the cause of a large number of car accidents in the United States. Studies done by

More information

Indoor Mobile Robot Navigation and Obstacle Avoidance Using a 3D Camera and Laser Scanner

Indoor Mobile Robot Navigation and Obstacle Avoidance Using a 3D Camera and Laser Scanner AARMS Vol. 15, No. 1 (2016) 51 59. Indoor Mobile Robot Navigation and Obstacle Avoidance Using a 3D Camera and Laser Scanner Peter KUCSERA 1 Thanks to the developing sensor technology in mobile robot navigation

More information

Zürich. Roland Siegwart Margarita Chli Martin Rufli Davide Scaramuzza. ETH Master Course: L Autonomous Mobile Robots Summary

Zürich. Roland Siegwart Margarita Chli Martin Rufli Davide Scaramuzza. ETH Master Course: L Autonomous Mobile Robots Summary Roland Siegwart Margarita Chli Martin Rufli Davide Scaramuzza ETH Master Course: 151-0854-00L Autonomous Mobile Robots Summary 2 Lecture Overview Mobile Robot Control Scheme knowledge, data base mission

More information

Vision Based Parking Space Classification

Vision Based Parking Space Classification 1 Vision Based Parking Space Classification Ananth Nallamuthu, Sandeep Lokala, Department of ECE, Clemson University. Abstract The problem of Vacant Parking space detection from static images using computer

More information

Building Reliable 2D Maps from 3D Features

Building Reliable 2D Maps from 3D Features Building Reliable 2D Maps from 3D Features Dipl. Technoinform. Jens Wettach, Prof. Dr. rer. nat. Karsten Berns TU Kaiserslautern; Robotics Research Lab 1, Geb. 48; Gottlieb-Daimler- Str.1; 67663 Kaiserslautern;

More information

Ch 22 Inspection Technologies

Ch 22 Inspection Technologies Ch 22 Inspection Technologies Sections: 1. Inspection Metrology 2. Contact vs. Noncontact Inspection Techniques 3. Conventional Measuring and Gaging Techniques 4. Coordinate Measuring Machines 5. Surface

More information

Performance Evaluation of Monitoring System Using IP Camera Networks

Performance Evaluation of Monitoring System Using IP Camera Networks 1077 Performance Evaluation of Monitoring System Using IP Camera Networks Maysoon Hashim Ismiaal Department of electronic and communications, faculty of engineering, university of kufa Abstract Today,

More information

Autonomous Robot Navigation: Using Multiple Semi-supervised Models for Obstacle Detection

Autonomous Robot Navigation: Using Multiple Semi-supervised Models for Obstacle Detection Autonomous Robot Navigation: Using Multiple Semi-supervised Models for Obstacle Detection Adam Bates University of Colorado at Boulder Abstract: This paper proposes a novel approach to efficiently creating

More information

ROBOT TEAMS CH 12. Experiments with Cooperative Aerial-Ground Robots

ROBOT TEAMS CH 12. Experiments with Cooperative Aerial-Ground Robots ROBOT TEAMS CH 12 Experiments with Cooperative Aerial-Ground Robots Gaurav S. Sukhatme, James F. Montgomery, and Richard T. Vaughan Speaker: Jeff Barnett Paper Focus Heterogeneous Teams for Surveillance

More information

A Vision System for Automatic State Determination of Grid Based Board Games

A Vision System for Automatic State Determination of Grid Based Board Games A Vision System for Automatic State Determination of Grid Based Board Games Michael Bryson Computer Science and Engineering, University of South Carolina, 29208 Abstract. Numerous programs have been written

More information

CSc Topics in Computer Graphics 3D Photography

CSc Topics in Computer Graphics 3D Photography CSc 83010 Topics in Computer Graphics 3D Photography Tuesdays 11:45-1:45 1:45 Room 3305 Ioannis Stamos istamos@hunter.cuny.edu Office: 1090F, Hunter North (Entrance at 69 th bw/ / Park and Lexington Avenues)

More information

Overview. EECS 124, UC Berkeley, Spring 2008 Lecture 23: Localization and Mapping. Statistical Models

Overview. EECS 124, UC Berkeley, Spring 2008 Lecture 23: Localization and Mapping. Statistical Models Introduction ti to Embedded dsystems EECS 124, UC Berkeley, Spring 2008 Lecture 23: Localization and Mapping Gabe Hoffmann Ph.D. Candidate, Aero/Astro Engineering Stanford University Statistical Models

More information

ME 597/747 Autonomous Mobile Robots. Mid Term Exam. Duration: 2 hour Total Marks: 100

ME 597/747 Autonomous Mobile Robots. Mid Term Exam. Duration: 2 hour Total Marks: 100 ME 597/747 Autonomous Mobile Robots Mid Term Exam Duration: 2 hour Total Marks: 100 Instructions: Read the exam carefully before starting. Equations are at the back, but they are NOT necessarily valid

More information

Vision-Based Navigation Solution for Autonomous Indoor Obstacle Avoidance Flight

Vision-Based Navigation Solution for Autonomous Indoor Obstacle Avoidance Flight Vision-Based Navigation Solution for Autonomous Indoor Obstacle Avoidance Flight Kirill E. Shilov 1, Vladimir V. Afanasyev 2 and Pavel A. Samsonov 3 1 Moscow Institute of Physics and Technology (MIPT),

More information

Field-of-view dependent registration of point clouds and incremental segmentation of table-tops using time-offlight

Field-of-view dependent registration of point clouds and incremental segmentation of table-tops using time-offlight Field-of-view dependent registration of point clouds and incremental segmentation of table-tops using time-offlight cameras Dipl.-Ing. Georg Arbeiter Fraunhofer Institute for Manufacturing Engineering

More information

Real-time Door Detection based on AdaBoost learning algorithm

Real-time Door Detection based on AdaBoost learning algorithm Real-time Door Detection based on AdaBoost learning algorithm Jens Hensler, Michael Blaich, and Oliver Bittel University of Applied Sciences Konstanz, Germany Laboratory for Mobile Robots Brauneggerstr.

More information

Watchmaker precision for robotic placement of automobile body parts

Watchmaker precision for robotic placement of automobile body parts FlexPlace Watchmaker precision for robotic placement of automobile body parts Staff Report ABB s commitment to adding value for customers includes a constant quest for innovation and improvement new ideas,

More information

Southern Illinois University Edwardsville

Southern Illinois University Edwardsville Southern Illinois University Edwardsville 2014-2015 I certify that the design and engineering of the vehicle Roadrunner by the SIUE Team Roadrunner has been significant and equivalent to what might be

More information

Matching Evaluation of 2D Laser Scan Points using Observed Probability in Unstable Measurement Environment

Matching Evaluation of 2D Laser Scan Points using Observed Probability in Unstable Measurement Environment Matching Evaluation of D Laser Scan Points using Observed Probability in Unstable Measurement Environment Taichi Yamada, and Akihisa Ohya Abstract In the real environment such as urban areas sidewalk,

More information

Terrain Roughness Identification for High-Speed UGVs

Terrain Roughness Identification for High-Speed UGVs Proceedings of the International Conference of Control, Dynamic Systems, and Robotics Ottawa, Ontario, Canada, May 15-16 2014 Paper No. 11 Terrain Roughness Identification for High-Speed UGVs Graeme N.

More information

Active2012 HOSEI UNIVERSITY

Active2012 HOSEI UNIVERSITY Active2012 HOSEI UNIVERSITY Faculty of Science and Engineering, Hosei University 3-7-2 Kajinocho Koganei, Tokyo 194-8584, Japan E-mail; ikko@hosei.ac.jp Faculty Advisor Statement I hereby certify that

More information

The NAO Robot, a case of study Robotics Franchi Alessio Mauro

The NAO Robot, a case of study Robotics Franchi Alessio Mauro The NAO Robot, a case of study Robotics 2013-2014 Franchi Alessio Mauro alessiomauro.franchi@polimi.it Who am I? Franchi Alessio Mauro Master Degree in Computer Science Engineer at Politecnico of Milan

More information

Scan-point Planning and 3-D Map Building for a 3-D Laser Range Scanner in an Outdoor Environment

Scan-point Planning and 3-D Map Building for a 3-D Laser Range Scanner in an Outdoor Environment Scan-point Planning and 3-D Map Building for a 3-D Laser Range Scanner in an Outdoor Environment Keiji NAGATANI 1, Takayuki Matsuzawa 1, and Kazuya Yoshida 1 Tohoku University Summary. During search missions

More information

Stochastic Road Shape Estimation, B. Southall & C. Taylor. Review by: Christopher Rasmussen

Stochastic Road Shape Estimation, B. Southall & C. Taylor. Review by: Christopher Rasmussen Stochastic Road Shape Estimation, B. Southall & C. Taylor Review by: Christopher Rasmussen September 26, 2002 Announcements Readings for next Tuesday: Chapter 14-14.4, 22-22.5 in Forsyth & Ponce Main Contributions

More information

Efficient SLAM Scheme Based ICP Matching Algorithm Using Image and Laser Scan Information

Efficient SLAM Scheme Based ICP Matching Algorithm Using Image and Laser Scan Information Proceedings of the World Congress on Electrical Engineering and Computer Systems and Science (EECSS 2015) Barcelona, Spain July 13-14, 2015 Paper No. 335 Efficient SLAM Scheme Based ICP Matching Algorithm

More information

A Modular Software Framework for Eye-Hand Coordination in Humanoid Robots

A Modular Software Framework for Eye-Hand Coordination in Humanoid Robots A Modular Software Framework for Eye-Hand Coordination in Humanoid Robots Jurgen Leitner, Simon Harding, Alexander Forster and Peter Corke Presentation: Hana Fusman Introduction/ Overview The goal of their

More information

Final Report. EEL 5666 Intelligent Machines Design Laboratory

Final Report. EEL 5666 Intelligent Machines Design Laboratory Final Report EEL 5666 Intelligent Machines Design Laboratory TAs: Mike Pridgen & Thomas Vermeer Instructors: Dr. A. Antonio Arroyo & Dr. Eric M. Schwartz Hao (Hardy) He Dec 08 th, 2009 Table of Contents

More information

3D Convolutional Neural Networks for Landing Zone Detection from LiDAR

3D Convolutional Neural Networks for Landing Zone Detection from LiDAR 3D Convolutional Neural Networks for Landing Zone Detection from LiDAR Daniel Mataruna and Sebastian Scherer Presented by: Sabin Kafle Outline Introduction Preliminaries Approach Volumetric Density Mapping

More information

Image Fusion of Video Images and Geo-localization for UAV Applications K.Senthil Kumar 1, Kavitha.G 2, Subramanian.

Image Fusion of Video Images and Geo-localization for UAV Applications K.Senthil Kumar 1, Kavitha.G 2, Subramanian. Image Fusion of Video Images and Geo-localization for UAV Applications K.Senthil Kumar 1, Kavitha.G 2, Subramanian.R 3 and Marwan 4 Abstract We present in this paper a very fine method for determining

More information

Indoor-Outdoor Navigation System for Visually-Impaired Pedestrians: Preliminary Evaluation of Position Measurement and Obstacle Display

Indoor-Outdoor Navigation System for Visually-Impaired Pedestrians: Preliminary Evaluation of Position Measurement and Obstacle Display Indoor-Outdoor Navigation System for Visually-Impaired Pedestrians: Preliminary Evaluation of Position Measurement and Obstacle Display Takeshi KURATA 12, Masakatsu KOUROGI 1, Tomoya ISHIKAWA 1, Yoshinari

More information