1998 IEEE International Conference on Intelligent Vehicles 213
|
|
- Amber Lewis
- 5 years ago
- Views:
Transcription
1 Navigation by Integrating Iconic and GPS Information Shigang Li and Akira Hayashi Faculty of Information Sciences Hiroshima City University Asaminami-ku, Hiroshima, , Japan Abstract This paper proposes a method ofrobot navigation in outdoor environments by using GPS (Global Postioning System) information and panoramic views. Since an outdoor environment is very large scaled, it is very dicult to determine the global position based ona visual sensor. Although a global position can be obtained from a GPS navigator, it is dicult to locate a robot accurately since it has a distance error about meter. Our system is equipped with a GPS navigator and a camera. A coarse-to-ne method isused to generate an outdoor environment map and locate a mobile robot. First, a robot nds its approximate position based on the GPS information. Then, it identies its location from the image information. Experimental results in outdoor environments show the eectiveness of this method. 1 Introduction Making a map of environments and navigation by using the acquired map is an important problem for a mobile robot. Although many researches have been done so far, not so many researches cope with outdoor environments in contrast with indoor environments [6] [2] [5]. The scale of an outdoor environment isvery large. The data measured from a camera sensor is very noisy. Therefore, the qualitative methods by using landmarks are proposed. For example, viewpoint position is determined given relative visual angles between landmarks [7]; a distinct landmark which can be identied from very dierent points of view is used for recognition [1]. However, how to select landmarks from scenes is a problem. To cope with this problem, [8] proposes a datadriven method to describe route scenes by using panoramic views when a robot moves along a route. Further, the landmark with distinctive scene features can be selected automatically based upon the range and iconic cues of segmented objects along route [3], and a robot can recognize the route when it moves along the same route again by matching the memorized objects with incoming new scenes [4]. However, if an outdoor environment is very large and includes some similar scenes, it is dicult to identify its location in the environment only by using the visual cues. Even a human being may be lost in this situation. How a robot knows that it returns to the starting point from a visual sensor is a dicult problem, especially in a very-large-scale outdoor environment. On the other hand, the global position can be obtained simply from a GPS (Global Positioning System) navigator. This paper proposes a method of robot navigation in outdoor environments by using GPS (Global Postioning System) information and panoramic views. Since an outdoor environment is very large scaled, it is very dicult to determine the global position based on a visual sensor. Although a global position can be obtained from GPS information, it is dicult to locate a robot position accurately since it has a distance error about meter. Our system is equipped with a GPS navigator and a camera. A coarse-to-ne method is used to generate an outdoor environment map and locate a mobile robot. First, a robot nds its approximate position based on the GPS information. Then, it identies its location from the image information. An experimental result in outdoor environments shows the eectiveness of this method. 2 Our system Our system consists of a car, a GPS navigator and a camera. The experimental environment is the campus of Hiroshima City University which is shown in Figure. 1. The contours enclosing the area is the route used in our experiment. A human drives the car along the route in the direction of counter clockwise. Mean IEEE International Conference on Intelligent Vehicles 213
2 while the camera is mounted on the left side of the car and its optical axis is orthogonal with the car moving direction. Thus, we can acquire the GPS information and image sequences simultaneously when the car moves along a route. Height[m] E132 º 25.4 N34 º 26.4 N34 º 26.2 Latitude N34 º 26.0 N34 º25.8 E132 º 25.2 E132 º 25.0 Longitude E132 º 24.8 Figure 2: The 3D map of the route plotted by using the data of latitude, longitude and height. 2.2 Panoramic views Figure 1: The top view of a part of the campus of Hiroshima City University, where the contours enclosing the area is the route used in our experiment. Suppose a camera moving along a smooth path with its optical axis orthogonal to the moving direction. If we set a vertical slit with a single pixel width in an image and arranges 1D images through the slit along the time axis, we get a wide view of environment called a panoramic view [3]. Figure. 3 shows the panoramic view taken along the route, in which intersections are enclosed by rectangles. 2.1 GPS navigator The GPS system is fully operated and maintained by U.S. Department of Defence. The GPS navigator receives signals from a group of satellites in space to show the position anywhere in the world, 24 hours a day. So, it is very convenient to use it for navigation. The HG-7 type of GPS navigator made by Marine Electronics, Inc. is used in our experiment. The latitude, longitude, height and speed at a position can be acquired from the GPS navigator. Figure. 2 shows the 3D map of the route in Figure. 1, which is plotted by using the data of latitude, longitude and height. Since the GPS navigator has a distance error about meter, the 3D curve in Figure. 2 is not closed although the car returns to the starting point. Thus, it is dicult to use a GPS navigator to localize our car accurately. For example, it may be that you really have moved through an intersection, but from the GPS navigator data the intersection seems forward; as a result, you may fail to make a turn at the intersection. To cope with such problems, we use visual information taken from a camera. 3 Making environment map by fusing GPS information and panoramic views When the car runs along a route, we can acquire the GPS information and panoramic views simultaneously. We synchronize the two devices (the GPS navigaotr and the camera) according to time. By plotting the panoramic view along the route in terms of GPS information, we can obtain Figure. 4. Figure. 5 shows the view segments which is zoomed up at the starting point. Due to the position error from the GPS navigator, the same starting point is plotted at dierent positions. However, as shown in Figure. 5, the same scene is viewed again if the car returns to the starting point. Thus, we could correct the position error according to the iconic information of panoramic views. Our method is as follows. The car runs along a route and acquires the position information from a GPS navigator and a panoramic view from a camera simultaneously. When the car arrives at a position which is near to that passed before within the distance error of the GPS device, the incoming scene is compared with the memorized panoramic view IEEE International Conference on Intelligent Vehicles 214
3 Figure 3: The panoramic view taken along the route IEEE International Conference on Intelligent Vehicles 215
4 The Dynamic programming method is used to match the incoming scenes with the memorized panoramic view [4]. If the score of matching is high enough, we can know that the car return to the same position and fuse them to obtain a closed route. If the matching is failed, we can know that the car is runing another route which is near enough to that tranversed before. Here, we give more detailed explaination on the dynamic programming method. The objective is to match the memorized panoramic view segment with the incoming scenes. Instead of matching the whole memorized view segment, we sample the view segment and obtain a sequence of slit images, fs1;s2; :::; S j ; :::; S n g, which is shown on the lower left in Figure. 6. The upper right image in Figure. 6 is the new panoramic view taken from incoming scenes. We Figure 5: The view zoomed up at the starting point. compute of the similarity, C(S j ;P m;j ), between the slit image, S j, of the memorized view segment and the slit image, P m;j, of the new panoramic view by SSD method. Where n is the number of the sequence of slit images. P m;j is the slit image from the new incoming C(S j ;P m;j )= panoramic which is within the error range of the GPS Xq01 f( r 0 i;sj ri;pm;j T i=0 i;s j 0 ) 2 +( g 0 i;sj gi;pm;j Ti;pm;j 0 T i;s j 0 ) 2 +( b 0 i;sj bi;pm;j Ti;pm;j 0 T i;s j 0 ) 2 navigator. The bigger the error range of the GPS navigator is, the bigger m is. We carry it out by the dy- Ti;pm;j 0 g namic programming method [4]. Figure. 6 shows the Where, r i;s j, g i;s j, b i;s j are the R, G, B values of ith pixel at S j, respectively; ri;pm;j 0, g0 i;pm;j and b0 i;pm;j are matching result of memorized slit images with the incoming panoramic view when the car returns to the the R, G, B values of ith pixel at p m;j, respectively; T i;s j = r i;s j + g i;s j + b i;s j ;T0 i;pm;j = r0 i;pm;j + g0 i;pm;j + starting point. The matching result is shown in the lower right. b 0 i;pm;j ; q is the pixel number of the slit image. The closed route found after fusing GPS information and panoramic views is shown in Figure. 7. Fig- If the sequence of slit images, S1;S2; :::; S j ; :::; S n, appear in the new incoming panoramic with the same ure. 8 shows the view segments which is zoomed up order, the matching succeeds. We realize the matching at the starting point. We distribute the position error by minimizing the following evaluation function. along the route. nx We can use this map for car navigation. min f C(S j ;P m;j )g fpm;jg j=1 Figure 4: Plotting the panoramic view along the route in term of GPS information. Figure 7: The closed route found after fusing GPS information and panoramic views IEEE International Conference on Intelligent Vehicles 216
5 left in Figure. 10. Here the same dynamic programming method with Figure. 6 is used. Thus, the car can clear its position error accumulated so far when it identies such a location, and is expected to arrive its destination correctly. 5 Conclusions Figure 8: The view zoomed up at the starting point after the fusion. 4 Navigation based upon the built map Now we consider how a car makes its navigation based upon the built map. As a concrete task, let the car nd a given intersection. Figure. 9 shows the GPS information and the view segment at intersetion 7 in Figure. 3. GPS N W H283[m] Figure 9: The GPS information and the view segment at intersetion 7. Instead of matching the whole view segment with the incoming scene, we sample the view segment and obtain a sequence of slit images which is shown on the This paper proposes a method of robot (car) navigation in outdoor environments by using GPS (Global Postioning System) information and panoramic views. A coarse-to-ne method is used to generate an outdoor environment map and locate a mobile robot. First, a robot nds its approximate position based on the GPS information. Then, it identies its destination from the image information. An experimental result in outdoor environments shows the eectiveness of this method. References [1] D. Dai and D. T. Lawton, \Range-free qualitative navigation," Proc. of IEEE Int. Conf. on Robotics & Automation, pp , [2] J. Horn and G. Schmidt, \Continuous localization for long-range indoor navigation of mobile robots," Proc. of IEEE Int. Conf. on Robotics & Automation, pp , [3] S. Li and S. Tsuji, \Selecting distinctive scene features for landmarks", Proc. IEEE Int. Conf. Robotics & Automation, pp , [4] S. Li, S. Tsuji and A. Hayashi, \Qualitative Representation of Outdoor Environment along Route", Proc. ICPR, Vol.1, pp , [5] S. Li, J. Nagata and S. Tsuji, \A navigation system based upon panoramic representation," Proc. of IROS, Vol. 17, pp , [6] M. J. Mataric, \Integration of representation into goal-driven behavior-based robot," IEEE Tran. on Robotics & Automation, Vol. 8, No. 3, pp , [7] K. T. Sutherland and W. B. Thompson, \Inexact navigation," Proc. of IEEE Int. Conf. on Robotics & Automation, pp. 1-7, [8] J. Y. Zheng and S. Tsuji, \Panoramic representation for route recognition by a mobile robot", Int. J. Computer Vision,, 9:(1), pp , IEEE International Conference on Intelligent Vehicles 217
6 Figure 6: The matching result of memorized panoramic view with the incoming slit images when the car returns to the starting point. Figure 10: The matching result of a given intersection with the incoming scenes IEEE International Conference on Intelligent Vehicles 218
Image-Based Memory of Environment. homing uses a similar idea that the agent memorizes. [Hong 91]. However, the agent nds diculties in arranging its
Image-Based Memory of Environment Hiroshi ISHIGURO Department of Information Science Kyoto University Kyoto 606-01, Japan E-mail: ishiguro@kuis.kyoto-u.ac.jp Saburo TSUJI Faculty of Systems Engineering
More informationVISUAL NAVIGATION SYSTEM ON WINDSHIELD HEAD-UP DISPLAY
VISUAL NAVIGATION SYSTEM ON WINDSHIELD HEAD-UP DISPLAY Akihiko SATO *, Itaru KITAHARA, Yoshinari KAMEDA, Yuichi OHTA Department of Intelligent Interaction Technologies, Graduate School of Systems and Information
More informationProc. 14th Int. Conf. on Intelligent Autonomous Systems (IAS-14), 2016
Proc. 14th Int. Conf. on Intelligent Autonomous Systems (IAS-14), 2016 Outdoor Robot Navigation Based on View-based Global Localization and Local Navigation Yohei Inoue, Jun Miura, and Shuji Oishi Department
More informationMeasurement of Pedestrian Groups Using Subtraction Stereo
Measurement of Pedestrian Groups Using Subtraction Stereo Kenji Terabayashi, Yuki Hashimoto, and Kazunori Umeda Chuo University / CREST, JST, 1-13-27 Kasuga, Bunkyo-ku, Tokyo 112-8551, Japan terabayashi@mech.chuo-u.ac.jp
More informationRealtime Omnidirectional Stereo for Obstacle Detection and Tracking in Dynamic Environments
Proc. 2001 IEEE/RSJ Int. Conf. on Intelligent Robots and Systems pp. 31-36, Maui, Hawaii, Oct./Nov. 2001. Realtime Omnidirectional Stereo for Obstacle Detection and Tracking in Dynamic Environments Hiroshi
More informationIntelligent Outdoor Navigation of a Mobile Robot Platform Using a Low Cost High Precision RTK-GPS and Obstacle Avoidance System
Intelligent Outdoor Navigation of a Mobile Robot Platform Using a Low Cost High Precision RTK-GPS and Obstacle Avoidance System Under supervision of: Prof. Dr. -Ing. Klaus-Dieter Kuhnert Dipl.-Inform.
More informationAcquisition of Qualitative Spatial Representation by Visual Observation
Acquisition of Qualitative Spatial Representation by Visual Observation Takushi Sogo Hiroshi Ishiguro Toru Ishida Department of Social Informatics, Kyoto University Kyoto 606-8501, Japan sogo@kuis.kyoto-u.ac.jp,
More informationOptical Flow-Based Person Tracking by Multiple Cameras
Proc. IEEE Int. Conf. on Multisensor Fusion and Integration in Intelligent Systems, Baden-Baden, Germany, Aug. 2001. Optical Flow-Based Person Tracking by Multiple Cameras Hideki Tsutsui, Jun Miura, and
More informationVISUAL NAVIGATION SYSTEM ON WINDSHIELD HEAD-UP DISPLAY
VISUAL NAVIGATION SYSTEM ON WINDSHIELD HEAD-UP DISPLAY Akihiko SATO *, Itaru KITAHARA, Yoshinari KAMEDA, Yuichi OHTA Department of Intelligent Interaction Technologies, Graduate School of Systems and Information
More informationDevelopment of Low-Cost Compact Omnidirectional Vision Sensors and their applications
Development of Low-Cost Compact Omnidirectional Vision Sensors and their applications Hiroshi ISHIGURO Department of Electrical & Computer Engineering, University of California, San Diego (9500 Gilman
More informationBehavior Learning for a Mobile Robot with Omnidirectional Vision Enhanced by an Active Zoom Mechanism
Behavior Learning for a Mobile Robot with Omnidirectional Vision Enhanced by an Active Zoom Mechanism Sho ji Suzuki, Tatsunori Kato, Minoru Asada, and Koh Hosoda Dept. of Adaptive Machine Systems, Graduate
More informationDETECTION OF 3D POINTS ON MOVING OBJECTS FROM POINT CLOUD DATA FOR 3D MODELING OF OUTDOOR ENVIRONMENTS
DETECTION OF 3D POINTS ON MOVING OBJECTS FROM POINT CLOUD DATA FOR 3D MODELING OF OUTDOOR ENVIRONMENTS Tsunetake Kanatani,, Hideyuki Kume, Takafumi Taketomi, Tomokazu Sato and Naokazu Yokoya Hyogo Prefectural
More informationGPS mini Watch User Manual Introduction. Getting Started. Caution: Step 1) Know your Watch:
Watch User Manual Introduction Thank you for purchasing the GPS Watch. This GPS Watch is packed with personal Training features like speed, trip time, laps, etc. Watch features include but not limited
More informationDominant plane detection using optical flow and Independent Component Analysis
Dominant plane detection using optical flow and Independent Component Analysis Naoya OHNISHI 1 and Atsushi IMIYA 2 1 School of Science and Technology, Chiba University, Japan Yayoicho 1-33, Inage-ku, 263-8522,
More informationEpipolar geometry-based ego-localization using an in-vehicle monocular camera
Epipolar geometry-based ego-localization using an in-vehicle monocular camera Haruya Kyutoku 1, Yasutomo Kawanishi 1, Daisuke Deguchi 1, Ichiro Ide 1, Hiroshi Murase 1 1 : Nagoya University, Japan E-mail:
More informationDepartment of Electrical Engineering, Keio University Hiyoshi Kouhoku-ku Yokohama 223, Japan
Shape Modeling from Multiple View Images Using GAs Satoshi KIRIHARA and Hideo SAITO Department of Electrical Engineering, Keio University 3-14-1 Hiyoshi Kouhoku-ku Yokohama 223, Japan TEL +81-45-563-1141
More informationAR Cultural Heritage Reconstruction Based on Feature Landmark Database Constructed by Using Omnidirectional Range Sensor
AR Cultural Heritage Reconstruction Based on Feature Landmark Database Constructed by Using Omnidirectional Range Sensor Takafumi Taketomi, Tomokazu Sato, and Naokazu Yokoya Graduate School of Information
More informationGenerating Dynamic Projection Images for Scene Representation and Understanding
COMPUTER VISION AND IMAGE UNDERSTANDING Vol. 72, No. 3, December, pp. 237 256, 1998 ARTICLE NO. IV980678 Generating Dynamic Projection Images for Scene Representation and Understanding Jiang Yu Zheng and
More informationURBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES
URBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES An Undergraduate Research Scholars Thesis by RUI LIU Submitted to Honors and Undergraduate Research Texas A&M University in partial fulfillment
More informationFOR ROUTE UNDERSTANDING
PANORAMIC REPRESENTATION OF SCENES FOR ROUTE UNDERSTANDING Abstract Jiang Yu Zheng and Saburo Tsuji Department of Control Engineering, Osaka University Toyonaka, Osaka 560, Japan This work tackles route
More informationRobot Localization based on Geo-referenced Images and G raphic Methods
Robot Localization based on Geo-referenced Images and G raphic Methods Sid Ahmed Berrabah Mechanical Department, Royal Military School, Belgium, sidahmed.berrabah@rma.ac.be Janusz Bedkowski, Łukasz Lubasiński,
More informationCamera Calibration for a Robust Omni-directional Photogrammetry System
Camera Calibration for a Robust Omni-directional Photogrammetry System Fuad Khan 1, Michael Chapman 2, Jonathan Li 3 1 Immersive Media Corporation Calgary, Alberta, Canada 2 Ryerson University Toronto,
More informationDEVELOPMENT OF POSITION MEASUREMENT SYSTEM FOR CONSTRUCTION PILE USING LASER RANGE FINDER
S17- DEVELOPMENT OF POSITION MEASUREMENT SYSTEM FOR CONSTRUCTION PILE USING LASER RANGE FINDER Fumihiro Inoue 1 *, Takeshi Sasaki, Xiangqi Huang 3, and Hideki Hashimoto 4 1 Technica Research Institute,
More informationSecurity Management System Object Counting
Security Management System Object Counting This document describes the steps for using the Object Counting module available in the Security Management System software. (A) Object Counting Configuration
More informationRecognizing Buildings in Urban Scene of Distant View ABSTRACT
Recognizing Buildings in Urban Scene of Distant View Peilin Liu, Katsushi Ikeuchi and Masao Sakauchi Institute of Industrial Science, University of Tokyo, Japan 7-22-1 Roppongi, Minato-ku, Tokyo 106, Japan
More informationHow to setup an accurate perimeter guarding system with DeepinMind NVR HIKVISION TECHNICAL SUPPORT TEAM
How to setup an accurate perimeter guarding system with DeepinMind NVR HIKVISION TECHNICAL SUPPORT TEAM 2017-11-13 1 ids-9632nxi-i8/16s How to setup an accurate perimeter guarding system with DeepinMind
More information3D Tracking Using Two High-Speed Vision Systems
3D Tracking Using Two High-Speed Vision Systems Yoshihiro NAKABO 1, Idaku ISHII 2, Masatoshi ISHIKAWA 3 1 University of Tokyo, Tokyo, Japan, nakabo@k2.t.u-tokyo.ac.jp 2 Tokyo University of Agriculture
More information07ATC 22 Image Exploitation System for Airborne Surveillance
07ATC 22 Image Exploitation System for Airborne Surveillance Karunakaran. P, Frederick Mathews, S.H.Padmanabhan, T.K.Sateesh HCL Technologies, India Copyright 2007 SAE International ABSTRACT The operational
More informationMOTION. Feature Matching/Tracking. Control Signal Generation REFERENCE IMAGE
Head-Eye Coordination: A Closed-Form Solution M. Xie School of Mechanical & Production Engineering Nanyang Technological University, Singapore 639798 Email: mmxie@ntuix.ntu.ac.sg ABSTRACT In this paper,
More information3D Reconstruction of Outdoor Environments from Omnidirectional Range and Color Images
3D Reconstruction of Outdoor Environments from Omnidirectional Range and Color Images Toshihiro ASAI, Masayuki KANBARA, Naokazu YOKOYA Graduate School of Information Science, Nara Institute of Science
More informationReal-time target tracking using a Pan and Tilt platform
Real-time target tracking using a Pan and Tilt platform Moulay A. Akhloufi Abstract In recent years, we see an increase of interest for efficient tracking systems in surveillance applications. Many of
More informationSegmentation and Tracking of Partial Planar Templates
Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract
More informationAN APPROACH OF SEMIAUTOMATED ROAD EXTRACTION FROM AERIAL IMAGE BASED ON TEMPLATE MATCHING AND NEURAL NETWORK
AN APPROACH OF SEMIAUTOMATED ROAD EXTRACTION FROM AERIAL IMAGE BASED ON TEMPLATE MATCHING AND NEURAL NETWORK Xiangyun HU, Zuxun ZHANG, Jianqing ZHANG Wuhan Technique University of Surveying and Mapping,
More informationVol. 21 No. 6, pp ,
Vol. 21 No. 6, pp.69 696, 23 69 3 3 3 Map Generation of a Mobile Robot by Integrating Omnidirectional Stereo and Laser Range Finder Yoshiro Negishi 3, Jun Miura 3 and Yoshiaki Shirai 3 This paper describes
More information2 ATTILA FAZEKAS The tracking model of the robot car The schematic picture of the robot car can be seen on Fig.1. Figure 1. The main controlling task
NEW OPTICAL TRACKING METHODS FOR ROBOT CARS Attila Fazekas Debrecen Abstract. In this paper new methods are proposed for intelligent optical tracking of robot cars the important tools of CIM (Computer
More informationDevelopment of a video-rate range finder using dynamic threshold method for characteristic point detection
Engineering Industrial & Management Engineering fields Year 1999 Development of a video-rate range finder using dynamic threshold method for characteristic point detection Yutaka Tanaka Nobuo Takeda Akio
More informationAn Angle Estimation to Landmarks for Autonomous Satellite Navigation
5th International Conference on Environment, Materials, Chemistry and Power Electronics (EMCPE 2016) An Angle Estimation to Landmarks for Autonomous Satellite Navigation Qing XUE a, Hongwen YANG, Jian
More informationEmploying a Fish-Eye for Scene Tunnel Scanning
Employing a Fish-Eye for Scene Tunnel Scanning Jiang Yu Zheng 1, Shigang Li 2 1 Dept. of Computer Science, Indiana University Purdue University Indianapolis, 723 W. Michigan St. Indianapolis, IN46202,
More informationCS4670: Computer Vision
CS4670: Computer Vision Noah Snavely Lecture 6: Feature matching and alignment Szeliski: Chapter 6.1 Reading Last time: Corners and blobs Scale-space blob detector: Example Feature descriptors We know
More informationDept. of Adaptive Machine Systems, Graduate School of Engineering Osaka University, Suita, Osaka , Japan
An Application of Vision-Based Learning for a Real Robot in RoboCup - A Goal Keeping Behavior for a Robot with an Omnidirectional Vision and an Embedded Servoing - Sho ji Suzuki 1, Tatsunori Kato 1, Hiroshi
More informationRobot localization method based on visual features and their geometric relationship
, pp.46-50 http://dx.doi.org/10.14257/astl.2015.85.11 Robot localization method based on visual features and their geometric relationship Sangyun Lee 1, Changkyung Eem 2, and Hyunki Hong 3 1 Department
More informationMobile Robot Navigation Using Omnidirectional Vision
Mobile Robot Navigation Using Omnidirectional Vision László Mornailla, Tamás Gábor Pekár, Csaba Gergő Solymosi, Zoltán Vámossy John von Neumann Faculty of Informatics, Budapest Tech Bécsi út 96/B, H-1034
More informationRecovering light directions and camera poses from a single sphere.
Title Recovering light directions and camera poses from a single sphere Author(s) Wong, KYK; Schnieders, D; Li, S Citation The 10th European Conference on Computer Vision (ECCV 2008), Marseille, France,
More informationA 100Hz Real-time Sensing System of Textured Range Images
A 100Hz Real-time Sensing System of Textured Range Images Hidetoshi Ishiyama Course of Precision Engineering School of Science and Engineering Chuo University 1-13-27 Kasuga, Bunkyo-ku, Tokyo 112-8551,
More informationGesture Recognition using Temporal Templates with disparity information
8- MVA7 IAPR Conference on Machine Vision Applications, May 6-8, 7, Tokyo, JAPAN Gesture Recognition using Temporal Templates with disparity information Kazunori Onoguchi and Masaaki Sato Hirosaki University
More informationLocalization of Wearable Users Using Invisible Retro-reflective Markers and an IR Camera
Localization of Wearable Users Using Invisible Retro-reflective Markers and an IR Camera Yusuke Nakazato, Masayuki Kanbara and Naokazu Yokoya Graduate School of Information Science, Nara Institute of Science
More informationCamera Parameters Estimation from Hand-labelled Sun Sositions in Image Sequences
Camera Parameters Estimation from Hand-labelled Sun Sositions in Image Sequences Jean-François Lalonde, Srinivasa G. Narasimhan and Alexei A. Efros {jlalonde,srinivas,efros}@cs.cmu.edu CMU-RI-TR-8-32 July
More informationHand-Eye Calibration from Image Derivatives
Hand-Eye Calibration from Image Derivatives Abstract In this paper it is shown how to perform hand-eye calibration using only the normal flow field and knowledge about the motion of the hand. The proposed
More informationReal-time Generation and Presentation of View-dependent Binocular Stereo Images Using a Sequence of Omnidirectional Images
Real-time Generation and Presentation of View-dependent Binocular Stereo Images Using a Sequence of Omnidirectional Images Abstract This paper presents a new method to generate and present arbitrarily
More informationAdaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision
Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision Zhiyan Zhang 1, Wei Qian 1, Lei Pan 1 & Yanjun Li 1 1 University of Shanghai for Science and Technology, China
More informationVision-Motion Planning with Uncertainty
Vision-Motion Planning with Uncertainty Jun MIURA Yoshiaki SHIRAI Dept. of Mech. Eng. for Computer-Controlled Machinery, Osaka University, Suita, Osaka 565, Japan jun@ccm.osaka-u.ac.jp Abstract This paper
More informationVehicle Ego-localization by Matching In-vehicle Camera Images to an Aerial Image
Vehicle Ego-localization by Matching In-vehicle Camera Images to an Aerial Image Masafumi NODA 1,, Tomokazu TAKAHASHI 1,2, Daisuke DEGUCHI 1, Ichiro IDE 1, Hiroshi MURASE 1, Yoshiko KOJIMA 3 and Takashi
More informationEfficient Path Finding Method Based Evaluation Function in Large Scene Online Games and Its Application
Journal of Information Hiding and Multimedia Signal Processing c 2017 ISSN 2073-4212 Ubiquitous International Volume 8, Number 3, May 2017 Efficient Path Finding Method Based Evaluation Function in Large
More informationEstimating Camera Position And Posture by Using Feature Landmark Database
Estimating Camera Position And Posture by Using Feature Landmark Database Motoko Oe 1, Tomokazu Sato 2 and Naokazu Yokoya 2 1 IBM Japan 2 Nara Institute of Science and Technology, Japan Abstract. Estimating
More informationSherlock 7 Technical Resource. Search Geometric
Sherlock 7 Technical Resource DALSA Corp., Industrial Products (IPD) www.goipd.com 978.670.2002 (U.S.A.) Document Revision: September 24, 2007 Search Geometric Search utilities A common task in machine
More informationCollecting outdoor datasets for benchmarking vision based robot localization
Collecting outdoor datasets for benchmarking vision based robot localization Emanuele Frontoni*, Andrea Ascani, Adriano Mancini, Primo Zingaretti Department of Ingegneria Infromatica, Gestionale e dell
More informationCamera Registration in a 3D City Model. Min Ding CS294-6 Final Presentation Dec 13, 2006
Camera Registration in a 3D City Model Min Ding CS294-6 Final Presentation Dec 13, 2006 Goal: Reconstruct 3D city model usable for virtual walk- and fly-throughs Virtual reality Urban planning Simulation
More informationOBSTACLE DETECTION USING STRUCTURED BACKGROUND
OBSTACLE DETECTION USING STRUCTURED BACKGROUND Ghaida Al Zeer, Adnan Abou Nabout and Bernd Tibken Chair of Automatic Control, Faculty of Electrical, Information and Media Engineering University of Wuppertal,
More informationChapter 11 Arc Extraction and Segmentation
Chapter 11 Arc Extraction and Segmentation 11.1 Introduction edge detection: labels each pixel as edge or no edge additional properties of edge: direction, gradient magnitude, contrast edge grouping: edge
More informationA Novel Real-Time Feature Matching Scheme
Sensors & Transducers, Vol. 165, Issue, February 01, pp. 17-11 Sensors & Transducers 01 by IFSA Publishing, S. L. http://www.sensorsportal.com A Novel Real-Time Feature Matching Scheme Ying Liu, * Hongbo
More informationetrex quick start manual for use with models 20 and 30
etrex quick start manual for use with models 20 and 30 Getting Started Device Overview warning See the Important Safety and Product Information guide in the product box for product warnings and other important
More informationDetection of Small-Waving Hand by Distributed Camera System
Detection of Small-Waving Hand by Distributed Camera System Kenji Terabayashi, Hidetsugu Asano, Takeshi Nagayasu, Tatsuya Orimo, Mutsumi Ohta, Takaaki Oiwa, and Kazunori Umeda Department of Mechanical
More informationSummarization of Egocentric Moving Videos for Generating Walking Route Guidance
Summarization of Egocentric Moving Videos for Generating Walking Route Guidance Masaya Okamoto and Keiji Yanai Department of Informatics, The University of Electro-Communications 1-5-1 Chofugaoka, Chofu-shi,
More informationPreceding vehicle detection and distance estimation. lane change, warning system.
Preceding vehicle detection and distance estimation for lane change warning system U. Iqbal, M.S. Sarfraz Computer Vision Research Group (COMVis) Department of Electrical Engineering, COMSATS Institute
More informationAUTOMATIC RECONSTRUCTION OF LARGE-SCALE VIRTUAL ENVIRONMENT FOR INTELLIGENT TRANSPORTATION SYSTEMS SIMULATION
AUTOMATIC RECONSTRUCTION OF LARGE-SCALE VIRTUAL ENVIRONMENT FOR INTELLIGENT TRANSPORTATION SYSTEMS SIMULATION Khairil Azmi, Shintaro Ono, Masataka Kagesawa, Katsushi Ikeuchi Institute of Industrial Science,
More informationMonocular Vision Based Autonomous Navigation for Arbitrarily Shaped Urban Roads
Proceedings of the International Conference on Machine Vision and Machine Learning Prague, Czech Republic, August 14-15, 2014 Paper No. 127 Monocular Vision Based Autonomous Navigation for Arbitrarily
More information(a) (b) (c) Fig. 1. Omnidirectional camera: (a) principle; (b) physical construction; (c) captured. of a local vision system is more challenging than
An Omnidirectional Vision System that finds and tracks color edges and blobs Felix v. Hundelshausen, Sven Behnke, and Raul Rojas Freie Universität Berlin, Institut für Informatik Takustr. 9, 14195 Berlin,
More informationSatGuide Symbian User Manual. SatGuide Symbian OS USER MANUAL
SatGuide Symbian User Manual SatGuide Symbian OS USER MANUAL CONTENTS 1 - INTRODUCTION...1 WHAT IS SATGUIDE?...1 SATGUIDE FEATURES...1 2 - SMART PRACTICES WHILE DRIVING...2 DRIVE SAFE, CALL SMART SM...2
More informationUAV Position and Attitude Sensoring in Indoor Environment Using Cameras
UAV Position and Attitude Sensoring in Indoor Environment Using Cameras 1 Peng Xu Abstract There are great advantages of indoor experiment for UAVs. Test flights of UAV in laboratory is more convenient,
More informationHandheld Augmented Reality. Reto Lindegger
Handheld Augmented Reality Reto Lindegger lreto@ethz.ch 1 AUGMENTED REALITY 2 A Definition Three important characteristics: Combines real and virtual environment Interactive in real-time Registered in
More informationThree-Dimensional Measurement of Objects in Liquid with an Unknown Refractive Index Using Fisheye Stereo Camera
Three-Dimensional Measurement of Objects in Liquid with an Unknown Refractive Index Using Fisheye Stereo Camera Kazuki Sakamoto, Alessandro Moro, Hiromitsu Fujii, Atsushi Yamashita, and Hajime Asama Abstract
More informationBROWN UNIVERSITY Department of Computer Science Master's Project CS-95-M17
BROWN UNIVERSITY Department of Computer Science Master's Project CS-95-M17 "Robotic Object Recognition: Utilizing Multiple Views to Recognize Partially Occluded Objects" by Neil A. Jacobson ,'. Robotic
More informationConstructing a 3D Object Model from Multiple Visual Features
Constructing a 3D Object Model from Multiple Visual Features Jiang Yu Zheng Faculty of Computer Science and Systems Engineering Kyushu Institute of Technology Iizuka, Fukuoka 820, Japan Abstract This work
More informationRecognition of Finger Pointing Direction Using Color Clustering and Image Segmentation
September 14-17, 2013, Nagoya University, Nagoya, Japan Recognition of Finger Pointing Direction Using Color Clustering and Image Segmentation Hidetsugu Asano 1, Takeshi Nagayasu 2, Tatsuya Orimo 1, Kenji
More informationAutomatic 3-D 3 D Model Acquisition from Range Images. Michael K. Reed and Peter K. Allen Computer Science Department Columbia University
Automatic 3-D 3 D Model Acquisition from Range Images Michael K. Reed and Peter K. Allen Computer Science Department Columbia University Introduction Objective: given an arbitrary object or scene, construct
More informationBritish Machine Vision Conference 2 The established approach for automatic model construction begins by taking surface measurements from a number of v
Segmentation of Range Data into Rigid Subsets using Planar Surface Patches A. P. Ashbrook, R. B. Fisher, C. Robertson and N. Wergi Department of Articial Intelligence The University of Edinburgh 5, Forrest
More informationDense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera
Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera Tomokazu Satoy, Masayuki Kanbaray, Naokazu Yokoyay and Haruo Takemuraz ygraduate School of Information
More information3D MODELING OF OUTDOOR SCENES BY INTEGRATING STOP-AND-GO AND CONTINUOUS SCANNING OF RANGEFINDER
3D MODELING OF OUTDOOR SCENES BY INTEGRATING STOP-AND-GO AND CONTINUOUS SCANNING OF RANGEFINDER Toshihiro ASAI, Masayuki KANBARA, Naokazu YOKOYA Graduate School of Information Science, Nara Institute of
More informationLong-term motion estimation from images
Long-term motion estimation from images Dennis Strelow 1 and Sanjiv Singh 2 1 Google, Mountain View, CA, strelow@google.com 2 Carnegie Mellon University, Pittsburgh, PA, ssingh@cmu.edu Summary. Cameras
More informationTeam Aware Perception System using Stereo Vision and Radar
Team Aware Perception System using Stereo Vision and Radar System Development Review 03/ 08 / 2017 Amit Agarwal Harry Golash Yihao Qian Menghan Zhang Zihao (Theo) Zhang Project Description Develop a standalone
More informationAN EFFICIENT BINARY CORNER DETECTOR. P. Saeedi, P. Lawrence and D. Lowe
AN EFFICIENT BINARY CORNER DETECTOR P. Saeedi, P. Lawrence and D. Lowe Department of Electrical and Computer Engineering, Department of Computer Science University of British Columbia Vancouver, BC, V6T
More informationECE 172A: Introduction to Intelligent Systems: Machine Vision, Fall Midterm Examination
ECE 172A: Introduction to Intelligent Systems: Machine Vision, Fall 2008 October 29, 2008 Notes: Midterm Examination This is a closed book and closed notes examination. Please be precise and to the point.
More informationv9 Quick Start Guide
v9 Quick Start Guide overview Driving Screen Most of your time using CoPIlot Truck will be spent on this screen. Let s take a moment and get familiar with the items you can interact with. Zoom Out Zoom
More informationФУНДАМЕНТАЛЬНЫЕ НАУКИ. Информатика 9 ИНФОРМАТИКА MOTION DETECTION IN VIDEO STREAM BASED ON BACKGROUND SUBTRACTION AND TARGET TRACKING
ФУНДАМЕНТАЛЬНЫЕ НАУКИ Информатика 9 ИНФОРМАТИКА UDC 6813 OTION DETECTION IN VIDEO STREA BASED ON BACKGROUND SUBTRACTION AND TARGET TRACKING R BOGUSH, S ALTSEV, N BROVKO, E IHAILOV (Polotsk State University
More informationA Path Tracking Method For Autonomous Mobile Robots Based On Grid Decomposition
A Path Tracking Method For Autonomous Mobile Robots Based On Grid Decomposition A. Pozo-Ruz*, C. Urdiales, A. Bandera, E. J. Pérez and F. Sandoval Dpto. Tecnología Electrónica. E.T.S.I. Telecomunicación,
More informationIntegration of Multiple-baseline Color Stereo Vision with Focus and Defocus Analysis for 3D Shape Measurement
Integration of Multiple-baseline Color Stereo Vision with Focus and Defocus Analysis for 3D Shape Measurement Ta Yuan and Murali Subbarao tyuan@sbee.sunysb.edu and murali@sbee.sunysb.edu Department of
More informationLaser Eye a new 3D sensor for active vision
Laser Eye a new 3D sensor for active vision Piotr Jasiobedzki1, Michael Jenkin2, Evangelos Milios2' Brian Down1, John Tsotsos1, Todd Campbell3 1 Dept. of Computer Science, University of Toronto Toronto,
More information1216 P a g e 2.1 TRANSLATION PARAMETERS ESTIMATION. If f (x, y) F(ξ,η) then. f(x,y)exp[j2π(ξx 0 +ηy 0 )/ N] ) F(ξ- ξ 0,η- η 0 ) and
An Image Stitching System using Featureless Registration and Minimal Blending Ch.Rajesh Kumar *,N.Nikhita *,Santosh Roy *,V.V.S.Murthy ** * (Student Scholar,Department of ECE, K L University, Guntur,AP,India)
More informationScene Text Detection Using Machine Learning Classifiers
601 Scene Text Detection Using Machine Learning Classifiers Nafla C.N. 1, Sneha K. 2, Divya K.P. 3 1 (Department of CSE, RCET, Akkikkvu, Thrissur) 2 (Department of CSE, RCET, Akkikkvu, Thrissur) 3 (Department
More informationLost! Leveraging the Crowd for Probabilistic Visual Self-Localization
Lost! Leveraging the Crowd for Probabilistic Visual Self-Localization Marcus A. Brubaker (Toyota Technological Institute at Chicago) Andreas Geiger (Karlsruhe Institute of Technology & MPI Tübingen) Raquel
More informationPanoramic Vision and LRF Sensor Fusion Based Human Identification and Tracking for Autonomous Luggage Cart
Panoramic Vision and LRF Sensor Fusion Based Human Identification and Tracking for Autonomous Luggage Cart Mehrez Kristou, Akihisa Ohya and Shin ichi Yuta Intelligent Robot Laboratory, University of Tsukuba,
More informationBabyTigers-98: Osaka Legged Robot Team
BabyTigers-98: saka Legged Robot Team Noriaki Mitsunaga and Minoru Asada and Chizuko Mishima Dept. of Adaptive Machine Systems, saka niversity, Suita, saka, 565-0871, Japan Abstract. The saka Legged Robot
More informationMulti-View Omni-Directional Imaging
Multi-View Omni-Directional Imaging Tuesday, December 19, 2000 Moshe Ben-Ezra, Shmuel Peleg Abstract This paper describes a novel camera design or the creation o multiple panoramic images, such that each
More informationUnconstrained License Plate Detection Using the Hausdorff Distance
SPIE Defense & Security, Visual Information Processing XIX, Proc. SPIE, Vol. 7701, 77010V (2010) Unconstrained License Plate Detection Using the Hausdorff Distance M. Lalonde, S. Foucher, L. Gagnon R&D
More informationzūmo 590 Quick Start Manual
zūmo 590 Quick Start Manual January 2014 190-01706-01_0A Printed in Taiwan Getting Started WARNING See the Important Safety and Product Information guide in the product box for product warnings and other
More informationCar License Plate Detection Based on Line Segments
, pp.99-103 http://dx.doi.org/10.14257/astl.2014.58.21 Car License Plate Detection Based on Line Segments Dongwook Kim 1, Liu Zheng Dept. of Information & Communication Eng., Jeonju Univ. Abstract. In
More informationResearch on Evaluation Method of Video Stabilization
International Conference on Advanced Material Science and Environmental Engineering (AMSEE 216) Research on Evaluation Method of Video Stabilization Bin Chen, Jianjun Zhao and i Wang Weapon Science and
More information3D Modeling of Outdoor Environments by Integrating Omnidirectional Range and Color Images
3D Modeling of Outdoor Environments by Integrating Omnidirectional Range and Color Images Toshihiro ASAI, Masayuki KANBARA, Naokazu YOKOYA Nara Institute of Science and Technology 8916-5 Takayama Ikoma
More informationTransactions on Information and Communications Technologies vol 16, 1996 WIT Press, ISSN
ransactions on Information and Communications echnologies vol 6, 996 WI Press, www.witpress.com, ISSN 743-357 Obstacle detection using stereo without correspondence L. X. Zhou & W. K. Gu Institute of Information
More informationTime-to-Contact from Image Intensity
Time-to-Contact from Image Intensity Yukitoshi Watanabe Fumihiko Sakaue Jun Sato Nagoya Institute of Technology Gokiso, Showa, Nagoya, 466-8555, Japan {yukitoshi@cv.,sakaue@,junsato@}nitech.ac.jp Abstract
More information