1998 IEEE International Conference on Intelligent Vehicles 213

Size: px
Start display at page:

Download "1998 IEEE International Conference on Intelligent Vehicles 213"

Transcription

1 Navigation by Integrating Iconic and GPS Information Shigang Li and Akira Hayashi Faculty of Information Sciences Hiroshima City University Asaminami-ku, Hiroshima, , Japan Abstract This paper proposes a method ofrobot navigation in outdoor environments by using GPS (Global Postioning System) information and panoramic views. Since an outdoor environment is very large scaled, it is very dicult to determine the global position based ona visual sensor. Although a global position can be obtained from a GPS navigator, it is dicult to locate a robot accurately since it has a distance error about meter. Our system is equipped with a GPS navigator and a camera. A coarse-to-ne method isused to generate an outdoor environment map and locate a mobile robot. First, a robot nds its approximate position based on the GPS information. Then, it identies its location from the image information. Experimental results in outdoor environments show the eectiveness of this method. 1 Introduction Making a map of environments and navigation by using the acquired map is an important problem for a mobile robot. Although many researches have been done so far, not so many researches cope with outdoor environments in contrast with indoor environments [6] [2] [5]. The scale of an outdoor environment isvery large. The data measured from a camera sensor is very noisy. Therefore, the qualitative methods by using landmarks are proposed. For example, viewpoint position is determined given relative visual angles between landmarks [7]; a distinct landmark which can be identied from very dierent points of view is used for recognition [1]. However, how to select landmarks from scenes is a problem. To cope with this problem, [8] proposes a datadriven method to describe route scenes by using panoramic views when a robot moves along a route. Further, the landmark with distinctive scene features can be selected automatically based upon the range and iconic cues of segmented objects along route [3], and a robot can recognize the route when it moves along the same route again by matching the memorized objects with incoming new scenes [4]. However, if an outdoor environment is very large and includes some similar scenes, it is dicult to identify its location in the environment only by using the visual cues. Even a human being may be lost in this situation. How a robot knows that it returns to the starting point from a visual sensor is a dicult problem, especially in a very-large-scale outdoor environment. On the other hand, the global position can be obtained simply from a GPS (Global Positioning System) navigator. This paper proposes a method of robot navigation in outdoor environments by using GPS (Global Postioning System) information and panoramic views. Since an outdoor environment is very large scaled, it is very dicult to determine the global position based on a visual sensor. Although a global position can be obtained from GPS information, it is dicult to locate a robot position accurately since it has a distance error about meter. Our system is equipped with a GPS navigator and a camera. A coarse-to-ne method is used to generate an outdoor environment map and locate a mobile robot. First, a robot nds its approximate position based on the GPS information. Then, it identies its location from the image information. An experimental result in outdoor environments shows the eectiveness of this method. 2 Our system Our system consists of a car, a GPS navigator and a camera. The experimental environment is the campus of Hiroshima City University which is shown in Figure. 1. The contours enclosing the area is the route used in our experiment. A human drives the car along the route in the direction of counter clockwise. Mean IEEE International Conference on Intelligent Vehicles 213

2 while the camera is mounted on the left side of the car and its optical axis is orthogonal with the car moving direction. Thus, we can acquire the GPS information and image sequences simultaneously when the car moves along a route. Height[m] E132 º 25.4 N34 º 26.4 N34 º 26.2 Latitude N34 º 26.0 N34 º25.8 E132 º 25.2 E132 º 25.0 Longitude E132 º 24.8 Figure 2: The 3D map of the route plotted by using the data of latitude, longitude and height. 2.2 Panoramic views Figure 1: The top view of a part of the campus of Hiroshima City University, where the contours enclosing the area is the route used in our experiment. Suppose a camera moving along a smooth path with its optical axis orthogonal to the moving direction. If we set a vertical slit with a single pixel width in an image and arranges 1D images through the slit along the time axis, we get a wide view of environment called a panoramic view [3]. Figure. 3 shows the panoramic view taken along the route, in which intersections are enclosed by rectangles. 2.1 GPS navigator The GPS system is fully operated and maintained by U.S. Department of Defence. The GPS navigator receives signals from a group of satellites in space to show the position anywhere in the world, 24 hours a day. So, it is very convenient to use it for navigation. The HG-7 type of GPS navigator made by Marine Electronics, Inc. is used in our experiment. The latitude, longitude, height and speed at a position can be acquired from the GPS navigator. Figure. 2 shows the 3D map of the route in Figure. 1, which is plotted by using the data of latitude, longitude and height. Since the GPS navigator has a distance error about meter, the 3D curve in Figure. 2 is not closed although the car returns to the starting point. Thus, it is dicult to use a GPS navigator to localize our car accurately. For example, it may be that you really have moved through an intersection, but from the GPS navigator data the intersection seems forward; as a result, you may fail to make a turn at the intersection. To cope with such problems, we use visual information taken from a camera. 3 Making environment map by fusing GPS information and panoramic views When the car runs along a route, we can acquire the GPS information and panoramic views simultaneously. We synchronize the two devices (the GPS navigaotr and the camera) according to time. By plotting the panoramic view along the route in terms of GPS information, we can obtain Figure. 4. Figure. 5 shows the view segments which is zoomed up at the starting point. Due to the position error from the GPS navigator, the same starting point is plotted at dierent positions. However, as shown in Figure. 5, the same scene is viewed again if the car returns to the starting point. Thus, we could correct the position error according to the iconic information of panoramic views. Our method is as follows. The car runs along a route and acquires the position information from a GPS navigator and a panoramic view from a camera simultaneously. When the car arrives at a position which is near to that passed before within the distance error of the GPS device, the incoming scene is compared with the memorized panoramic view IEEE International Conference on Intelligent Vehicles 214

3 Figure 3: The panoramic view taken along the route IEEE International Conference on Intelligent Vehicles 215

4 The Dynamic programming method is used to match the incoming scenes with the memorized panoramic view [4]. If the score of matching is high enough, we can know that the car return to the same position and fuse them to obtain a closed route. If the matching is failed, we can know that the car is runing another route which is near enough to that tranversed before. Here, we give more detailed explaination on the dynamic programming method. The objective is to match the memorized panoramic view segment with the incoming scenes. Instead of matching the whole memorized view segment, we sample the view segment and obtain a sequence of slit images, fs1;s2; :::; S j ; :::; S n g, which is shown on the lower left in Figure. 6. The upper right image in Figure. 6 is the new panoramic view taken from incoming scenes. We Figure 5: The view zoomed up at the starting point. compute of the similarity, C(S j ;P m;j ), between the slit image, S j, of the memorized view segment and the slit image, P m;j, of the new panoramic view by SSD method. Where n is the number of the sequence of slit images. P m;j is the slit image from the new incoming C(S j ;P m;j )= panoramic which is within the error range of the GPS Xq01 f( r 0 i;sj ri;pm;j T i=0 i;s j 0 ) 2 +( g 0 i;sj gi;pm;j Ti;pm;j 0 T i;s j 0 ) 2 +( b 0 i;sj bi;pm;j Ti;pm;j 0 T i;s j 0 ) 2 navigator. The bigger the error range of the GPS navigator is, the bigger m is. We carry it out by the dy- Ti;pm;j 0 g namic programming method [4]. Figure. 6 shows the Where, r i;s j, g i;s j, b i;s j are the R, G, B values of ith pixel at S j, respectively; ri;pm;j 0, g0 i;pm;j and b0 i;pm;j are matching result of memorized slit images with the incoming panoramic view when the car returns to the the R, G, B values of ith pixel at p m;j, respectively; T i;s j = r i;s j + g i;s j + b i;s j ;T0 i;pm;j = r0 i;pm;j + g0 i;pm;j + starting point. The matching result is shown in the lower right. b 0 i;pm;j ; q is the pixel number of the slit image. The closed route found after fusing GPS information and panoramic views is shown in Figure. 7. Fig- If the sequence of slit images, S1;S2; :::; S j ; :::; S n, appear in the new incoming panoramic with the same ure. 8 shows the view segments which is zoomed up order, the matching succeeds. We realize the matching at the starting point. We distribute the position error by minimizing the following evaluation function. along the route. nx We can use this map for car navigation. min f C(S j ;P m;j )g fpm;jg j=1 Figure 4: Plotting the panoramic view along the route in term of GPS information. Figure 7: The closed route found after fusing GPS information and panoramic views IEEE International Conference on Intelligent Vehicles 216

5 left in Figure. 10. Here the same dynamic programming method with Figure. 6 is used. Thus, the car can clear its position error accumulated so far when it identies such a location, and is expected to arrive its destination correctly. 5 Conclusions Figure 8: The view zoomed up at the starting point after the fusion. 4 Navigation based upon the built map Now we consider how a car makes its navigation based upon the built map. As a concrete task, let the car nd a given intersection. Figure. 9 shows the GPS information and the view segment at intersetion 7 in Figure. 3. GPS N W H283[m] Figure 9: The GPS information and the view segment at intersetion 7. Instead of matching the whole view segment with the incoming scene, we sample the view segment and obtain a sequence of slit images which is shown on the This paper proposes a method of robot (car) navigation in outdoor environments by using GPS (Global Postioning System) information and panoramic views. A coarse-to-ne method is used to generate an outdoor environment map and locate a mobile robot. First, a robot nds its approximate position based on the GPS information. Then, it identies its destination from the image information. An experimental result in outdoor environments shows the eectiveness of this method. References [1] D. Dai and D. T. Lawton, \Range-free qualitative navigation," Proc. of IEEE Int. Conf. on Robotics & Automation, pp , [2] J. Horn and G. Schmidt, \Continuous localization for long-range indoor navigation of mobile robots," Proc. of IEEE Int. Conf. on Robotics & Automation, pp , [3] S. Li and S. Tsuji, \Selecting distinctive scene features for landmarks", Proc. IEEE Int. Conf. Robotics & Automation, pp , [4] S. Li, S. Tsuji and A. Hayashi, \Qualitative Representation of Outdoor Environment along Route", Proc. ICPR, Vol.1, pp , [5] S. Li, J. Nagata and S. Tsuji, \A navigation system based upon panoramic representation," Proc. of IROS, Vol. 17, pp , [6] M. J. Mataric, \Integration of representation into goal-driven behavior-based robot," IEEE Tran. on Robotics & Automation, Vol. 8, No. 3, pp , [7] K. T. Sutherland and W. B. Thompson, \Inexact navigation," Proc. of IEEE Int. Conf. on Robotics & Automation, pp. 1-7, [8] J. Y. Zheng and S. Tsuji, \Panoramic representation for route recognition by a mobile robot", Int. J. Computer Vision,, 9:(1), pp , IEEE International Conference on Intelligent Vehicles 217

6 Figure 6: The matching result of memorized panoramic view with the incoming slit images when the car returns to the starting point. Figure 10: The matching result of a given intersection with the incoming scenes IEEE International Conference on Intelligent Vehicles 218

Image-Based Memory of Environment. homing uses a similar idea that the agent memorizes. [Hong 91]. However, the agent nds diculties in arranging its

Image-Based Memory of Environment. homing uses a similar idea that the agent memorizes. [Hong 91]. However, the agent nds diculties in arranging its Image-Based Memory of Environment Hiroshi ISHIGURO Department of Information Science Kyoto University Kyoto 606-01, Japan E-mail: ishiguro@kuis.kyoto-u.ac.jp Saburo TSUJI Faculty of Systems Engineering

More information

VISUAL NAVIGATION SYSTEM ON WINDSHIELD HEAD-UP DISPLAY

VISUAL NAVIGATION SYSTEM ON WINDSHIELD HEAD-UP DISPLAY VISUAL NAVIGATION SYSTEM ON WINDSHIELD HEAD-UP DISPLAY Akihiko SATO *, Itaru KITAHARA, Yoshinari KAMEDA, Yuichi OHTA Department of Intelligent Interaction Technologies, Graduate School of Systems and Information

More information

Proc. 14th Int. Conf. on Intelligent Autonomous Systems (IAS-14), 2016

Proc. 14th Int. Conf. on Intelligent Autonomous Systems (IAS-14), 2016 Proc. 14th Int. Conf. on Intelligent Autonomous Systems (IAS-14), 2016 Outdoor Robot Navigation Based on View-based Global Localization and Local Navigation Yohei Inoue, Jun Miura, and Shuji Oishi Department

More information

Measurement of Pedestrian Groups Using Subtraction Stereo

Measurement of Pedestrian Groups Using Subtraction Stereo Measurement of Pedestrian Groups Using Subtraction Stereo Kenji Terabayashi, Yuki Hashimoto, and Kazunori Umeda Chuo University / CREST, JST, 1-13-27 Kasuga, Bunkyo-ku, Tokyo 112-8551, Japan terabayashi@mech.chuo-u.ac.jp

More information

Realtime Omnidirectional Stereo for Obstacle Detection and Tracking in Dynamic Environments

Realtime Omnidirectional Stereo for Obstacle Detection and Tracking in Dynamic Environments Proc. 2001 IEEE/RSJ Int. Conf. on Intelligent Robots and Systems pp. 31-36, Maui, Hawaii, Oct./Nov. 2001. Realtime Omnidirectional Stereo for Obstacle Detection and Tracking in Dynamic Environments Hiroshi

More information

Intelligent Outdoor Navigation of a Mobile Robot Platform Using a Low Cost High Precision RTK-GPS and Obstacle Avoidance System

Intelligent Outdoor Navigation of a Mobile Robot Platform Using a Low Cost High Precision RTK-GPS and Obstacle Avoidance System Intelligent Outdoor Navigation of a Mobile Robot Platform Using a Low Cost High Precision RTK-GPS and Obstacle Avoidance System Under supervision of: Prof. Dr. -Ing. Klaus-Dieter Kuhnert Dipl.-Inform.

More information

Acquisition of Qualitative Spatial Representation by Visual Observation

Acquisition of Qualitative Spatial Representation by Visual Observation Acquisition of Qualitative Spatial Representation by Visual Observation Takushi Sogo Hiroshi Ishiguro Toru Ishida Department of Social Informatics, Kyoto University Kyoto 606-8501, Japan sogo@kuis.kyoto-u.ac.jp,

More information

Optical Flow-Based Person Tracking by Multiple Cameras

Optical Flow-Based Person Tracking by Multiple Cameras Proc. IEEE Int. Conf. on Multisensor Fusion and Integration in Intelligent Systems, Baden-Baden, Germany, Aug. 2001. Optical Flow-Based Person Tracking by Multiple Cameras Hideki Tsutsui, Jun Miura, and

More information

VISUAL NAVIGATION SYSTEM ON WINDSHIELD HEAD-UP DISPLAY

VISUAL NAVIGATION SYSTEM ON WINDSHIELD HEAD-UP DISPLAY VISUAL NAVIGATION SYSTEM ON WINDSHIELD HEAD-UP DISPLAY Akihiko SATO *, Itaru KITAHARA, Yoshinari KAMEDA, Yuichi OHTA Department of Intelligent Interaction Technologies, Graduate School of Systems and Information

More information

Development of Low-Cost Compact Omnidirectional Vision Sensors and their applications

Development of Low-Cost Compact Omnidirectional Vision Sensors and their applications Development of Low-Cost Compact Omnidirectional Vision Sensors and their applications Hiroshi ISHIGURO Department of Electrical & Computer Engineering, University of California, San Diego (9500 Gilman

More information

Behavior Learning for a Mobile Robot with Omnidirectional Vision Enhanced by an Active Zoom Mechanism

Behavior Learning for a Mobile Robot with Omnidirectional Vision Enhanced by an Active Zoom Mechanism Behavior Learning for a Mobile Robot with Omnidirectional Vision Enhanced by an Active Zoom Mechanism Sho ji Suzuki, Tatsunori Kato, Minoru Asada, and Koh Hosoda Dept. of Adaptive Machine Systems, Graduate

More information

DETECTION OF 3D POINTS ON MOVING OBJECTS FROM POINT CLOUD DATA FOR 3D MODELING OF OUTDOOR ENVIRONMENTS

DETECTION OF 3D POINTS ON MOVING OBJECTS FROM POINT CLOUD DATA FOR 3D MODELING OF OUTDOOR ENVIRONMENTS DETECTION OF 3D POINTS ON MOVING OBJECTS FROM POINT CLOUD DATA FOR 3D MODELING OF OUTDOOR ENVIRONMENTS Tsunetake Kanatani,, Hideyuki Kume, Takafumi Taketomi, Tomokazu Sato and Naokazu Yokoya Hyogo Prefectural

More information

GPS mini Watch User Manual Introduction. Getting Started. Caution: Step 1) Know your Watch:

GPS mini Watch User Manual Introduction. Getting Started. Caution: Step 1) Know your Watch: Watch User Manual Introduction Thank you for purchasing the GPS Watch. This GPS Watch is packed with personal Training features like speed, trip time, laps, etc. Watch features include but not limited

More information

Dominant plane detection using optical flow and Independent Component Analysis

Dominant plane detection using optical flow and Independent Component Analysis Dominant plane detection using optical flow and Independent Component Analysis Naoya OHNISHI 1 and Atsushi IMIYA 2 1 School of Science and Technology, Chiba University, Japan Yayoicho 1-33, Inage-ku, 263-8522,

More information

Epipolar geometry-based ego-localization using an in-vehicle monocular camera

Epipolar geometry-based ego-localization using an in-vehicle monocular camera Epipolar geometry-based ego-localization using an in-vehicle monocular camera Haruya Kyutoku 1, Yasutomo Kawanishi 1, Daisuke Deguchi 1, Ichiro Ide 1, Hiroshi Murase 1 1 : Nagoya University, Japan E-mail:

More information

Department of Electrical Engineering, Keio University Hiyoshi Kouhoku-ku Yokohama 223, Japan

Department of Electrical Engineering, Keio University Hiyoshi Kouhoku-ku Yokohama 223, Japan Shape Modeling from Multiple View Images Using GAs Satoshi KIRIHARA and Hideo SAITO Department of Electrical Engineering, Keio University 3-14-1 Hiyoshi Kouhoku-ku Yokohama 223, Japan TEL +81-45-563-1141

More information

AR Cultural Heritage Reconstruction Based on Feature Landmark Database Constructed by Using Omnidirectional Range Sensor

AR Cultural Heritage Reconstruction Based on Feature Landmark Database Constructed by Using Omnidirectional Range Sensor AR Cultural Heritage Reconstruction Based on Feature Landmark Database Constructed by Using Omnidirectional Range Sensor Takafumi Taketomi, Tomokazu Sato, and Naokazu Yokoya Graduate School of Information

More information

Generating Dynamic Projection Images for Scene Representation and Understanding

Generating Dynamic Projection Images for Scene Representation and Understanding COMPUTER VISION AND IMAGE UNDERSTANDING Vol. 72, No. 3, December, pp. 237 256, 1998 ARTICLE NO. IV980678 Generating Dynamic Projection Images for Scene Representation and Understanding Jiang Yu Zheng and

More information

URBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES

URBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES URBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES An Undergraduate Research Scholars Thesis by RUI LIU Submitted to Honors and Undergraduate Research Texas A&M University in partial fulfillment

More information

FOR ROUTE UNDERSTANDING

FOR ROUTE UNDERSTANDING PANORAMIC REPRESENTATION OF SCENES FOR ROUTE UNDERSTANDING Abstract Jiang Yu Zheng and Saburo Tsuji Department of Control Engineering, Osaka University Toyonaka, Osaka 560, Japan This work tackles route

More information

Robot Localization based on Geo-referenced Images and G raphic Methods

Robot Localization based on Geo-referenced Images and G raphic Methods Robot Localization based on Geo-referenced Images and G raphic Methods Sid Ahmed Berrabah Mechanical Department, Royal Military School, Belgium, sidahmed.berrabah@rma.ac.be Janusz Bedkowski, Łukasz Lubasiński,

More information

Camera Calibration for a Robust Omni-directional Photogrammetry System

Camera Calibration for a Robust Omni-directional Photogrammetry System Camera Calibration for a Robust Omni-directional Photogrammetry System Fuad Khan 1, Michael Chapman 2, Jonathan Li 3 1 Immersive Media Corporation Calgary, Alberta, Canada 2 Ryerson University Toronto,

More information

DEVELOPMENT OF POSITION MEASUREMENT SYSTEM FOR CONSTRUCTION PILE USING LASER RANGE FINDER

DEVELOPMENT OF POSITION MEASUREMENT SYSTEM FOR CONSTRUCTION PILE USING LASER RANGE FINDER S17- DEVELOPMENT OF POSITION MEASUREMENT SYSTEM FOR CONSTRUCTION PILE USING LASER RANGE FINDER Fumihiro Inoue 1 *, Takeshi Sasaki, Xiangqi Huang 3, and Hideki Hashimoto 4 1 Technica Research Institute,

More information

Security Management System Object Counting

Security Management System Object Counting Security Management System Object Counting This document describes the steps for using the Object Counting module available in the Security Management System software. (A) Object Counting Configuration

More information

Recognizing Buildings in Urban Scene of Distant View ABSTRACT

Recognizing Buildings in Urban Scene of Distant View ABSTRACT Recognizing Buildings in Urban Scene of Distant View Peilin Liu, Katsushi Ikeuchi and Masao Sakauchi Institute of Industrial Science, University of Tokyo, Japan 7-22-1 Roppongi, Minato-ku, Tokyo 106, Japan

More information

How to setup an accurate perimeter guarding system with DeepinMind NVR HIKVISION TECHNICAL SUPPORT TEAM

How to setup an accurate perimeter guarding system with DeepinMind NVR HIKVISION TECHNICAL SUPPORT TEAM How to setup an accurate perimeter guarding system with DeepinMind NVR HIKVISION TECHNICAL SUPPORT TEAM 2017-11-13 1 ids-9632nxi-i8/16s How to setup an accurate perimeter guarding system with DeepinMind

More information

3D Tracking Using Two High-Speed Vision Systems

3D Tracking Using Two High-Speed Vision Systems 3D Tracking Using Two High-Speed Vision Systems Yoshihiro NAKABO 1, Idaku ISHII 2, Masatoshi ISHIKAWA 3 1 University of Tokyo, Tokyo, Japan, nakabo@k2.t.u-tokyo.ac.jp 2 Tokyo University of Agriculture

More information

07ATC 22 Image Exploitation System for Airborne Surveillance

07ATC 22 Image Exploitation System for Airborne Surveillance 07ATC 22 Image Exploitation System for Airborne Surveillance Karunakaran. P, Frederick Mathews, S.H.Padmanabhan, T.K.Sateesh HCL Technologies, India Copyright 2007 SAE International ABSTRACT The operational

More information

MOTION. Feature Matching/Tracking. Control Signal Generation REFERENCE IMAGE

MOTION. Feature Matching/Tracking. Control Signal Generation REFERENCE IMAGE Head-Eye Coordination: A Closed-Form Solution M. Xie School of Mechanical & Production Engineering Nanyang Technological University, Singapore 639798 Email: mmxie@ntuix.ntu.ac.sg ABSTRACT In this paper,

More information

3D Reconstruction of Outdoor Environments from Omnidirectional Range and Color Images

3D Reconstruction of Outdoor Environments from Omnidirectional Range and Color Images 3D Reconstruction of Outdoor Environments from Omnidirectional Range and Color Images Toshihiro ASAI, Masayuki KANBARA, Naokazu YOKOYA Graduate School of Information Science, Nara Institute of Science

More information

Real-time target tracking using a Pan and Tilt platform

Real-time target tracking using a Pan and Tilt platform Real-time target tracking using a Pan and Tilt platform Moulay A. Akhloufi Abstract In recent years, we see an increase of interest for efficient tracking systems in surveillance applications. Many of

More information

Segmentation and Tracking of Partial Planar Templates

Segmentation and Tracking of Partial Planar Templates Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract

More information

AN APPROACH OF SEMIAUTOMATED ROAD EXTRACTION FROM AERIAL IMAGE BASED ON TEMPLATE MATCHING AND NEURAL NETWORK

AN APPROACH OF SEMIAUTOMATED ROAD EXTRACTION FROM AERIAL IMAGE BASED ON TEMPLATE MATCHING AND NEURAL NETWORK AN APPROACH OF SEMIAUTOMATED ROAD EXTRACTION FROM AERIAL IMAGE BASED ON TEMPLATE MATCHING AND NEURAL NETWORK Xiangyun HU, Zuxun ZHANG, Jianqing ZHANG Wuhan Technique University of Surveying and Mapping,

More information

Vol. 21 No. 6, pp ,

Vol. 21 No. 6, pp , Vol. 21 No. 6, pp.69 696, 23 69 3 3 3 Map Generation of a Mobile Robot by Integrating Omnidirectional Stereo and Laser Range Finder Yoshiro Negishi 3, Jun Miura 3 and Yoshiaki Shirai 3 This paper describes

More information

2 ATTILA FAZEKAS The tracking model of the robot car The schematic picture of the robot car can be seen on Fig.1. Figure 1. The main controlling task

2 ATTILA FAZEKAS The tracking model of the robot car The schematic picture of the robot car can be seen on Fig.1. Figure 1. The main controlling task NEW OPTICAL TRACKING METHODS FOR ROBOT CARS Attila Fazekas Debrecen Abstract. In this paper new methods are proposed for intelligent optical tracking of robot cars the important tools of CIM (Computer

More information

Development of a video-rate range finder using dynamic threshold method for characteristic point detection

Development of a video-rate range finder using dynamic threshold method for characteristic point detection Engineering Industrial & Management Engineering fields Year 1999 Development of a video-rate range finder using dynamic threshold method for characteristic point detection Yutaka Tanaka Nobuo Takeda Akio

More information

An Angle Estimation to Landmarks for Autonomous Satellite Navigation

An Angle Estimation to Landmarks for Autonomous Satellite Navigation 5th International Conference on Environment, Materials, Chemistry and Power Electronics (EMCPE 2016) An Angle Estimation to Landmarks for Autonomous Satellite Navigation Qing XUE a, Hongwen YANG, Jian

More information

Employing a Fish-Eye for Scene Tunnel Scanning

Employing a Fish-Eye for Scene Tunnel Scanning Employing a Fish-Eye for Scene Tunnel Scanning Jiang Yu Zheng 1, Shigang Li 2 1 Dept. of Computer Science, Indiana University Purdue University Indianapolis, 723 W. Michigan St. Indianapolis, IN46202,

More information

CS4670: Computer Vision

CS4670: Computer Vision CS4670: Computer Vision Noah Snavely Lecture 6: Feature matching and alignment Szeliski: Chapter 6.1 Reading Last time: Corners and blobs Scale-space blob detector: Example Feature descriptors We know

More information

Dept. of Adaptive Machine Systems, Graduate School of Engineering Osaka University, Suita, Osaka , Japan

Dept. of Adaptive Machine Systems, Graduate School of Engineering Osaka University, Suita, Osaka , Japan An Application of Vision-Based Learning for a Real Robot in RoboCup - A Goal Keeping Behavior for a Robot with an Omnidirectional Vision and an Embedded Servoing - Sho ji Suzuki 1, Tatsunori Kato 1, Hiroshi

More information

Robot localization method based on visual features and their geometric relationship

Robot localization method based on visual features and their geometric relationship , pp.46-50 http://dx.doi.org/10.14257/astl.2015.85.11 Robot localization method based on visual features and their geometric relationship Sangyun Lee 1, Changkyung Eem 2, and Hyunki Hong 3 1 Department

More information

Mobile Robot Navigation Using Omnidirectional Vision

Mobile Robot Navigation Using Omnidirectional Vision Mobile Robot Navigation Using Omnidirectional Vision László Mornailla, Tamás Gábor Pekár, Csaba Gergő Solymosi, Zoltán Vámossy John von Neumann Faculty of Informatics, Budapest Tech Bécsi út 96/B, H-1034

More information

Recovering light directions and camera poses from a single sphere.

Recovering light directions and camera poses from a single sphere. Title Recovering light directions and camera poses from a single sphere Author(s) Wong, KYK; Schnieders, D; Li, S Citation The 10th European Conference on Computer Vision (ECCV 2008), Marseille, France,

More information

A 100Hz Real-time Sensing System of Textured Range Images

A 100Hz Real-time Sensing System of Textured Range Images A 100Hz Real-time Sensing System of Textured Range Images Hidetoshi Ishiyama Course of Precision Engineering School of Science and Engineering Chuo University 1-13-27 Kasuga, Bunkyo-ku, Tokyo 112-8551,

More information

Gesture Recognition using Temporal Templates with disparity information

Gesture Recognition using Temporal Templates with disparity information 8- MVA7 IAPR Conference on Machine Vision Applications, May 6-8, 7, Tokyo, JAPAN Gesture Recognition using Temporal Templates with disparity information Kazunori Onoguchi and Masaaki Sato Hirosaki University

More information

Localization of Wearable Users Using Invisible Retro-reflective Markers and an IR Camera

Localization of Wearable Users Using Invisible Retro-reflective Markers and an IR Camera Localization of Wearable Users Using Invisible Retro-reflective Markers and an IR Camera Yusuke Nakazato, Masayuki Kanbara and Naokazu Yokoya Graduate School of Information Science, Nara Institute of Science

More information

Camera Parameters Estimation from Hand-labelled Sun Sositions in Image Sequences

Camera Parameters Estimation from Hand-labelled Sun Sositions in Image Sequences Camera Parameters Estimation from Hand-labelled Sun Sositions in Image Sequences Jean-François Lalonde, Srinivasa G. Narasimhan and Alexei A. Efros {jlalonde,srinivas,efros}@cs.cmu.edu CMU-RI-TR-8-32 July

More information

Hand-Eye Calibration from Image Derivatives

Hand-Eye Calibration from Image Derivatives Hand-Eye Calibration from Image Derivatives Abstract In this paper it is shown how to perform hand-eye calibration using only the normal flow field and knowledge about the motion of the hand. The proposed

More information

Real-time Generation and Presentation of View-dependent Binocular Stereo Images Using a Sequence of Omnidirectional Images

Real-time Generation and Presentation of View-dependent Binocular Stereo Images Using a Sequence of Omnidirectional Images Real-time Generation and Presentation of View-dependent Binocular Stereo Images Using a Sequence of Omnidirectional Images Abstract This paper presents a new method to generate and present arbitrarily

More information

Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision

Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision Zhiyan Zhang 1, Wei Qian 1, Lei Pan 1 & Yanjun Li 1 1 University of Shanghai for Science and Technology, China

More information

Vision-Motion Planning with Uncertainty

Vision-Motion Planning with Uncertainty Vision-Motion Planning with Uncertainty Jun MIURA Yoshiaki SHIRAI Dept. of Mech. Eng. for Computer-Controlled Machinery, Osaka University, Suita, Osaka 565, Japan jun@ccm.osaka-u.ac.jp Abstract This paper

More information

Vehicle Ego-localization by Matching In-vehicle Camera Images to an Aerial Image

Vehicle Ego-localization by Matching In-vehicle Camera Images to an Aerial Image Vehicle Ego-localization by Matching In-vehicle Camera Images to an Aerial Image Masafumi NODA 1,, Tomokazu TAKAHASHI 1,2, Daisuke DEGUCHI 1, Ichiro IDE 1, Hiroshi MURASE 1, Yoshiko KOJIMA 3 and Takashi

More information

Efficient Path Finding Method Based Evaluation Function in Large Scene Online Games and Its Application

Efficient Path Finding Method Based Evaluation Function in Large Scene Online Games and Its Application Journal of Information Hiding and Multimedia Signal Processing c 2017 ISSN 2073-4212 Ubiquitous International Volume 8, Number 3, May 2017 Efficient Path Finding Method Based Evaluation Function in Large

More information

Estimating Camera Position And Posture by Using Feature Landmark Database

Estimating Camera Position And Posture by Using Feature Landmark Database Estimating Camera Position And Posture by Using Feature Landmark Database Motoko Oe 1, Tomokazu Sato 2 and Naokazu Yokoya 2 1 IBM Japan 2 Nara Institute of Science and Technology, Japan Abstract. Estimating

More information

Sherlock 7 Technical Resource. Search Geometric

Sherlock 7 Technical Resource. Search Geometric Sherlock 7 Technical Resource DALSA Corp., Industrial Products (IPD) www.goipd.com 978.670.2002 (U.S.A.) Document Revision: September 24, 2007 Search Geometric Search utilities A common task in machine

More information

Collecting outdoor datasets for benchmarking vision based robot localization

Collecting outdoor datasets for benchmarking vision based robot localization Collecting outdoor datasets for benchmarking vision based robot localization Emanuele Frontoni*, Andrea Ascani, Adriano Mancini, Primo Zingaretti Department of Ingegneria Infromatica, Gestionale e dell

More information

Camera Registration in a 3D City Model. Min Ding CS294-6 Final Presentation Dec 13, 2006

Camera Registration in a 3D City Model. Min Ding CS294-6 Final Presentation Dec 13, 2006 Camera Registration in a 3D City Model Min Ding CS294-6 Final Presentation Dec 13, 2006 Goal: Reconstruct 3D city model usable for virtual walk- and fly-throughs Virtual reality Urban planning Simulation

More information

OBSTACLE DETECTION USING STRUCTURED BACKGROUND

OBSTACLE DETECTION USING STRUCTURED BACKGROUND OBSTACLE DETECTION USING STRUCTURED BACKGROUND Ghaida Al Zeer, Adnan Abou Nabout and Bernd Tibken Chair of Automatic Control, Faculty of Electrical, Information and Media Engineering University of Wuppertal,

More information

Chapter 11 Arc Extraction and Segmentation

Chapter 11 Arc Extraction and Segmentation Chapter 11 Arc Extraction and Segmentation 11.1 Introduction edge detection: labels each pixel as edge or no edge additional properties of edge: direction, gradient magnitude, contrast edge grouping: edge

More information

A Novel Real-Time Feature Matching Scheme

A Novel Real-Time Feature Matching Scheme Sensors & Transducers, Vol. 165, Issue, February 01, pp. 17-11 Sensors & Transducers 01 by IFSA Publishing, S. L. http://www.sensorsportal.com A Novel Real-Time Feature Matching Scheme Ying Liu, * Hongbo

More information

etrex quick start manual for use with models 20 and 30

etrex quick start manual for use with models 20 and 30 etrex quick start manual for use with models 20 and 30 Getting Started Device Overview warning See the Important Safety and Product Information guide in the product box for product warnings and other important

More information

Detection of Small-Waving Hand by Distributed Camera System

Detection of Small-Waving Hand by Distributed Camera System Detection of Small-Waving Hand by Distributed Camera System Kenji Terabayashi, Hidetsugu Asano, Takeshi Nagayasu, Tatsuya Orimo, Mutsumi Ohta, Takaaki Oiwa, and Kazunori Umeda Department of Mechanical

More information

Summarization of Egocentric Moving Videos for Generating Walking Route Guidance

Summarization of Egocentric Moving Videos for Generating Walking Route Guidance Summarization of Egocentric Moving Videos for Generating Walking Route Guidance Masaya Okamoto and Keiji Yanai Department of Informatics, The University of Electro-Communications 1-5-1 Chofugaoka, Chofu-shi,

More information

Preceding vehicle detection and distance estimation. lane change, warning system.

Preceding vehicle detection and distance estimation. lane change, warning system. Preceding vehicle detection and distance estimation for lane change warning system U. Iqbal, M.S. Sarfraz Computer Vision Research Group (COMVis) Department of Electrical Engineering, COMSATS Institute

More information

AUTOMATIC RECONSTRUCTION OF LARGE-SCALE VIRTUAL ENVIRONMENT FOR INTELLIGENT TRANSPORTATION SYSTEMS SIMULATION

AUTOMATIC RECONSTRUCTION OF LARGE-SCALE VIRTUAL ENVIRONMENT FOR INTELLIGENT TRANSPORTATION SYSTEMS SIMULATION AUTOMATIC RECONSTRUCTION OF LARGE-SCALE VIRTUAL ENVIRONMENT FOR INTELLIGENT TRANSPORTATION SYSTEMS SIMULATION Khairil Azmi, Shintaro Ono, Masataka Kagesawa, Katsushi Ikeuchi Institute of Industrial Science,

More information

Monocular Vision Based Autonomous Navigation for Arbitrarily Shaped Urban Roads

Monocular Vision Based Autonomous Navigation for Arbitrarily Shaped Urban Roads Proceedings of the International Conference on Machine Vision and Machine Learning Prague, Czech Republic, August 14-15, 2014 Paper No. 127 Monocular Vision Based Autonomous Navigation for Arbitrarily

More information

(a) (b) (c) Fig. 1. Omnidirectional camera: (a) principle; (b) physical construction; (c) captured. of a local vision system is more challenging than

(a) (b) (c) Fig. 1. Omnidirectional camera: (a) principle; (b) physical construction; (c) captured. of a local vision system is more challenging than An Omnidirectional Vision System that finds and tracks color edges and blobs Felix v. Hundelshausen, Sven Behnke, and Raul Rojas Freie Universität Berlin, Institut für Informatik Takustr. 9, 14195 Berlin,

More information

SatGuide Symbian User Manual. SatGuide Symbian OS USER MANUAL

SatGuide Symbian User Manual. SatGuide Symbian OS USER MANUAL SatGuide Symbian User Manual SatGuide Symbian OS USER MANUAL CONTENTS 1 - INTRODUCTION...1 WHAT IS SATGUIDE?...1 SATGUIDE FEATURES...1 2 - SMART PRACTICES WHILE DRIVING...2 DRIVE SAFE, CALL SMART SM...2

More information

UAV Position and Attitude Sensoring in Indoor Environment Using Cameras

UAV Position and Attitude Sensoring in Indoor Environment Using Cameras UAV Position and Attitude Sensoring in Indoor Environment Using Cameras 1 Peng Xu Abstract There are great advantages of indoor experiment for UAVs. Test flights of UAV in laboratory is more convenient,

More information

Handheld Augmented Reality. Reto Lindegger

Handheld Augmented Reality. Reto Lindegger Handheld Augmented Reality Reto Lindegger lreto@ethz.ch 1 AUGMENTED REALITY 2 A Definition Three important characteristics: Combines real and virtual environment Interactive in real-time Registered in

More information

Three-Dimensional Measurement of Objects in Liquid with an Unknown Refractive Index Using Fisheye Stereo Camera

Three-Dimensional Measurement of Objects in Liquid with an Unknown Refractive Index Using Fisheye Stereo Camera Three-Dimensional Measurement of Objects in Liquid with an Unknown Refractive Index Using Fisheye Stereo Camera Kazuki Sakamoto, Alessandro Moro, Hiromitsu Fujii, Atsushi Yamashita, and Hajime Asama Abstract

More information

BROWN UNIVERSITY Department of Computer Science Master's Project CS-95-M17

BROWN UNIVERSITY Department of Computer Science Master's Project CS-95-M17 BROWN UNIVERSITY Department of Computer Science Master's Project CS-95-M17 "Robotic Object Recognition: Utilizing Multiple Views to Recognize Partially Occluded Objects" by Neil A. Jacobson ,'. Robotic

More information

Constructing a 3D Object Model from Multiple Visual Features

Constructing a 3D Object Model from Multiple Visual Features Constructing a 3D Object Model from Multiple Visual Features Jiang Yu Zheng Faculty of Computer Science and Systems Engineering Kyushu Institute of Technology Iizuka, Fukuoka 820, Japan Abstract This work

More information

Recognition of Finger Pointing Direction Using Color Clustering and Image Segmentation

Recognition of Finger Pointing Direction Using Color Clustering and Image Segmentation September 14-17, 2013, Nagoya University, Nagoya, Japan Recognition of Finger Pointing Direction Using Color Clustering and Image Segmentation Hidetsugu Asano 1, Takeshi Nagayasu 2, Tatsuya Orimo 1, Kenji

More information

Automatic 3-D 3 D Model Acquisition from Range Images. Michael K. Reed and Peter K. Allen Computer Science Department Columbia University

Automatic 3-D 3 D Model Acquisition from Range Images. Michael K. Reed and Peter K. Allen Computer Science Department Columbia University Automatic 3-D 3 D Model Acquisition from Range Images Michael K. Reed and Peter K. Allen Computer Science Department Columbia University Introduction Objective: given an arbitrary object or scene, construct

More information

British Machine Vision Conference 2 The established approach for automatic model construction begins by taking surface measurements from a number of v

British Machine Vision Conference 2 The established approach for automatic model construction begins by taking surface measurements from a number of v Segmentation of Range Data into Rigid Subsets using Planar Surface Patches A. P. Ashbrook, R. B. Fisher, C. Robertson and N. Wergi Department of Articial Intelligence The University of Edinburgh 5, Forrest

More information

Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera

Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera Tomokazu Satoy, Masayuki Kanbaray, Naokazu Yokoyay and Haruo Takemuraz ygraduate School of Information

More information

3D MODELING OF OUTDOOR SCENES BY INTEGRATING STOP-AND-GO AND CONTINUOUS SCANNING OF RANGEFINDER

3D MODELING OF OUTDOOR SCENES BY INTEGRATING STOP-AND-GO AND CONTINUOUS SCANNING OF RANGEFINDER 3D MODELING OF OUTDOOR SCENES BY INTEGRATING STOP-AND-GO AND CONTINUOUS SCANNING OF RANGEFINDER Toshihiro ASAI, Masayuki KANBARA, Naokazu YOKOYA Graduate School of Information Science, Nara Institute of

More information

Long-term motion estimation from images

Long-term motion estimation from images Long-term motion estimation from images Dennis Strelow 1 and Sanjiv Singh 2 1 Google, Mountain View, CA, strelow@google.com 2 Carnegie Mellon University, Pittsburgh, PA, ssingh@cmu.edu Summary. Cameras

More information

Team Aware Perception System using Stereo Vision and Radar

Team Aware Perception System using Stereo Vision and Radar Team Aware Perception System using Stereo Vision and Radar System Development Review 03/ 08 / 2017 Amit Agarwal Harry Golash Yihao Qian Menghan Zhang Zihao (Theo) Zhang Project Description Develop a standalone

More information

AN EFFICIENT BINARY CORNER DETECTOR. P. Saeedi, P. Lawrence and D. Lowe

AN EFFICIENT BINARY CORNER DETECTOR. P. Saeedi, P. Lawrence and D. Lowe AN EFFICIENT BINARY CORNER DETECTOR P. Saeedi, P. Lawrence and D. Lowe Department of Electrical and Computer Engineering, Department of Computer Science University of British Columbia Vancouver, BC, V6T

More information

ECE 172A: Introduction to Intelligent Systems: Machine Vision, Fall Midterm Examination

ECE 172A: Introduction to Intelligent Systems: Machine Vision, Fall Midterm Examination ECE 172A: Introduction to Intelligent Systems: Machine Vision, Fall 2008 October 29, 2008 Notes: Midterm Examination This is a closed book and closed notes examination. Please be precise and to the point.

More information

v9 Quick Start Guide

v9 Quick Start Guide v9 Quick Start Guide overview Driving Screen Most of your time using CoPIlot Truck will be spent on this screen. Let s take a moment and get familiar with the items you can interact with. Zoom Out Zoom

More information

ФУНДАМЕНТАЛЬНЫЕ НАУКИ. Информатика 9 ИНФОРМАТИКА MOTION DETECTION IN VIDEO STREAM BASED ON BACKGROUND SUBTRACTION AND TARGET TRACKING

ФУНДАМЕНТАЛЬНЫЕ НАУКИ. Информатика 9 ИНФОРМАТИКА MOTION DETECTION IN VIDEO STREAM BASED ON BACKGROUND SUBTRACTION AND TARGET TRACKING ФУНДАМЕНТАЛЬНЫЕ НАУКИ Информатика 9 ИНФОРМАТИКА UDC 6813 OTION DETECTION IN VIDEO STREA BASED ON BACKGROUND SUBTRACTION AND TARGET TRACKING R BOGUSH, S ALTSEV, N BROVKO, E IHAILOV (Polotsk State University

More information

A Path Tracking Method For Autonomous Mobile Robots Based On Grid Decomposition

A Path Tracking Method For Autonomous Mobile Robots Based On Grid Decomposition A Path Tracking Method For Autonomous Mobile Robots Based On Grid Decomposition A. Pozo-Ruz*, C. Urdiales, A. Bandera, E. J. Pérez and F. Sandoval Dpto. Tecnología Electrónica. E.T.S.I. Telecomunicación,

More information

Integration of Multiple-baseline Color Stereo Vision with Focus and Defocus Analysis for 3D Shape Measurement

Integration of Multiple-baseline Color Stereo Vision with Focus and Defocus Analysis for 3D Shape Measurement Integration of Multiple-baseline Color Stereo Vision with Focus and Defocus Analysis for 3D Shape Measurement Ta Yuan and Murali Subbarao tyuan@sbee.sunysb.edu and murali@sbee.sunysb.edu Department of

More information

Laser Eye a new 3D sensor for active vision

Laser Eye a new 3D sensor for active vision Laser Eye a new 3D sensor for active vision Piotr Jasiobedzki1, Michael Jenkin2, Evangelos Milios2' Brian Down1, John Tsotsos1, Todd Campbell3 1 Dept. of Computer Science, University of Toronto Toronto,

More information

1216 P a g e 2.1 TRANSLATION PARAMETERS ESTIMATION. If f (x, y) F(ξ,η) then. f(x,y)exp[j2π(ξx 0 +ηy 0 )/ N] ) F(ξ- ξ 0,η- η 0 ) and

1216 P a g e 2.1 TRANSLATION PARAMETERS ESTIMATION. If f (x, y) F(ξ,η) then. f(x,y)exp[j2π(ξx 0 +ηy 0 )/ N] ) F(ξ- ξ 0,η- η 0 ) and An Image Stitching System using Featureless Registration and Minimal Blending Ch.Rajesh Kumar *,N.Nikhita *,Santosh Roy *,V.V.S.Murthy ** * (Student Scholar,Department of ECE, K L University, Guntur,AP,India)

More information

Scene Text Detection Using Machine Learning Classifiers

Scene Text Detection Using Machine Learning Classifiers 601 Scene Text Detection Using Machine Learning Classifiers Nafla C.N. 1, Sneha K. 2, Divya K.P. 3 1 (Department of CSE, RCET, Akkikkvu, Thrissur) 2 (Department of CSE, RCET, Akkikkvu, Thrissur) 3 (Department

More information

Lost! Leveraging the Crowd for Probabilistic Visual Self-Localization

Lost! Leveraging the Crowd for Probabilistic Visual Self-Localization Lost! Leveraging the Crowd for Probabilistic Visual Self-Localization Marcus A. Brubaker (Toyota Technological Institute at Chicago) Andreas Geiger (Karlsruhe Institute of Technology & MPI Tübingen) Raquel

More information

Panoramic Vision and LRF Sensor Fusion Based Human Identification and Tracking for Autonomous Luggage Cart

Panoramic Vision and LRF Sensor Fusion Based Human Identification and Tracking for Autonomous Luggage Cart Panoramic Vision and LRF Sensor Fusion Based Human Identification and Tracking for Autonomous Luggage Cart Mehrez Kristou, Akihisa Ohya and Shin ichi Yuta Intelligent Robot Laboratory, University of Tsukuba,

More information

BabyTigers-98: Osaka Legged Robot Team

BabyTigers-98: Osaka Legged Robot Team BabyTigers-98: saka Legged Robot Team Noriaki Mitsunaga and Minoru Asada and Chizuko Mishima Dept. of Adaptive Machine Systems, saka niversity, Suita, saka, 565-0871, Japan Abstract. The saka Legged Robot

More information

Multi-View Omni-Directional Imaging

Multi-View Omni-Directional Imaging Multi-View Omni-Directional Imaging Tuesday, December 19, 2000 Moshe Ben-Ezra, Shmuel Peleg Abstract This paper describes a novel camera design or the creation o multiple panoramic images, such that each

More information

Unconstrained License Plate Detection Using the Hausdorff Distance

Unconstrained License Plate Detection Using the Hausdorff Distance SPIE Defense & Security, Visual Information Processing XIX, Proc. SPIE, Vol. 7701, 77010V (2010) Unconstrained License Plate Detection Using the Hausdorff Distance M. Lalonde, S. Foucher, L. Gagnon R&D

More information

zūmo 590 Quick Start Manual

zūmo 590 Quick Start Manual zūmo 590 Quick Start Manual January 2014 190-01706-01_0A Printed in Taiwan Getting Started WARNING See the Important Safety and Product Information guide in the product box for product warnings and other

More information

Car License Plate Detection Based on Line Segments

Car License Plate Detection Based on Line Segments , pp.99-103 http://dx.doi.org/10.14257/astl.2014.58.21 Car License Plate Detection Based on Line Segments Dongwook Kim 1, Liu Zheng Dept. of Information & Communication Eng., Jeonju Univ. Abstract. In

More information

Research on Evaluation Method of Video Stabilization

Research on Evaluation Method of Video Stabilization International Conference on Advanced Material Science and Environmental Engineering (AMSEE 216) Research on Evaluation Method of Video Stabilization Bin Chen, Jianjun Zhao and i Wang Weapon Science and

More information

3D Modeling of Outdoor Environments by Integrating Omnidirectional Range and Color Images

3D Modeling of Outdoor Environments by Integrating Omnidirectional Range and Color Images 3D Modeling of Outdoor Environments by Integrating Omnidirectional Range and Color Images Toshihiro ASAI, Masayuki KANBARA, Naokazu YOKOYA Nara Institute of Science and Technology 8916-5 Takayama Ikoma

More information

Transactions on Information and Communications Technologies vol 16, 1996 WIT Press, ISSN

Transactions on Information and Communications Technologies vol 16, 1996 WIT Press,   ISSN ransactions on Information and Communications echnologies vol 6, 996 WI Press, www.witpress.com, ISSN 743-357 Obstacle detection using stereo without correspondence L. X. Zhou & W. K. Gu Institute of Information

More information

Time-to-Contact from Image Intensity

Time-to-Contact from Image Intensity Time-to-Contact from Image Intensity Yukitoshi Watanabe Fumihiko Sakaue Jun Sato Nagoya Institute of Technology Gokiso, Showa, Nagoya, 466-8555, Japan {yukitoshi@cv.,sakaue@,junsato@}nitech.ac.jp Abstract

More information