Super Wide Viewing for Tele-operation

Size: px
Start display at page:

Download "Super Wide Viewing for Tele-operation"

Transcription

1 Super Wide Viewing for Tele-operation Hajime Nagahara*, Yasushi Yagi**, and Masahiko Yachida* *Graduate School of Engineering Science Osaka University 1-3, Machikaneyama-cho, Toyonaka, Osaka, , Japan ** The institute of Scientific and Industrial Research, Osaka University 8-1, Mihogaoka, Ibaraki, Osaka, , Japan Abstract: - In this paper, we propose a tele-operation system with a super wide field of view (FOV). Our teleoperation system has an omnidirectional image sensor for capturing a panoramic image and a personal spherical screen display for projecting a wide-angle image. The first original contribution is applying our system to navigating the mobile robot by tele-operation and evaluating efficiency of wide FOV. The second original contribution is a new propose of a super wide FOV. Generally, FOVs of a common display system and a HMD system are restricted and it is unable to see the outer of FOV of these displays. The idea of super wide FOV is that the wide FOV is nonlinearly compressed to display FOV. Therefore, the operator can look behind even if the display is not a panoramic large screen display. Key-Words: - Wide filed of view, Tele-opration, Robot, Navigation. 1 Introduction The tele-operation that human operator directs the robot from the remote site has been studied and produced [1], because human does not have to go hazardous environments such as space, volcano, sea and nuclear plant. To complete the tasks effectively and safely, tele-operation systems, called telepresence [2][3] and tele-existence [4][5], are required high quality and realistic feed back information, as if the operator felt actual presence at robot site. Visual feedback is the most important factor of tele-presence. Considerable factors of visual reality are plasticity related to vergence and focus controls, texture related to resolution and dynamic range of images, timedelay caused by camera rotation and communication, and extensity caused by field of view of display. In this paper, we focus to extensity caused by field of view (FOV) of display. It is well known that peripheral vision included wide FOV would influence postural control of human [6]. In addition, a FOV over 80 degree is required for feeling immersion and reality by human operator [7]. Takahashi et al [8] measured the influence of wide FOV head mounted display to human attitude control. They compared 140 degree FOV. This research suggests that the wide FOV is one of important factor for human attitude control. Caldwell et al. [9] reported that the narrow FOV disimproved task efficiency and reality in a navigation task. However, they compared only 30 and 60 degree FOV. Almost previous visual tele-presence systems use the ordinary narrow FOV display such as CRT. The view field can be changed by rotating a camera to get the large FOV information. However, a rotating camera arises a time-delay problem. One solution of enlarging the FOV without such a time-delay problem is to use an omnidirectional image sensor on tele-presence system. Onoe et al. [10] proposed the tele-operation system with omnidirectional image sensor. The system presents the transformed perspective image on a common Head Mounted Display (HMD) corresponding to an operator's head motion. The operator control led the mobile robot in indoor environment. FOV of the common HMD is narrow less than 60 degree, therefore, it is not enough wide angle to present peripheral vision information and the report did not evaluate easefulness of the system. However, they built the virtual tour system [11] by projecting panorama movies on the cylindrical large screen named CYLINDRA and the cubic screen system like CAVE [12]. The first research suggested us that the 360-degree panoramic image is suitable for tele-operation system. The second one suggested us that the wide field of view increases existence and presence. In this paper, we propose a tele-operation system with a super wide FOV. Our tele-operation system has an omnidirectional image sensor named HyperOmni Vision [13] for capturing a panoramic

2 Slave system Omnidirectional image sensor NTSC Master system PC for Imaging PentiumII 400MHz Transmitter Receiver Spherical display Robot RS232C Radio modem RS232C RS232C Radio modem Joystick PC for robot control PentiumII 266 MHz Figure 1. System configuration image and a personal spherical screen display for projecting a wide-angle image. The first original contribution is applying our system to navigating the mobile robot by teleoperation and evaluating efficiency of wide FOV by experiments. The second original contribution is a new proposal of a super wide FOV. Generally, FOVs of a common display system and a HMD system are restricted and it is unable to see the outer of FOV of these displays. Therefore, we propose the idea to project super wide FOV information over the display capacity. Actually, FOV of the input image is transformed to the display system, nonlinearly. The operator can look behind even if the display is not the panoramic large screen display such as CYLINDRA and CAVE. 2 Tele-operation system Fig. 1 shows the configuration of the tele-operation system used in our experiments. The slave system shown in fig. 2-a is a mobile robot (B12: Real World Interface inc.) with an omnidirectional image sensor, HyperOmni Vision. The omnidirectional image sensor composed of a hyperboloidal mirror and a common NTSC video camera unit (EVI-310: Sony). It can capture a 360-degree FOV on a single input image. The input image is transmitted to the master system by a wireless NTSC video transmitter. The mobile robot B12 has three wheels steering and three wheels drive. These three wheels are driven synchronously and remain parallel at all times. The robot is controlled by the joystick on the master system through a wireless RS232C modem. a: Slave system b: Master system Figure 2. Tele-operation system for navigation The master system shown in fig. 2-b consists of a spherical screen display (VisionStarion: Elumens inc.), a joystick, PCs for an image processing and robot controle, a wireless RS232C modem and a wireless NTSC video receiver. The display can project 120hx65v FOV image by SXGA(1280x1024 pixels) resolution. Where, 120hx65v means that horizontal and vertical fields of view are 120 and 65 degree, respectively. Afterward, we represent horizontal and vertical fields of view like this in this paper. The input image is transformed to a spherical image by the PC on the master system. The transformed spherical image is projected on the spherical screen display. Incidentally, the omnidirectional image sensor has a single center of projection. The input image can transform to any desired coordinates without distortion. Therefore, the complete distortionless image is shown to the operator on the spherical screen display. Equations 2-4 are relationships between the input image coordinate of the omnidirectional image sensor and the spherical image coordinate. Here, p(x, y) and P(i, a) indicate to the input image coordinates and spherical coordinates longitude and attitude around the sensor. b and c are coefficients of the hyperboloidal mirror shape described in equation 1. f is the focal length of camera. More details of the image formation are in reference [13]. X 2 + Y 2 a - Z 2 2 b 2 =- 1,(Z >0) (1) a 2 + b 2 = c 2 (c r = 2 - b 2 ) f (b 2 + c 2 ) tan a +2bc tan 2 a +1 (2) x = r cos i (3) y = r sin i (4) Fig. 3 and 4 show the input image of the

3 Figure 3. An omnidirectional image Figure hx65v degree FOV with linear transformation Figure hx65v degree FOV Figure hx65v degree FOV with nonlinear transformation omnidirectional image sensor and the projected image on the spherical screen display, respectively. 3 Wide field of view over the display capacity Generally, FOVs of a common display system and a HMD system are restricted and it is unable to see the outer of FOV of a common display system. A new proposal of a super wide FOV is that FOV of the input image is nonlinearly transformed to the display FOV. It is like compression the input image to display size. Therefore, the operator can look behind even if the display is not the panoramic large screen display such as CYRINDRA and CAVE. A linear transformation described in equation 5 is a simple way to transform FOV. Here, i and i' are azimuth angles on input image and output image. H view and H disp are horizontal sizes of input FOV and that of display screen FOV, respectively. Fig. 5 shows the sample of the linear FOV transformation image. However, the operator would be confused which way to heading in remote environment, because of discordance between the real observed azimuth and the transformed direction on the display. Incidentally, the operator would not be confused at the peripheral area, because human peripheral vision is not usually used for details analysis but also rough sensing such as motion detection. Therefore, our system nonlinearly transform the input image to the display image by the equation 6. In this system, the central area is done by a linear transformation and the peripheral area is done by a nonlinear transformation. From preliminary experiments, the center area of 60 degree should keep the direction accordance. The outer area is transformed by quadratic curve shown in fig. 7. In this figure, the horizontal FOV of display

4 Figure 7. FOV transformation screen H disp is set on 120[degree] and the horizontal FOV of input H view is set on 180, 270 and 360[degree]. This figure also shows the azimuth accordant case when the horizontal FOV of input is within that of display screen FOV (120[degree]). Fig. 6 shows an example of a transformed image by nonlinear transformation. Figure 8. The experimental scene i = H view H i' disp (5) i' -c (i' +i c ) 2 (i' # - i c ) i = * i' (- i c < i' <i c ) (6) i' +c (i' +i c ) 2 (i' $ i c ) 2(H view - H disp ) c = (H disp -2ic) 2 4 Experiments on real environments We evaluated the efficiency of wide FOV to robot navigation. We prepare four navigation courses such as a slalom, a crank and parkings. The robot is 35[cm] wide. The motion freedom of the mobile robot is translation and rotation. The maximum speeds of rotation and translation are set on 31.6[deg/s] and 14.1[cm/s], respectively. In case of slalom and crank, the robot can do pinwheeling. However, in case of two types of parking tasks, the robot movement is same as car; steering and driving. We used the nine different types of FOV setups (six sizes of FOV for linear and three sizes for nonlinear transformation). 60hx45v is the setup to assume a common video camera FOV. The transformed image is projected on the spherical screen display that has 120hx65v FOV. The setups over the FOV 120hx65v are applied to the nonlinear transformation and also the linear transformation to Figure 9. Results in real environments compare them. The angular resolutions are same in all FOV setups. Figs. 4-6 show the sample images displayed on the spherical screen display. Ten subjects tried every combination between task and FOV. Fig. 8 shows the experimental scene in the clank course. The time taken to perform the course together with the number of collisions was recorded. Fig. 9 shows a total result of all experiments. In this figure, horizontal axis indicates the field of view and vertical axis indicate the average, standard deviation of accomplished time and average number of collisions. Hairline and bold line indicate the linear transformation and the proposed nonlinear transformation methods, respectively. The averages and the standard deviations of accomplished time and the number of the average collisions were decreased when FOV was larger. It means task efficiency was improved by wide FOV. It shows the advantage for using wide FOV display and image sensor such as an omnidirectional image sensor. Over the 120hx65v setup, the average and the standard deviation were increased in case of the linear FOV transformation method. This was caused by the direction discordance.

5 Field of view[deg] Field of view[deg] Figure 11. Results of 2D navigation task in virtual environments Figure 10. Simulator Simulater view For example, if robot rotating 90[degree] in the remote environment, transformed image shows only 30[degree] rotation on display FOV in 360hx65v setup. Many subjects report that linearly transformed images were unnatural and difficult to understand the relation between robot heading direction and obstacles around the robots in the remote environments. Next, the accomplished time of the proposed nonlinear FOV transformation method was decreased until 270hx65v setup, and it was faster than linear FOV transformation. It shows that the nonlinear transformation method got more improvement of the tasks to keep the direction accordance and to present the wide FOV information over the display capacity. As shown in figure 9, totally 270hx65v with nonlinear FOV transformation was the best setup in sight of both the time and the collision on the tested setups and tasks. 5 Experiments on virtual environments In this section, we also evaluated the efficiency of wide FOV to robot navigation in virtual environments. It is effective to use the virtual environments for the evaluation, because it is easy to set up a complicated environment and tasks, such as 3D navigation and spacious environment. We constructed a simulator of robot navigation with a wide field of view for the experiments. The simulator was constructed by using OpenGL graphics library. Figure 10 shows a view of the simulator with 270hx65v FOV set up as an example. We use two Field of view[deg] Field of view[deg] Figure 12. Results of 3D navigation task in virtual environments types of tasks; 2D navigation task (similar to experiments of real environments in section 4 and 3D navigation task (assumed like an exploration of submarine). Both tasks are to move from a start to a goal point with avoiding the obstacles. The 3D navigation task required to overcome and to pass through under the obstacle objects. Motion of elevation was added in the 3D navigation task. We employed eight subjects for 2D and 3D navigation tasks. Experiments were carried out with the different view and FOV; 60hx45v and 120hx65v, nonlinear 180hx65v and 270hx65v, 360hx65v. Additionally the omnidirectional input image was also compared in these experiments, because some previous research [14] directly used the omnidirectional input image for tele-presence application. Figures 11 and 12 show the resultant time to perform the tasks and the number of collisions for 2D and 3D navigation tasks. The resultant times also show the standard deviations to indicate a scattering of samples. These results also show short time to accomplish the task and small number of collisions about 180hx65v and 270hx65v in the both tasks. It means that wide FOV is effective to a remote navigation task. The operation with an omnidirectional input image is also good for 2D task. However, the result with the omnidirectional image is

6 worst in the case of 3D task. It shows that the omnidirectional input image is easy to recognize the environment around the robot in the 2D task that only requires motion about horizontal plane. On the other hands, it is difficult to operate with the omnidirectional input image, because the omnidirectional image is hard to recognize the object height. So proposed wide FOV display method is effective on more general cases. 6 Conclusions To accomplish the tasks effectively and safely on tele-operation, high quality and realistic feed back information are displayed to the remote operator as if operator is in robot site. We think that FOV is the one of important factor of reality on visual feed back system on tele-operation. In this paper, we focus on the effect of FOV to operate the mobile robot. We construct the wide field of view tele-operation system. The slave robot system has the omnidirectional image sensor to take the panoramic image and the master system has spherical display that can display 120hx65v wide field of view distortionless image. Moreover, we propose FOV transformation method to present super wide field of view information over the display capacity. We evaluated efficiency of the tele-operation system and the nonlinear FOV transformation method. From experimental results both simulation and real environments, we confirmed the effect of wide field of view to accomplish the navigation task effectively and safely. The proposed nonlinear FOV transformation method was also effective at the setup of 270hx65v degree. However, we ignored another problem on tele-operating system such as communication delay etc. We will pursue to estimate more realistic situation and tasks. References: [1] C. R. Weisbin and D. Lavery, ``NASA Rover and Telerobotics Technology Program'', IEEE Robotics and Automation Magazine, pp.14-21, December, [2] S. S. Fisher et al.: ``Virtual environment display system'', ACM Workshop on Interactive 3D Graphics, pp.1-11, [3] J. D. Hightower, E. H. Spain et al.: ``Telepresence: A Hybrid approach to highperformance robots'', Proc. Int. Conf. Advanced Robotics, pp , [4] S. Tachi et al.: ``Tele-existance(I)-Design and evaluation of a visual display with sensation of presence-'', Proc. 5th Symposium Theory and Practice of Robots and Manipulators, pp , [5] S. Tachi and H. Arai: ``Study on teleexistance(ii)-three dimentional color dispaly with sensastion of presence'', Proc. Int. Conf. Advanced Robotics, pp , [6] J. D. Dickinson and J. A. Leonald: ``The role of peripheral vision in static balancing'', EGONOMICS, Vol.10, pp , [7] T. A. Furness: ``Creating better virtual worlds'', Technical Report HITL-M-89-3, [8] M. Takahashi, K. Arai and K. Yamamoto: ``Wide field of view using a 4LCD HMD is effective for postural control'', Second International Conference on Psychophysiology in Ergonomics, [9] D. G. Caldwell, K. Reddy, O. Kocak and A. Wardle: ``Sensory Requirements and Performance Assessment of Tele-Presence Controlled Robots'', Proc. IEEE Int. Conf. Robotics and Automation, pp , [10] Y. Onoe, K. Yamazawa, H. Takemura and N. Yokoya: ``Telepresence by real-time view dependent image generation from omnidirectional video stream'', Computer Vision and Image Understanding, Vol.71, No.2 pp , [11] Y. Manabe, K. Sato, K. Yamazawa, N. Yokoya and K. Chihara: ``Reproduction of motion for immersive mixed environments'', Proc. 1st Int. Conf. Image and Graphics, pp , [12] C. Cruz-Neira, D. J. Sandin and T. A. DeFanti: ``Surround-screen Projection-based Virtual Reality: The Design and Implementation of the CAVE'', Proc. ACM SIGGRAPH'93, pp , [13] K. Yamazawa, Y. Yagi and M. Yachida: ``New real-time omnidirectional image sensor with hyperboloidal mirror'', Proc. 8th Scandinavian Conf. Image Analysis, Vol.2, pp , [14] J. Baldwin, A. Basu, and H. Zhang: ``Panoramic Video with Predictive Windows for Telepresence Applications'', Proc. IEEE Int. Conf. Robotics and Automations, pp , 1999.

Real-time Generation and Presentation of View-dependent Binocular Stereo Images Using a Sequence of Omnidirectional Images

Real-time Generation and Presentation of View-dependent Binocular Stereo Images Using a Sequence of Omnidirectional Images Real-time Generation and Presentation of View-dependent Binocular Stereo Images Using a Sequence of Omnidirectional Images Abstract This paper presents a new method to generate and present arbitrarily

More information

Networked Telepresence System Using Web Browsers and Omni-directional Video Streams

Networked Telepresence System Using Web Browsers and Omni-directional Video Streams Networked Telepresence System Using Web Browsers and Omni-directional Video Streams Tomoya Ishikawa a, Kazumasa Yamazawa a, Tomokazu Sato a, Sei Ikeda a, Yutaka Nakamura a, Kazutoshi Fujikawa a, Hideki

More information

Non-isotropic Omnidirectional Imaging System for an Autonomous Mobile Robot

Non-isotropic Omnidirectional Imaging System for an Autonomous Mobile Robot Non-isotropic Omnidirectional Imaging System for an Autonomous Mobile Robot Kazuaki Kondo and Yasushi Yagi The Institute of Scientific and Industrial Research Osaka University Mihogaoka 8- Ibaragishi Osaka,

More information

Realtime Omnidirectional Stereo for Obstacle Detection and Tracking in Dynamic Environments

Realtime Omnidirectional Stereo for Obstacle Detection and Tracking in Dynamic Environments Proc. 2001 IEEE/RSJ Int. Conf. on Intelligent Robots and Systems pp. 31-36, Maui, Hawaii, Oct./Nov. 2001. Realtime Omnidirectional Stereo for Obstacle Detection and Tracking in Dynamic Environments Hiroshi

More information

A Networked Telepresence System with an Internet Automobile and its Transmission Issue

A Networked Telepresence System with an Internet Automobile and its Transmission Issue A Networked Telepresence System with an Internet Automobile and its Transmission Issue Kazutoshi Fujikawa 1 Kazumasa Yamazawa 2 Tomoya Ishikawa 2 Yuichi Butsuen 2 Tomokazu Sato 2 Yutaka Nakaramura 1 Hideki

More information

Development of Low-Cost Compact Omnidirectional Vision Sensors and their applications

Development of Low-Cost Compact Omnidirectional Vision Sensors and their applications Development of Low-Cost Compact Omnidirectional Vision Sensors and their applications Hiroshi ISHIGURO Department of Electrical & Computer Engineering, University of California, San Diego (9500 Gilman

More information

A System for Visualization and Summarization of Omnidirectional Surveillance Video

A System for Visualization and Summarization of Omnidirectional Surveillance Video A System for Visualization and Summarization of Omnidirectional Surveillance Video Noboru Babaguchi Yoshihiko Fujimoto Kazumasa Yamazawa + Naokazu Yokoya + * ISIR, Osaka University, Ibaraki, Osaka 567-0047,

More information

Dept. of Adaptive Machine Systems, Graduate School of Engineering Osaka University, Suita, Osaka , Japan

Dept. of Adaptive Machine Systems, Graduate School of Engineering Osaka University, Suita, Osaka , Japan An Application of Vision-Based Learning for a Real Robot in RoboCup - A Goal Keeping Behavior for a Robot with an Omnidirectional Vision and an Embedded Servoing - Sho ji Suzuki 1, Tatsunori Kato 1, Hiroshi

More information

Construction of an Immersive Mixed Environment Using an Omnidirectional Stereo Image Sensor

Construction of an Immersive Mixed Environment Using an Omnidirectional Stereo Image Sensor Construction of an Immersive Mixed Environment Using an Omnidirectional Stereo Image Sensor Jun Shimamura, Naokazu Yokoya, Haruo Takemura and Kazumasa Yamazawa Graduate School of Information Science Nara

More information

Outdoor Scene Reconstruction from Multiple Image Sequences Captured by a Hand-held Video Camera

Outdoor Scene Reconstruction from Multiple Image Sequences Captured by a Hand-held Video Camera Outdoor Scene Reconstruction from Multiple Image Sequences Captured by a Hand-held Video Camera Tomokazu Sato, Masayuki Kanbara and Naokazu Yokoya Graduate School of Information Science, Nara Institute

More information

Behavior Learning for a Mobile Robot with Omnidirectional Vision Enhanced by an Active Zoom Mechanism

Behavior Learning for a Mobile Robot with Omnidirectional Vision Enhanced by an Active Zoom Mechanism Behavior Learning for a Mobile Robot with Omnidirectional Vision Enhanced by an Active Zoom Mechanism Sho ji Suzuki, Tatsunori Kato, Minoru Asada, and Koh Hosoda Dept. of Adaptive Machine Systems, Graduate

More information

Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera

Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera Tomokazu Satoy, Masayuki Kanbaray, Naokazu Yokoyay and Haruo Takemuraz ygraduate School of Information

More information

Visual Servoing Utilizing Zoom Mechanism

Visual Servoing Utilizing Zoom Mechanism IEEE Int. Conf. on Robotics and Automation 1995, pp.178 183, Nagoya, May. 12 16, 1995 1 Visual Servoing Utilizing Zoom Mechanism Koh HOSODA, Hitoshi MORIYAMA and Minoru ASADA Dept. of Mechanical Engineering

More information

A Stereo Vision-based Mixed Reality System with Natural Feature Point Tracking

A Stereo Vision-based Mixed Reality System with Natural Feature Point Tracking A Stereo Vision-based Mixed Reality System with Natural Feature Point Tracking Masayuki Kanbara y, Hirofumi Fujii z, Haruo Takemura y and Naokazu Yokoya y ygraduate School of Information Science, Nara

More information

1 Introduction. 2 Real-time Omnidirectional Stereo

1 Introduction. 2 Real-time Omnidirectional Stereo J. of Robotics and Mechatronics, Vol. 14, 00 (to appear) Recognizing Moving Obstacles for Robot Navigation using Real-time Omnidirectional Stereo Vision Hiroshi Koyasu, Jun Miura, and Yoshiaki Shirai Dept.

More information

Image-Based Memory of Environment. homing uses a similar idea that the agent memorizes. [Hong 91]. However, the agent nds diculties in arranging its

Image-Based Memory of Environment. homing uses a similar idea that the agent memorizes. [Hong 91]. However, the agent nds diculties in arranging its Image-Based Memory of Environment Hiroshi ISHIGURO Department of Information Science Kyoto University Kyoto 606-01, Japan E-mail: ishiguro@kuis.kyoto-u.ac.jp Saburo TSUJI Faculty of Systems Engineering

More information

Motion Planning for Dynamic Knotting of a Flexible Rope with a High-speed Robot Arm

Motion Planning for Dynamic Knotting of a Flexible Rope with a High-speed Robot Arm The 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems October 18-22, 2010, Taipei, Taiwan Motion Planning for Dynamic Knotting of a Flexible Rope with a High-speed Robot Arm Yuji

More information

Construction Tele-robot System With Virtual Reality

Construction Tele-robot System With Virtual Reality Construction Tele-robot System With Virtual Reality Hironao Yamada Department of Mechanical & Systems Engineering Gifu University Gifu, Japan yamada@gifu-u.ac.jp Ni Tao Virtual System Laboratory Gifu University

More information

Simultaneous Tele-visualization of Construction Machine and Environment Using Body Mounted Cameras

Simultaneous Tele-visualization of Construction Machine and Environment Using Body Mounted Cameras Simultaneous Tele-visualization of Construction Machine and Environment Using Body Mounted Cameras Wei Sun, Soichiro Iwataki, Ren Komatsu, Hiromitsu Fujii, Atsushi Yamashita, and Hajime Asama Abstract

More information

Occlusion Detection of Real Objects using Contour Based Stereo Matching

Occlusion Detection of Real Objects using Contour Based Stereo Matching Occlusion Detection of Real Objects using Contour Based Stereo Matching Kenichi Hayashi, Hirokazu Kato, Shogo Nishida Graduate School of Engineering Science, Osaka University,1-3 Machikaneyama-cho, Toyonaka,

More information

3D Environment Measurement Using Binocular Stereo and Motion Stereo by Mobile Robot with Omnidirectional Stereo Camera

3D Environment Measurement Using Binocular Stereo and Motion Stereo by Mobile Robot with Omnidirectional Stereo Camera 3D Environment Measurement Using Binocular Stereo and Motion Stereo by Mobile Robot with Omnidirectional Stereo Camera Shinichi GOTO Department of Mechanical Engineering Shizuoka University 3-5-1 Johoku,

More information

Live Video Integration for High Presence Virtual World

Live Video Integration for High Presence Virtual World Live Video Integration for High Presence Virtual World Tetsuro OGI, Toshio YAMADA Gifu MVL Research Center, TAO IML, The University of Tokyo 2-11-16, Yayoi, Bunkyo-ku, Tokyo 113-8656, Japan Michitaka HIROSE

More information

1-2 Feature-Based Image Mosaicing

1-2 Feature-Based Image Mosaicing MVA'98 IAPR Workshop on Machine Vision Applications, Nov. 17-19, 1998, Makuhari, Chibq Japan 1-2 Feature-Based Image Mosaicing Naoki Chiba, Hiroshi Kano, Minoru Higashihara, Masashi Yasuda, and Masato

More information

Partial Calibration and Mirror Shape Recovery for Non-Central Catadioptric Systems

Partial Calibration and Mirror Shape Recovery for Non-Central Catadioptric Systems Partial Calibration and Mirror Shape Recovery for Non-Central Catadioptric Systems Abstract In this paper we present a method for mirror shape recovery and partial calibration for non-central catadioptric

More information

Localization of Wearable Users Using Invisible Retro-reflective Markers and an IR Camera

Localization of Wearable Users Using Invisible Retro-reflective Markers and an IR Camera Localization of Wearable Users Using Invisible Retro-reflective Markers and an IR Camera Yusuke Nakazato, Masayuki Kanbara and Naokazu Yokoya Graduate School of Information Science, Nara Institute of Science

More information

Fisheye Camera s Intrinsic Parameter Estimation Using Trajectories of Feature Points Obtained from Camera Rotation

Fisheye Camera s Intrinsic Parameter Estimation Using Trajectories of Feature Points Obtained from Camera Rotation Fisheye Camera s Intrinsic Parameter Estimation Using Trajectories of Feature Points Obtained from Camera Rotation Akihiko Hishigi, Yuki Tanaka, Gakuto Masuyama, and Kazunori Umeda Abstract This paper

More information

Room-sized Immersive Projection Display for Tele-immersion Environment

Room-sized Immersive Projection Display for Tele-immersion Environment Room-sized Immersive Projection Display for Tele-immersion Environment Tetsuro Ogi, Masahiro Hayashi, Mitsutaka Sakai University of Tsukuba tetsu@cc.tsukuba.ac.jp Abstract Although an immersive projection

More information

Relative Posture Estimation Using High Frequency Markers

Relative Posture Estimation Using High Frequency Markers The 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems October 18-22, 2010, Taipei, Taiwan Relative Posture Estimation Using High Frequency Markers Yuya Ono, Yoshio Iwai and Hiroshi

More information

A 100Hz Real-time Sensing System of Textured Range Images

A 100Hz Real-time Sensing System of Textured Range Images A 100Hz Real-time Sensing System of Textured Range Images Hidetoshi Ishiyama Course of Precision Engineering School of Science and Engineering Chuo University 1-13-27 Kasuga, Bunkyo-ku, Tokyo 112-8551,

More information

POSE ESTIMATION OF CONSTRUCTION MATERIALS USING MULTIPLE ID DEVICES FOR CONSTRUCTION AUTOMATION

POSE ESTIMATION OF CONSTRUCTION MATERIALS USING MULTIPLE ID DEVICES FOR CONSTRUCTION AUTOMATION POSE ESTIMATION OF CONSTRUCTION MATERIALS USING MULTIPLE ID DEVICES FOR CONSTRUCTION AUTOMATION Tomohiro UMETANI Graduate School of Design and Architecture Nagoya City University 2-1-10 Kitachikusa, Chikusa,

More information

SYSTEM FOR ACTIVE VIDEO OBSERVATION OVER THE INTERNET

SYSTEM FOR ACTIVE VIDEO OBSERVATION OVER THE INTERNET SYSTEM FOR ACTIVE VIDEO OBSERVATION OVER THE INTERNET Borut Batagelj, Peter Peer, Franc Solina University of Ljubljana Faculty of Computer and Information Science Computer Vision Laboratory Tržaška 25,

More information

AR Cultural Heritage Reconstruction Based on Feature Landmark Database Constructed by Using Omnidirectional Range Sensor

AR Cultural Heritage Reconstruction Based on Feature Landmark Database Constructed by Using Omnidirectional Range Sensor AR Cultural Heritage Reconstruction Based on Feature Landmark Database Constructed by Using Omnidirectional Range Sensor Takafumi Taketomi, Tomokazu Sato, and Naokazu Yokoya Graduate School of Information

More information

Security Monitoring around a Video Surveillance Car with a Pair of Two-camera Omnidirectional

Security Monitoring around a Video Surveillance Car with a Pair of Two-camera Omnidirectional Security Monitoring around a Video Surveillance Car with a Pair of Two-camera Omnidirectional Imaging Devices Pei-Hsuan Yuan Institute of Multimedia Engineering National Chiao Tung University Hsinchu,

More information

Projection simulator to support design development of spherical immersive display

Projection simulator to support design development of spherical immersive display Projection simulator to support design development of spherical immersive display Wataru Hashimoto, Yasuharu Mizutani, and Satoshi Nishiguchi Faculty of Information Sciences and Technology, Osaka Institute

More information

Remote Reality Demonstration

Remote Reality Demonstration Remote Reality Demonstration Terrance E. Boult EECS Dept., 19 Memorial Drive West Lehigh Univ., Bethlehem, PA 18015 tboult@eecs.lehigh.edu Fax: 610 758 6279 Contact Author: T.Boult Submission category:

More information

Natural Viewing 3D Display

Natural Viewing 3D Display We will introduce a new category of Collaboration Projects, which will highlight DoCoMo s joint research activities with universities and other companies. DoCoMo carries out R&D to build up mobile communication,

More information

Estimation of Camera Motion with Feature Flow Model for 3D Environment Modeling by Using Omni-Directional Camera

Estimation of Camera Motion with Feature Flow Model for 3D Environment Modeling by Using Omni-Directional Camera Estimation of Camera Motion with Feature Flow Model for 3D Environment Modeling by Using Omni-Directional Camera Ryosuke Kawanishi, Atsushi Yamashita and Toru Kaneko Abstract Map information is important

More information

Virtual Interaction System Based on Optical Capture

Virtual Interaction System Based on Optical Capture Sensors & Transducers 203 by IFSA http://www.sensorsportal.com Virtual Interaction System Based on Optical Capture Peng CHEN, 2 Xiaoyang ZHOU, 3 Jianguang LI, Peijun WANG School of Mechanical Engineering,

More information

Optical Flow-Based Person Tracking by Multiple Cameras

Optical Flow-Based Person Tracking by Multiple Cameras Proc. IEEE Int. Conf. on Multisensor Fusion and Integration in Intelligent Systems, Baden-Baden, Germany, Aug. 2001. Optical Flow-Based Person Tracking by Multiple Cameras Hideki Tsutsui, Jun Miura, and

More information

Partial Calibration and Mirror Shape Recovery for Non-Central Catadioptric Systems

Partial Calibration and Mirror Shape Recovery for Non-Central Catadioptric Systems Partial Calibration and Mirror Shape Recovery for Non-Central Catadioptric Systems Nuno Gonçalves and Helder Araújo Institute of Systems and Robotics - Coimbra University of Coimbra Polo II - Pinhal de

More information

Real-time Security Monitoring around a Video Surveillance Vehicle with a Pair of Two-camera Omni-imaging Devices

Real-time Security Monitoring around a Video Surveillance Vehicle with a Pair of Two-camera Omni-imaging Devices Real-time Security Monitoring around a Video Surveillance Vehicle with a Pair of Two-camera Omni-imaging Devices Pei-Hsuan Yuan, Kuo-Feng Yang and Wen-Hsiang Tsai, Senior Member, IEEE Abstract A pair of

More information

Ray tracing based fast refraction method for an object seen through a cylindrical glass

Ray tracing based fast refraction method for an object seen through a cylindrical glass 20th International Congress on Modelling and Simulation, Adelaide, Australia, 1 6 December 2013 www.mssanz.org.au/modsim2013 Ray tracing based fast refraction method for an object seen through a cylindrical

More information

3D Terrain Sensing System using Laser Range Finder with Arm-Type Movable Unit

3D Terrain Sensing System using Laser Range Finder with Arm-Type Movable Unit 3D Terrain Sensing System using Laser Range Finder with Arm-Type Movable Unit 9 Toyomi Fujita and Yuya Kondo Tohoku Institute of Technology Japan 1. Introduction A 3D configuration and terrain sensing

More information

DEVELOPMENT OF TELE-ROBOTIC INTERFACE SYSTEM FOR THE HOT-LINE MAINTENANCE. Chang-Hyun Kim, Min-Soeng Kim, Ju-Jang Lee,1

DEVELOPMENT OF TELE-ROBOTIC INTERFACE SYSTEM FOR THE HOT-LINE MAINTENANCE. Chang-Hyun Kim, Min-Soeng Kim, Ju-Jang Lee,1 DEVELOPMENT OF TELE-ROBOTIC INTERFACE SYSTEM FOR THE HOT-LINE MAINTENANCE Chang-Hyun Kim, Min-Soeng Kim, Ju-Jang Lee,1 Dept. of Electrical Engineering and Computer Science Korea Advanced Institute of Science

More information

SYNTHETIC VISION AND EMOTION CALCULATION IN INTELLIGENT VIRTUAL HUMAN MODELING

SYNTHETIC VISION AND EMOTION CALCULATION IN INTELLIGENT VIRTUAL HUMAN MODELING SYNTHETIC VISION AND EMOTION CALCULATION IN INTELLIGENT VIRTUAL HUMAN MODELING Y. Zhao, J. Kang and D. K. Wright School of Engineering and Design, Brunel University, Uxbridge, Middlesex, UB8 3PH, UK ABSTRACT

More information

Ego-Mot ion and Omnidirectional Cameras*

Ego-Mot ion and Omnidirectional Cameras* Ego-Mot ion and Omnidirectional Cameras* Joshua Gluckman and Shree K. Nayar Department of Computer Science Columbia University New York, New York 10027 Abstract Recent research in image sensors has produced

More information

Proc. 14th Int. Conf. on Intelligent Autonomous Systems (IAS-14), 2016

Proc. 14th Int. Conf. on Intelligent Autonomous Systems (IAS-14), 2016 Proc. 14th Int. Conf. on Intelligent Autonomous Systems (IAS-14), 2016 Outdoor Robot Navigation Based on View-based Global Localization and Local Navigation Yohei Inoue, Jun Miura, and Shuji Oishi Department

More information

CANAL FOLLOWING USING AR DRONE IN SIMULATION

CANAL FOLLOWING USING AR DRONE IN SIMULATION CANAL FOLLOWING USING AR DRONE IN SIMULATION ENVIRONMENT Ali Ahmad, Ahmad Aneeque Khalid Department of Electrical Engineering SBA School of Science & Engineering, LUMS, Pakistan {14060006, 14060019}@lums.edu.pk

More information

(a) (b) (c) Fig. 1. Omnidirectional camera: (a) principle; (b) physical construction; (c) captured. of a local vision system is more challenging than

(a) (b) (c) Fig. 1. Omnidirectional camera: (a) principle; (b) physical construction; (c) captured. of a local vision system is more challenging than An Omnidirectional Vision System that finds and tracks color edges and blobs Felix v. Hundelshausen, Sven Behnke, and Raul Rojas Freie Universität Berlin, Institut für Informatik Takustr. 9, 14195 Berlin,

More information

Sketch-based Interface for Crowd Animation

Sketch-based Interface for Crowd Animation Sketch-based Interface for Crowd Animation Masaki Oshita 1, Yusuke Ogiwara 1 1 Kyushu Institute of Technology 680-4 Kawazu, Iizuka, Fukuoka, 820-8502, Japan oshita@ces.kyutech.ac.p ogiwara@cg.ces.kyutech.ac.p

More information

Augmenting Reality with Projected Interactive Displays

Augmenting Reality with Projected Interactive Displays Augmenting Reality with Projected Interactive Displays Claudio Pinhanez IBM T.J. Watson Research Center, P.O. Box 218 Yorktown Heights, N.Y. 10598, USA Abstract. This paper examines a steerable projection

More information

DEVELOPMENT OF REAL TIME 3-D MEASUREMENT SYSTEM USING INTENSITY RATIO METHOD

DEVELOPMENT OF REAL TIME 3-D MEASUREMENT SYSTEM USING INTENSITY RATIO METHOD DEVELOPMENT OF REAL TIME 3-D MEASUREMENT SYSTEM USING INTENSITY RATIO METHOD Takeo MIYASAKA and Kazuo ARAKI Graduate School of Computer and Cognitive Sciences, Chukyo University, Japan miyasaka@grad.sccs.chukto-u.ac.jp,

More information

Human trajectory tracking using a single omnidirectional camera

Human trajectory tracking using a single omnidirectional camera Human trajectory tracking using a single omnidirectional camera Atsushi Kawasaki, Dao Huu Hung and Hideo Saito Graduate School of Science and Technology Keio University 3-14-1, Hiyoshi, Kohoku-Ku, Yokohama,

More information

Fast Local Planner for Autonomous Helicopter

Fast Local Planner for Autonomous Helicopter Fast Local Planner for Autonomous Helicopter Alexander Washburn talexan@seas.upenn.edu Faculty advisor: Maxim Likhachev April 22, 2008 Abstract: One challenge of autonomous flight is creating a system

More information

Geolocation with FW 6.4x & Video Security Client Geolocation with FW 6.4x & Video Security Client 2.1 Technical Note

Geolocation with FW 6.4x & Video Security Client Geolocation with FW 6.4x & Video Security Client 2.1 Technical Note Geolocation with FW 6.4x & Video Security Client 2.1 1 10 Geolocation with FW 6.4x & Video Security Client 2.1 Technical Note Geolocation with FW 6.4x & Video Security Client 2.1 2 10 Table of contents

More information

MOTION TRAJECTORY PLANNING AND SIMULATION OF 6- DOF MANIPULATOR ARM ROBOT

MOTION TRAJECTORY PLANNING AND SIMULATION OF 6- DOF MANIPULATOR ARM ROBOT MOTION TRAJECTORY PLANNING AND SIMULATION OF 6- DOF MANIPULATOR ARM ROBOT Hongjun ZHU ABSTRACT:In order to better study the trajectory of robot motion, a motion trajectory planning and simulation based

More information

Mobile Robot Navigation Using Omnidirectional Vision

Mobile Robot Navigation Using Omnidirectional Vision Mobile Robot Navigation Using Omnidirectional Vision László Mornailla, Tamás Gábor Pekár, Csaba Gergő Solymosi, Zoltán Vámossy John von Neumann Faculty of Informatics, Budapest Tech Bécsi út 96/B, H-1034

More information

Automatic Disparity Control in Stereo Panoramas (OmniStereo) Λ

Automatic Disparity Control in Stereo Panoramas (OmniStereo) Λ Automatic Disparity Control in Stereo Panoramas (OmniStereo) Λ Yael Pritch Moshe Ben-Ezra Shmuel Peleg School of Computer Science and Engineering The Hebrew University of Jerusalem 91904 Jerusalem, ISRAEL

More information

DEVELOPMENT OF POSITION MEASUREMENT SYSTEM FOR CONSTRUCTION PILE USING LASER RANGE FINDER

DEVELOPMENT OF POSITION MEASUREMENT SYSTEM FOR CONSTRUCTION PILE USING LASER RANGE FINDER S17- DEVELOPMENT OF POSITION MEASUREMENT SYSTEM FOR CONSTRUCTION PILE USING LASER RANGE FINDER Fumihiro Inoue 1 *, Takeshi Sasaki, Xiangqi Huang 3, and Hideki Hashimoto 4 1 Technica Research Institute,

More information

Extrinsic Camera Parameter Recovery from Multiple Image Sequences Captured by an Omni-directional Multi-camera System

Extrinsic Camera Parameter Recovery from Multiple Image Sequences Captured by an Omni-directional Multi-camera System Extrinsic Camera Parameter Recovery from Multiple Image Sequences Captured by an Omni-directional Multi-camera System Tomokazu Sato, Sei Ikeda, and Naokazu Yokoya Graduate School of Information Science,

More information

Measurement of Pedestrian Groups Using Subtraction Stereo

Measurement of Pedestrian Groups Using Subtraction Stereo Measurement of Pedestrian Groups Using Subtraction Stereo Kenji Terabayashi, Yuki Hashimoto, and Kazunori Umeda Chuo University / CREST, JST, 1-13-27 Kasuga, Bunkyo-ku, Tokyo 112-8551, Japan terabayashi@mech.chuo-u.ac.jp

More information

Robotics Project. Final Report. Computer Science University of Minnesota. December 17, 2007

Robotics Project. Final Report. Computer Science University of Minnesota. December 17, 2007 Robotics Project Final Report Computer Science 5551 University of Minnesota December 17, 2007 Peter Bailey, Matt Beckler, Thomas Bishop, and John Saxton Abstract: A solution of the parallel-parking problem

More information

A Simple Interface for Mobile Robot Equipped with Single Camera using Motion Stereo Vision

A Simple Interface for Mobile Robot Equipped with Single Camera using Motion Stereo Vision A Simple Interface for Mobile Robot Equipped with Single Camera using Motion Stereo Vision Stephen Karungaru, Atsushi Ishitani, Takuya Shiraishi, and Minoru Fukumi Abstract Recently, robot technology has

More information

Compositing a bird's eye view mosaic

Compositing a bird's eye view mosaic Compositing a bird's eye view mosaic Robert Laganiere School of Information Technology and Engineering University of Ottawa Ottawa, Ont KN 6N Abstract This paper describes a method that allows the composition

More information

VISUAL NAVIGATION SYSTEM ON WINDSHIELD HEAD-UP DISPLAY

VISUAL NAVIGATION SYSTEM ON WINDSHIELD HEAD-UP DISPLAY VISUAL NAVIGATION SYSTEM ON WINDSHIELD HEAD-UP DISPLAY Akihiko SATO *, Itaru KITAHARA, Yoshinari KAMEDA, Yuichi OHTA Department of Intelligent Interaction Technologies, Graduate School of Systems and Information

More information

Variable-resolution Velocity Roadmap Generation Considering Safety Constraints for Mobile Robots

Variable-resolution Velocity Roadmap Generation Considering Safety Constraints for Mobile Robots Variable-resolution Velocity Roadmap Generation Considering Safety Constraints for Mobile Robots Jingyu Xiang, Yuichi Tazaki, Tatsuya Suzuki and B. Levedahl Abstract This research develops a new roadmap

More information

APPLICATIONS FOR COORDINATED MOTION IN RADOME TESTING

APPLICATIONS FOR COORDINATED MOTION IN RADOME TESTING APPLICATIONS FOR COORDINATED MOTION IN RADOME TESTING Scott McBride Evan Langman Marion Baggett MI Technologies, LLC 4500 River Green Parkway, Suite 200 Duluth, GA 30096 ABSTRACT Traditional data collection

More information

Tele-operation Construction Robot Control System with Virtual Reality Technology

Tele-operation Construction Robot Control System with Virtual Reality Technology Available online at www.sciencedirect.com Procedia Engineering 15 (2011) 1071 1076 Advanced in Control Engineering and Information Science Tele-operation Construction Robot Control System with Virtual

More information

A Fast Linear Registration Framework for Multi-Camera GIS Coordination

A Fast Linear Registration Framework for Multi-Camera GIS Coordination A Fast Linear Registration Framework for Multi-Camera GIS Coordination Karthik Sankaranarayanan James W. Davis Dept. of Computer Science and Engineering Ohio State University Columbus, OH 4320 USA {sankaran,jwdavis}@cse.ohio-state.edu

More information

Omni Stereo Vision of Cooperative Mobile Robots

Omni Stereo Vision of Cooperative Mobile Robots Omni Stereo Vision of Cooperative Mobile Robots Zhigang Zhu*, Jizhong Xiao** *Department of Computer Science **Department of Electrical Engineering The City College of the City University of New York (CUNY)

More information

Development of Vision System on Humanoid Robot HRP-2

Development of Vision System on Humanoid Robot HRP-2 Development of Vision System on Humanoid Robot HRP-2 Yutaro Fukase Institute of Technology, Shimizu Corporation, Japan fukase@shimz.co.jp Junichiro Maeda Institute of Technology, Shimizu Corporation, Japan

More information

Introduction to 3D Concepts

Introduction to 3D Concepts PART I Introduction to 3D Concepts Chapter 1 Scene... 3 Chapter 2 Rendering: OpenGL (OGL) and Adobe Ray Tracer (ART)...19 1 CHAPTER 1 Scene s0010 1.1. The 3D Scene p0010 A typical 3D scene has several

More information

VISUAL NAVIGATION SYSTEM ON WINDSHIELD HEAD-UP DISPLAY

VISUAL NAVIGATION SYSTEM ON WINDSHIELD HEAD-UP DISPLAY VISUAL NAVIGATION SYSTEM ON WINDSHIELD HEAD-UP DISPLAY Akihiko SATO *, Itaru KITAHARA, Yoshinari KAMEDA, Yuichi OHTA Department of Intelligent Interaction Technologies, Graduate School of Systems and Information

More information

cse 252c Fall 2004 Project Report: A Model of Perpendicular Texture for Determining Surface Geometry

cse 252c Fall 2004 Project Report: A Model of Perpendicular Texture for Determining Surface Geometry cse 252c Fall 2004 Project Report: A Model of Perpendicular Texture for Determining Surface Geometry Steven Scher December 2, 2004 Steven Scher SteveScher@alumni.princeton.edu Abstract Three-dimensional

More information

Pose Estimation of Construction Materials by Acquisition of Multiple IDs of Devices

Pose Estimation of Construction Materials by Acquisition of Multiple IDs of Devices 11 Pose Estimation of Construction Materials by Acquisition of Multiple IDs of Devices Tomohiro Umetani 1, Kenji Inoue 2 and Tatsuo Arai 3 1 Department of Intelligence and Informatics, Konan University

More information

MOTION. Feature Matching/Tracking. Control Signal Generation REFERENCE IMAGE

MOTION. Feature Matching/Tracking. Control Signal Generation REFERENCE IMAGE Head-Eye Coordination: A Closed-Form Solution M. Xie School of Mechanical & Production Engineering Nanyang Technological University, Singapore 639798 Email: mmxie@ntuix.ntu.ac.sg ABSTRACT In this paper,

More information

Cooperative Conveyance of an Object with Tethers by Two Mobile Robots

Cooperative Conveyance of an Object with Tethers by Two Mobile Robots Proceeding of the 11th World Congress in Mechanism and Machine Science April 1-4, 2004, Tianjin, China China Machine Press, edited by Tian Huang Cooperative Conveyance of an Object with Tethers by Two

More information

Estimating Camera Position And Posture by Using Feature Landmark Database

Estimating Camera Position And Posture by Using Feature Landmark Database Estimating Camera Position And Posture by Using Feature Landmark Database Motoko Oe 1, Tomokazu Sato 2 and Naokazu Yokoya 2 1 IBM Japan 2 Nara Institute of Science and Technology, Japan Abstract. Estimating

More information

Continuous Valued Q-learning for Vision-Guided Behavior Acquisition

Continuous Valued Q-learning for Vision-Guided Behavior Acquisition Continuous Valued Q-learning for Vision-Guided Behavior Acquisition Yasutake Takahashi, Masanori Takeda, and Minoru Asada Dept. of Adaptive Machine Systems Graduate School of Engineering Osaka University

More information

On Road Vehicle Detection using Shadows

On Road Vehicle Detection using Shadows On Road Vehicle Detection using Shadows Gilad Buchman Grasp Lab, Department of Computer and Information Science School of Engineering University of Pennsylvania, Philadelphia, PA buchmag@seas.upenn.edu

More information

DETECTION OF 3D POINTS ON MOVING OBJECTS FROM POINT CLOUD DATA FOR 3D MODELING OF OUTDOOR ENVIRONMENTS

DETECTION OF 3D POINTS ON MOVING OBJECTS FROM POINT CLOUD DATA FOR 3D MODELING OF OUTDOOR ENVIRONMENTS DETECTION OF 3D POINTS ON MOVING OBJECTS FROM POINT CLOUD DATA FOR 3D MODELING OF OUTDOOR ENVIRONMENTS Tsunetake Kanatani,, Hideyuki Kume, Takafumi Taketomi, Tomokazu Sato and Naokazu Yokoya Hyogo Prefectural

More information

Stereo Rectification for Equirectangular Images

Stereo Rectification for Equirectangular Images Stereo Rectification for Equirectangular Images Akira Ohashi, 1,2 Fumito Yamano, 1 Gakuto Masuyama, 1 Kazunori Umeda, 1 Daisuke Fukuda, 2 Kota Irie, 3 Shuzo Kaneko, 2 Junya Murayama, 2 and Yoshitaka Uchida

More information

Virtual Human with regard to Physical contact and Eye contact

Virtual Human with regard to Physical contact and Eye contact Virtual Human with regard to Physical contact and Eye contact Asami Takayama 1, Yusuke Sugimoto 1, Akio Okuie1, Tomoya Suzuki 1 and Kiyotaka Kato 1 1 Tokyo Univ. of Science,1-3 Kagurazaka, Shinjuku-ku,Tokyo

More information

A Framework to Evaluate Omnidirectional Video Coding Schemes

A Framework to Evaluate Omnidirectional Video Coding Schemes 2015 IEEE International Symposium on Mixed and Augmented Reality A Framework to Evaluate Omnidirectional Video Coding Schemes Matt Yu Haricharan Lakshman Department of Electrical Engineering Stanford University

More information

Calibration of Lens Distortion for Super-Wide-Angle Stereo Vision

Calibration of Lens Distortion for Super-Wide-Angle Stereo Vision Calibration of Lens Distortion for Super-Wide-Angle Stereo Vision Hajime KAWANISHI, Yoshitaka HARA, Takashi TSUBOUCHI, Akihisa OHYA Graduate School of Systems and Information Engineering, University of

More information

Wide area tracking method for augmented reality supporting nuclear power plant maintenance work

Wide area tracking method for augmented reality supporting nuclear power plant maintenance work Journal of Marine Science and Application, Vol.6, No.1, January 2006, PP***-*** Wide area tracking method for augmented reality supporting nuclear power plant maintenance work ISHII Hirotake 1, YAN Weida

More information

Precise Omnidirectional Camera Calibration

Precise Omnidirectional Camera Calibration Precise Omnidirectional Camera Calibration Dennis Strelow, Jeffrey Mishler, David Koes, and Sanjiv Singh Carnegie Mellon University {dstrelow, jmishler, dkoes, ssingh}@cs.cmu.edu Abstract Recent omnidirectional

More information

THE POSITION AND ORIENTATION MEASUREMENT OF GONDOLA USING A VISUAL CAMERA

THE POSITION AND ORIENTATION MEASUREMENT OF GONDOLA USING A VISUAL CAMERA THE POSITION AND ORIENTATION MEASUREMENT OF GONDOLA USING A VISUAL CAMERA Hwadong Sun 1, Dong Yeop Kim 1 *, Joon Ho Kwon 2, Bong-Seok Kim 1, and Chang-Woo Park 1 1 Intelligent Robotics Research Center,

More information

High-speed Three-dimensional Mapping by Direct Estimation of a Small Motion Using Range Images

High-speed Three-dimensional Mapping by Direct Estimation of a Small Motion Using Range Images MECATRONICS - REM 2016 June 15-17, 2016 High-speed Three-dimensional Mapping by Direct Estimation of a Small Motion Using Range Images Shinta Nozaki and Masashi Kimura School of Science and Engineering

More information

JUST-IN-TIME PIXELS. Mark Mine and Gary Bishop. Department of Computer Science University of North Carolina Chapel Hill, NC

JUST-IN-TIME PIXELS. Mark Mine and Gary Bishop. Department of Computer Science University of North Carolina Chapel Hill, NC JUST-IN-TIME PIXELS Mark Mine and Gary Bishop Department of Computer Science University of North Carolina Chapel Hill, NC 27599-3175 Abstract This paper describes Just-In-Time Pixels, a technique for generating

More information

9th Intelligent Ground Vehicle Competition. Design Competition Written Report. Design Change Report AMIGO

9th Intelligent Ground Vehicle Competition. Design Competition Written Report. Design Change Report AMIGO 9th Intelligent Ground Vehicle Competition Design Competition Written Report Design Change Report AMIGO AMIGO means the friends who will join to the IGV Competition. Watanabe Laboratory Team System Control

More information

The Uncertainty of Parallel Model Coordinate Measuring Machine

The Uncertainty of Parallel Model Coordinate Measuring Machine MAPAN - Journal of The Metrology Society of of Parallel India, Vol. Model 26, Coordinate No. 1, 2011; Measuring pp. 47-53 Machine ORIGINAL ARTICLE The of Parallel Model Coordinate Measuring Machine KOSUKE

More information

MR-Mirror: A Complex of Real and Virtual Mirrors

MR-Mirror: A Complex of Real and Virtual Mirrors MR-Mirror: A Complex of Real and Virtual Mirrors Hideaki Sato 1, Itaru Kitahara 1, and Yuichi Ohta 1 1 Department of Intelligent Interaction Technologies, Graduate School of Systems and Information Engineering,

More information

Vol. 21 No. 6, pp ,

Vol. 21 No. 6, pp , Vol. 21 No. 6, pp.69 696, 23 69 3 3 3 Map Generation of a Mobile Robot by Integrating Omnidirectional Stereo and Laser Range Finder Yoshiro Negishi 3, Jun Miura 3 and Yoshiaki Shirai 3 This paper describes

More information

Proc. Int. Symp. Robotics, Mechatronics and Manufacturing Systems 92 pp , Kobe, Japan, September 1992

Proc. Int. Symp. Robotics, Mechatronics and Manufacturing Systems 92 pp , Kobe, Japan, September 1992 Proc. Int. Symp. Robotics, Mechatronics and Manufacturing Systems 92 pp.957-962, Kobe, Japan, September 1992 Tracking a Moving Object by an Active Vision System: PANTHER-VZ Jun Miura, Hideharu Kawarabayashi,

More information

Using Artificial Neural Networks for Prediction Of Dynamic Human Motion

Using Artificial Neural Networks for Prediction Of Dynamic Human Motion ABSTRACT Using Artificial Neural Networks for Prediction Of Dynamic Human Motion Researchers in robotics and other human-related fields have been studying human motion behaviors to understand and mimic

More information

Published Technical Disclosure. Camera-Mirror System on a Remotely Operated Vehicle or Machine Authors: Harald Staab, Carlos Martinez and Biao Zhang

Published Technical Disclosure. Camera-Mirror System on a Remotely Operated Vehicle or Machine Authors: Harald Staab, Carlos Martinez and Biao Zhang Background: Camera-Mirror System on a Remotely Operated Vehicle or Machine Authors: Harald Staab, Carlos Martinez and Biao Zhang This innovation relates to situational awareness for a remote operator of

More information

1998 IEEE International Conference on Intelligent Vehicles 213

1998 IEEE International Conference on Intelligent Vehicles 213 Navigation by Integrating Iconic and GPS Information Shigang Li and Akira Hayashi Faculty of Information Sciences Hiroshima City University Asaminami-ku, Hiroshima, 731-31, Japan li@im.hiroshima-cu.ac.jp

More information

Mathematics of a Multiple Omni-Directional System

Mathematics of a Multiple Omni-Directional System Mathematics of a Multiple Omni-Directional System A. Torii A. Sugimoto A. Imiya, School of Science and National Institute of Institute of Media and Technology, Informatics, Information Technology, Chiba

More information

Chapter 3 Image Registration. Chapter 3 Image Registration

Chapter 3 Image Registration. Chapter 3 Image Registration Chapter 3 Image Registration Distributed Algorithms for Introduction (1) Definition: Image Registration Input: 2 images of the same scene but taken from different perspectives Goal: Identify transformation

More information