Dept. of Adaptive Machine Systems, Graduate School of Engineering Osaka University, Suita, Osaka , Japan

Similar documents
Behavior Learning for a Mobile Robot with Omnidirectional Vision Enhanced by an Active Zoom Mechanism

Continuous Valued Q-learning for Vision-Guided Behavior Acquisition

BabyTigers-98: Osaka Legged Robot Team

Visual Attention Control by Sensor Space Segmentation for a Small Quadruped Robot based on Information Criterion

Realtime Omnidirectional Stereo for Obstacle Detection and Tracking in Dynamic Environments

Development of Low-Cost Compact Omnidirectional Vision Sensors and their applications

Visual Tracking of Unknown Moving Object by Adaptive Binocular Visual Servoing

A Real Time Vision System for Robotic Soccer

Visual Servoing Utilizing Zoom Mechanism

Motion Control in Dynamic Multi-Robot Environments

Proc. Int. Symp. Robotics, Mechatronics and Manufacturing Systems 92 pp , Kobe, Japan, September 1992

Chapter 3 Image Registration. Chapter 3 Image Registration

CIS 580, Machine Perception, Spring 2015 Homework 1 Due: :59AM

Eagle Knights: RoboCup Small Size League. Dr. Alfredo Weitzenfeld ITAM - Mexico

Robust and Accurate Detection of Object Orientation and ID without Color Segmentation

Robust Color Choice for Small-size League RoboCup Competition

(a) (b) (c) Fig. 1. Omnidirectional camera: (a) principle; (b) physical construction; (c) captured. of a local vision system is more challenging than

NuBot Team Description Paper 2013

Improving Vision-Based Distance Measurements using Reference Objects

Measurement of Pedestrian Groups Using Subtraction Stereo

Ball tracking with velocity based on Monte-Carlo localization

Non-isotropic Omnidirectional Imaging System for an Autonomous Mobile Robot

Image-Based Memory of Environment. homing uses a similar idea that the agent memorizes. [Hong 91]. However, the agent nds diculties in arranging its

126 ACTION AND PERCEPTION

Particle-Filter-Based Self-Localization Using Landmarks and Directed Lines

Self-calibration of a pair of stereo cameras in general position

Motion Recognition and Generation for Humanoid based on Visual-Somatic Field Mapping

A Framework for 3D Pushbroom Imaging CUCS

Horus: Object Orientation and Id without Additional Markers

Omni-Directional Visual Servoing for Human-Robot Interaction

Real-time Generation and Presentation of View-dependent Binocular Stereo Images Using a Sequence of Omnidirectional Images

Acquisition of Qualitative Spatial Representation by Visual Observation

An Efficient Need-Based Vision System in Variable Illumination Environment of Middle Size RoboCup

Optical Flow-Based Person Tracking by Multiple Cameras

Panoramic 3D Reconstruction Using Rotational Stereo Camera with Simple Epipolar Constraints

Soccer Server and Researches on Multi-Agent Systems. NODA, Itsuki and MATSUBARA, Hitoshi ETL Umezono, Tsukuba, Ibaraki 305, Japan

A 100Hz Real-time Sensing System of Textured Range Images

Chapter 23. Geometrical Optics (lecture 1: mirrors) Dr. Armen Kocharian

Catadioptric camera model with conic mirror

Development of Vision System on Humanoid Robot HRP-2

SYSTEM FOR ACTIVE VIDEO OBSERVATION OVER THE INTERNET

Super Wide Viewing for Tele-operation

Tracking Multiple Objects in 3D. Coimbra 3030 Coimbra target projection in the same image position (usually

3D Environment Measurement Using Binocular Stereo and Motion Stereo by Mobile Robot with Omnidirectional Stereo Camera

Fisheye Camera s Intrinsic Parameter Estimation Using Trajectories of Feature Points Obtained from Camera Rotation

Decision Algorithm for Pool Using Fuzzy System

3D Terrain Sensing System using Laser Range Finder with Arm-Type Movable Unit

A Fast Linear Registration Framework for Multi-Camera GIS Coordination

Perceptual Attention Through Image Feature Generation Based on Visio-Motor Map Learning

Chapter 26 Geometrical Optics

Precise Omnidirectional Camera Calibration

Autodesk Fusion 360 Training: The Future of Making Things Attendee Guide

arxiv:cs/ v1 [cs.cv] 24 Mar 2003

A Fuzzy Reinforcement Learning for a Ball Interception Problem

Mathematics of a Multiple Omni-Directional System

Three-Dimensional Measurement of Objects in Liquid with an Unknown Refractive Index Using Fisheye Stereo Camera

Autonomous Programming FTC Challenge Workshops VCU School of Engineering September 24, 2016 Presented by: Team 8297 Geared UP!

Vol. 21 No. 6, pp ,

Grasping Known Objects with Aldebaran Nao

Perspective Projection [2 pts]

SHARPKUNGFU TEAM DESCRIPTION 2006

Chapter 26 Geometrical Optics

600 M. Jamzad, A.R. Hadjkhodabakhshi and V.S. Mirrokni Figure 1. A picture of the middle size RoboCup eld in is given in [3]. Detailed theories

Motion Sketch: Acquisition of Visual Motion Guided Behaviors

Estimating Camera Position And Posture by Using Feature Landmark Database

Reflection and Image Formation by Mirrors

Visuo-Motor Learning for Face-to-face Pass between Heterogeneous Humanoids

Omni Stereo Vision of Cooperative Mobile Robots

State Estimation for Continuous-Time Systems with Perspective Outputs from Discrete Noisy Time-Delayed Measurements

Ego-Mot ion and Omnidirectional Cameras*

2. TARGET PHOTOS FOR ANALYSIS

Image Feature Generation by Visio-Motor Map Learning towards Selective Attention

THE UNIVERSITY OF AUCKLAND

Topics and things to know about them:

Optimized Design of 3D Laser Triangulation Systems

DD2429 Computational Photography :00-19:00

Chapter 12 3D Localisation and High-Level Processing

Baseline Detection and Localization for Invisible Omnidirectional Cameras

Optical design of COrE+

Terrain Data Real-time Analysis Based on Point Cloud for Mars Rover

A Simple Interface for Mobile Robot Equipped with Single Camera using Motion Stereo Vision

Production of Video Images by Computer Controlled Cameras and Its Application to TV Conference System

Sensor and information fusion applied to a Robotic Soccer Team

Motion Planning for Dynamic Knotting of a Flexible Rope with a High-speed Robot Arm

DESIGN AND IMPLEMENTATION OF VISUAL FEEDBACK FOR AN ACTIVE TRACKING

Automatic Generation of Indoor VR-Models by a Mobile Robot with a Laser Range Finder and a Color Camera

Hidefumi Wakamatsu, Yuusuke Tanaka, Akira Tsumaya, Keiichi Shirase, and Eiji Arai

METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS

Light: Geometric Optics

arxiv: v1 [cs.cv] 28 Sep 2018

Estimation of Camera Motion with Feature Flow Model for 3D Environment Modeling by Using Omni-Directional Camera

3D Rendering and Ray Casting

1 Introduction. 2 Real-time Omnidirectional Stereo

DEVELOPMENT OF REAL TIME 3-D MEASUREMENT SYSTEM USING INTENSITY RATIO METHOD

CIS 580, Machine Perception, Spring 2016 Homework 2 Due: :59AM

A Robust Two Feature Points Based Depth Estimation Method 1)

Eagle Knights 2007: Four-Legged League

SketchUp. SketchUp. Google SketchUp. Using SketchUp. The Tool Set

Time-to-Contact from Image Intensity

DEVELOPMENT OF POSITION MEASUREMENT SYSTEM FOR CONSTRUCTION PILE USING LASER RANGE FINDER

Transcription:

An Application of Vision-Based Learning for a Real Robot in RoboCup - A Goal Keeping Behavior for a Robot with an Omnidirectional Vision and an Embedded Servoing - Sho ji Suzuki 1, Tatsunori Kato 1, Hiroshi Ishizuka 1, Yasutake Takahashi 1, Eiji Uchibe 1, and Minoru Asada 1 Dept. of Adaptive Machine Systems, Graduate School of Engineering Osaka University, Suita, Osaka 565-0871, Japan Abstract. We have applied Q-learning method so that a mobile robot acquire a shooting behavior of a soccer game. Thought the acquired behavior is tested in RoboCup-97, the performance is limited because of the narrow visual angle of the robot. Therefore, we install an omnidirectional vision into a mobile robot to enlarge a visual angle and apply Q-learning to behavior acquisition of the robot. An attention control method is also proposed for an omnidirectional vision system by means of an active zoom mechanism to reduce the learning time. We select a a goal defending behavior as a task and perform computer simulation and real robot experiments to show the validity of the proposed system. 1 Introduction One of major issues in RoboCup is to make a robot adapt itself to changes in dynamic environments. We have approached to this problem with a reinforcement learning method, Q-learning. We have selected a shooting behavior as an example task by a mobile robot with a single camera [6] and tested four robots in RoboCup-97 [7]. The performance of the behavior shown there is limited by several reasons; noises in image, noises in motor control, limitation of image processing, perceptual aliasing, hidden states, and so on. Especially, the robots easily lose the rolling ball and is difficult to find out, since they have only fixed cameras and their view angles are narrow. To enlarge a visual angle we have developed two types of robots. One has an omnidirectional vision system which can capture the whole view all around the imaging system at any instant in time. Another one has an active camera which can control the pan angle of the camera. Figure 1 shows our robots for RoboCup-98 which are categorized into three types; a fixed camera type used in RoboCup-97 (upper left), an omnidirectional vision type (upper right), and an active camera type (lower). The one with an omnidirectional vision will be used as a keeper since it can look at the ball and the goal simultaneously. Other two types of robots will be used as an attacker. In this paper, we particularly describe about learning of a goal defending behavior by a keeper robot with an ominidirectional vision system.

Fig. 1. Robots of University Osaka Trackies team Several applications of the ominidirectional vision have been proposed, such as autonomous navigation [1], visual surveillance and guidance [2], video conference, virtual reality, and site modeling [3]. These methods have focused on its opto-geometric features to reconstruct 3-D scene structure. Our approach differs from their applications in two fold: we don to reconstruct any geometric structure from the omnidirectional views. Rather, we use it as a sensory system for a goal defending mobile robot. As a method of behavior acquisition, Q-learning is used with a state space consisting of ball and goal images. We introduce an omnidirectional vision system with an active zoom mechanism to accelerate the learning. We implement a zoom servoing [4] by which the target image can be captured at the constant position if the target moves on the ground plane. It is realized by controlling focal length of the camera. Due to the active zoom mechanism, the target motion in the radius direction can be canceled, and only circular motions around the image center can be observed. This simplifies the image processing and target tracking. We apply the learning method [6] to a mobile robot with the enhanced omnidirectional vision system and the state space is designed for it. The rest of paper consists as follows. First, we examine the relationship between the target position in the omnidirectional view in terms of the focal length and the dis-

Front Back tance between the robot and the target. Then, we set up a control low to realize a zoom servoing, Finally, we design the state space for the robot to acquire the desired behavior based on the reinforcement learning scheme. 2 The Task and the Robot System for a Goal Keeping Behavior Right Left (a) overview (b) action Fig. 2. The robot Our robot is shown in Figure 2(a) where an omnidirectional vision system is installed onto the 2-DOFs non-holonomic vehicle so that its optical axis can be coincident with the axis of vehicle rotation. The robot is controlled by a remote host computer via radio link. Figure 2(b) shows actions of the robot; moving toward left and right directions, turning to left and right, and combination of moving and turning. The remote computer sends motor commands to control the robot motion. An omnidirectional vision systems consists of a conic mirror and a TV camera [1] of which optical axis is aligned with the vertical axis of the mirror as shown Figure 3(a). The projection onto the image plane is determined by the the shape of the mirror and the camera configuration parameters (height of the camera, distance between the mirror and the lens, and focal length) which can be designed according to the purpose. Our omnidirectional vision system has a hyperbolic mirror and a sample of its image is shown in Figure 3(b). The omnidirectional image is transmitted to the remote computer via video transmitter and processed there.

virtual axis mirror image plane (a) projection (b) image Fig. 3. A sample of an omnidirectional vision We set up a simplified soccer game according to the RoboCup context [5]. The task of the robot is to block a ball in front of the goal, that is, a goalie task (see Figure 4). In order to keep the goal, the robot has to track the moving ball and move to appropriate position. Fig. 4. The task of the robot 3 Learning of a Robot with an Omnidirectional Vision and an Embeded Zoom Control 3.1 Active Zoom Control on the Omnidirectional Vision The coordinate system and parameters are shown in Figure 5(a). Let P (R, θ, Z) and p(r, θ) denote a point in the environment and a projected point of P in the

h f c c image plane, respectively. We assume that the object is on the ground plane, therefore Z becomes a constant and P is uniquely projected onto p. The relation between P and p is given by, Z = R tan α + c + h, tan γ = b2 + c 2 b 2 c 2 tan α + 2bc 1 c 2 b 2 cos α, and (1) r = f tan γ, R where a and b are the parameters of the hyperbolic mirror, 2 a Z2 2 b = 1, 2 and c = a 2 + b 2. In our system these parameters are α 2 = 233.3[mm], β 2 = 1135.7[mm], and h = 250[mm]. Z mirror OM Q 14 12 r = 35 [pix] r = 40 [pix] r = 45 [pix] p ( r, z ) 10 image plane P ( R, Z ) f [mm] 8 6 OC 4 2 O R 0 50 100 150 200 250 300 R [cm] (a) The projection of the hyperbolic mirror (b) The relation between distances and the focal length Fig. 5. The basics of the hyperbolic mirror We add an attention control on an omnidirectional vision by controlling focal length of the camera in order to reduce the learning time for behavior acquisition. In general, an attention control is realized by tracking an object in the image plane, which is implemented by controlling pan and tilt angles of the camera. However, in an omnidirectional vision, matching of the object with the target image is complicated since the projection is not simple. Therefore we propose an attention control by seeing the object in a certain distance from the center in

the image plane. The change of the distance of the target in the image is tracked by changing focal length of the camera as shown in Figure 6. I rd I rd Z mirror OM p ( r, z ) f p ( r, z ) image plane P ( R, Z ) OC O R Fig. 6. Attention control on the omnidirectional vision Figure 5(b) shows the relation between r the distance of the object in the image from the center, R the distance of the object in the environment, and f the focal length of the camera. We control the focal length of the camera with a following equation; u f = K( I r d I r), (2) where u f is the focal length of the camera, I r d is the desired distance in the image, I r is the current distance in the image. For example, if I r is smaller than I r d it comes closer by increasing f as shown in Figure 6. 3.2 Learning of a Goal Keepint Behavior We use Q-learning, one of major reinforcement learning methods, to acquire a goal defending behavior. The state space needs to be defined from the image observed from the robot [6]. We define the substates as shown in the first column in Table 1. The second column shows the numbers of quantisation for each substates. In addition, the numbers of the quantisation of the substates without the attention control are shown. For the case without attention control, the states are defined in terms of the direction and the distance of the ball and the goal in the image. We define 8 substates for the direction of the ball as shown in Figure 7(a) and 3 substates (far, medium and near) for the distance as shown in Figure 7(b). In addition we define temporal changes of the direction and the distance, respectively. They are

clock wise/counter clock wise/no change and farther/nearer/no change, respectively. The numbers of the substaes for the direction of the goal is 8, the same quantisation for the ball, and the that of the distance is two (far and near). The total number of states is 3456(= 8 3 3 3 8 2). For the case with attention control, substates of the distance and the temporal change of the distance of the ball are not necessary since the distance in the image is constant. The direction of the ball and the goal are defined in the same manner as above. The distance of the goal is far and near, however, the observed image of the goal changes when the attention control is used. The distance of the goal can be represented by a monotonic function in terms of the actual distance between the robot and the ball. The total number of states is 320(= 8 3 2 8). with attention control without attention control direction of the ball in the image 8 8 change of the direction of the ball 3 3 distance of the ball in the image 3 change of the distance of the ball 3 direction of the goal in the image 8 8 distance of the goal in the image 2 2 total number of the state 3456 320 Table 1. Substates Left Front Back Near Right Middle Far (a) directions (b) distances Fig. 7. Substates

4 Experiments We performed a computer simulation to acquire a goal defending behavior. Figure 8 shows the environment and the initial positions of the robot and the ball. The environment is built according to the RoboCup middle league regulations. The size of the field is 4575[mm] in width and 4110[mm] in length which is equivalent to the half size of the field. The goal size is 1500[mm] in width and 600[mm] in height and the diameter of the ball is 200[mm]. The goal and the ball are painted in blue and red respectively for easy detection. Fig. 8. Initial position The ball is located on a half circle defined by the center of the goal and two corners, and the robot is located inside the circle randomly. The ball rolls toward the goal at a constant velocity. One trial terminates when the ball comes into the goal or goes out from the field. Figure 9 shows the task success rate with the learned behavior. When the attention control is used the robot learn quicker than the case without the attention control. After learning in the simulation, the acquired behavior is implemented on the real robot. Figures 10(a)-(f) show a sequence of the real behavior, when the robot succeeded in blocking the ball in front of the goal. 5 Conclusions We have proposed an attention control for an omnidirectional vision by controlling focal length of the camera and implemented it on a mobile robot. We have applied Q-learning method for acquisition of a goal defending behavior and

Shoot Block Rate[%] 100 80 60 40 20 with Attention control without Attention control 0 0 10000 20000 30000 40000 Trial times Fig. 9. Result (a) (b) (c) (d) (e) (f) Fig. 10. A sequence of behavior shown that the attention control effectively worked to reduce the learning time. We will test our method in RoboCup-98. In this paper, we have shown a case that an embedded servo worked effectively for learning of the robot. However, we have not considered on a trade-off between installing a servoing mechanism and reduction of the learning time. This is the future work. References 1. Y. Yagi and S. Kawato. Panoramic scene analysis with conic projection. In Proc.

of IEEE/RSJ International Conference on Intelligent Robots and Systems 1990 (IROS 90), 1990. 2. H. Ishiguro. Distributed vision systems: A perceptual information infrastructure for robot navigation. In Proc. of IJCAI-97, pages 36 41, 1997. 3. V. N. Peri and S. Nayar. Generation of perspective and panoramic video from omnidorectinal video. In Proc. of 1997 Image Understanding Workshop, pages 243 245, 1997. 4. K. Hosoda, H. Moriyama, and M. Asada. Visual servoing utilizing zoom mechanism. In Proc. of IEEE Int. Conf. on Robotics and Automation, pages 178 183, 1995. 5. H. Kitano, M. Asada, Y. Kuniyoshi, I. Noda, E. Osawa, and H. Matsubara. robocup: A challenge problem of Al. AI Magazine, 18:73 85, 1997. 6. M. Asada, S. Noda, S. Tawaratsumida, and K. Hosoda. Purposive Behavior Acquisition for a Real Robot by Vision-Based Reinforcement Learning. Machine Learning, 23:279 303, 1996. 7. S. Suzuki, Y. Takahashi, E. Uchibe, M. Nakamura, C. Mishima, and M. Asada. Vision-Based Learning Towards RoboCup: Osaka University Trackies. RoboCup-97: Robot Soccer World Cup I, Springer, pp.305 319, 1997. This article was processed using the LaT E X macro package with LLNCS style