BabyTigers-98: Osaka Legged Robot Team

Size: px
Start display at page:

Download "BabyTigers-98: Osaka Legged Robot Team"

Transcription

1 BabyTigers-98: saka Legged Robot Team Noriaki Mitsunaga and Minoru Asada and Chizuko Mishima Dept. of Adaptive Machine Systems, saka niversity, Suita, saka, , Japan Abstract. The saka Legged Robot Team, BabyTigers-98, attended the First Sony Legged Robot Competition and Demonstration which was held at La Cite La illeta, a science and technology museum in Paris in conjunction with the second Robot Soccer World Cup, RoboCup-98, July, This article describes our approach to the competition. The main feature of our system is to apply a direct teaching method to behavior acquisition for four-legged robots as one of the robot learning issues. Since the robots recognize objects in the field by color information, the color calibration is another issue. To cope with color changes due to lighting conditions, we developed an interactive color calibration tool with graphic user interface. Based on the color calibration and the learning by teaching methods, the robot successfully exhibits basic skills such as ball approaching and kicking. 1 Introduction The saka Legged Robot Team, BabyTigers-98, attended the First Sony Legged Robot Competition and Demonstration which was held at La Cite La illeta, a science and technology museum in Paris in conjunction with the Second Robot Soccer World Cup Competition and Conference, RoboCup-98, July 2-9, The final goal of the Legged Robot Project in our group is to establish the methodology to acquire behaviors for team cooperation in the RoboCup context from the interactions between the legged robots through multi sensormotor coordinations. The desired behaviors can be categorized into three levels; a basic level, a basic cooperation level, and a higher team cooperation level. In this article, we briefly explain our first step for the first level skill acquisition with preliminary results. That is behavior acquisition by direct teaching. The most fundamental feature of the legged robot is that it moves by its four legs (12 DFs), which is quite different from conventional mobile robots (2 or 3 DFs). From a viewpoint of sensor-motor learning and development, multi modal information and multi DFs control should be established simultaneously, that is, affecting each other, the sensory information is abstracted and the multi joint motions are well coordinated at the same time [1]. However, it seems very difficult for artificial systems to develop the both together. ur goal is to design such a method. From our experiences on robot learning, we realized that the number of trials by real robot is limited and a good trade-off between computer simulations and real robot experiences is essential for good performance. However, the computer

2 simulation of the legged robots seems difficult to build, then we decided to adopt a direct teaching method in order to reduce the number of trials by real robot. Since the robot detects objects in the field using color information such as a orange ball, aqua-blue and yellow goals, color calibration is another issue for the robot to robustly detect objects in the field and then to learn by teaching. Since color calibration during the game seems difficult to realize due to the limitation of on-board computation power, we adopt an off-line color calibration method. To make the calibration process much easier, we developed an interactive color calibration tool with graphic user interface. This article is structured as follows. First, we show an interactive off-line color calibration system we developed. Next, we describe the method of robot learning by direct teaching. Finally, we show some experimental results and discuss the future issues. 2 ision System 2.1 Color detection In the legged league field [3], 7 colors (aqua-blue, yellow, orange, blue, red, pink, green) are used and robots need to detect all of them. The Sony s legged robot has the color detection facility in hardware that can handle up to 8 colors at frame rate (60-80[ms]). To detect colors with this facility, we need to specify each color in terms of subspace in Y color space. Y subspace is expressed in a table called Color Detection Table(CDT). In this table, Y are equally separated into 32 level and in each Y level we specify one rectangle (u mini, v mini ),(u maxi, v maxi ) (i = 1,..., 32). The color observed by the robots dynamically changes due to the robot and/or object motions. The reason is that it depends on various kinds of parameters such as, the color of object itself, the color of light source, the surface material of object, the surface orientation of object, the angle between the view line of the camera and the beam of the light and so on. In order to cope with color changes, Y subspaces should be large enough to involve the changes, but small not to include different colors. Then we developed an interactive tool so that we can make and modify CDTs that can handle expansion and reduction of the CDT easily. Basic procedures are: 1. To make a CDT, we specify multiple pixels that should be treated as the same color in an image taken by the legged robot. Each pixel has a pair of (y, u, v) value or a point in Y space. According to the value of y, each point is classified into 32 Y level. In each y level, (u mini, v mini ),(u maxi, v maxi ) (i = 1,..., 32) are set to indicate the bounding box that includes all points in that y level (see Fig.1(a)). 2. To add a point in Y to the CDT which should be treated as the same color, expand the bounding box of the corresponding y level to include the point (see Fig.1(b)).

3 Points to be included at this Y level. Points already included at this Y level. Previous bounding box at this Y level. New points to be added at this Y level. Bounding box at this Y level. New bounding box at this Y level. (a) Make a bounding box of the CDT (b) Add a point to the CDT Fig. 1. Making a CDT first time and expansion of a CDT 3. To delete a point in Y from the CDT which should not be treated as the same color, shrink the bounding box of the corresponding y level to exclude the point. To exclude the point, we have four options (see Fig.2) and we take the one which is the least reduction of the area (in this case Fig.2(a)). Previous bounding box at this Y level. Point to be removed from this Y level ne candidate of new rectangle. (a) (b) (c) (d) Fig. 2. Deleting a point from CDT(four options) We developed a GI tool written in Tcl/TK to realize the above procedures easily. The appearance of the tool and emulated color detection is shown in Fig.3. We show an example of CDT construction with images shown in Fig.4, which shows ball images observed by the robot from different positions. First we specify some pixels in Fig.4(a) and make a CDT. The distribution in

4 (a) making color detection table with the tool (b) emulation of color detection Fig. 3. GI tool to make CDTs Y space projected to plane (at y level 16) is shown in Fig.5 (the bounding box is also shown). Distributions in Y space of the ball color in the picture Fig.4(a) and (b) are shown in Fig.6. It indicates that the distribution in Fig.4(a) does not cover the one in Fig.4(b). That is, the CDT constructed by Fig.4(a) does not cover the color distribution of the ball in Fig.4(b) (see Figs.8(a) and (b)). Then, we specify some pixels that do not seem to be detected in Fig.4(b) and add to the CDT. Fig.7 shows the expansion of the bounding box in y level 16. After all of addition, the robot can detect Fig.4(b) as Fig.8(c). 2.2 bject detection Since the legged league field includes multiple objects, some of which have the same color, and observed images are noisy, object detection is carried out with noise reduction based on a priori knowledge of the environment. First, using the color detected image, pixels in the same color are connected into regions by their 8-neighbors. At the same time, area sizes, centroids and bounding boxes of the regions are calculated. Next, the small regions less than 4 pixels are deleted because the minimum size of the smallest object (ball) in the image plane is 4. Then, the order of object detection is determined, consid-

5 (a) ball picture (A) (b) ball picture (B) Fig. 4. Pictures of a ball points in y level 16 points other than in y level 16 bounding box of y level Fig. 5. Distribution of the ball color (A) in space points included in the ball picture A points included only in the ball picture B Y Fig. 6. Distribution of the ball color in Y space

6 points in picture (A) old boundary box points to add new boundary box Fig. 7. Distribution of the ball color in space y level 16 (a) ball picture (A): color detection with the CDT made with the ball picture (A) (b) ball picture (B): color detection with the CDT made with the ball picture (A) (c) ball picture (B): color detection with the CDT made with the ball picture (A) and (B) Fig. 8. Result of color detection with CDTs (emulation by software) ering the importance of object, easiness of color discrimination and visible constraints of multiple object colors. Constraints means that maximum number of objects that can be observed simultaneously is known, orange(1), aqua-blue(3), yellow(3), green(3), pink(3), blue(3) and red(3). This can also be used to noise reduction. The order we used is as follows. 1. ball 2. poles: (top color)pink (bottom color)aqua-blue aqua-blue pink pink yellow green pink pink green

7 3. goals: aqua-blue goal yellow goal 4. robots: blue robots red robots The color of robot s body and head was blue or red, of which saturation was low and difficult to detect. That is why we put them in low priority in object detection. 3 Behavior Acquisition In the competition, we assigned the roles of attacker and goal keeper to three robots, that is, two attackers and one goal keeper. In the following, we explain how the behaviors for the goal keeper and the attackers are obtained. 3.1 Goal-keeper We adopted a strategy for the goal keeper to save the goal by positioning itself in front of the goal. The angles of the visible area is 53 degrees in width and 41 degrees in height. It means that the robot cannot see as many goals and landmark poles to determine its position. The result of color detection is small in size (88 59 pixels), and the poles are small (diameter is 0.1[m] and height is 0.2[m]). So, the robot rotates its head to enlarge the visible field, and decides its movement depending on rough estimation of its position. Since the landmark poles near the opponent goal is difficult for it to observe, the robot compares the angles between its own goal and four poles near own goal in front of its own goal, and determines which direction to move. Further, the robot tracks the ball for a while when it is near the ball. 3.2 Attacker The behavior of two attackers is obtained by direct teaching. More precisely, the robot learns its behavior by classifying the data taught by a human trainer from a viewpoint of information theory. A human trainer can teach the robot via serial console with PC. First, the robot collects test data, pairs of an action command given by human trainer and the sensory information during the action execution in every 300ms. Next, C4.5 [2] is applied to the test data to extract rule sets. Then, the validity of the rule sets are checked against test data by applying the rule sets. Specifications of the data for shooting skill acquisition are as follows: action command forward, backward, left-shift, right-shift, left-rotation, and right rotation (20 degs/sec): these abstracted action commands will be decomposed into more primitive motor commands in future.

8 sensory information head direction (rad.) and image features of both the ball and the goal in the observed 88x59 image: area (pixels), position (x-y coordinates), bounding rectangle (x-y coordinates of corners), height, and width. See Figure 9. training position and sampling rate: initial positions of direct teaching by serial line connection are evenly distributed in the field heading the goal, and the sampling rate is 300ms. ne trial from the initial position to the goal takes about 10 seconds. In the experiments, 740 pairs of an action command and sensory information for training data to make rule sets, and 500 pairs for test data to check the validity of the rule sets are given. The both data are obtained in the same manner starting from the similar initial positions, but individual pairs are different from each other trial by trial. The number of rules obtained is about 30, and typical ones for forward(f), left-rotation (LR), and right-shift (RS) are shown below. Figure 9 shows the typical situation of right-shift motion. Forward: BallArea>56 Left- BallArea>49 Right- BallArea<=40 HeadDir>-.14 Rotation: HeadDir>.52 Shift: BallYcen>8 HeadDir<=.37 HeadDir<=1.10 BallXmin>3 GoalXmin<=11 BallWidth<=11 HeadDir<=-.24 GoalArea<=665 GoalXmax<=64 HeadDir Bxmin Gxmax (Bxc,Byc) Fig. 9. Typical situation to take Right-Shift motion The following table indicates a confusion matrix showing where the missclassification of the training cases occur. Due to the wrong teaching by a human trainer (actually, she has not got used to teaching), this matrix includes many

9 miss-classifications, especially LEFT-TRN does not have any correct classifications. We are planning to skill up teaching and also to use more training data to construct robust classification. Generalization is one of the big issue of direct teaching, and EBL seems one alternative to solve the problem. These are under investigation. classified as: LEFT LEFT-TRN RIGHT RIGHT-TRN FRWARD BACK Conclusion This paper has described the implementation of vision system and strategies used in BabyTigers-98. Although it was not the best, the resultant behaviors were encouraging. The followings are future issues. 1. The current color calibration system includes an off-line interactive process, therefore, it cannot cope with dynamic changes in color during the match. Full autonomous on-line color calibration system should be developed. 2. Direct teaching seems useful under the condition that the robot sensation can be informed to the human trainer. However, it does not generally true, and often perceptual aliasing happens. As a short range research issue, the robot should reconstruct its state space so that the instructions given by the trainer can be consistent and effective. As a longer one, both the robot and the trainer should learn each other, that is, the robot build the trainers model and vice versa. Through the model developing process, the robot learns how to behave according to the instructions, and the trainer learns how to teach it. 3. Currently, only the basic skills have been obtained and not the cooperative ones yet. Direct teaching should be useful for such behavior acquisition. Mutual learning (co-evolution) process is indispensable to obtain the cooperative and competitive behavior in dynamic multi-agent environment. References 1. Asada, M., An Agent and an Environment: A iew on Having Bodies - A Case Study on Behavior Learning for ision-based Mobile Robot, Proceedings of 1996 IRS Workshop on Towards Real Autonomy, pp.19-24, Quinlan, J. R., C4.5: Programs for Machine Learning, The Morgan Kaufmann Series in Machine Learning, eloso, M., ther, W., Fujita, M., Asada, M. and Kitano, H., Playing Soccer with Legged Robot, Proceedings of the 1998 IEEE/RSJ International Conference on Intelligent Robots and Systems, ol. 1, pp , This article was processed using the LaT E X macro package with LLNCS style

Visual Attention Control by Sensor Space Segmentation for a Small Quadruped Robot based on Information Criterion

Visual Attention Control by Sensor Space Segmentation for a Small Quadruped Robot based on Information Criterion Visual Attention Control by Sensor Space Segmentation for a Small Quadruped Robot based on Information Criterion Noriaki Mitsunaga and Minoru Asada Dept. of Adaptive Machine Systems, Osaka University,

More information

Dept. of Adaptive Machine Systems, Graduate School of Engineering Osaka University, Suita, Osaka , Japan

Dept. of Adaptive Machine Systems, Graduate School of Engineering Osaka University, Suita, Osaka , Japan An Application of Vision-Based Learning for a Real Robot in RoboCup - A Goal Keeping Behavior for a Robot with an Omnidirectional Vision and an Embedded Servoing - Sho ji Suzuki 1, Tatsunori Kato 1, Hiroshi

More information

Using Layered Color Precision for a Self-Calibrating Vision System

Using Layered Color Precision for a Self-Calibrating Vision System ROBOCUP2004 SYMPOSIUM, Instituto Superior Técnico, Lisboa, Portugal, July 4-5, 2004. Using Layered Color Precision for a Self-Calibrating Vision System Matthias Jüngel Institut für Informatik, LFG Künstliche

More information

Behavior Learning for a Mobile Robot with Omnidirectional Vision Enhanced by an Active Zoom Mechanism

Behavior Learning for a Mobile Robot with Omnidirectional Vision Enhanced by an Active Zoom Mechanism Behavior Learning for a Mobile Robot with Omnidirectional Vision Enhanced by an Active Zoom Mechanism Sho ji Suzuki, Tatsunori Kato, Minoru Asada, and Koh Hosoda Dept. of Adaptive Machine Systems, Graduate

More information

Continuous Valued Q-learning for Vision-Guided Behavior Acquisition

Continuous Valued Q-learning for Vision-Guided Behavior Acquisition Continuous Valued Q-learning for Vision-Guided Behavior Acquisition Yasutake Takahashi, Masanori Takeda, and Minoru Asada Dept. of Adaptive Machine Systems Graduate School of Engineering Osaka University

More information

(a) (b) (c) Fig. 1. Omnidirectional camera: (a) principle; (b) physical construction; (c) captured. of a local vision system is more challenging than

(a) (b) (c) Fig. 1. Omnidirectional camera: (a) principle; (b) physical construction; (c) captured. of a local vision system is more challenging than An Omnidirectional Vision System that finds and tracks color edges and blobs Felix v. Hundelshausen, Sven Behnke, and Raul Rojas Freie Universität Berlin, Institut für Informatik Takustr. 9, 14195 Berlin,

More information

Particle-Filter-Based Self-Localization Using Landmarks and Directed Lines

Particle-Filter-Based Self-Localization Using Landmarks and Directed Lines Particle-Filter-Based Self-Localization Using Landmarks and Directed Lines Thomas Röfer 1, Tim Laue 1, and Dirk Thomas 2 1 Center for Computing Technology (TZI), FB 3, Universität Bremen roefer@tzi.de,

More information

Perceptual Attention Through Image Feature Generation Based on Visio-Motor Map Learning

Perceptual Attention Through Image Feature Generation Based on Visio-Motor Map Learning Perceptual Attention Through Image Feature Generation Based on Visio-Motor Map Learning Minoru Asada and Takashi Minato Dept. of Adaptive Machine Systems Graduate School of Engineering Osaka University

More information

Horus: Object Orientation and Id without Additional Markers

Horus: Object Orientation and Id without Additional Markers Computer Science Department of The University of Auckland CITR at Tamaki Campus (http://www.citr.auckland.ac.nz) CITR-TR-74 November 2000 Horus: Object Orientation and Id without Additional Markers Jacky

More information

Realtime Object Recognition Using Decision Tree Learning

Realtime Object Recognition Using Decision Tree Learning Realtime Object Recognition Using Decision Tree Learning Dirk Wilking 1 and Thomas Röfer 2 1 Chair for Computer Science XI, Embedded Software Group, RWTH Aachen wilking@informatik.rwth-aachen.de 2 Center

More information

Image Feature Generation by Visio-Motor Map Learning towards Selective Attention

Image Feature Generation by Visio-Motor Map Learning towards Selective Attention Image Feature Generation by Visio-Motor Map Learning towards Selective Attention Takashi Minato and Minoru Asada Dept of Adaptive Machine Systems Graduate School of Engineering Osaka University Suita Osaka

More information

Ball tracking with velocity based on Monte-Carlo localization

Ball tracking with velocity based on Monte-Carlo localization Book Title Book Editors IOS Press, 23 1 Ball tracking with velocity based on Monte-Carlo localization Jun Inoue a,1, Akira Ishino b and Ayumi Shinohara c a Department of Informatics, Kyushu University

More information

Eagle Knights 2007: Four-Legged League

Eagle Knights 2007: Four-Legged League Eagle Knights 2007: Four-Legged League Alfredo Weitzenfeld 1, Alonso Martínez 1, Bernardo Muciño 1, Gabriela Serrano 1, Carlos Ramos 1, and Carlos Rivera 1 1 Robotics Laboratory Lab, ITAM, Rio Hondo 1,

More information

NuBot Team Description Paper 2013

NuBot Team Description Paper 2013 NuBot Team Description Paper 2013 Zhiwen Zeng, Dan Xiong, Qinghua Yu, Kaihong Huang, Shuai Cheng, Huimin Lu, Xiangke Wang, Hui Zhang, Xun Li, Zhiqiang Zheng College of Mechatronics and Automation, National

More information

An Efficient Need-Based Vision System in Variable Illumination Environment of Middle Size RoboCup

An Efficient Need-Based Vision System in Variable Illumination Environment of Middle Size RoboCup An Efficient Need-Based Vision System in Variable Illumination Environment of Middle Size RoboCup Mansour Jamzad and Abolfazal Keighobadi Lamjiri Sharif University of Technology Department of Computer

More information

Robust Color Choice for Small-size League RoboCup Competition

Robust Color Choice for Small-size League RoboCup Competition Robust Color Choice for Small-size League RoboCup Competition Qiang Zhou Limin Ma David Chelberg David Parrott School of Electrical Engineering and Computer Science, Ohio University Athens, OH 45701, U.S.A.

More information

Robot Controller. Sensor Processing Module. Locomotion Module. Sensors. Environment

Robot Controller. Sensor Processing Module. Locomotion Module. Sensors. Environment Evolution of Controllers from a High-Level Simulator to a High DOF Robot G. S. Hornby 1, S. Takamura 2, O. Hanagata 3, M. Fujita 3, and J. Pollack 1 1 Computer Science Dept., Brandeis University, Waltham,

More information

Nao Devils Dortmund. Team Description Paper for RoboCup Matthias Hofmann, Ingmar Schwarz, and Oliver Urbann

Nao Devils Dortmund. Team Description Paper for RoboCup Matthias Hofmann, Ingmar Schwarz, and Oliver Urbann Nao Devils Dortmund Team Description Paper for RoboCup 2017 Matthias Hofmann, Ingmar Schwarz, and Oliver Urbann Robotics Research Institute Section Information Technology TU Dortmund University 44221 Dortmund,

More information

Towards a Calibration-Free Robot: The ACT Algorithm for Automatic Online Color Training

Towards a Calibration-Free Robot: The ACT Algorithm for Automatic Online Color Training Towards a Calibration-Free Robot: The ACT Algorithm for Automatic Online Color Training Patrick Heinemann, Frank Sehnke, Felix Streichert, and Andreas Zell Wilhelm-Schickard-Institute, Department of Computer

More information

Machine Learning on Physical Robots

Machine Learning on Physical Robots Machine Learning on Physical Robots Alfred P. Sloan Research Fellow Department or Computer Sciences The University of Texas at Austin Research Question To what degree can autonomous intelligent agents

More information

A Real Time Vision System for Robotic Soccer

A Real Time Vision System for Robotic Soccer A Real Time Vision System for Robotic Soccer Chunmiao Wang, Hui Wang, William Y. C. Soh, Han Wang Division of Control & Instrumentation, School of EEE, Nanyang Technological University, 4 Nanyang Avenue,

More information

Acquiring Observation Models through Reverse Plan Monitoring

Acquiring Observation Models through Reverse Plan Monitoring Acquiring Observation Models through Reverse Plan Monitoring Sonia Chernova, Elisabeth Crawford, and Manuela Veloso Computer Science Department, Carnegie Mellon University, Pittsburgh PA 15213, USA Abstract.

More information

Motion Control in Dynamic Multi-Robot Environments

Motion Control in Dynamic Multi-Robot Environments Motion Control in Dynamic Multi-Robot Environments Michael Bowling mhb@cs.cmu.edu Manuela Veloso mmv@cs.cmu.edu Computer Science Department Carnegie Mellon University Pittsburgh, PA 15213-3890 Abstract

More information

AN ADAPTIVE COLOR SEGMENTATION ALGORITHM FOR SONY LEGGED ROBOTS

AN ADAPTIVE COLOR SEGMENTATION ALGORITHM FOR SONY LEGGED ROBOTS AN ADAPTIE COLOR SEGMENTATION ALGORITHM FOR SONY LEGGED ROBOTS Bo Li, Huosheng Hu, Libor Spacek Department of Computer Science, University of Essex Wivenhoe Park, Colchester CO4 3SQ, United Kingdom Email:

More information

UAV Position and Attitude Sensoring in Indoor Environment Using Cameras

UAV Position and Attitude Sensoring in Indoor Environment Using Cameras UAV Position and Attitude Sensoring in Indoor Environment Using Cameras 1 Peng Xu Abstract There are great advantages of indoor experiment for UAVs. Test flights of UAV in laboratory is more convenient,

More information

Canny Edge Based Self-localization of a RoboCup Middle-sized League Robot

Canny Edge Based Self-localization of a RoboCup Middle-sized League Robot Canny Edge Based Self-localization of a RoboCup Middle-sized League Robot Yoichi Nakaguro Sirindhorn International Institute of Technology, Thammasat University P.O. Box 22, Thammasat-Rangsit Post Office,

More information

Document downloaded from: The final publication is available at: Copyright.

Document downloaded from: The final publication is available at: Copyright. Document downloaded from: http://hdl.handle.net/1459.1/62843 The final publication is available at: https://doi.org/1.17/978-3-319-563-8_4 Copyright Springer International Publishing Switzerland, 213 Evaluation

More information

The University of Pennsylvania Robocup 2011 SPL Nao Soccer Team

The University of Pennsylvania Robocup 2011 SPL Nao Soccer Team The University of Pennsylvania Robocup 2011 SPL Nao Soccer Team General Robotics Automation, Sensing and Perception (GRASP) Laboratory University of Pennsylvania, Philadelphia, PA 19104 Abstract. This

More information

Combining Edge Detection and Colour Segmentation in the Four-Legged League

Combining Edge Detection and Colour Segmentation in the Four-Legged League Combining Edge Detection and Colour Segmentation in the Four-Legged League Craig L. Murch and Stephan K. Chalup School of Electrical Engineering & Computer Science The University of Newcastle, Callaghan

More information

126 ACTION AND PERCEPTION

126 ACTION AND PERCEPTION Motion Sketch: Acquisition of Visual Motion Guided Behaviors Takayuki Nakamura and Minoru Asada Dept. of Mechanical Eng. for Computer-Controlled Machinery, Osaka University, Suita 565 JAPAN E-mail: nakamura@robotics.ccm.eng.osaka-u.ac.jp

More information

Proc. Int. Symp. Robotics, Mechatronics and Manufacturing Systems 92 pp , Kobe, Japan, September 1992

Proc. Int. Symp. Robotics, Mechatronics and Manufacturing Systems 92 pp , Kobe, Japan, September 1992 Proc. Int. Symp. Robotics, Mechatronics and Manufacturing Systems 92 pp.957-962, Kobe, Japan, September 1992 Tracking a Moving Object by an Active Vision System: PANTHER-VZ Jun Miura, Hideharu Kawarabayashi,

More information

An Interactive Technique for Robot Control by Using Image Processing Method

An Interactive Technique for Robot Control by Using Image Processing Method An Interactive Technique for Robot Control by Using Image Processing Method Mr. Raskar D. S 1., Prof. Mrs. Belagali P. P 2 1, E&TC Dept. Dr. JJMCOE., Jaysingpur. Maharashtra., India. 2 Associate Prof.

More information

Eagle Knights: RoboCup Small Size League. Dr. Alfredo Weitzenfeld ITAM - Mexico

Eagle Knights: RoboCup Small Size League. Dr. Alfredo Weitzenfeld ITAM - Mexico Eagle Knights: RoboCup Small Size League Dr. Alfredo Weitzenfeld ITAM - Mexico Playing Field and Robot Size camera 15 cm 4 m 18 cm 4 m 5.5 m Distributed Robot Control Eagle Knights (2003-2006) 2003 2004

More information

Visual Servoing Utilizing Zoom Mechanism

Visual Servoing Utilizing Zoom Mechanism IEEE Int. Conf. on Robotics and Automation 1995, pp.178 183, Nagoya, May. 12 16, 1995 1 Visual Servoing Utilizing Zoom Mechanism Koh HOSODA, Hitoshi MORIYAMA and Minoru ASADA Dept. of Mechanical Engineering

More information

using an omnidirectional camera, sufficient information for controlled play can be collected. Another example for the use of omnidirectional cameras i

using an omnidirectional camera, sufficient information for controlled play can be collected. Another example for the use of omnidirectional cameras i An Omnidirectional Vision System that finds and tracks color edges and blobs Felix v. Hundelshausen, Sven Behnke, and Raul Rojas Freie Universität Berlin, Institut für Informatik Takustr. 9, 14195 Berlin,

More information

Task analysis based on observing hands and objects by vision

Task analysis based on observing hands and objects by vision Task analysis based on observing hands and objects by vision Yoshihiro SATO Keni Bernardin Hiroshi KIMURA Katsushi IKEUCHI Univ. of Electro-Communications Univ. of Karlsruhe Univ. of Tokyo Abstract In

More information

Robust Color Classification for Robot Soccer

Robust Color Classification for Robot Soccer Robust Color Classification for Robot Soccer Ingo Dahm, Sebastian Deutsch, Matthias Hebbel, André Osterhues Computer Engineering Institute E-mail: {name.familyname}@uni-dortmund.de University of Dortmund,

More information

Jolly Pochie 2005 in the Four Legged Robot League

Jolly Pochie 2005 in the Four Legged Robot League Jolly Pochie 2005 in the Four Legged Robot League Jun Inoue 1, Hayato Kobayashi 1, Akira Ishino 2, and Ayumi Shinohara 3 1 Department of Informatics, Kyushu University {j-inoue, h-koba}@i.kyushu-u.ac.jp

More information

Robust and Accurate Detection of Object Orientation and ID without Color Segmentation

Robust and Accurate Detection of Object Orientation and ID without Color Segmentation 0 Robust and Accurate Detection of Object Orientation and ID without Color Segmentation Hironobu Fujiyoshi, Tomoyuki Nagahashi and Shoichi Shimizu Chubu University Japan Open Access Database www.i-techonline.com

More information

3D Environment Measurement Using Binocular Stereo and Motion Stereo by Mobile Robot with Omnidirectional Stereo Camera

3D Environment Measurement Using Binocular Stereo and Motion Stereo by Mobile Robot with Omnidirectional Stereo Camera 3D Environment Measurement Using Binocular Stereo and Motion Stereo by Mobile Robot with Omnidirectional Stereo Camera Shinichi GOTO Department of Mechanical Engineering Shizuoka University 3-5-1 Johoku,

More information

Shape Modeling of A String And Recognition Using Distance Sensor

Shape Modeling of A String And Recognition Using Distance Sensor Proceedings of the 24th IEEE International Symposium on Robot and Human Interactive Communication Kobe, Japan, Aug 31 - Sept 4, 2015 Shape Modeling of A String And Recognition Using Distance Sensor Keisuke

More information

Visuo-Motor Learning for Face-to-face Pass between Heterogeneous Humanoids

Visuo-Motor Learning for Face-to-face Pass between Heterogeneous Humanoids Visuo-Motor Learning for Face-to-face Pass between Heterogeneous Humanoids Masaki Ogino a,, Masaaki Kikuchi a, Minoru Asada a,b a Dept. of Adaptive Machine Systems b HANDAI Frontier Research Center Graduate

More information

Decision Algorithm for Pool Using Fuzzy System

Decision Algorithm for Pool Using Fuzzy System Decision Algorithm for Pool Using Fuzzy System S.C.Chua 1 W.C.Tan 2 E.K.Wong 3 V.C.Koo 4 1 Faculty of Engineering & Technology Tel: +60-06-252-3007, Fax: +60-06-231-6552, E-mail: scchua@mmu.edu.my 2 Faculty

More information

Optical Flow-Based Person Tracking by Multiple Cameras

Optical Flow-Based Person Tracking by Multiple Cameras Proc. IEEE Int. Conf. on Multisensor Fusion and Integration in Intelligent Systems, Baden-Baden, Germany, Aug. 2001. Optical Flow-Based Person Tracking by Multiple Cameras Hideki Tsutsui, Jun Miura, and

More information

SHARPKUNGFU TEAM DESCRIPTION 2006

SHARPKUNGFU TEAM DESCRIPTION 2006 SHARPKUNGFU TEAM DESCRIPTION 2006 Qining Wang, Chunxia Rong, Yan Huang, Guangming Xie, Long Wang Intelligent Control Laboratory, College of Engineering, Peking University, Beijing 100871, China http://www.mech.pku.edu.cn/robot/fourleg/

More information

Motion Recognition and Generation for Humanoid based on Visual-Somatic Field Mapping

Motion Recognition and Generation for Humanoid based on Visual-Somatic Field Mapping Motion Recognition and Generation for Humanoid based on Visual-Somatic Field Mapping 1 Masaki Ogino, 1 Shigeo Matsuyama, 1 Jun ichiro Ooga, and 1, Minoru Asada 1 Dept. of Adaptive Machine Systems, HANDAI

More information

MOTION STEREO DOUBLE MATCHING RESTRICTION IN 3D MOVEMENT ANALYSIS

MOTION STEREO DOUBLE MATCHING RESTRICTION IN 3D MOVEMENT ANALYSIS MOTION STEREO DOUBLE MATCHING RESTRICTION IN 3D MOVEMENT ANALYSIS ZHANG Chun-sen Dept of Survey, Xi an University of Science and Technology, No.58 Yantazhonglu, Xi an 710054,China -zhchunsen@yahoo.com.cn

More information

Jo-Car2 Autonomous Mode. Path Planning (Cost Matrix Algorithm)

Jo-Car2 Autonomous Mode. Path Planning (Cost Matrix Algorithm) Chapter 8.2 Jo-Car2 Autonomous Mode Path Planning (Cost Matrix Algorithm) Introduction: In order to achieve its mission and reach the GPS goal safely; without crashing into obstacles or leaving the lane,

More information

Object Recognition in Robot Football Using a one Dimensional Image

Object Recognition in Robot Football Using a one Dimensional Image Object Recognition in Robot Football Using a one Dimensional Image Hatice Köse and H. Levent Akin Bogaziçi University, Department of Computer Engineering, 80815 Bebek, Istanbul, TURKEY {kose,akin}@boun.edu.tr

More information

An Illumination Identification System for the AIBO Robot

An Illumination Identification System for the AIBO Robot An Illumination Identification System for the AIBO Robot Michael S. Zehmeister, Yang Sok Kim and Byeong Ho Kang School of Computing, University of Tasmania Sandy Bay, Tasmania 7001, Australia {yangsokk,

More information

Soccer Server and Researches on Multi-Agent Systems. NODA, Itsuki and MATSUBARA, Hitoshi ETL Umezono, Tsukuba, Ibaraki 305, Japan

Soccer Server and Researches on Multi-Agent Systems. NODA, Itsuki and MATSUBARA, Hitoshi ETL Umezono, Tsukuba, Ibaraki 305, Japan Soccer Server and Researches on Multi-Agent Systems NODA, Itsuki and MATSUBARA, Hitoshi ETL 1-1-4 Umezono, Tsukuba, Ibaraki 305, Japan fnoda,matsubarg@etl.go.jp Abstract We have developed Soccer Server,

More information

PASS EVALUATING IN SIMULATED SOCCER DOMAIN USING ANT-MINER ALGORITHM

PASS EVALUATING IN SIMULATED SOCCER DOMAIN USING ANT-MINER ALGORITHM PASS EVALUATING IN SIMULATED SOCCER DOMAIN USING ANT-MINER ALGORITHM Mohammad Ali Darvish Darab Qazvin Azad University Mechatronics Research Laboratory, Qazvin Azad University, Qazvin, Iran ali@armanteam.org

More information

Expanding gait identification methods from straight to curved trajectories

Expanding gait identification methods from straight to curved trajectories Expanding gait identification methods from straight to curved trajectories Yumi Iwashita, Ryo Kurazume Kyushu University 744 Motooka Nishi-ku Fukuoka, Japan yumi@ieee.org Abstract Conventional methods

More information

A Simple Interface for Mobile Robot Equipped with Single Camera using Motion Stereo Vision

A Simple Interface for Mobile Robot Equipped with Single Camera using Motion Stereo Vision A Simple Interface for Mobile Robot Equipped with Single Camera using Motion Stereo Vision Stephen Karungaru, Atsushi Ishitani, Takuya Shiraishi, and Minoru Fukumi Abstract Recently, robot technology has

More information

EE368 Project Report CD Cover Recognition Using Modified SIFT Algorithm

EE368 Project Report CD Cover Recognition Using Modified SIFT Algorithm EE368 Project Report CD Cover Recognition Using Modified SIFT Algorithm Group 1: Mina A. Makar Stanford University mamakar@stanford.edu Abstract In this report, we investigate the application of the Scale-Invariant

More information

Real time game field limits recognition for robot self-localization using collinearity in Middle-Size RoboCup Soccer

Real time game field limits recognition for robot self-localization using collinearity in Middle-Size RoboCup Soccer Real time game field limits recognition for robot self-localization using collinearity in Middle-Size RoboCup Soccer Fernando Ribeiro (1) Gil Lopes (2) (1) Department of Industrial Electronics, Guimarães,

More information

3D Object Model Acquisition from Silhouettes

3D Object Model Acquisition from Silhouettes 4th International Symposium on Computing and Multimedia Studies 1 3D Object Model Acquisition from Silhouettes Masaaki Iiyama Koh Kakusho Michihiko Minoh Academic Center for Computing and Media Studies

More information

Visual Tracking of Unknown Moving Object by Adaptive Binocular Visual Servoing

Visual Tracking of Unknown Moving Object by Adaptive Binocular Visual Servoing Visual Tracking of Unknown Moving Object by Adaptive Binocular Visual Servoing Minoru Asada, Takamaro Tanaka, and Koh Hosoda Adaptive Machine Systems Graduate School of Engineering Osaka University, Suita,

More information

FutBotIII: Towards a Robust Centralized Vision System for RoboCup Small League

FutBotIII: Towards a Robust Centralized Vision System for RoboCup Small League RoboCup-99 Team Descriptions Small Robots League, Team FutBotIII, pages 43 47 http: /www.ep.liu.se/ea/cis/1999/006/07/ 43 FutBotIII: Towards a Robust Centralized Vision System for RoboCup Small League

More information

ECE 172A: Introduction to Intelligent Systems: Machine Vision, Fall Midterm Examination

ECE 172A: Introduction to Intelligent Systems: Machine Vision, Fall Midterm Examination ECE 172A: Introduction to Intelligent Systems: Machine Vision, Fall 2008 October 29, 2008 Notes: Midterm Examination This is a closed book and closed notes examination. Please be precise and to the point.

More information

Introducing Robotics Vision System to a Manufacturing Robotics Course

Introducing Robotics Vision System to a Manufacturing Robotics Course Paper ID #16241 Introducing Robotics Vision System to a Manufacturing Robotics Course Dr. Yuqiu You, Ohio University c American Society for Engineering Education, 2016 Introducing Robotics Vision System

More information

3D Terrain Sensing System using Laser Range Finder with Arm-Type Movable Unit

3D Terrain Sensing System using Laser Range Finder with Arm-Type Movable Unit 3D Terrain Sensing System using Laser Range Finder with Arm-Type Movable Unit 9 Toyomi Fujita and Yuya Kondo Tohoku Institute of Technology Japan 1. Introduction A 3D configuration and terrain sensing

More information

Fast and Robust Edge-Based Localization in the Sony Four-Legged Robot League

Fast and Robust Edge-Based Localization in the Sony Four-Legged Robot League Fast and Robust Edge-Based Localization in the Sony Four-Legged Robot League Thomas Röfer 1 and Matthias Jüngel 2 1 Bremer Institut für Sichere Systeme, Technologie-Zentrum Informatik, FB 3, Universität

More information

Structure-Based Color Learning on a Mobile Robot under Changing Illumination

Structure-Based Color Learning on a Mobile Robot under Changing Illumination To appear in Autonomous Robots Journal, 2007. 1 Structure-Based Color Learning on a Mobile Robot under Changing Illumination Mohan Sridharan and Peter Stone Abstract A central goal of robotics and AI is

More information

A deformable model driven method for handling clothes

A deformable model driven method for handling clothes A deformable model driven method for handling clothes Yasuyo Kita Fuminori Saito Nobuyuki Kita Intelligent Systems Institute, National Institute of Advanced Industrial Science and Technology (AIST) AIST

More information

Multi Robot Object Tracking and Self Localization Using Visual Percept Relations

Multi Robot Object Tracking and Self Localization Using Visual Percept Relations Multi Robot Object Tracking and Self Localization Using Visual Percept Relations Daniel Göhring and Hans-Dieter Burkhard Department of Computer Science Artificial Intelligence Laboratory Humboldt-Universität

More information

Towards Autonomous Vision Self-Calibration for Soccer Robots

Towards Autonomous Vision Self-Calibration for Soccer Robots Towards Autonomous Vision Self-Calibration for Soccer Robots Gerd Mayer Hans Utz Gerhard Kraetzschmar University of Ulm, James-Franck-Ring, 89069 Ulm, Germany Abstract The ability to autonomously adapt

More information

3D Digitization of a Hand-held Object with a Wearable Vision Sensor

3D Digitization of a Hand-held Object with a Wearable Vision Sensor 3D Digitization of a Hand-held Object with a Wearable Vision Sensor Sotaro TSUKIZAWA, Kazuhiko SUMI, and Takashi MATSUYAMA tsucky@vision.kuee.kyoto-u.ac.jp sumi@vision.kuee.kyoto-u.ac.jp tm@i.kyoto-u.ac.jp

More information

FLY THROUGH VIEW VIDEO GENERATION OF SOCCER SCENE

FLY THROUGH VIEW VIDEO GENERATION OF SOCCER SCENE FLY THROUGH VIEW VIDEO GENERATION OF SOCCER SCENE Naho INAMOTO and Hideo SAITO Keio University, Yokohama, Japan {nahotty,saito}@ozawa.ics.keio.ac.jp Abstract Recently there has been great deal of interest

More information

Reduced Image Noise on Shape Recognition Using Singular Value Decomposition for Pick and Place Robotic Systems

Reduced Image Noise on Shape Recognition Using Singular Value Decomposition for Pick and Place Robotic Systems Reduced Image Noise on Shape Recognition Using Singular Value Decomposition for Pick and Place Robotic Systems Angelo A. Beltran Jr. 1, Christian Deus T. Cayao 2, Jay-K V. Delicana 3, Benjamin B. Agraan

More information

Toward an Undergraduate League for RoboCup.

Toward an Undergraduate League for RoboCup. Toward an Undergraduate League for RoboCup. John Anderson, Jacky Baltes, David Livingston, Elizabeth Sklar, and Jonah Tower Department of Computer Science Columbia University New York, NY, 10027, USA sklar,jpt2002@cs.columbia.edu

More information

Embodiment based Object Recognition for Vision-Based Mobile Agents

Embodiment based Object Recognition for Vision-Based Mobile Agents Embodiment based Object Recognition for Vision-Based Mobile Agents Kazunori Terada, Takayuki Nakamura, Hideaki Takeda and Tsukasa Ogasawara Dept. of Information Systems, Nara Institute of Science and Technology

More information

THE RoboCup 1 is a scientific annual competition aimed

THE RoboCup 1 is a scientific annual competition aimed X WORKSHOP DE AGENTES FÍSICOS, SEPTIEMBRE 2009, CÁCERES 1 Visual Goal Detection for the RoboCup Standard Platform League José M. Cañas, Domenec Puig, Eduardo Perdices and Tomás González Abstract This paper

More information

ROBOCUP 1 is an international robotics competition that

ROBOCUP 1 is an international robotics competition that JOURNAL OF PHYSICAL AGENTS, VOL. 4, NO. 1, JANUARY 2010 11 Recognition of Standard Platform RoboCup Goals José M. Cañas, Eduardo Perdices, Tomás González and Domenec Puig Abstract RoboCup provides a common

More information

OBSTACLE DETECTION USING STRUCTURED BACKGROUND

OBSTACLE DETECTION USING STRUCTURED BACKGROUND OBSTACLE DETECTION USING STRUCTURED BACKGROUND Ghaida Al Zeer, Adnan Abou Nabout and Bernd Tibken Chair of Automatic Control, Faculty of Electrical, Information and Media Engineering University of Wuppertal,

More information

A Hybrid Software Platform for Sony AIBO Robots

A Hybrid Software Platform for Sony AIBO Robots A Hybrid Software Platform for Sony AIBO Robots Dragos Golubovic, Bo Li, Huosheng Hu Department of Computer Science, University of Essex, Wivenhoe Park, Colchester CO4 3SQ, United Kingdom Email: {dgolub,

More information

DETAILED VISION DOCUMENTATION

DETAILED VISION DOCUMENTATION DETAILED VISION DOCUMENTATION Real-time Visual Perception for Autonomous Robotic Soccer Agents RoboCup 99 Thomas Charles Karpati Prof. Rafaello D Andrea Dr. Jin-Woo Lee Revision 2. Revision Date: 5/26/99

More information

Natural Viewing 3D Display

Natural Viewing 3D Display We will introduce a new category of Collaboration Projects, which will highlight DoCoMo s joint research activities with universities and other companies. DoCoMo carries out R&D to build up mobile communication,

More information

NAO-Team Humboldt 2009

NAO-Team Humboldt 2009 NAO-Team Humboldt 2009 The RoboCup NAO Team of Humboldt-Universität zu Berlin Alexander Borisov, Alireza Ferdowsizadeh, Christian Mohr, Heinrich Mellmann, Martin Martius, Thomas Krause, Tobias Hermann,

More information

Real-Time Human Detection using Relational Depth Similarity Features

Real-Time Human Detection using Relational Depth Similarity Features Real-Time Human Detection using Relational Depth Similarity Features Sho Ikemura, Hironobu Fujiyoshi Dept. of Computer Science, Chubu University. Matsumoto 1200, Kasugai, Aichi, 487-8501 Japan. si@vision.cs.chubu.ac.jp,

More information

Towards Cognitive Agents: Embodiment based Object Recognition for Vision-Based Mobile Agents Kazunori Terada 31y, Takayuki Nakamura 31, Hideaki Takeda

Towards Cognitive Agents: Embodiment based Object Recognition for Vision-Based Mobile Agents Kazunori Terada 31y, Takayuki Nakamura 31, Hideaki Takeda Towards Cognitive Agents: Embodiment based Object Recognition for Vision-Based Mobile Agents Kazunori Terada 31y, Takayuki Nakamura 31, Hideaki Takeda 3231 and Tsukasa Ogasawara 31 31 Dept. of Information

More information

COSC160: Detection and Classification. Jeremy Bolton, PhD Assistant Teaching Professor

COSC160: Detection and Classification. Jeremy Bolton, PhD Assistant Teaching Professor COSC160: Detection and Classification Jeremy Bolton, PhD Assistant Teaching Professor Outline I. Problem I. Strategies II. Features for training III. Using spatial information? IV. Reducing dimensionality

More information

Selective Visual Attention for Object Detection on a Legged Robot

Selective Visual Attention for Object Detection on a Legged Robot In Gerhard Lakemeyer, Elizabeth Sklar, Domenico Sorenti, and Tomoichi Takahashi, editors, RoboCup-2006, Springer Verlag, 2007. Selective Visual Attention for Object Detection on a Legged Robot Daniel Stronger

More information

Robot Localization based on Geo-referenced Images and G raphic Methods

Robot Localization based on Geo-referenced Images and G raphic Methods Robot Localization based on Geo-referenced Images and G raphic Methods Sid Ahmed Berrabah Mechanical Department, Royal Military School, Belgium, sidahmed.berrabah@rma.ac.be Janusz Bedkowski, Łukasz Lubasiński,

More information

3D Corner Detection from Room Environment Using the Handy Video Camera

3D Corner Detection from Room Environment Using the Handy Video Camera 3D Corner Detection from Room Environment Using the Handy Video Camera Ryo HIROSE, Hideo SAITO and Masaaki MOCHIMARU : Graduated School of Science and Technology, Keio University, Japan {ryo, saito}@ozawa.ics.keio.ac.jp

More information

EE368 Project: Visual Code Marker Detection

EE368 Project: Visual Code Marker Detection EE368 Project: Visual Code Marker Detection Kahye Song Group Number: 42 Email: kahye@stanford.edu Abstract A visual marker detection algorithm has been implemented and tested with twelve training images.

More information

Improving Percept Reliability in the Sony Four-Legged Robot League

Improving Percept Reliability in the Sony Four-Legged Robot League Improving Percept Reliability in the Sony Four-Legged Robot League Walter Nisticò 1 and Thomas Röfer 2 1 Institute for Robot Research (IRF), Universität Dortmund walter.nistico@udo.edu 2 Center for Computing

More information

Lighting- and Occlusion-robust View-based Teaching/Playback for Model-free Robot Programming

Lighting- and Occlusion-robust View-based Teaching/Playback for Model-free Robot Programming Lighting- and Occlusion-robust View-based Teaching/Playback for Model-free Robot Programming *Yusuke MAEDA (Yokohama National University) Yoshito SAITO (Ricoh Corp) Background Conventional Teaching/Playback

More information

Learning to Kick the Ball Using Back to Reality

Learning to Kick the Ball Using Back to Reality Learning to Kick the Ball Using Back to Reality Juan Cristóbal Zagal and Javier Ruiz-del-Solar Department of Electrical Engineering, Universidad de Chile, Av. Tupper 2007, 6513027 Santiago, Chile {jzagal,

More information

Proc. 14th Int. Conf. on Intelligent Autonomous Systems (IAS-14), 2016

Proc. 14th Int. Conf. on Intelligent Autonomous Systems (IAS-14), 2016 Proc. 14th Int. Conf. on Intelligent Autonomous Systems (IAS-14), 2016 Outdoor Robot Navigation Based on View-based Global Localization and Local Navigation Yohei Inoue, Jun Miura, and Shuji Oishi Department

More information

Machine Learning Through Mimicry and Association

Machine Learning Through Mimicry and Association Georgia Southern University Digital Commons@Georgia Southern University Honors Program Theses Student Research Papers 2016 Machine Learning Through Mimicry and Association Nickolas S. Holcomb Georgia Southern

More information

Towards selective attention: generating image features by learning a visuo-motor map

Towards selective attention: generating image features by learning a visuo-motor map Robotics and Autonomous Systems 45 (2003) 2 22 Towards selective attention: generating image features by learning a visuo-motor map Takashi Minato, Minoru Asada Department of Adaptive Machine Systems,

More information

Acquiring a Robust Case Base for the Robot Soccer Domain

Acquiring a Robust Case Base for the Robot Soccer Domain Acquiring a Robust Case Base for the Robot Soccer Domain Raquel Ros, Josep Lluís Arcos IIIA - Artificial Intelligence Research Institute CSIC - Spanish Council for Scientific Research Campus UAB, 08193

More information

A Simple Automated Void Defect Detection for Poor Contrast X-ray Images of BGA

A Simple Automated Void Defect Detection for Poor Contrast X-ray Images of BGA Proceedings of the 3rd International Conference on Industrial Application Engineering 2015 A Simple Automated Void Defect Detection for Poor Contrast X-ray Images of BGA Somchai Nuanprasert a,*, Sueki

More information

beestanbul RoboCup 3D Simulation League Team Description Paper 2011

beestanbul RoboCup 3D Simulation League Team Description Paper 2011 beestanbul RoboCup 3D Simulation League Team Description Paper 2011 Shahriar Asta, Tuna Sonmez, Onuralp Ulusoy, Alper Alimoglu, Mustafa Ersen, Ozdemir Sozmen, and Sanem Sariel-Talay Artificial Intelligence

More information

Autonomous Navigation in Unknown Environments via Language Grounding

Autonomous Navigation in Unknown Environments via Language Grounding Autonomous Navigation in Unknown Environments via Language Grounding Koushik (kbhavani) Aditya (avmandal) Sanjay (svnaraya) Mentor Jean Oh Introduction As robots become an integral part of various domains

More information

Hand-Eye Calibration from Image Derivatives

Hand-Eye Calibration from Image Derivatives Hand-Eye Calibration from Image Derivatives Abstract In this paper it is shown how to perform hand-eye calibration using only the normal flow field and knowledge about the motion of the hand. The proposed

More information

Mouse Pointer Tracking with Eyes

Mouse Pointer Tracking with Eyes Mouse Pointer Tracking with Eyes H. Mhamdi, N. Hamrouni, A. Temimi, and M. Bouhlel Abstract In this article, we expose our research work in Human-machine Interaction. The research consists in manipulating

More information

Acquisition of Qualitative Spatial Representation by Visual Observation

Acquisition of Qualitative Spatial Representation by Visual Observation Acquisition of Qualitative Spatial Representation by Visual Observation Takushi Sogo Hiroshi Ishiguro Toru Ishida Department of Social Informatics, Kyoto University Kyoto 606-8501, Japan sogo@kuis.kyoto-u.ac.jp,

More information

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS Cognitive Robotics Original: David G. Lowe, 004 Summary: Coen van Leeuwen, s1460919 Abstract: This article presents a method to extract

More information