MULTI-ROBOT research has gained a broad attention. A Novel Way to Implement Self-localization in a Multi-robot Experimental Platform

Similar documents
Autonomous Mobile Robot Design

EKF Localization and EKF SLAM incorporating prior information

Data Association for SLAM

Localization of Multiple Robots with Simple Sensors

ROBOT TEAMS CH 12. Experiments with Cooperative Aerial-Ground Robots

GPS denied Navigation Solutions

Geometrical Feature Extraction Using 2D Range Scanner

ROBUST LOCALISATION OF AUTOMATED GUIDED VEHICLES FOR COMPUTER-INTEGRATED MANUFACTURING ENVIRONMENTS. R.C. Dixon 1 *, G.Bright 2 & R.

Neuro-adaptive Formation Maintenance and Control of Nonholonomic Mobile Robots

Low Cost solution for Pose Estimation of Quadrotor

MATLAB Based Interactive Music Player using XBOX Kinect

Evaluation of Moving Object Tracking Techniques for Video Surveillance Applications

Optimization of the Simultaneous Localization and Map-Building Algorithm for Real-Time Implementation

/10/$ IEEE 4048

Mobile Robotics. Mathematics, Models, and Methods. HI Cambridge. Alonzo Kelly. Carnegie Mellon University UNIVERSITY PRESS

L12. EKF-SLAM: PART II. NA568 Mobile Robotics: Methods & Algorithms

Low Cost Motion Capture

POTENTIAL ACTIVE-VISION CONTROL SYSTEMS FOR UNMANNED AIRCRAFT

Pictures at an Exhibition

Collaborative Multi-Vehicle Localization and Mapping in Marine Environments

Introduction to SLAM Part II. Paul Robertson

AMR 2011/2012: Final Projects

Robotics. Lecture 7: Simultaneous Localisation and Mapping (SLAM)

Navigation of Multiple Mobile Robots Using Swarm Intelligence

Efficient SLAM Scheme Based ICP Matching Algorithm Using Image and Laser Scan Information

RoboCupRescue - Simulation League Team RescueRobots Freiburg (Germany)

Appearance-Based Place Recognition Using Whole-Image BRISK for Collaborative MultiRobot Localization

arxiv: v1 [cs.ro] 2 Sep 2017

An Efficient Method for Solving the Direct Kinematics of Parallel Manipulators Following a Trajectory

Human Detection. A state-of-the-art survey. Mohammad Dorgham. University of Hamburg

LOCAL AND GLOBAL DESCRIPTORS FOR PLACE RECOGNITION IN ROBOTICS

Expanding gait identification methods from straight to curved trajectories

UNIVERSITÀ DEGLI STUDI DI GENOVA MASTER S THESIS

Model-Based 3D Object Tracking Using an Extended-Extended Kalman Filter and Graphics Rendered Measurements

Preliminary Results in Tracking Mobile Targets Using Range Sensors from Multiple Robots

Machine learning based automatic extrinsic calibration of an onboard monocular camera for driving assistance applications on smart mobile devices

Monocular SLAM for a Small-Size Humanoid Robot

Model-based Real-Time Estimation of Building Occupancy During Emergency Egress

Semi-Supervised PCA-based Face Recognition Using Self-Training

Dept. of Mechanical and Materials Engineering University of Western Ontario London, Ontario, Canada

CS231A Course Project Final Report Sign Language Recognition with Unsupervised Feature Learning

Estimation of Altitude and Vertical Velocity for Multirotor Aerial Vehicle using Kalman Filter

Segmentation and Tracking of Partial Planar Templates

Shape Modeling of A String And Recognition Using Distance Sensor

Overview. EECS 124, UC Berkeley, Spring 2008 Lecture 23: Localization and Mapping. Statistical Models

Dealing with Scale. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

Robust and Accurate Detection of Object Orientation and ID without Color Segmentation

URBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES

Robotic Mapping. Outline. Introduction (Tom)

EE631 Cooperating Autonomous Mobile Robots

Non-rigid body Object Tracking using Fuzzy Neural System based on Multiple ROIs and Adaptive Motion Frame Method

Honours Project Proposal. Luke Ross Supervisor: Dr. Karen Bradshaw Department of Computer Science, Rhodes University

Inverse KKT Motion Optimization: A Newton Method to Efficiently Extract Task Spaces and Cost Parameters from Demonstrations

A Reactive Bearing Angle Only Obstacle Avoidance Technique for Unmanned Ground Vehicles

Visual Odometry for Non-Overlapping Views Using Second-Order Cone Programming

Tracking Multiple Mobile Targets Using Cooperative Unmanned Aerial Vehicles

High-precision, consistent EKF-based visual-inertial odometry

Improving Door Detection for Mobile Robots by fusing Camera and Laser-Based Sensor Data

Robust Real-Time Local Laser Scanner Registration with Uncertainty Estimation

Announcements. Exam #2 next Thursday (March 13) Covers material from Feb. 11 through March 6

Simultaneous local and global state estimation for robotic navigation

Motion Control in Dynamic Multi-Robot Environments

A Comparison of Position Estimation Techniques Using Occupancy Grids

AWireless sensor network has been successfully applied to. Vision-Based Coordinated Localization for Mobile Sensor Networks

DEVELOPMENT OF POSITION MEASUREMENT SYSTEM FOR CONSTRUCTION PILE USING LASER RANGE FINDER

A Simple Interface for Mobile Robot Equipped with Single Camera using Motion Stereo Vision

Proceedings of the 2016 Winter Simulation Conference T. M. K. Roeder, P. I. Frazier, R. Szechtman, E. Zhou, T. Huschka, and S. E. Chick, eds.

Texture Image Segmentation using FCM

Placement and Motion Planning Algorithms for Robotic Sensing Systems

Matching Evaluation of 2D Laser Scan Points using Observed Probability in Unstable Measurement Environment

Fusion of Radar and EO-sensors for Surveillance

Investigation of Methods for Target State Estimation Using Vision Sensors

Canny Edge Based Self-localization of a RoboCup Middle-sized League Robot

An Indoor Mobile Robot Localization Method Based on Information Fusion

Occluded Facial Expression Tracking

Puzzle games (like Rubik s cube) solver

Subpixel Corner Detection Using Spatial Moment 1)

EE368 Project Report CD Cover Recognition Using Modified SIFT Algorithm

Localization from Pairwise Distance Relationships using Kernel PCA

Implementation of Storing Data Algorithm for Tracking Targets using LABVIEW

State Estimation for Continuous-Time Systems with Perspective Outputs from Discrete Noisy Time-Delayed Measurements

COMPARISON OF ROBOT NAVIGATION METHODS USING PERFORMANCE METRICS

Robotics. Lecture 8: Simultaneous Localisation and Mapping (SLAM)

A MATLAB TOOL FOR DEVELOPMENT AND TESTING OF TRACK INITIATION AND MULTIPLE TARGET TRACKING ALGORITHMS

MTRX4700: Experimental Robotics

Three-Dimensional Off-Line Path Planning for Unmanned Aerial Vehicle Using Modified Particle Swarm Optimization

From Theory to Practice: Distributed Coverage Control Experiments with Groups of Robots

Multi-view Surface Inspection Using a Rotating Table

A System of Image Matching and 3D Reconstruction

Stable Vision-Aided Navigation for Large-Area Augmented Reality

Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview

Discovery of the Source of Contaminant Release

METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS

Visual Bearing-Only Simultaneous Localization and Mapping with Improved Feature Matching

DETC APPROXIMATE MOTION SYNTHESIS OF SPHERICAL KINEMATIC CHAINS

Dense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm

Non-Homogeneous Swarms vs. MDP s A Comparison of Path Finding Under Uncertainty

Autonomous Navigation in Complex Indoor and Outdoor Environments with Micro Aerial Vehicles

Smoothing and Mapping using Multiple Robots

Today MAPS AND MAPPING. Features. process of creating maps. More likely features are things that can be extracted from images:

Transcription:

21 American Control Conference Marriott Waterfront, Baltimore, MD, USA June 3-July 2, 21 FrC16.5 A Novel Way to Implement Self-localization in a Multi-robot Experimental Platform Sheng Zhao and Manish Kumar, Member, ASME Abstract Self-localization is an important part of multirobot experiments. The traditional way to obtain the global pose (position and orientation) of a robot is to attach a unique pattern to the robot. A centralized sensor, such as a camera, is typically used to identify the pattern and estimate its pose directly. However, the pattern-based localization scheme lends itself a poor scalability property because, in presence of a large number of robots, the recognition of all patterns may result into a computational bottleneck. Moreover, the implementation of this technique in a decentralized framework is questionable. Additionally, the practicality of designing a large number of unique patterns and attaching them on often miniature robots present some challenges. To overcome the use of a pattern, a novel method has been proposed in this paper for the robots to obtain their global poses without using any pattern. In this method, we run Extended Kalman Filter (EKF) on each robot to fuse the global position data with the control input data to estimate the orientation. In a multi-robot setting, this becomes challenging because the positional data is untagged and robots do not have a priori knowledge about which data pertains to their own position. To overcome this issue, we propose a method in which each robot can identify its own track from the other s tracks by comparing the average errors over time between prediction and measurement in the EKFs that are run on each candidate trajectory. Therefore, instead of trying to identify robots in a centralized manner for localization, we distribute the task to each robot and let them estimate their pose on their own. Extensive simulations and experiments have been conducted and the results are provided to show the effectiveness of this method. I. INTRODUCTION MULTI-ROBOT research has gained a broad attention over the recent years [1, 2]. Comparing to the single robot system, multiple robots working in a team can be more robust to individual failures and more efficient in time to accomplish a job. The multi-robot system has wide potential applications including large area surveillance, search and rescue operation, perimeter protection, and intruder detection [3]. A lot of theoretical advancement has been made in multirobot cooperative control area, and a preferred way to verify the theories has been the use of simulation. However, experimental validation of theories is highly desirable because the real world environment presents several challenges including sensor limitations, noise in sensory information, communication error, and time delays which are difficult to simulate in a Manuscript received September 22, 29. S. Zhao is with the University of Cincinnati, Cincinnati, OH 45221 USA (513-763-9924) zhaose@mail.uc.edu M. Kumar is with University of Cincinnati, Cincinnati, OH 45221 USA kumarmu@ucmail.uc.edu computer. There have been some efforts in developing multirobot experimental testbed. For example, the University of Pennsylvania has built their own miniature robots called Scarab [4] which are equipped with onboard cameras and laser sensors. Three IR emitters on the top of the robot emit lights at certain frequencies. Overhead cameras can capture these patterns to identify each robot and estimate the pose directly. The other examples of multi-robot experimental testbeds include the one developed at MIT. They also have designed miniature robots called SwarmBot [5] which are equipped with specially designed IR sensor rings that can measure the relative positions of the neighboring robots. A single IR emitter on the top of each robot is used to identify the robots and provide their ground truth global position. The heading of robots in the global coordinate system is estimated by combining local estimates of bearing to neighbor and the global estimates of neighbor s positions [5]. It can be noted that both of the above testbeds require unique patterns for identification of the robots. Other than IR emitters, visual patterns such as circular codes in SwisTrack [6] or ARTag in Augmented Reality system [7] have been widely used to identify robots and estimate the position and the heading. Actually, due to the fact that the pattern recognition algorithms are already mature in literature, using patterns has become a popular method to identify robot and provide the global pose in an indoor experiment setup. The benefit of using a pattern is that it provides a unique ID and can be used to estimate the heading of a robot quite easily. However, using patterns has its own drawbacks. Firstly, the pattern based localization system is poorly scalable to the number of robots. When the number of robots is large (hundreds or thousands), creating a unique pattern for each of them is very difficult from a practical point of view. Furthermore, compared with the simple blob extraction algorithm for image processing, the identification of patterns usually require heavy computational load and the time taken to recognize all patterns will increase dramatically with the increase in the number of robots. Secondly, the pattern based localization system won t work in the situations where it is infeasible to attach any pattern on the robot. Usually, either the IR emitters or other visual patterns take extra space to install. Some robots like miniature UAV don t have the room to install patterns like that. Furthermore, pattern based localization schemes usually require a central sensor, and is difficult to implement in a decentralized fashion. Attaching a pattern to a robot amounts to its identification, a concept which against the true spirits of many multi-robot control and swarming algorithms. 978-1-4244-7425-7/1/$26. 21 AACC 6834

This paper investigates if we can carry out the selflocalization without using a pattern. We propose a method which distributes the self-localization task to each robot and let them estimate their position and heading on their own. This provides a decentralized and scalable way to solve the self-localization problem in a multi-robot scenario. In this paper, we utilize an overhead camera to provide positional information of all the robots without tagging them. There is no orientation information provided by the camera. This information is fed to each robot which uses our proposed algorithm to carry out data association and tracking. The challenges in accomplishing the self-localization in multi-robot setting arise from two aspects. First is the estimation of robot s heading. Since we do not use any pattern, the camera does not provide heading information. Then the problem boils down to the estimation of heading based upon only positional observation. The second challenge is the data association. Without a unique ID to the data, it is difficult to associate data to the robot, i.e., to determine which position data from a set of data points belongs to a particular robot. The following sections of the paper will introduce the proposed method. First, the next section formulates the EKF used to estimate the heading and track the robot in a single robot scenario. Next, the third section introduces the data association and self-localization technique in a multi-robot scenario. Finally, we demonstrate the effectiveness of our algorithm via experimental data which is followed by the conclusion and a discussion of the future work. II. TRACKING IN A SINGLE ROBOT SCENARIO In this section, we formulate an Extended Kalman Filter (EKF) to track a robot s position and heading. EKF is broadly used for nonlinear system in estimation and tracking problems [8]. In our problem, we define the state of the robot to include its two-dimensional position x and y, and heading θ. Therefore the state X at each sampling time step k is defined as: X k = x k y k θ k (1) Now we can write down the discrete system equations: Dynamic equation: Observation equation: X k = f(x k,u k,w) (2) = FX k +B k U k +W (3) Z k = HX k +V (4) Where F = B k = 1 1 1 (5) T s cos(θ k ) T s sin(θ k ) T s [ ] 1 H = 1 [ ] vk U k = ω k (6) The Z k R 2 is the observation of the robot s position, i.e., x k and y k. U k R 2 is the motion command or control input of the robot that is composed of v k, the translation velocity, and ω k, the turning rate. T s is the sampling time. W R 3 and V R 2 are the system noise and observation noise respectively. We assume they obey the Gaussian distribution. The covariance matrices of the system and observation noise are represented by as Q R 3 3 and R R 2 2 respectively. As we can see, due to the nonlinear property of matrix B k, the system is nonlinear. Hence an EKF is formulated to track the state of the system. The update equations for the EKF are listed below: The prediction equations are: (7) (8) ˆX k k = F ˆX k k +B k U k (9) P k k = A k P k k A T k +Q (1) The measurement update equations are: K k = P k k H T (HP k k H T +R) (11) ˆX k k = ˆX k k +K k (Z k H ˆX k k ) (12) P k k = (I K k H)P k k (13) where Z k H ˆX k k is called measurement residual, and A k is the Jacobian matrix of partial derivatives of f with respect to X [9]: A [i,j] = f [i] X [j] (X k,u k, ) (14) Applying this to (2), we get the A k as: A k = 1 v k T s sin(θ k ) 1 v k T s cos(θ k ) (15) 1 By using this EKF we can estimate the heading θ k and position of the robot at the same time. III. TRACKING AND SELF-LOCALIZATION IN A MULTI-ROBOT SCENARIO Tracking and self-localization in a multi-robot scenario present new challenges. When there are multiple robots in the platform, each robot will receive all the other robots positions at the same time. In order to use the EKF developed in the last section to track a robot s position and heading, the robot needs to associate to its own observation data from 6835

the set of data received at each time step. This would be easy if the state of the system is fully observable. In other words, if we can observe the heading θ, we can associate the data by simply choosing the measurement with the smallest measurement residual at each time step if we assume the commandu k is different for each robot. However, our system is partially observable (see (4)). Since we don t have the precise knowledge of the current heading which plays a critical role in prediction and data association, we cannot perform the association task on a stepwise basis. Fig. 1 shows that the incorrect estimation of the initial heading of a robot can possibly lead to incorrect identification of its own track. Hence, the data association should be carried out in a different way in our case. Data association is a typical problem in multi-target tracking area which involves finding the correspondence between the data received at current time step with the previous estimates. In this paper we assume that, for every robot s track, there is only one potential data among the received data at each time step that can be associated. So the data association problem in this paper can be divided into two sub-problems: One is how to generate all the potential tracks from the data. Then the next one is how to identify a robot s own track among these candidate tracks. To solve the first sub-problem, we use the Nearest-Neighbor algorithm (NN) [1] to help in the association process. This algorithm uses Euclidean distance as the metric of correlation between two data points x 1 and x 2 : Correlation(x 1,x 2 ) = x 1 x 2 (16) Firstly, we initiate a potential track on each measurement data obtained at the first time step. Then at each time step a new measurement data arrive, we apply NN algorithm on the current measurement Z k and the previous measurement Z k on each potential track then associate the current measurement data to the track with the smallest Correlation(Z k,z k ). In this way, the measurement can be associated at the beginning to generate all potential tracks. However, till now the tracks are still not associated to robots. To overcome the track identification problem, we need some unique information for each robot and the only unique information each robot has is the motion command or control input U k. Relying on this unique information, we can successfully distinguish a robot s track from other tracks by running one EKF on each candidate track. In the proposed technique, each robot maintains one EKF for each track and applies its own control input data on each EKF. Then, the predicted position is compared with the current measurement associated to this track. If the track matches the motion command of the robot, the average measurement residual over time should be the smallest one among all other EKFs. This happens because of the fact that if the data association is carried out properly, the estimation of heading will converge and eventually the average measurement residual will also Case 1: has a correct guess on its initial heading Trajectory 1 (robot 1) Case 2: has an incorrect guess on its initial heading Trajectory 1 (robot 1) Measurement at time step k+1 Trajectory 2 (robot 2) Predicted position at time step k+1 Then it will successfully identify Measurement at time step k trajectory 1 as its own trajectory Trajectory 2 (robot 2) Then it will choose trajectory 2 as its own trajectory which actually belongs to robot 2 Fig. 1: Incorrect estimation of initial heading can lead to a wrong choice of trajectory if the decision is based on a single comparison. converge to the smallest value. Hence, we define: e l k = Z l k H ˆX l k k (17) Where e l k is the prediction and measurement error (or norm of measurement residual) at time step k. Zk l is the associated measurement and H ˆX k k l is the predicted position. The superscript l indicates different candidate tracks. The average prediction and measurement error on candidate track l over time is defined as: E l = 1 T e l k (18) T k=1 E l is in fact the metric reflecting the error in the estimation of heading and so chosen to decide the correct track of the robot. AfterT steps, we choose the track, among all potential tracks, with the smallest E l as the correct track of the robot. The choice of T is based on trial and error. To summarize, the tracking and self-localization in multiple robot scenario includes two phases. In the first phase, we use NN algorithm to generate all possible tracks and run one EKF on each track with the commanded input data. After T steps, we compare the average error E l on each track and choose the track that has the minimum E l as the correct track. The second phase starts after the successful initiation of track. In this phase, we apply the NN algorithm in a stepwise fashion to associate the new measurements to this initialized EKF at each time step. Since the correct track has been initiated and the estimation of heading has converged, the prediction is much accurate now. Hence the NN algorithm, in second phase, is used to associate data based oncorrelation(zk l,h ˆX k k l ). The illustration of the whole algorithm is shown in Fig. 2. IV. EXPERIMENTAL SETUP In this paper, we use Khepera III robots for the experimental validation. Overhead camera (Cognex Insight 56) is used to capture images and transfer them through Ethernet to the desktop for further processing. On the desktop, we 6836

Repeat fortsteps.2 Measure ment Data NN Generate all potential trajectories Run EKF on each trajectory Calculate error l e k on each trajectory.4.6.8 Associate to the correct Measurement NN Measurement Data Keep the one with minimum l E Calculate average error l E for each trajectory Fig. 2: The scheme of the self-localization algorithm in one robot. 15 1.8.2.4.6.8 1 1.2 1.4 1.6 1.8 Fig. 4: The trajectory of a single robot use SwisTrack open source image processing software to extract the positions of robots. This can be done using simple blob extraction algorithm. In order to obtain the ground truth of the global heading (for comparison purposes) for validating our method, we attach the circular patterns [6] on the top of robot. The processing speed on the desktop is 4ms per frame. After the SwisTrack obtains all robots positions and headings, it broadcasts position data to each robot via wireless network. Each robot logs the measurement it receives and the local odometry data (representing the commanded control input) to the desktop at each time step. Then we apply EKF on these data offline using Matlab. The sampling time T s is approximately.5s. In the following experiments, the control program running in each robot is random motion program with obstacle avoidance. The maximum translational velocity is.4m/s and the maximum turning rate is.3rad/s. Player/Stage [11], an open source software, is used as the control and communication the robot. The schematic of the whole overall system control and communication scheme is shown in Fig. 3. A. Experiment 1 (Single robot) In the first experiment, only one robot is used to test the performance of the proposed EKF s tracking method. The trajectory of robot is shown in Fig. 4. As we can see from Fig. 5, the estimation error of the heading converges and remains mostly between ±5 degrees. Also, the estimation converges to the correct heading in 1 time steps. We Fig. 3: The overlook of the system setup error(degree) error(degree) 5 5 5 1 15 2 25 3 35 4 Step 2 15 1 5 5 5 2 5 1 15 2 25 3 35 4 Step Fig. 5: The error in the heading estimation (in degree) can also notice that there are few steps have larger error that represents outliers in data because of errors in image processing. B. Experiment 2 (Multiple robots) The second experiment is conducted using four robots. In this experiment, at each time step, each robot receives all the robots positions. The problem is to determine which among the complete data is a robot s own position, and to estimate the track that includes its heading. During the first phase of our algorithm, each robot needs to maintain four EKFs (one for each trajectory candidate). The trajectories of four robots are shown in Fig. 6. If we plot the average error given by (18) for each EKF track in each robot (Fig. 8), we can see that as time goes by, one of the tracks average errors converges to the smallest value as compared with other tracks. Therefore the robot can obtain the correct track among several tracks by comparing the average errors for each EKF track. In this experiment, after 5 steps, all robots finish the first phase of the algorithm. At the end of this phase, the robot will terminate all other EKF tracks and keep the correct one. In the second phase, each robot uses the nearest neighbor algorithm to associate data to this EKF track. An important aspect for the successful phase 1 is that the 6837

.2.4.6.8 1 1.2 1.4 1.6 1.8.2.2.2.2.4.4.4.4.6.6.6.6.8.8.8.8.8.8.2.4.6.8 1 1.2 1.4 1.6 1.8.8.2.4.6.8 1 1.2 1.4 1.6 1.8 Fig. 6: The trajectories of four robots.8.2.4.6.8 1 1.2 1.4 1.6 1.8.2.4 T = 5.2.4 T = 1.2.4 T = 15.2.4 T = 2.6.8.6.8.6.8.6.8.8.2.4.6.8 1 1.2 1.4 1.6 1.8.8.2.4.6.8 1 1.2 1.4 1.6 1.8.8.2.4.6.8 1 1.2 1.4 1.6 1.8 Fig. 7: The trajectories of four robots at 5, 1, 15 and 2 time step.8.2.4.6.8 1 1.2 1.4 1.6 1.8 command control input for each robot should be different to some extent from the others. For example, if multiple robots run in a straight line or a circle or exhibit some similar motions, the first phase can possibly fail or take a longer time to finish. Experiment 3 is conducted to illustrate this aspect. C. Experiment 3 (Worst case) The robots trajectories are shown in Fig. 7. As we can see, robot 3 and 1 are moving in a similar trajectory (start moving in a straight line and then turn left). Hence, it can be seen from Fig. 9c that it takes approximately 15 time steps for robot 3 s trajectory error to be separable from others. So is the case with robot 1 (Fig. 9a). However, for robots 2 and 4, which follow very distinct trajectories, the trajectory errors converge quickly (Fig. 9b and d). This is intuitive because the control inputs, which play a crucial role in trajectory identification, are similar for robots 1 and 3. V. CONCLUSION AND FUTURE WORK In this paper, we proposed a novel method to localize robots in a multi-robot control experiment. The proposed method is based on Extended Kalman Filter (EKF) in which each robot estimates its global position and heading by fusing global position measurements and control inputs. It was shown that the heading of the robot can be estimated pretty accurately without even its direct measurement. The error of the estimated heading is mostly within ±5 degrees of error. In the multi-robot scenario, each robot associates data at each time step to create all possible tracks based upon all measurements and then identify its own track among these tracks by comparing the average error between the predicted position and the observation for each EKF track. This method helps initiate the track, and once initiated, NN algorithm is used to associate the further data to maintain the track. The major advantage of the proposed technique is that it can be applied in a decentralized manner and is scalable. It was seen that the ability of the proposed method to initiate the correct track depends upon distinct control inputs of the robots. In other words, the motion of the robots should be different enough for them to identify their tracks in a short time. Hence, a specially designed random motion at the beginning can be used to aid in the track initiation process. Future work includes applying the proposed method in a decentralized experimental framework in which the robots use the onboard sensors to carry out the self-localization and tracking. REFERENCES [1] T. Arai, E. Pagello, and L. Parker, Editorial: Advances in multi-robot systems, IEEE Transactions on Robotics and Automation, vol. 18, no. 5, pp. 655 661, 22. [2] L. Bayindir and E. Sahin, A Review of Studies in Swarm Robotics, Turkish Journal of Electrical Engineering, vol. 15, no. 2, 27. [3] L. Parker, Distributed intelligence: Overview of the field and its application in multi-robot systems, Journal of Physical Agents, vol. 2, no. 1, p. 5, 28. [4] N. Michael, J. Fink, and V. Kumar, Experimental testbed for large multirobot teams, IEEE Robotics and Automation Magazine, vol. 15, no. 1, pp. 53 61, 28. [5] J. McLurkin, Analysis and implementation of distributed algorithms for multi-robot systems, Ph.D. dissertation, Massachusetts Institute of Technology, 28. [6] T. Lochmatter, P. Roduit, C. Cianci, N. Correll, J. Jacot, and A. Martinoli, Swistrack-a flexible open source tracking software for multiagent systems, in IEEE/RSJ International Conference on Intelligent Robots and Systems, 28. [7] M. Fiala, ARTag, a Fiducial Marker System Using Digital Techniques, in Proceedings of the 25 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 5), vol. 2. IEEE Computer Society, 25, p. 596. [8] G. Pulford, Taxonomy of multiple target tracking methods, IEEE Proceedings-Radar, Sonar and Navigation, vol. 152, p. 291, 25. [9] G. Welch and G. Bishop, An introduction to the Kalman filter, University of North Carolina at Chapel Hill, Chapel Hill, NC, Tech. Rep., 1995. [1] Y. Bar-Shalom and T. Fortmann, Tracking and data association. Academic Pr, 1988. [11] B. Gerkey, R. Vaughan, and A. Howard, The player/stage project: Tools for multi-robot and distributed sensor systems, in Proceedings of the 11th international conference on advanced robotics. Citeseer, 23, pp. 317 323. 6838

.6 Robot1.7 Robot1.5.6.4.5.3.4.3.2.2.1.1 5 1 15 2 25 3 (a) 5 1 15 2 25 3 (a).9 Robot2.7 Robot2.8.6.7.6.5.5.4.4.3.3.2.2.1.1 5 1 15 2 25 3 (b) 5 1 15 2 25 3 (b).6 Robot3.1 Robot3.9.5.8.4.7.3.6.5.4.2.3.1.2.1 5 1 15 2 25 3 (c) 5 1 15 2 25 3 (c).5 Robot4.8 Robot4.45.7.4.35.6.3.25.2.5.4.3.15.1.2.5.1 5 1 15 2 25 3 Fig. 8: The average prediction-measurement errors E l for experiment 2 (d) 5 1 15 2 25 3 Fig. 9: The average prediction-measurement errors E l for experiment 3 (d) 6839