Probabilistic Double-Distance Algorithm of Search after Static or Moving Target by Autonomous Mobile Agent

Size: px
Start display at page:

Download "Probabilistic Double-Distance Algorithm of Search after Static or Moving Target by Autonomous Mobile Agent"

Transcription

1 2010 IEEE 26-th Convention of Electrical and Electronics Engineers in Israel Probabilistic Double-Distance Algorithm of Search after Static or Moving Target by Autonomous Mobile Agent Eugene Kagan Dept. Industrial Engineering Gal Goren Dept. Mechanical Engineering Irad Ben-Gal Dept. Industrial Engineering Abstract We propose a real-time algorithm of search and path planning after a static or a moving target in a discrete probability space. The search is conducted by an autonomous mobile agent that is given an initial probability distribution of the target's location, and at each search step obtains information regarding target's location in the agent's local neighborhood. The suggested algorithm implements a decision-making procedure of a probabilistic version of local search with estimated global distances and results in agent's path over the domain. The suggested algorithm finds efficiently both static and moving targets, as well as targets that change their movement patterns during the search. Additional information regarding the target locations, which is unknown at the beginning of the search, can be integrated in the search in real-time, as well. It is found that for the search after a static target, the algorithm actions depend on the global estimation at all stages of the search, while for the search after a moving target the global estimations mostly affect the initial search steps. Preliminary analysis shows that for the search after a static target the obtained average number of steps is close to optimal, while for the Markovian target the average number of steps is at least in the bounds that are provided by known search methods. Index terms Search and screening, static and moving target, autonomous mobile agent 1. INTRODUCTION Consider a target moving in a bounded discrete domain and a mobile agent that searches the target. The agent starts with an initial probability distribution of the target's possible location over the domain. At each search step the agent obtains information regarding target's location in the agent's local neighborhood, and terminates the search when the target is located in this neighborhood. The target movements are governed by a definite Markov process that is known to the agent, while the target is not informed about the agent's behavior. The goal of the search agent is to find the target in minimum number of steps. In the presented formulation, the problem of search after static or mobile target was originated by Koopman [14] during the Second World War in response to the German submarine threat in the Atlantic [20]. The initial studies of the problem dealt with the probabilistic search for a static target or a target that moves relatively slow, while in the last decades the main interest is focused on search for targets with comparable velocity to that of the search agent, where the location probabilities of target s change over time [5], [8], [21]. There are two main approaches of the considered search problem. Following the first approach, an optimal path of the agent is created off-line by use of global optimization methods. It is further assumed that the search effort [14] is infinitely divisible over the points of the domain and that the search period is defined and finite. The path-planning is based on considerations of optimal allocations that represent optimal distributions of search efforts over the domain with respect to the location probabilities and the search agent abilities [4]. For cases of search in continuous time and space the existence of optimal allocations has been proved, and optimal paths of the agent were obtained [19]. Regarding discrete time and space search, existence of optimal allocations were proven in the most general cases, while the algorithms that generate such allocations were obtained only for a search after a static target [1], [19]. For the search after a moving target, optimal algorithms have been built for the concave detection probabilities function [2] and for any detection probability function and some necessary (but not sufficient) conditions of optimality [22], [23]. In variants of the above problem, methods of partially observable Markov decision processes were applied and resulted in optimal allocations when the evaluation function is piecewise linear and convex [7] and when dynamic restrictions on search efforts are applied [18]. The second approach to the above problem addresses a realtime search, such that the path-planning is generated during the search process. Decision-making regarding each next step of the search agent is conducted on the basis of the obtained information, and after choosing the next step, the agent updates the available information about target s locations. Such approach follows the line of A* algorithms and local optimization for the search over graphs or grid domains. The basic real-time A* algorithm results in a path of the search agent in the graph, and implements a search after a static target [15]. Later, a similar approach was applied in the moving-target search algorithm [9]. For the decision-making, both algorithms use local distances and heuristic distance /10/$ IEEE

2 estimations without a utilization of probabilistic information regarding target s location. Further studies of the real-time algorithms and methods of learning lead to a number of different procedures, some of which implemented heuristic learning methods of target location probabilities [3], [17]. A certain progress in the unification of off-line and realtime search was obtained by the use of information theory methods and application of real-time search procedures with information metrics [10], [11]. The obtained algorithms used admissible informational distances both for local decisionmaking and as global distance estimations and demonstrated Huffman-like optimal and near-optimal results in the search after static and moving targets. The decision-making in these algorithms is conducted by the use of partitions of the considered domain on the basis of the target s location probabilities. In further developments of these informational algorithms, a topology of the domain was specified [12] and an optimal algorithm of search after static target by a number of cooperative and non-cooperative searchers was found [13]. Although the significant advancement that has been achieved in the development of real-time search algorithms and the considerable progress in the studies of optimal allocating algorithms, no effective procedures for real-time allocating and on-line path-planning have been reported so far. The objective of this work is to develop and to analyze a real-time algorithm of search after static and moving targets that implements methods of allocating and of on-line pathplanning. The suggested algorithm is based on a probabilistic version of a local search with estimated global distances. Decision-making at each step of the search applies a probabilistic distance estimation that utilizes distance estimations and the characteristics of target's movement. The algorithm is applicable for search after static or moving targets and for the search after a target that changes its movement pattern during the search process. Additionally, obtained information regarding the target locations, which was unknown to the agent at the beginning of the search, can be taken into account in real-time, as well. 2. PROBLEM FORMULATION AND BACKGROUND Let ={,,, } be a set that represents a rectangle geographical domain of the size (, ), =. Each point, = 1,2,,, represents a cell, in which the target can be located at any time moment = 0, 1, 2,, and is called location. For simplicity, we index the points of X in linear order such that for the Cartesian coordinates (, ) of the point, its index is given by =( 1) +, where = 1,, and =1,,. Assume that at each time moment t there is a defined probability mass function : [0, 1] that specifies the probabilities ( ) =Pr{ = }, = 1, 2,,, of target's location in the points of the set X, where denotes the target's location at time moment t. Probabilities ( ) are called location probabilities, and for any t it holds true that 0 ( ) 1and ( ) =1. The set X with the defined probability mass function is the sample space. The target's movement is governed by a discrete Markov process with transition probabilities matrix =, where = Pr = = and =1for each = 1,2,..,. If matrix is a unit matrix, then the target is static, otherwise, we say that it is moving. Given target location probabilities (),, at time moment t, location probabilities at the next time moment +1are calculated as ( ) = ( ), =1,2,..,. The search agent moves over the sample space and looks for a target by testing a neighboring observed area of a certain radius >0. The ability of the agent to detect the target in the observed area is defined by a detection function that, for any observed area, specifies a probability (,) of detecting the target in the area given that the target is located in,, and the agent applies on a search effort [14], [19]. In this paper, we assume that for any search effort [ 0, ), the probability (,) =( ) is defined as an indicator (, ) of the area, i.e., ( ) =1 if and ( ) =0 otherwise. In addition, we assume that for all time moments t the observed areas have the same size. The search process is conducted as follows [10], [11]. The search agent deals with the estimated location probabilities ( ), that for the initial time moment =0are equal to the initial location probabilities ( ) = ( ),, = 1, 2,,. At each time moment t, the agent chooses an observed area and observes it. If an observation result in ( ) =1, the search terminates. Otherwise, the agent updates estimated location probabilities ( ) over the sample space X according to the Bayes rule. The updated probabilities are called observed probabilities and are denoted by ( ). To the observed probabilities, the agent applies some known or estimated rule of the target movement and obtains the estimated location probabilities ( ), = 1, 2,,, for the next time moment +1. E.g., if the target is moving according to a Markov process with a transition probabilities matrix, then the estimated location probabilities ( ) for the next time moment +1 are defined as ( ) = ( ), =1,2,..,. Given the estimated probabilities (),, the search agent applies a decision making regarding the next observed area to select and the process continues. The goal is to find a policy for choosing the areas such that the search procedure terminates in a minimal average (over possible target moves) number of steps, while the centers of the areas form a continuous trajectory over the sample space X with respect to a Manhattan metric. An existence of such policy in the most general form and the convergence of the search procedure are provided by the A*-type algorithms [9], [15] and their informational successors [10], [11], while the existence of optimal allocations are supported by Washburn theorems [23]. Here, we present a real-time algorithm that, as it follows from simulative results, creates the required trajectory of the agent

3 3. DOUBLE-DISTANCE SEARCH ALGORITHM As indicated above, the suggested algorithm is based on a probabilistic version of local search with estimated global distances, while decision-making utilizes distance estimations and the characteristics of target's movement. We assume that at each time moment t, both the search agent and the target can apply one of the following movements: ={,,,, }, where stands for move north, - move south, - move east, - move west and - stay in the current point. These are the simplest movements that correspond to the Manhattan metric; for other metrics additional movements can be considered. The target s choice of the movements is conducted according to its transition probabilities matrix, while the agent s choice follows the decision that is obtained according to the result ( ) of the observation of the area around the agent. The outline of the suggested search algorithm includes the following steps. Given initial target location probabilities ( ), = 1, 2,,, and transition probabilities matrix, 1. The search agent starts with estimated probabilities ( ) = ( ), = 1, 2,,, and an initial observed area. 2. If ( ) =1, i.e. the target is found and the process terminates. Otherwise, the following steps are conducted. 3. While ( ) 1,do 3.1. Calculate probabilities ( ), = 1, 2,, : For all do: if then set ( ) =0. For all do: set ( ) = ( ) ( ) Calculate probabilities ( ), = 1,2,,: For all do: set ( ) = ( ) Calculate weights for the movements : For all do: set = (),where is an area that will be observed if the agent will move according to the movement Estimate global probabilities if need: If =0for each k then Set = (,). else Set = Apply local decision-making Set = () Move to the area Set =+1and obtain observation result ( ). 4. Return obtained area, for which ( ) =1. As indicated above, at each time moment t, the target moves to its next location according to the transition probabilities matrix. Since the target is not aware of the agent activities up to the successful termination of the search, the target s movements do not depend on the searcher actions. The algorithm includes two key functions: global probabilities estimation function ( ) and local decision-making function ( ). The first function resolves an uncertainty for local decisionmaking, while the second conducts local decision-making and points to the next observed area. Let us start with the ( ) function. In the algorithm we implement local decision-making that is based on a discrete analogous to the maximum gradient over expected probabilities ( ) from the current location of the search agent to possible observed areas. In the other words, the agent chooses the next feasible observed area according to the defined movements V for which the observation will maximally affect the probabilities ( ) as they are determined in line 3.4 of the algorithm. The distances between the probability mass functions, where k corresponds to the indices of the movements, and the probability mass function are calculated according to the Ornstein distance metric [16] which is defined as follows: (,) = ( ) ( ). Given the set of possible movements V, the outline of the local decision-making function is the following. (): 1. For all movements do: 1.1. Determine according to Set = () For all do: if then set ( ) = For all do: set ( ) =( ) ( ) For all do: if then set ( ) = For all do: set ( ) =( ) ( ) Set ( ) = (,) + (1 )(,). 2. Set = argmax{ ( ) over all }; ties are broken randomly. 3. Return A. The calculations of the function lines have the same meaning as in the corresponding lines of the algorithm. In particular, line 1.2 defines the weights for the movements, while in the lines observed probabilities for successful and unsuccessful observations are calculated. In line 1.7 the Ornstein distance averaged over successful and unsuccessful observations is calculated, and in line 2 the area, for which this distance is maximal, is chosen. Local decision-making function acts on given probabilities ( ), = 1, 2,,. If there is no uncertainty as it is defined in the line 3.4 of the algorithm, then this function is applied to the probabilities ( ). If, in opposite, for all possible movements, the probabilities ( ) are equal to zero (see line 3.3 of the algorithm), then to resolve the uncertainty global probabilities estimation is applied. The probabilities estimation is implemented by the use of the function ( ) that starts from the probabilities ( ) and results in the probabilities ( ), = 1, 2,,. Probabilities ( ) are calculated by the use of the target s transition probabilities matrix as target location probabilities after a certain number of steps m, while the calculation starts from the agent s expected probabilities ( ), = 1, 2,,. The indicated number of steps m is defined by the use of the estimated target's location and by the entropy measure [6] that specifies the lack of information available to the searcher. Since the search agent observes the areas A that include equal

4 number of points, the entropy is calculated by using the size s of the areas. Then, the required or missing amount of information is defined as follows: ( ) =log () ( )log ( ). Notice that the logarithms are calculated on the base size s of the observed areas A. Such a calculation follows an assumption that the areas A have the same size and that the observations are error free. The estimated target's location is defined by a center of distribution that is calculated as a "center of gravity" for the probabilities ( ), = 1, 2,,. E.g., over the rectangle domain of the size (, ), =, coordinates (,), the estimated target's location, are calculated as follows: (,): 1. Set =0and =0. 2. For =1,, do: For =1,, do: Set = + ( ), =( 1) For =1,, do: For =1,, do: Set = + ( ), =( 1) Set = () and = (). 5. Return =(,). Denote by y the current location of the search agent. Then, by the use of information ( ) and the defined estimated target's location, function ( ) is outlined as follows. (,): 1. Calculate estimates target location: Set = (,). 2. Calculate distance between and agent's location y: 3. Set = (, ).Calculate number of estimated steps: = ( ) (). 4. For all do: set ( ) = ( ). 5. Do m times: 5.1. For all do: set ( ) =( ) For all do: set ( ) = ( ). 6. Return. Note that () stands for calculating the closest integer number of the real number b, and () stands for calculating the closest integer number that is greater than b. For two points =(, ) and = (, ), function (, ) calculates a geographical distance between the points. In the considered algorithm of search in discrete space, the distance is defined by the Manhattan metric: (, ) = +, which is applied in the numerical simulations below. 4. NUMERICAL ANALYSIS AND SIMULATION RESULTS The performance of the suggested algorithm was studied by the use of simulated trials with different initial distributions of the target's location and different types of target's movement. The algorithm was implemented in the form of MatLab scripts. Below, we present an example of such simulation results that have been obtained for search over a square domain with size = 2500 points. Each trial included 1000 sessions. For each session an initial target location probabilities were generated by a combination of ten binormal distributions with random centers and unit standard deviations. Initial locations of the target were chosen randomly according to the initial location probabilities. The trials were executed both for static target and for two types of the Markovian target. In each session the search was executed up to finding the target by the agent. The results of the trials over 1000 sessions for each type of target movement are presented in Table I. Target type Radius = 1, Obs. Area 3 3=9 TABLE I SIMULATION RESULTS Radius = 2, Obs. area 5 5=25 Radius = 3, Obs. area 7 7=49 Static Markov Brown In the table, stands for the average number of search steps and for the standard deviation. The possible movements of the Markovian target are: move north, move south, move east, move west and stay in the current point, while the Brownian target was prohibited to stay in its current position. In all trials the average distance between the initial target location and the agent's starting point was approximately 37.0 and standard deviation was approximately A relation between statistics of the initial distance and statistics of the number of search steps is discussed below. As expected, from the simulation results it follows that the average number of search steps strongly depends on the radius of the observed area; however both for static and for moving targets this dependence is not linearly proportional. In addition, it can be seen that the search after the Markovian target results in a lower average number of steps (at least for small observed areas) than the Brownian equivalent. This property is similar to the property of general real-time local search algorithms in a probabilistic space [10], [11]. Let us note on the algorithm's efficiency. Consider a static target search with the observed area =1, while the agent at each time moment covers a square area A of 9 points and obtains observation result () {0, 1}. The target's location probabilities distribution is combined by ten equivalent binormal distributions with unit standard deviations, i.e., the target is located in one of the areas that can be checked by the agent in one time moment, while the agent is located exactly in the center of this area. The average initial distance between the search agent and the target in this scenario is that presents an average cost of finding the target in one of ten areas. The minimal number of search steps for finding the target in one of the equiprobable areas is bounded by the entropy of the number of areas. Hence, the minimal average

5 number of search steps in the considered scenario is log (10) = The obtained result (see Table I) = is rather close to this bound. Now let us take into account that the agent moves sequentially, so it finds the target in the area of its location in two steps. The standard deviation of the initial distances is Hence, the theoretical standard deviation of search steps is log (10) 2 = that is close to the obtained standard deviation (see Table I) = The same statistical results are obtained in the simulations for the other initial target location probabilities and the sizes of the observed area. To evaluate the effectiveness of the algorithm for the search after a moving target, let us consider its approximate comparison with the known near-optimal Forward and Backward (FAB) algorithm [23]. In particular, it is known [23] that for the search after a target over 67 cells the FAB algorithm requires steps, while at each step the searcher checks a single cell and obtains an observation result with an error of In comparison, for the considered case the sample space contains 2500 points, while the agent observes an area that contains 9 points. Hence, an approximate evaluation of the number of steps required by the FAB algorithm for this search is (2500 (9 67) ) = This value is greater than the average number of steps = (see Table I) that is required by the suggested algorithm. However, note that the FAB algorithm allows erroneous detections, while in the suggested algorithms the detections have to be accurate. The obtained preliminary results support a claim that the suggested algorithm is close to optimal in the case of static target search and for the search after a moving target the obtained average number of steps is in the bounds that are provided by known search methods. Notice again, that the algorithm can be applied to various types of target movements and for both static and moving targets. Additional simulations were conducted for different types of local decision-making, such as integrating additional weighted global distance estimations to the expected target's location. Nonetheless, the obtained results for both the static and the moving targets, were found to be worse or statistically equivalent to the results presented in Table I. 5. CONCLUSION In the paper, we suggested a real-time algorithm of search after static and moving targets in discrete time and space. The algorithm inputs are the initial target's location probabilities over the considered domain and the transition matrix of possible target's moves and its output is the path trajectory of the agent that minimizes the average number of search steps. The algorithm is based on a probabilistic version of local search with estimated global distances. Decision-making at each step of the search applies a probabilistic distance estimation that utilizes distance estimations characterize the target's movement. The suggested algorithm finds efficiently both static and moving targets, as well as targets that change their movement patterns during the search. Nevertheless, for a search after a static target, the algorithm steps depend on global estimation at most stages of the search, while for a search after a moving target, global estimation mostly affect the initial search steps. Preliminary analysis and simulation results show that for the search after a static target the obtained average number of steps is close to optimal, while for the search after a moving target it is in the bounds that are provided by known nearoptimal search methods. REFERENCES [1] R. Ahlswege, and I. Wegener, Search Problems. John Wiley & Sons: New York, [2] S. S. Brown, Optimal search for a moving target in discrete time and space, Operations Research, vol. 28, no. 6, 1980, pp [3] V. Bulitko, and G. Lee, Learning in real-time search: a unifying framework, J. Artificial Intelligence Research, vol. 25, 2006, pp [4] A. P. Ciervo, Search for moving targets, Pacific-Sierra Research Corporation, Santa-Monica, CA, PSR Report 619 for Office of Naval Research, Arlington, Virginia, [5] D. C. Cooper, J. R. Frost, and R. Quincy Robe, Compatibility of land SAR procedures with search theory, US Department of Homeland Security, Washington, [6] T. M. Cover, and J. A. Thomas, Elements of Information Theory. John Wiley & Sons: New York, etc., [7] J. N. Eagle, The optimal search for a moving target when the search path is constrained, Operations Research, vol. 32, 1984, pp [8] J. R. Frost, and L. D. Stone, Review of Search Theory: Advances and Applications to Search and Rescue Decision Support, US Coast Guard Research and Development Center, Groton, [9] T. Ishida, and R. E. Korf, Moving target search: a real-time search for changing goals, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 17, no. 6, 1995, pp [10] E. Kagan, and I. Ben-Gal, An informational search for a moving target, in Proc IEEE 24-th Conv. of Electrical and Electronics Engineers in Israel, e-publ. [11] E. Kagan, and I. Ben-Gal, A MDP test for dynamic targets via informational distance measure, in Proc Conf. IIERM, pp [12] E. Kagan, and I. Ben-Gal, Topology of the sample space in the problem of informational search for a moving target, in Proc Conf. IIERM, pp [13] E. Kagan, and I. Ben-Gal, Search after a static by multiple searchers by the use of informational distance measures, in Proc Conf. IE&M, p. 72 & e-publ. [14] B. O. Koopman, The theory of search, I-III, Operations research, vol. 4, 1956, pp , ; vol. 5, 1957, pp [15] R. E. Korf, Real-time heuristic search, Artificial intelligence, vol. 42, no. 2-3, 1990, pp [16] D. S. Ornstein, Ergodic Theory, Randomness, and Dynamical Systems. Yale University Press: New Haven and London, [17] M. Shimbo, and T. Ishida, Controlling the learning process of real-time heuristic search, Artificial Intelligence, vol. 146, 2003, pp [18] S. Singh, and V. Krishnamurthy, The optimal search for a moving target when the search path is constrained: the infinite-horizon case, IEEE Trans. Automatic Control, vol. 48, no. 3, 2003, pp [19] L. D. Stone, Theory of Optimal Search. Academic Press: New York, San Francisco, London, [20] L. D. Stone, The process of search planning: current approaches and continuing problems, Operations Research, vol. 31, no. 2, 1983, pp [21] L. D. Stone, What's happened in search theory since the 1975 Lanchester prize? Operations Research, vol. 37, no. 3, 1989, pp [22] A. R. Washburn, Search and Detection. ORSA Books: Arlington, VA, [23] A. R. Washburn, Search for a moving target: the FAB algorithm, Operations Research, vol. 31, no. 4, 1983, pp

CSE151 Assignment 2 Markov Decision Processes in the Grid World

CSE151 Assignment 2 Markov Decision Processes in the Grid World CSE5 Assignment Markov Decision Processes in the Grid World Grace Lin A484 gclin@ucsd.edu Tom Maddock A55645 tmaddock@ucsd.edu Abstract Markov decision processes exemplify sequential problems, which are

More information

Predictive Autonomous Robot Navigation

Predictive Autonomous Robot Navigation Predictive Autonomous Robot Navigation Amalia F. Foka and Panos E. Trahanias Institute of Computer Science, Foundation for Research and Technology-Hellas (FORTH), Heraklion, Greece and Department of Computer

More information

A SIMULATED ANNEALING ALGORITHM FOR SOME CLASS OF DISCRETE-CONTINUOUS SCHEDULING PROBLEMS. Joanna Józefowska, Marek Mika and Jan Węglarz

A SIMULATED ANNEALING ALGORITHM FOR SOME CLASS OF DISCRETE-CONTINUOUS SCHEDULING PROBLEMS. Joanna Józefowska, Marek Mika and Jan Węglarz A SIMULATED ANNEALING ALGORITHM FOR SOME CLASS OF DISCRETE-CONTINUOUS SCHEDULING PROBLEMS Joanna Józefowska, Marek Mika and Jan Węglarz Poznań University of Technology, Institute of Computing Science,

More information

15-780: MarkovDecisionProcesses

15-780: MarkovDecisionProcesses 15-780: MarkovDecisionProcesses J. Zico Kolter Feburary 29, 2016 1 Outline Introduction Formal definition Value iteration Policy iteration Linear programming for MDPs 2 1988 Judea Pearl publishes Probabilistic

More information

Real-Time Adaptive A*

Real-Time Adaptive A* Real-Time Adaptive A* Sven Koenig Computer Science Department University of Southern California Los Angeles, CA - skoenig@usc.edu Maxim Likhachev Computer Science Department Carnegie Mellon University

More information

The strong chromatic number of a graph

The strong chromatic number of a graph The strong chromatic number of a graph Noga Alon Abstract It is shown that there is an absolute constant c with the following property: For any two graphs G 1 = (V, E 1 ) and G 2 = (V, E 2 ) on the same

More information

Constraint Satisfaction Problems (CSPs)

Constraint Satisfaction Problems (CSPs) Constraint Satisfaction Problems (CSPs) CPSC 322 CSP 1 Poole & Mackworth textbook: Sections 4.0-4.2 Lecturer: Alan Mackworth September 28, 2012 Problem Type Static Sequential Constraint Satisfaction Logic

More information

On the Complexity of the Policy Improvement Algorithm. for Markov Decision Processes

On the Complexity of the Policy Improvement Algorithm. for Markov Decision Processes On the Complexity of the Policy Improvement Algorithm for Markov Decision Processes Mary Melekopoglou Anne Condon Computer Sciences Department University of Wisconsin - Madison 0 West Dayton Street Madison,

More information

INDIAN INSTITUTE OF MANAGEMENT CALCUTTA WORKING PAPER SERIES. WPS No. 644/ August A Markov-based Diurnal Mobility Model for 3G Cellular Networks

INDIAN INSTITUTE OF MANAGEMENT CALCUTTA WORKING PAPER SERIES. WPS No. 644/ August A Markov-based Diurnal Mobility Model for 3G Cellular Networks INDIAN INSTITUTE OF MANAGEMENT CALCUTTA WORKING PAPER SERIES WPS No. 644/ August 2009 A Markov-based Diurnal Mobility Model for 3G Cellular Networks by Samir K Sadhukhan SSA, IIM Calcutta, Diamond Harbour

More information

Study on Indoor and Outdoor environment for Mobile Ad Hoc Network: Random Way point Mobility Model and Manhattan Mobility Model

Study on Indoor and Outdoor environment for Mobile Ad Hoc Network: Random Way point Mobility Model and Manhattan Mobility Model Study on and Outdoor for Mobile Ad Hoc Network: Random Way point Mobility Model and Manhattan Mobility Model Ibrahim khider,prof.wangfurong.prof.yinweihua,sacko Ibrahim khider, Communication Software and

More information

Interactive Treatment Planning in Cancer Radiotherapy

Interactive Treatment Planning in Cancer Radiotherapy Interactive Treatment Planning in Cancer Radiotherapy Mohammad Shakourifar Giulio Trigila Pooyan Shirvani Ghomi Abraham Abebe Sarah Couzens Laura Noreña Wenling Shang June 29, 212 1 Introduction Intensity

More information

Figure 1: An example of a hypercube 1: Given that the source and destination addresses are n-bit vectors, consider the following simple choice of rout

Figure 1: An example of a hypercube 1: Given that the source and destination addresses are n-bit vectors, consider the following simple choice of rout Tail Inequalities Wafi AlBalawi and Ashraf Osman Department of Computer Science and Electrical Engineering, West Virginia University, Morgantown, WV fwafi,osman@csee.wvu.edug 1 Routing in a Parallel Computer

More information

roving the Learning Efficiencies of ealtime Searc

roving the Learning Efficiencies of ealtime Searc roving the Learning Efficiencies of ealtime Searc Toru Ishida Masashi Shimbo Department of Information Science Kyoto University Kyoto, 606-01, JAPAN {ishida, shimbo}@kuis.kyoto-u.ac.jp Abstract The capability

More information

Self-formation, Development and Reproduction of the Artificial System

Self-formation, Development and Reproduction of the Artificial System Solid State Phenomena Vols. 97-98 (4) pp 77-84 (4) Trans Tech Publications, Switzerland Journal doi:.48/www.scientific.net/ssp.97-98.77 Citation (to be inserted by the publisher) Copyright by Trans Tech

More information

Distributed and Asynchronous Policy Iteration for Bounded Parameter Markov Decision Processes

Distributed and Asynchronous Policy Iteration for Bounded Parameter Markov Decision Processes Distributed and Asynchronous Policy Iteration for Bounded Parameter Markov Decision Processes Willy Arthur Silva Reis 1, Karina Valdivia Delgado 2, Leliane Nunes de Barros 1 1 Departamento de Ciência da

More information

Partially Observable Markov Decision Processes. Silvia Cruciani João Carvalho

Partially Observable Markov Decision Processes. Silvia Cruciani João Carvalho Partially Observable Markov Decision Processes Silvia Cruciani João Carvalho MDP A reminder: is a set of states is a set of actions is the state transition function. is the probability of ending in state

More information

An Improved Policy Iteratioll Algorithm for Partially Observable MDPs

An Improved Policy Iteratioll Algorithm for Partially Observable MDPs An Improved Policy Iteratioll Algorithm for Partially Observable MDPs Eric A. Hansen Computer Science Department University of Massachusetts Amherst, MA 01003 hansen@cs.umass.edu Abstract A new policy

More information

6. Concluding Remarks

6. Concluding Remarks [8] K. J. Supowit, The relative neighborhood graph with an application to minimum spanning trees, Tech. Rept., Department of Computer Science, University of Illinois, Urbana-Champaign, August 1980, also

More information

ADAPTIVE TILE CODING METHODS FOR THE GENERALIZATION OF VALUE FUNCTIONS IN THE RL STATE SPACE A THESIS SUBMITTED TO THE FACULTY OF THE GRADUATE SCHOOL

ADAPTIVE TILE CODING METHODS FOR THE GENERALIZATION OF VALUE FUNCTIONS IN THE RL STATE SPACE A THESIS SUBMITTED TO THE FACULTY OF THE GRADUATE SCHOOL ADAPTIVE TILE CODING METHODS FOR THE GENERALIZATION OF VALUE FUNCTIONS IN THE RL STATE SPACE A THESIS SUBMITTED TO THE FACULTY OF THE GRADUATE SCHOOL OF THE UNIVERSITY OF MINNESOTA BY BHARAT SIGINAM IN

More information

Branch & Bound (B&B) and Constraint Satisfaction Problems (CSPs)

Branch & Bound (B&B) and Constraint Satisfaction Problems (CSPs) Branch & Bound (B&B) and Constraint Satisfaction Problems (CSPs) Alan Mackworth UBC CS 322 CSP 1 January 25, 2013 P&M textbook 3.7.4 & 4.0-4.2 Lecture Overview Recap Branch & Bound Wrap up of search module

More information

1 Project: Fast Trajectory Replanning for Computer Games

1 Project: Fast Trajectory Replanning for Computer Games Project: Fast rajectory Replanning for omputer Games Figure : otal nnihilation by avedog onsider characters in real-time computer games, such as otal nnihilation shown in Figure. o make them easy to control,

More information

Detection and Mitigation of Cyber-Attacks using Game Theory

Detection and Mitigation of Cyber-Attacks using Game Theory Detection and Mitigation of Cyber-Attacks using Game Theory João P. Hespanha Kyriakos G. Vamvoudakis Correlation Engine COAs Data Data Data Data Cyber Situation Awareness Framework Mission Cyber-Assets

More information

Revision of a Floating-Point Genetic Algorithm GENOCOP V for Nonlinear Programming Problems

Revision of a Floating-Point Genetic Algorithm GENOCOP V for Nonlinear Programming Problems 4 The Open Cybernetics and Systemics Journal, 008,, 4-9 Revision of a Floating-Point Genetic Algorithm GENOCOP V for Nonlinear Programming Problems K. Kato *, M. Sakawa and H. Katagiri Department of Artificial

More information

Tabu search and genetic algorithms: a comparative study between pure and hybrid agents in an A-teams approach

Tabu search and genetic algorithms: a comparative study between pure and hybrid agents in an A-teams approach Tabu search and genetic algorithms: a comparative study between pure and hybrid agents in an A-teams approach Carlos A. S. Passos (CenPRA) carlos.passos@cenpra.gov.br Daniel M. Aquino (UNICAMP, PIBIC/CNPq)

More information

Formal Model. Figure 1: The target concept T is a subset of the concept S = [0, 1]. The search agent needs to search S for a point in T.

Formal Model. Figure 1: The target concept T is a subset of the concept S = [0, 1]. The search agent needs to search S for a point in T. Although this paper analyzes shaping with respect to its benefits on search problems, the reader should recognize that shaping is often intimately related to reinforcement learning. The objective in reinforcement

More information

Exploring Unknown Environments with Real-Time Search or Reinforcement Learning

Exploring Unknown Environments with Real-Time Search or Reinforcement Learning Exploring Unknown Environments with Real-Time Search or Reinforcement Learning Sven Koenig College of Computing, Georgia Institute of Technology skoenig@ccgatechedu Abstract Learning Real-Time A* (LRTA*)

More information

Planning and Control: Markov Decision Processes

Planning and Control: Markov Decision Processes CSE-571 AI-based Mobile Robotics Planning and Control: Markov Decision Processes Planning Static vs. Dynamic Predictable vs. Unpredictable Fully vs. Partially Observable Perfect vs. Noisy Environment What

More information

Evaluating Classifiers

Evaluating Classifiers Evaluating Classifiers Charles Elkan elkan@cs.ucsd.edu January 18, 2011 In a real-world application of supervised learning, we have a training set of examples with labels, and a test set of examples with

More information

Probabilistic Graphical Models

Probabilistic Graphical Models School of Computer Science Probabilistic Graphical Models Theory of Variational Inference: Inner and Outer Approximation Eric Xing Lecture 14, February 29, 2016 Reading: W & J Book Chapters Eric Xing @

More information

INTRODUCTION TO HEURISTIC SEARCH

INTRODUCTION TO HEURISTIC SEARCH INTRODUCTION TO HEURISTIC SEARCH What is heuristic search? Given a problem in which we must make a series of decisions, determine the sequence of decisions which provably optimizes some criterion. What

More information

Approximately Uniform Random Sampling in Sensor Networks

Approximately Uniform Random Sampling in Sensor Networks Approximately Uniform Random Sampling in Sensor Networks Boulat A. Bash, John W. Byers and Jeffrey Considine Motivation Data aggregation Approximations to COUNT, SUM, AVG, MEDIAN Existing work does not

More information

One Dim~nsional Representation Of Two Dimensional Information For HMM Based Handwritten Recognition

One Dim~nsional Representation Of Two Dimensional Information For HMM Based Handwritten Recognition One Dim~nsional Representation Of Two Dimensional Information For HMM Based Handwritten Recognition Nafiz Arica Dept. of Computer Engineering, Middle East Technical University, Ankara,Turkey nafiz@ceng.metu.edu.

More information

Q-learning with linear function approximation

Q-learning with linear function approximation Q-learning with linear function approximation Francisco S. Melo and M. Isabel Ribeiro Institute for Systems and Robotics [fmelo,mir]@isr.ist.utl.pt Conference on Learning Theory, COLT 2007 June 14th, 2007

More information

Estimating the Information Rate of Noisy Two-Dimensional Constrained Channels

Estimating the Information Rate of Noisy Two-Dimensional Constrained Channels Estimating the Information Rate of Noisy Two-Dimensional Constrained Channels Mehdi Molkaraie and Hans-Andrea Loeliger Dept. of Information Technology and Electrical Engineering ETH Zurich, Switzerland

More information

H = {(1,0,0,...),(0,1,0,0,...),(0,0,1,0,0,...),...}.

H = {(1,0,0,...),(0,1,0,0,...),(0,0,1,0,0,...),...}. II.4. Compactness 1 II.4. Compactness Note. Conway states on page 20 that the concept of compactness is an extension of benefits of finiteness to infinite sets. I often state this idea as: Compact sets

More information

Task Scheduling Using Probabilistic Ant Colony Heuristics

Task Scheduling Using Probabilistic Ant Colony Heuristics The International Arab Journal of Information Technology, Vol. 13, No. 4, July 2016 375 Task Scheduling Using Probabilistic Ant Colony Heuristics Umarani Srikanth 1, Uma Maheswari 2, Shanthi Palaniswami

More information

Statistical Testing of Software Based on a Usage Model

Statistical Testing of Software Based on a Usage Model SOFTWARE PRACTICE AND EXPERIENCE, VOL. 25(1), 97 108 (JANUARY 1995) Statistical Testing of Software Based on a Usage Model gwendolyn h. walton, j. h. poore and carmen j. trammell Department of Computer

More information

Markov Decision Processes. (Slides from Mausam)

Markov Decision Processes. (Slides from Mausam) Markov Decision Processes (Slides from Mausam) Machine Learning Operations Research Graph Theory Control Theory Markov Decision Process Economics Robotics Artificial Intelligence Neuroscience /Psychology

More information

Stability of Marriage and Vehicular Parking

Stability of Marriage and Vehicular Parking Stability of Marriage and Vehicular Parking Daniel Ayala, Ouri Wolfson, Bo Xu, Bhaskar DasGupta, and Jie Lin University of Illinois at Chicago Abstract. The proliferation of mobile devices, location-based

More information

MIT Programming Contest Team Contest 1 Problems 2008

MIT Programming Contest Team Contest 1 Problems 2008 MIT Programming Contest Team Contest 1 Problems 2008 October 5, 2008 1 Edit distance Given a string, an edit script is a set of instructions to turn it into another string. There are four kinds of instructions

More information

Discrete geometry. Lecture 2. Alexander & Michael Bronstein tosca.cs.technion.ac.il/book

Discrete geometry. Lecture 2. Alexander & Michael Bronstein tosca.cs.technion.ac.il/book Discrete geometry Lecture 2 Alexander & Michael Bronstein tosca.cs.technion.ac.il/book Numerical geometry of non-rigid shapes Stanford University, Winter 2009 The world is continuous, but the mind is discrete

More information

Marco Wiering Intelligent Systems Group Utrecht University

Marco Wiering Intelligent Systems Group Utrecht University Reinforcement Learning for Robot Control Marco Wiering Intelligent Systems Group Utrecht University marco@cs.uu.nl 22-11-2004 Introduction Robots move in the physical environment to perform tasks The environment

More information

1 Introduction and Examples

1 Introduction and Examples 1 Introduction and Examples Sequencing Problems Definition A sequencing problem is one that involves finding a sequence of steps that transforms an initial system state to a pre-defined goal state for

More information

Surrogate Gradient Algorithm for Lagrangian Relaxation 1,2

Surrogate Gradient Algorithm for Lagrangian Relaxation 1,2 Surrogate Gradient Algorithm for Lagrangian Relaxation 1,2 X. Zhao 3, P. B. Luh 4, and J. Wang 5 Communicated by W.B. Gong and D. D. Yao 1 This paper is dedicated to Professor Yu-Chi Ho for his 65th birthday.

More information

Points covered an odd number of times by translates

Points covered an odd number of times by translates Points covered an odd number of times by translates Rom Pinchasi August 5, 0 Abstract Let T be a fixed triangle and consider an odd number of translated copies of T in the plane. We show that the set of

More information

CHAPTER 2 CONVENTIONAL AND NON-CONVENTIONAL TECHNIQUES TO SOLVE ORPD PROBLEM

CHAPTER 2 CONVENTIONAL AND NON-CONVENTIONAL TECHNIQUES TO SOLVE ORPD PROBLEM 20 CHAPTER 2 CONVENTIONAL AND NON-CONVENTIONAL TECHNIQUES TO SOLVE ORPD PROBLEM 2.1 CLASSIFICATION OF CONVENTIONAL TECHNIQUES Classical optimization methods can be classified into two distinct groups:

More information

5/30/13. Guarding, Searching and Pursuing Evaders using Multiagent Systems. Todays topics. Example Scenario. Todays topics

5/30/13. Guarding, Searching and Pursuing Evaders using Multiagent Systems. Todays topics. Example Scenario. Todays topics //13 Guarding, Searching and Pursuing Evaders using Multiagent Systems Petter Ögren Computer Vision and Active Perception (CVAP) Airport Power plant Military base Port Factory Example Scenario This field

More information

A Simple and Strong Algorithm for Reconfiguration of Hexagonal Metamorphic Robots

A Simple and Strong Algorithm for Reconfiguration of Hexagonal Metamorphic Robots 50 A Simple and Strong Algorithm for Reconfiguration of Hexagonal Metamorphic Robots KwangEui Lee Department of Multimedia Engineering, Dongeui University, Busan, Korea Summary In this paper, we propose

More information

Optimal Configuration of Compute Nodes for Synthetic Aperture Radar Processing

Optimal Configuration of Compute Nodes for Synthetic Aperture Radar Processing Optimal Configuration of Compute Nodes for Synthetic Aperture Radar Processing Jeffrey T. Muehring and John K. Antonio Deptartment of Computer Science, P.O. Box 43104, Texas Tech University, Lubbock, TX

More information

Mobility Control for Complete Coverage in Wireless Sensor Networks

Mobility Control for Complete Coverage in Wireless Sensor Networks Mobility Control for Complete Coverage in Wireless Sensor Networks Zhen Jiang Computer Sci. Dept. West Chester University West Chester, PA 9383, USA zjiang@wcupa.edu Jie Wu Computer Sci. & Eng. Dept. Florida

More information

A new method for determination of a wave-ray trace based on tsunami isochrones

A new method for determination of a wave-ray trace based on tsunami isochrones Bull. Nov. Comp. Center, Math. Model. in Geoph., 13 (2010), 93 101 c 2010 NCC Publisher A new method for determination of a wave-ray trace based on tsunami isochrones An.G. Marchuk Abstract. A new method

More information

Topology and Topological Spaces

Topology and Topological Spaces Topology and Topological Spaces Mathematical spaces such as vector spaces, normed vector spaces (Banach spaces), and metric spaces are generalizations of ideas that are familiar in R or in R n. For example,

More information

A PARAMETRIC SIMPLEX METHOD FOR OPTIMIZING A LINEAR FUNCTION OVER THE EFFICIENT SET OF A BICRITERIA LINEAR PROBLEM. 1.

A PARAMETRIC SIMPLEX METHOD FOR OPTIMIZING A LINEAR FUNCTION OVER THE EFFICIENT SET OF A BICRITERIA LINEAR PROBLEM. 1. ACTA MATHEMATICA VIETNAMICA Volume 21, Number 1, 1996, pp. 59 67 59 A PARAMETRIC SIMPLEX METHOD FOR OPTIMIZING A LINEAR FUNCTION OVER THE EFFICIENT SET OF A BICRITERIA LINEAR PROBLEM NGUYEN DINH DAN AND

More information

Note Set 4: Finite Mixture Models and the EM Algorithm

Note Set 4: Finite Mixture Models and the EM Algorithm Note Set 4: Finite Mixture Models and the EM Algorithm Padhraic Smyth, Department of Computer Science University of California, Irvine Finite Mixture Models A finite mixture model with K components, for

More information

A Message Passing Strategy for Decentralized. Connectivity Maintenance in Multi-Agent Surveillance

A Message Passing Strategy for Decentralized. Connectivity Maintenance in Multi-Agent Surveillance A Message Passing Strategy for Decentralized Connectivity Maintenance in Multi-Agent Surveillance Derya Aksaray Boston University, Boston, MA, 225 A. Yasin Yazıcıoğlu 2 Massachusetts Institute of Technology,

More information

6. Learning Partitions of a Set

6. Learning Partitions of a Set 6. Learning Partitions of a Set Also known as clustering! Usually, we partition sets into subsets with elements that are somewhat similar (and since similarity is often task dependent, different partitions

More information

Bilinear Programming

Bilinear Programming Bilinear Programming Artyom G. Nahapetyan Center for Applied Optimization Industrial and Systems Engineering Department University of Florida Gainesville, Florida 32611-6595 Email address: artyom@ufl.edu

More information

Path Planning. Marcello Restelli. Dipartimento di Elettronica e Informazione Politecnico di Milano tel:

Path Planning. Marcello Restelli. Dipartimento di Elettronica e Informazione Politecnico di Milano   tel: Marcello Restelli Dipartimento di Elettronica e Informazione Politecnico di Milano email: restelli@elet.polimi.it tel: 02 2399 3470 Path Planning Robotica for Computer Engineering students A.A. 2006/2007

More information

Distributed minimum spanning tree problem

Distributed minimum spanning tree problem Distributed minimum spanning tree problem Juho-Kustaa Kangas 24th November 2012 Abstract Given a connected weighted undirected graph, the minimum spanning tree problem asks for a spanning subtree with

More information

Theoretical Concepts of Machine Learning

Theoretical Concepts of Machine Learning Theoretical Concepts of Machine Learning Part 2 Institute of Bioinformatics Johannes Kepler University, Linz, Austria Outline 1 Introduction 2 Generalization Error 3 Maximum Likelihood 4 Noise Models 5

More information

Partial Differential Equations

Partial Differential Equations Simulation in Computer Graphics Partial Differential Equations Matthias Teschner Computer Science Department University of Freiburg Motivation various dynamic effects and physical processes are described

More information

Automatic synthesis of switching controllers for linear hybrid systems: Reachability control

Automatic synthesis of switching controllers for linear hybrid systems: Reachability control Automatic synthesis of switching controllers for linear hybrid systems: Reachability control Massimo Benerecetti and Marco Faella Università di Napoli Federico II, Italy Abstract. We consider the problem

More information

A *69>H>N6 #DJGC6A DG C<>C::G>C<,8>:C8:H /DA 'D 2:6G, ()-"&"3 -"(' ( +-" " " % '.+ % ' -0(+$,

A *69>H>N6 #DJGC6A DG C<>C::G>C<,8>:C8:H /DA 'D 2:6G, ()-&3 -(' ( +-   % '.+ % ' -0(+$, The structure is a very important aspect in neural network design, it is not only impossible to determine an optimal structure for a given problem, it is even impossible to prove that a given structure

More information

Markov Decision Processes and Reinforcement Learning

Markov Decision Processes and Reinforcement Learning Lecture 14 and Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark Slides by Stuart Russell and Peter Norvig Course Overview Introduction Artificial Intelligence

More information

Multiagent Planning with Factored MDPs

Multiagent Planning with Factored MDPs Appeared in Advances in Neural Information Processing Systems NIPS-14, 2001. Multiagent Planning with Factored MDPs Carlos Guestrin Computer Science Dept Stanford University guestrin@cs.stanford.edu Daphne

More information

Approximate Linear Programming for Average-Cost Dynamic Programming

Approximate Linear Programming for Average-Cost Dynamic Programming Approximate Linear Programming for Average-Cost Dynamic Programming Daniela Pucci de Farias IBM Almaden Research Center 65 Harry Road, San Jose, CA 51 pucci@mitedu Benjamin Van Roy Department of Management

More information

COMP SCI 5400 SP2017 Exam 1 Key

COMP SCI 5400 SP2017 Exam 1 Key COMP SCI 5400 SP2017 Exam 1 Key This is a closed-book, closed-notes exam. The only items you are permitted to use are writing implements. Mark each sheet of paper you use with your name and the string

More information

arxiv: v1 [math.co] 4 Apr 2011

arxiv: v1 [math.co] 4 Apr 2011 arxiv:1104.0510v1 [math.co] 4 Apr 2011 Minimal non-extensible precolorings and implicit-relations José Antonio Martín H. Abstract. In this paper I study a variant of the general vertex coloring problem

More information

Stochastic branch & bound applying. target oriented branch & bound method to. optimal scenario tree reduction

Stochastic branch & bound applying. target oriented branch & bound method to. optimal scenario tree reduction Stochastic branch & bound applying target oriented branch & bound method to optimal scenario tree reduction Volker Stix Vienna University of Economics Department of Information Business Augasse 2 6 A-1090

More information

An Eternal Domination Problem in Grids

An Eternal Domination Problem in Grids Theory and Applications of Graphs Volume Issue 1 Article 2 2017 An Eternal Domination Problem in Grids William Klostermeyer University of North Florida, klostermeyer@hotmail.com Margaret-Ellen Messinger

More information

Heuristic Algorithms for Multiconstrained Quality-of-Service Routing

Heuristic Algorithms for Multiconstrained Quality-of-Service Routing 244 IEEE/ACM TRANSACTIONS ON NETWORKING, VOL 10, NO 2, APRIL 2002 Heuristic Algorithms for Multiconstrained Quality-of-Service Routing Xin Yuan, Member, IEEE Abstract Multiconstrained quality-of-service

More information

Artificial Intelligence p.1/49. n-queens. Artificial Intelligence p.2/49. Initial state: the empty board or a board with n random

Artificial Intelligence p.1/49. n-queens. Artificial Intelligence p.2/49. Initial state: the empty board or a board with n random Example: n-queens Put n queens on an n n board with no two queens on the same row, column, or diagonal A search problem! State space: the board with 0 to n queens Initial state: the empty board or a board

More information

Three-Dimensional Off-Line Path Planning for Unmanned Aerial Vehicle Using Modified Particle Swarm Optimization

Three-Dimensional Off-Line Path Planning for Unmanned Aerial Vehicle Using Modified Particle Swarm Optimization Three-Dimensional Off-Line Path Planning for Unmanned Aerial Vehicle Using Modified Particle Swarm Optimization Lana Dalawr Jalal Abstract This paper addresses the problem of offline path planning for

More information

Constructive floorplanning with a yield objective

Constructive floorplanning with a yield objective Constructive floorplanning with a yield objective Rajnish Prasad and Israel Koren Department of Electrical and Computer Engineering University of Massachusetts, Amherst, MA 13 E-mail: rprasad,koren@ecs.umass.edu

More information

Outline for today s lecture. Informed Search. Informed Search II. Review: Properties of greedy best-first search. Review: Greedy best-first search:

Outline for today s lecture. Informed Search. Informed Search II. Review: Properties of greedy best-first search. Review: Greedy best-first search: Outline for today s lecture Informed Search II Informed Search Optimal informed search: A* (AIMA 3.5.2) Creating good heuristic functions Hill Climbing 2 Review: Greedy best-first search: f(n): estimated

More information

Optimal Crane Scheduling

Optimal Crane Scheduling Optimal Crane Scheduling IonuŃ Aron Iiro Harjunkoski John Hooker Latife Genç Kaya March 2007 1 Problem Schedule 2 cranes to transfer material between locations in a manufacturing plant. For example, copper

More information

COS Lecture 13 Autonomous Robot Navigation

COS Lecture 13 Autonomous Robot Navigation COS 495 - Lecture 13 Autonomous Robot Navigation Instructor: Chris Clark Semester: Fall 2011 1 Figures courtesy of Siegwart & Nourbakhsh Control Structure Prior Knowledge Operator Commands Localization

More information

To earn the extra credit, one of the following has to hold true. Please circle and sign.

To earn the extra credit, one of the following has to hold true. Please circle and sign. CS 188 Spring 2011 Introduction to Artificial Intelligence Practice Final Exam To earn the extra credit, one of the following has to hold true. Please circle and sign. A I spent 3 or more hours on the

More information

Heuristic Search and Advanced Methods

Heuristic Search and Advanced Methods Heuristic Search and Advanced Methods Computer Science cpsc322, Lecture 3 (Textbook Chpt 3.6 3.7) May, 15, 2012 CPSC 322, Lecture 3 Slide 1 Course Announcements Posted on WebCT Assignment1 (due on Thurs!)

More information

Generalized Coordinates for Cellular Automata Grids

Generalized Coordinates for Cellular Automata Grids Generalized Coordinates for Cellular Automata Grids Lev Naumov Saint-Peterburg State Institute of Fine Mechanics and Optics, Computer Science Department, 197101 Sablinskaya st. 14, Saint-Peterburg, Russia

More information

A Game-Theoretic Framework for Congestion Control in General Topology Networks

A Game-Theoretic Framework for Congestion Control in General Topology Networks A Game-Theoretic Framework for Congestion Control in General Topology SYS793 Presentation! By:! Computer Science Department! University of Virginia 1 Outline 2 1 Problem and Motivation! Congestion Control

More information

Introduction to Machine Learning CMU-10701

Introduction to Machine Learning CMU-10701 Introduction to Machine Learning CMU-10701 Clustering and EM Barnabás Póczos & Aarti Singh Contents Clustering K-means Mixture of Gaussians Expectation Maximization Variational Methods 2 Clustering 3 K-

More information

Kapitel 4: Clustering

Kapitel 4: Clustering Ludwig-Maximilians-Universität München Institut für Informatik Lehr- und Forschungseinheit für Datenbanksysteme Knowledge Discovery in Databases WiSe 2017/18 Kapitel 4: Clustering Vorlesung: Prof. Dr.

More information

Heuristic Search in Cyclic AND/OR Graphs

Heuristic Search in Cyclic AND/OR Graphs From: AAAI-98 Proceedings. Copyright 1998, AAAI (www.aaai.org). All rights reserved. Heuristic Search in Cyclic AND/OR Graphs Eric A. Hansen and Shlomo Zilberstein Computer Science Department University

More information

Planar Graphs. 1 Graphs and maps. 1.1 Planarity and duality

Planar Graphs. 1 Graphs and maps. 1.1 Planarity and duality Planar Graphs In the first half of this book, we consider mostly planar graphs and their geometric representations, mostly in the plane. We start with a survey of basic results on planar graphs. This chapter

More information

The Randomized Shortest Path model in a nutshell

The Randomized Shortest Path model in a nutshell Panzacchi M, Van Moorter B, Strand O, Saerens M, Kivimäki I, Cassady St.Clair C., Herfindal I, Boitani L. (2015) Predicting the continuum between corridors and barriers to animal movements using Step Selection

More information

IE 5531: Engineering Optimization I

IE 5531: Engineering Optimization I IE 5531: Engineering Optimization I Lecture 3: Linear Programming, Continued Prof. John Gunnar Carlsson September 15, 2010 Prof. John Gunnar Carlsson IE 5531: Engineering Optimization I September 15, 2010

More information

Conflict based Backjumping for Constraints Optimization Problems

Conflict based Backjumping for Constraints Optimization Problems Conflict based Backjumping for Constraints Optimization Problems Roie Zivan and Amnon Meisels {zivanr,am}@cs.bgu.ac.il Department of Computer Science, Ben-Gurion University of the Negev, Beer-Sheva, 84-105,

More information

Learning Where You Are Going and from Whence You Came: h- and g-cost Learning in Real-Time Heuristic Search

Learning Where You Are Going and from Whence You Came: h- and g-cost Learning in Real-Time Heuristic Search Proceedings of the Twenty-Second International Joint Conference on Artificial Intelligence Learning Where You Are Going and from Whence You Came: h- and g-cost Learning in Real-Time Heuristic Search Nathan

More information

Final Exam. Introduction to Artificial Intelligence. CS 188 Spring 2010 INSTRUCTIONS. You have 3 hours.

Final Exam. Introduction to Artificial Intelligence. CS 188 Spring 2010 INSTRUCTIONS. You have 3 hours. CS 188 Spring 2010 Introduction to Artificial Intelligence Final Exam INSTRUCTIONS You have 3 hours. The exam is closed book, closed notes except a two-page crib sheet. Please use non-programmable calculators

More information

Modeling Robot Path Planning with CD++

Modeling Robot Path Planning with CD++ Modeling Robot Path Planning with CD++ Gabriel Wainer Department of Systems and Computer Engineering. Carleton University. 1125 Colonel By Dr. Ottawa, Ontario, Canada. gwainer@sce.carleton.ca Abstract.

More information

SIMULATION OF ARTIFICIAL SYSTEMS BEHAVIOR IN PARAMETRIC EIGHT-DIMENSIONAL SPACE

SIMULATION OF ARTIFICIAL SYSTEMS BEHAVIOR IN PARAMETRIC EIGHT-DIMENSIONAL SPACE 78 Proceedings of the 4 th International Conference on Informatics and Information Technology SIMULATION OF ARTIFICIAL SYSTEMS BEHAVIOR IN PARAMETRIC EIGHT-DIMENSIONAL SPACE D. Ulbikiene, J. Ulbikas, K.

More information

Supplementary material: Planning with Abstract Markov Decision Processes

Supplementary material: Planning with Abstract Markov Decision Processes Supplementary material: Planning with Abstract Markov Decision Processes Nakul Gopalan ngopalan@cs.brown.edu Marie desjardins mariedj@umbc.edu Michael L. Littman mlittman@cs.brown.edu James MacGlashan

More information

Maximum number of edges in claw-free graphs whose maximum degree and matching number are bounded

Maximum number of edges in claw-free graphs whose maximum degree and matching number are bounded Maximum number of edges in claw-free graphs whose maximum degree and matching number are bounded Cemil Dibek Tınaz Ekim Pinar Heggernes Abstract We determine the maximum number of edges that a claw-free

More information

NOT(Faster Implementation ==> Better Algorithm), A Case Study

NOT(Faster Implementation ==> Better Algorithm), A Case Study NOT(Faster Implementation ==> Better Algorithm), A Case Study Stephen Balakirsky and Thomas Kramer Intelligent Systems Division National Institute of Standards and Technology Gaithersburg, MD 0899-830

More information

Parameterization of Triangular Meshes with Virtual Boundaries

Parameterization of Triangular Meshes with Virtual Boundaries Parameterization of Triangular Meshes with Virtual Boundaries Yunjin Lee 1;Λ Hyoung Seok Kim 2;y Seungyong Lee 1;z 1 Department of Computer Science and Engineering Pohang University of Science and Technology

More information

A NEW MILP APPROACH FOR THE FACILITY LAYOUT DESIGN PROBLEM WITH RECTANGULAR AND L/T SHAPED DEPARTMENTS

A NEW MILP APPROACH FOR THE FACILITY LAYOUT DESIGN PROBLEM WITH RECTANGULAR AND L/T SHAPED DEPARTMENTS A NEW MILP APPROACH FOR THE FACILITY LAYOUT DESIGN PROBLEM WITH RECTANGULAR AND L/T SHAPED DEPARTMENTS Yossi Bukchin Michal Tzur Dept. of Industrial Engineering, Tel Aviv University, ISRAEL Abstract In

More information

COMPLETE AND SCALABLE MULTI-ROBOT PLANNING IN TUNNEL ENVIRONMENTS. Mike Peasgood John McPhee Christopher Clark

COMPLETE AND SCALABLE MULTI-ROBOT PLANNING IN TUNNEL ENVIRONMENTS. Mike Peasgood John McPhee Christopher Clark COMPLETE AND SCALABLE MULTI-ROBOT PLANNING IN TUNNEL ENVIRONMENTS Mike Peasgood John McPhee Christopher Clark Lab for Intelligent and Autonomous Robotics, Department of Mechanical Engineering, University

More information

Adaptive Robotics - Final Report Extending Q-Learning to Infinite Spaces

Adaptive Robotics - Final Report Extending Q-Learning to Infinite Spaces Adaptive Robotics - Final Report Extending Q-Learning to Infinite Spaces Eric Christiansen Michael Gorbach May 13, 2008 Abstract One of the drawbacks of standard reinforcement learning techniques is that

More information

Using Linear Programming for Bayesian Exploration in Markov Decision Processes

Using Linear Programming for Bayesian Exploration in Markov Decision Processes Using Linear Programming for Bayesian Exploration in Markov Decision Processes Pablo Samuel Castro and Doina Precup McGill University School of Computer Science {pcastr,dprecup}@cs.mcgill.ca Abstract A

More information