Best-First AND/OR Search for Graphical Models

Similar documents
Best-First AND/OR Search for Most Probable Explanations

Advancing AND/OR Search for Optimization Using Diverse Principles

AND/OR Cutset Conditioning

AND/OR Graph Search for Genetic Linkage Analysis

The Relationship Between AND/OR Search and Variable Elimination

Evaluating the Impact of AND/OR Search on 0-1 Integer Linear Programming

Evaluating the impact of AND/OR search on 0-1 integer linear programming

Towards Parallel Search for Optimization in Graphical Models

Compiling Constraint Networks into AND/OR Multi-Valued Decision Diagrams (AOMDDs)

Pushing Forward Marginal MAP with Best-First Search

From Exact to Anytime Solutions for Marginal MAP

On the Practical Significance of Hypertree vs Tree Width

Efficient AND/OR Search Algorithms for Exact MAP Inference Task over Graphical Models

AND/OR Multi-Valued Decision Diagrams for Constraint Networks

Evaluating Weighted DFS Branch and Bound over Graphical Models

Advances in Parallel Branch and Bound

Example: Bioinformatics. Soft Constraint Processing. From Optimal CSP to Soft CSP. Overview. From Optimal CSP to Soft CSP.

Stochastic Anytime Search for Bounding Marginal MAP

Anytime Best+Depth-First Search for Bounding Marginal MAP

AND/OR Multi-Valued Decision Diagrams (AOMDDs) for Graphical Models

Search Algorithms for Solving Queries on Graphical Models & the Importance of Pseudo-trees in their Complexity.

From Exact to Anytime Solutions for Marginal MAP

On the Space-Time Trade-off in Solving Constraint Satisfaction Problems*

Semi-Independent Partitioning: A Method for Bounding the Solution to COP s

Modern Exact and Approximate Combinatorial Optimization Algorithms: Max Product and Max Sum product

Look-ahead with Mini-Bucket Heuristics for MPE

DJAO: A Communication-Constrained DCOP algorithm that combines features of ADOPT and Action-GDL

A General Scheme for Automatic Generation of Search Heuristics from Specification Dependencies. Kalev Kask and Rina Dechter

Weighted heuristic anytime search: new schemes for optimization over graphical models

Mini-Buckets: A General Scheme for Generating Approximations in Automated Reasoning

Winning the PASCAL 2011 MAP Challenge with Enhanced AND/OR Branch-and-Bound

On the Space-Time Trade-off in Solving Constraint Satisfaction Problems*

Anytime AND/OR Depth-first Search for Combinatorial Optimization

AND/OR Search Spaces for Graphical Models

Secure Combinatorial Optimization using DFS-based Variable Elimination

Constraint Satisfaction Problems

Anytime Anyspace AND/OR Best-first Search for Bounding Marginal MAP

UNIVERSITY OF CALIFORNIA, IRVINE. Advancing Heuristics for Search over Graphical Models DISSERTATION

Towards Parallel Non Serial Dynamic Programming for Solving Hard Weighted CSP

Constraint Optimization Techniques for MultiObjective Branch and Bound Search

Dynamic heuristics for branch and bound search on tree-decomposition of Weighted CSPs

A Hybrid Recursive Multi-Way Number Partitioning Algorithm

Solving lexicographic multiobjective MIPs with Branch-Cut-Price

Integer Programming ISE 418. Lecture 7. Dr. Ted Ralphs

A Parameterized Local Consistency for Redundant Modeling in Weighted CSPs

Boolean Representations and Combinatorial Equivalence

MB-DPOP: A New Memory-Bounded Algorithm for Distributed Optimization

Constraint Optimisation Problems. Constraint Optimisation. Cost Networks. Branch and Bound. Dynamic Programming

Mathematical Programming Formulations, Constraint Programming

Topological parameters for time-space tradeoff

Artificial Intelligence

mywbut.com Uninformed Search

Utilizing Device Behavior in Structure-Based Diagnosis

Interleaved Depth-First Search *

CSE 473: Artificial Intelligence

Foundations of Artificial Intelligence

An iteration of Branch and Bound One iteration of Branch and Bound consists of the following four steps: Some definitions. Branch and Bound.

Constraint Satisfaction Problems

Constraint Satisfaction Problems

Outline. Best-first search

Lecture 6: Constraint Satisfaction Problems (CSPs)

Anytime Anyspace AND/OR Search for Bounding the Partition Function

Constraint Satisfaction Problems. Chapter 6

Optimization I : Brute force and Greedy strategy

Conflict Graphs for Combinatorial Optimization Problems

Model Checking I Binary Decision Diagrams

Recursive column generation for the Tactical Berth Allocation Problem

Evaluating the Impact of AND/OR Search on 0-1 Integer Linear Programming

Multi-Objective Russian Doll Search

An Efficient Approximate Algorithm for Winner Determination in Combinatorial Auctions

Consistency and Set Intersection

Integrating Probabilistic Reasoning with Constraint Satisfaction

Anytime AND/OR Best-First Search for Optimization in Graphical Models

Some Applications of Graph Bandwidth to Constraint Satisfaction Problems

General Methods and Search Algorithms

Overview. Overview. Example: Full Adder Diagnosis. Example: Full Adder Diagnosis. Valued CSP. Diagnosis using Bounded Search and Symbolic Inference

Constraint Satisfaction Problems

Search : Lecture 2. September 9, 2003

Constraint satisfaction search. Combinatorial optimization search.

Multi-Way Number Partitioning

Midterm Examination CS540-2: Introduction to Artificial Intelligence

Constraint Satisfaction. AI Slides (5e) c Lin

Column Generation Method for an Agent Scheduling Problem

The Size Robust Multiple Knapsack Problem

Homework 2: Multi-unit combinatorial auctions (due Nov. 7 before class)

Constraint Satisfaction Problems (CSPs)

CS 4100 // artificial intelligence

Loopy Belief Propagation

Constraint Satisfaction

6.034 Notes: Section 3.1

Processing Probabilistic and Deterministic Graphical Models. Rina Dechter DRAFT

!!" '!" Fall 2003 ICS 275A - Constraint Networks 2

Lecture 18. Questions? Monday, February 20 CS 430 Artificial Intelligence - Lecture 18 1

Constraint Satisfaction Problems (CSPs) Introduction and Backtracking Search

Incremental ARA*: An Incremental Anytime Search Algorithm for Moving-Target Search

3 No-Wait Job Shops with Variable Processing Times

An Improved Hybrid Genetic Algorithm for the Generalized Assignment Problem

Kalev Kask. Abstract. This paper evaluates the power of a new scheme that generates

A generic framework for solving CSPs integrating decomposition methods

Algorithms for Integer Programming

Transcription:

est-irst N/ Search for Graphical Models Radu Marinescu and Rina echter School of Information and omputer Science University of alifornia, Irvine, 92627 {radum,dechter}@ics.uci.edu bstract The paper presents and evaluates the power of best-first search over N/ search spaces in graphical models. The main virtue of the N/ representation is its sensitivity to the structure of the graphical model, which can translate into significant time savings. Indeed, in recent years depth-first N/ ranch-and-ound algorithms were shown to be very effective when exploring such search spaces, especially when using caching. Since best-first strategies are known to be superior to depth-first when memory is utilized, exploring the best-first control strategy is called for. In this paper we introduce two classes of best-first N/ search algorithms: those that explore a context-minimal N/ search graph and use static variable orderings, and those that use dynamic variable orderings but explore an N/ search tree. The superiority of the best-first search approach is demonstrated empirically on various real-world benchmarks. Introduction Graphical models such as belief networks or constraint networks are a widely used representation framework for reasoning with probabilistic and deterministic information. These models use graphs to capture conditional independencies between variables, allowing a concise representation of the knowledge as well as efficient graph-based query processing algorithms. Optimization problems such as finding the most likely state of a belief network or finding a solution that violates the least number of constraints can be defined within this framework and they are typically tackled with either search or inference algorithms. The N/ search space for graphical models (echter & Mateescu 26) is a framework for search that is sensitive to the independencies in the model, often resulting in reduced search spaces. The impact of the N/ search to optimization in graphical models was explored in recent years focusing exclusively on depth-first search. The N/ ranch-and-ound first introduced by (Marinescu & echter 25) traverses the N/ search tree in a depth-first manner. The memory intensive ranchand-ound algorithm (Marinescu & echter 26c) explores an N/ search graph, rather than a tree, by caching previously computed results and retrieving them opyright c 27, ssociation for the dvancement of rtificial Intelligence (www.aaai.org). ll rights reserved. when the same subproblems are encountered again. These algorithms were initially restricted to a static variable ordering. More recently, (Marinescu & echter 26b) showed how dynamic variable selection heuristics influence ranch-and-ound search over N/ trees. The depthfirst N/ search algorithms were shown to outperform dramatically state-of-the-art ranch-and-ound algorithms searching the traditional space. In this paper we focus on best-first search algorithms. We present a new N/ search algorithm that explores a context-minimal N/ search graph in a best-first rather than depth-first manner. Since variable selection can have a dramatic impact on search performance, we also introduce a best-first N/ search algorithm that explores the N/ search tree, rather than the graph, and combines the static N/ decomposition principle with dynamic variable selection heuristics. Under conditions of admissibility and monotonicity of the heuristic function, best-first search is known to expand the minimal number of nodes, at the expense of using additional memory (echter & Pearl 985). In practice, these savings in number of nodes may often translate into time savings as well. We focus the empirical evaluation on three common optimization problems: solving Weighted SPs (de Givry et al. 25), finding the Most Probable xplanation in belief networks (Pearl 988), and / Integer Linear Programming (Nemhauser & Wolsey 988). Our results show conclusively that the best-first N/ search approach outperforms significantly the depth-first N/ ranch-and- ound search algorithms on various benchmarks. ackground onstraint Optimization Problems finite onstraint Optimization Problem (OP) is a triple P = X,,, where X = {X,..., X n } is a set of variables, = {,..., n } is a set of finite domains and = {f,..., f m } is a set of cost functions. ost functions can be either soft or hard (constraints). Without loss of generality we assume that hard constraints are represented as (bi-valued) cost functions. llowed and forbidden tuples have cost and, respectively. The scope of function f i, denoted scope(f i ) X, is the set of arguments of f i. The goal is to find a complete value assignment to the vari- 7

N N N N N N (a) (b) N (c) N (d) igure : N/ Search Spaces for Graphical Models ables that minimizes the global cost function, namely to find x =argmin X m i= f i. Given a OP instance, its primal graph G associates each variable with a node and connects any two nodes whose variables appear in the scope of the same function. N/ Search Spaces for Graphical Models The usual way to do search is to instantiate variables, following a static/dynamic variable ordering. In the simplest case, this process defines an search tree, whose nodes represent partial assignments. This search space does not capture the structure of the underlying graphical model. However, to remedy this problem, N/ search spaces for graphical models were recently introduced by (echter & Mateescu 26). They are defined using a backbone pseudotree (reuder & Quinn 985). INITION (pseudo-tree) Given an undirected graph G = (V,), a directed rooted tree T = (V, ) defined on all its nodes is called pseudo-tree if any arc of G which is not included in is a back-arc, namely it connects a node to an ancestor in T. N/ Search Trees Given a OP instance P =(X,, ), its primal graph G and a pseudo-tree T of G, the associated N/ search tree, denoted S T, has alternating levels of nodes and N nodes. The nodes are labeled X i and correspond to the variables. The N nodes are labeled X i,x i and correspond to value assignments in the domains of the variables. The root of the N/ search tree is an node, labeled with the root of the pseudo-tree T. The children of an node X i are N nodes labeled with assignments X i,x i, consistent along the path from the root, path(x i,x i )=( X,x,..., X i,x i ). The children of an N node X i,x i are nodes labeled with the children of variable X i in T. Semantically, the states represent alternative solutions, whereas the N states represent problem decomposition into independent subproblems, all of which need be solved. When the pseudotree is a chain, the N/ search tree coincides with the regular search tree. solution tree Sol ST of S T is an N/ subtree such that: (i) it contains the root of S T ; (ii) if a nonterminal N node n S T is in Sol ST then all its children are in Sol ST ; (iii) if a nonterminal node n S T is in Sol ST then exactly one of its children is in Sol ST. xample igures (a) and (b) show the primal graph of a binary OP instance and its pseudo-tree together with the back-arcs (dotted lines). igure (c) shows the N/ search tree based on the pseudo-tree, for bi-valued variables. solution subtree is highlighted. rc Labels and Node Values The arcs from nodes X i to N nodes X i,x i in the N/ search tree S T are annotated by labels derived from the cost functions in. INITION 2 (label) The label l(x i, X i,x i ) of the arc from the node X i to the N node X i,x i is the sum of all the cost functions whose scope includes X i and is fully assigned along path(x i,x i ), evaluated at the values along the path. Given a labeled N/ search tree, each node can be associated with a value (echter & Mateescu 26). INITION 3 (value) The value v(n) of a node n S T is defined recursively as follows: (i) if n = X i,x i is a terminal N node then v(n) = ; (ii) if n = X i,x i is an internal N node then v(n) = n succ(n) v(n ); (iii) if n = X i is an internal node then v(n) = min n succ(n)(l(n, n )+v(n )), where succ(n) are the children of n in S T. It is easy to see that the value v(n) of a node in the N/ search tree S T is the minimal cost solution to the subproblem rooted at n, subject to the current variable instantiation along the path from the root to n. Ifn is the root of S T, then v(n) is the minimal cost solution to the initial problem (Marinescu & echter 25). N/ Search Graphs The N/ search tree may contain nodes that root identical subtrees (in particular, subproblems with identical optimal solutions) which can be unified. When unifiable nodes are merged, the search tree becomes a graph and its size becomes smaller. Some unifiable nodes can be identified based on their contexts. INITION 4 (context) Given a OP instance and the corresponding N/ search tree S T relative to a pseudotree T,thecontext of any N node X i,x i S T, denoted by context(x i ), is defined as the set of ancestors of X i in T, including X i, that are connected to descendants of X i. It is easy to verify that any two nodes having the same context represent the same subproblem. Therefore, we can solve P Xi, the subproblem rooted at X i, once and use its 72

optimal solution whenever the same subproblem is encountered again. The context-minimal N/ search graph based on pseudo-tree T, denoted G T, is obtained by merging all the N nodes that have the same context. It can be shown (echter & Mateescu 26) that the size of the largest context is bounded by the induced width w of the primal graph, extended with the pseudo-tree extra arcs, over the ordering given by the depth-first traversal of T (i.e. induced width of the pseudo-tree). Therefore, THM (complexity) The complexity of any search algorithm traversing a context-minimal N/ search graph is time and space O(exp(w )), where w is the induced width of the underlying pseudo-tree. xample 2 onsider the context-minimal N/ search graph in igure (d) of the pseudo-tree from igure (b). Its size is far smaller than that of the N/ tree from igure (c) (6 nodes vs. 36 nodes). The contexts of the nodes can be read from the pseudo-tree, as follows: context() = {}, context() = {,}, context() = {,}, context() = {}, context() = {,} and context( ) = {}. est-irst N/ Search In recent years, depth-first N/ ranch-and-ound algorithms were shown to be very effective, especially when using extensive caching (Marinescu & echter 25; 26c). Since best-first search is known to be superior among memory intensive search algorithms (echter & Pearl 985), the comparison with the best-first approach that exploits similar amounts of memory is warranted. In this section we introduce two new classes of best-first N/ search algorithms: one that explores a context-minimal N/ search graph and is restricted to a static variable ordering, and one that uses dynamic variable orderings but traverses an N/ search tree, rather than a graph. est-irst N/ Graph Search Our best-first N/ graph search algorithm, denoted by O, that traverses the context-minimal N/ search graph is described in lgorithm. It specializes Nillson s O algorithm (Nillson 98) to N/ spaces in graphical models. The algorithm interleaves forward expansion of the best partial solution tree with a cost revision step that updates estimated node values. irst, a top-down, graphgrowing operation (step 2.a) finds the best partial solution tree by tracing down through the marked arcs of the explicit N/ search graph G T. These previously computed marks indicate the current best partial solution tree from each node in G T. One of the nonterminal leaf nodes n of this best partial solution tree is then expanded, and a static heuristic estimate h(n i ), underestimating v(n i ),isassigned to its successors (step 2.b). The successors of an N node n = X j,x j are X j s children in the pseudotree, while the successors of an node n = X j correspond to X j s domain values. Notice that when expanding an node, the algorithm does not generates N children that are already present in the explicit search graph G T. ll these lgorithm : O ata: OP P =(X,, ), pseudo-tree T, root s. Result: Minimal cost solution to P.. reate explicit graph G T, consisting solely of the start node s. Set v(s) =h(s). 2. until s is labeled SOLV, do: (a) ompute a partial solution tree by tracing down the marked arcs in G T from s and select any nonterminal tip node n. (b) xpand node n and add any new successor node n i to G T. or each new node n i set v(n i)=h(n i). Label SOLV any of these successors that are terminal nodes. (c) reate a set S containing node n. (d) until S is empty, do: i. Remove from S a node m such that m has no descendants in G T still in S. ii. Revise the value v(m) as follows:. if m is an N node then v(m) = P m j succ(m) v(mj). If all the successor nodes are labeled SOLV, then label node m SOLV.. if m is an node then v(m) =min mj succ(m)(l(m, m j)+v(m j)) and mark the arc through which this minimum is achieved. If the marked successor is labeled SOLV, then label m SOLV. iii. If m has been marked SOLV or if the revised value v(m) is different than the previous one, then add to S all those parents of m such that m is one of their successors through a marked arc. 3. return v(s). identical N nodes in G T are easily recognized based on their contexts. The second operation in O is a bottom-up, cost revision, arc marking, SOLV-labeling procedure (step 2.c). Starting with the node just expanded n, the procedure revises its value v(n) (using the newly computed values of its successors) and marks the outgoing arcs on the estimated best path to terminal nodes. This revised value is then propagated upwards in the graph. The revised cost v(n) is an updated estimate of the cost of an optimal solution to the subproblem rooted at n. If we assume the monotone restriction on h, the algorithm considers only those ancestors that root best partial solution subtrees containing descendants with revised values. The optimal cost solution to the initial problem is obtained when the root node s is solved. ynamic Variable Orderings It is well known that variable selection may influence dramatically search performance. Recent work by (Marinescu & echter 26b) showed how several dynamic variable orderings affect depth-first ranch-and-ound search over N/ trees. One version, called N/ ranch-and- ound with Partial Variable Ordering (O+PVO) that orders dynamically the variables forming chains in the pseudotree was shown to outperform significantly static N/ search as well as state-of-the-art ranch-and-ound solvers for general OPs. Next, we extend the idea of partial variable ordering to best-first search on N/ trees. Note that O is restricted to a static variable ordering that corresponds to the pseudo-tree arrangement. The mechanism of identifying unifiable N nodes based solely on 73

spot5 n w toolbar3 O OM(i) OM(i) c h VO i=4 i=6 i=8 i= i=2 i=4 i=6 i=8 i= i=2 29 83 4 time 4.56.8 5.53 4.8.56 3.64 2.67 6.42 2.23.47 3.59 2.77 476 42 nodes 28,846 8,698 48,995 29,72 2,267,65 36,396 2,8 757 323 96 i=2 i=4 i=6 i=8 i= i=2 i=4 i=6 i=8 i= 54 68 time.3.6 546.89 8.42.23.6.69.69.4..6.69 283 33 nodes 2,939 688 5,94,5 98,72 2,477 59 2 3,96 2,74 63 32 68 i=6 i=8 i= i=2 i=4 i=6 i=8 i= i=2 i=4 44 9 time 5. 2.9 5.88 2.55.55.6 3.98.2.2.62.22 4. 7 42 nodes 6,25,35 88,79 529,2 23,565,74 598 232 6,399 5,4,33 576 84 i=6 i=8 i= i=2 i=4 i=6 i=8 i= i=2 i=4 48b 2 24 time - - - 757. 55.94 75.8 47.3 52.53 44.99 25.2 6.97 38.53 847 59 nodes 54,826,929 3,4,294 48,69 6,986 75,366 45,9 98,66 39,238 4,768 i=2 i=4 i=6 i=8 i= i=2 i=4 i=6 i=8 i= 53 44 9 time - 5. - - 89.39 29.72.42.25 5.28.56.59.42 639 39 nodes 44,495,545 2,442,998 4,5,474 256 22,967 6,4 9,929 9,86 44 i=6 i=8 i= i=2 i=4 i=6 i=8 i= i=2 i=4 55b 24 6 time - - - - - 8.48 367.93 42.73 29.25 3.2 54.9 373.72 72 98 nodes 8,95,473 6,2 44,723,223 8,256 3,692 5,758 Table : PU time in seconds and nodes visited to prove optimality for SPOT5 benchmarks. Time limit 3 hours. their contexts is hard to extend when variables are instantiated in a different order than that dictated by the pseudo-tree. est-first N/ search with Partial Variable Ordering (O+PVO) traverses an N/ search tree in a bestfirst manner and combines the static graph-based problem decomposition given by a pseudo-tree with a dynamic semantic variable selection heuristic. We illustrate the idea with an example. onsider the pseudo-tree from igure (b) inducing the following variable group ordering: {,}, {,}, {,}; which dictates that variables {,} should be considered before {,} and {,}. Variables in each group (or chain) can be dynamically ordered based on a second, independent heuristic. Notice that after variables {,} are instantiated, the problem decomposes into two independent components (represented by variables {,} and {,}, resp.) that can be solved separately. xperiments We evaluate the performance of the two classes of best-first N/ search algorithms on three common optimization problems: solving Weighted SPs (de Givry et al. 25), finding the Most Probable xplanation (MP) in belief networks (Pearl 988) and solving / Integer Linear Programs (Nemhauser & Wolsey 988). ll experiments were run on a 2.4GHz Pentium IV with 2G of RM. We report the average PU time (in seconds) and number of nodes visited, required for proving optimality of the solution. We also record the number of variables (n), the number of constraints (c), the depth of the pseudo-trees (h) and the induced width of the graphs (w ) obtained for the test instances. The pseudo-trees were generated using the min-fill heuristic, as described in (Marinescu & echter 25). The best performance points are highlighted. Weighted SPs or this domain we experimented with real-world scheduling and circuit diagnosis benchmarks. We consider the bestfirst N/ search algorithm guided by pre-compiled mini-bucket heuristics (Marinescu & echter 25) and denoted by OM(i). We compare it against the depth-first N/ ranch-and-ound algorithm with static minibucket heuristics and full caching introduced by (Marinescu & echter 26c) and denoted by OM(i). The parameter i represents the mini-bucket i-bound and controls the accuracy of the heuristic. oth algorithms traverse the context-minimal N/ search graph restricted to a static variable ordering determined by the pseudo-tree. We also report results obtained with the ranch-and- ound maintaining xistential irectional rc-onsistency () developed in (de Givry et al. 25) and denoted by toolbar3, and the N/ ranch-and-ound with and full dynamic variable ordering (O+VO) from (Marinescu & echter 26b). These algorithms instantiate variables dynamically, using the min-dom/deg heuristic which selects the variable with the smallest ratio of the domain size divided by the future degree. arth Observing Satellites The problem of scheduling an arth observing satellite is to select from a set of candidate photographs, the best subset such that a set of imperative constraints are satisfied and the total importance of the selected photographs is maximized. We experimented with problem instances from the SPOT5 benchmark (ensana, Lemaitre, & Verfaillie 999) which can be formulated as WSPs with binary and ternary constraints and domain sizes of 2 and 4 (instances 48b and 55b contain only binary constraints). Table shows the results for experiments with 6 scheduling problems. The columns are indexed by the mini-bucket i-bound. When comparing the best-first against the depthfirst N/ search algorithms with static mini-bucket heuristics we observe that, for relatively small i-bounds, OM(i) improves significantly (up to several orders of magnitude) in terms of both PU time and number of nodes visited. or example, on 55b, one of the hardest instances, OM(8) proves optimality in less than 3 seconds, whereas OM(8) exceeds the 3 hour time limit. This observation verifies the theory because best-first search is likely to expand the smallest number of nodes at the search frontier, especially when having relatively weak heuristic estimates. s the mini-bucket i-bound increases and the heuristics become strong enough to cut the search space substantially the difference between ranch-and-ound and best-first search decreases, because ranch-and-ound finds almost optimal solutions fast, and therefore will not explore 74

iscas89 n w toolbar3 O+ OM(i) OM(i) c h VO i=8 i= i=2 i=4 i=6 i=8 i= i=2 i=4 i=6 c432 432 27 time - - 422.8 4.9.89.89.64 39.33.52.3.38.67 432 45 nodes 2,945,23 337,574 6,254 6, 94 96,892 2,54,7 847 445 c88 88 27 time - -.66 9.66 3.6 59.35 4.78.36.9.8.9.44 883 67 nodes 56,56 446,893 69,38 36,24 78,268 4,454 2,792 2,23 2,862,589 s935 44 66 time - - 285.7 43.53-22.28 4.8 6.6.22.9.22 2.42 464 nodes 6,623,68 763,933 28,372 5, 25,493 4,87 3,39 2,26 883 s96 562 54 time - - 3347.38 53.3 2299.72 734.66 49.8 22.67 2.89 3.2 7.27 3.56 564 97 nodes 3,554,37 2,425,52,488,366 3,524,78 793,47 72,75 9,336 4,2 2,989 2,9 s238 54 59 time - - 897.37 682.99 28.5 248.27 2.64 34.9 29.4 2.3 6.64 4.63 543 94 nodes 8,386,634 7,43,223,35,933,22,658 59,635 37,96,25 53,95 26, 7,42 s494 66 48 time - - 364.8 5.64 27.64 6.92 9.2.44.59.95.5 3.8 66 69 nodes 953,945 7,279 8,895 23,3 2,4 5,694,472 2,3,476 985 Table 2: PU time in seconds and nodes visited to prove optimality for ISS 89 benchmarks. Time limit hour. ped n w Superlink.5 OM(i) OM(i) c h i=6 i=8 i= i=2 i=4 i=6 i=8 i= i=2 i=4 299 5 time 3.3 4.9 2.7.39.65.36.3 2.7.26.87.54 34 6 nodes 69,75 33,98 4,576 6,36 4,494 7,34 3,784,77 4,6 3,9 i= i=2 i=4 i=6 i=8 i= i=2 i=4 i=6 i=8 23 3 23 time 6,89 53.7 49.33 8.77 2.73 3.4 35.49 29.29.59 3.59 3.48 4 37 nodes 486,99 437,688 85,72 4,9 7,89 85,76 5,24 52,7,44 5,79 i=2 i=4 i=6 i=8 i=2 i=2 i=4 i=6 i=8 i=2 3 6 25 time 28,74 44.26 597.88 23.9 5.96 43.83 86.77 58.38 85.53 49.38 33.3 298 5 nodes,694,534 5,58,555,458,74,79,236 46,896 692,87 253,465 35,497 79,79 37,75 i=8 i= i=2 i=8 i= i=2 38 582 7 time 62.8 554.65 246. 272.69 34.4 26.94 3.7 727 59 nodes 8,986,648,868,672,42,976 348,723 583,4 242,429 i=6 i=8 i= i=2 i=6 i=8 i= i=2 5 479 8 time 76.6 44.29 2493.75 66.66 52. 78.53 36.3 2.75 38.52 57 58 nodes 28,2,843 5,729,294 43,234,32 24,886 4,289 25,57 5,766 Table 3: PU time in seconds and nodes visited to prove optimality for genetic linkage analysis. solutions whose cost is above the optimal one, like best-first search. Notice that toolbar3 and O+VO are able to solve relatively efficiently only the first 3 test instances. ISS 89 ircuits ISS 89 circuits are a common benchmark used in formal verification and diagnosis. or our purpose, we converted each of these circuits into a nonbinary WSP instance by removing flip-flops and buffers in a standard way, creating hard constraints for gates and uniform unary cost functions for inputs. The penalty costs were distributed uniformly randomly between and. Table 2 reports the results for experiments with 6 circuits. We observe again that OM(i) is the best performing algorithm. or instance, on the s96 circuit, OM() is 74 times faster and explores a search space 26 times smaller than OM(). In summary, the best-first N/ search algorithms with mini-bucket heuristics cause significant time savings especially for relatively small i-bounds which generate relatively weak heuristic estimates. The performance of the toolbar3 and O+VO algorithms that are designed specifically for WSP was very poor on this dataset and they were not able to solve the problems within the hour time limit. elief Networks The maximum likelihood haplotype problem in genetic linkage analysis is the task of finding a joint haplotype configuration for all members of the pedigree which maximizes the probability of data. It is equivalent to finding the most probable explanation of a belief network that represents the pedigree data (ishelson & Geiger 22). vailable at http://www.fm.vslib.cz/ kes/asic/iscas/ Table 3 displays the results obtained for 5 hard linkage analysis networks 2. In addition to the depth-first and bestfirst N/ search algorithms with pre-compiled minibucket heuristics (i.e. OM(i), OM(i)), we also ran Superlink.5, which is one of the most efficient solvers for genetic linkage analysis. We notice again the superiority of OM(i) over OM(i), especially for relatively small i-bounds. On some instances (e.g. ped, ped3), the best-first search algorithm OM(i) is several orders of magnitude faster than Superlink. / Integer Linear Programs or this domain we experimented with random combinatorial auctions which were drawn from the regions-upv distribution of the TS 2. test suite 3 and simulate the auction of radio spectrum for different geographical areas. In combinatorial auctions, an auctioneer has a set of goods to sell and the buyers submit a set of bids over subsets of goods. The winner determination problem is to label the bids as winning or loosing so as to maximize the sum of the accepted bid prices under the constraint that each good is allocated to at most one bid. The problem can be formulated as / ILP, as described in (Leyton-rown, Pearson, & Shoham 2). We consider two classes of best-first N/ search algorithms, as follows: O which explores the contextminimal N/ search graph, and O+PVO which explores a dynamic N/ search tree using partial variable orderings, respectively. We compare the best-first search 2 http://bioinfo.cs.technion.ac.il/superlink/ 3 http://cats.stanford.edu/ 75

variable ordering heuristics. The efficiency of the bestfirst N/ search approach compared to the depth-first N/ ranch-and-ound search is demonstrated empirically on various benchmarks including random as well as real-world problem instances. Our approach leaves room for further improvements. The space required by O can be enormous, due to the fact that all the nodes generated by the algorithm have to be saved prior to termination. Therefore, O can be extended to incorporate a memory bounding scheme similar to the one suggested in (hakrabati et al. 989). igure 2: Results for combinatorial auctions. algorithms against two depth-first N/ ranch-and- ound algorithms for / ILPs which were recently proposed by (Marinescu & echter 26a): N/ ranchand-ound with full context-based caching (O), and N/ ranch-and-ound with partial variable ordering (O+PVO), respectively. The guiding heuristic function is computed by solving the linear relaxation of the current subproblem with the SIMPLX method (we used the implementation from the lp solve5.5 library 4 ). or reference, we include results obtained with the classic ranch-and-ound algorithm () available from the lp solve library. The algorithms, O+PVO and O+PVO used a dynamic variable selection heuristic based on reduced costs (or dual values) which selects the next fractional variable with the smallest reduced cost. Since combinatorial auctions can be formulated as binary WSP instances (echter 23), we also ran toolbar3. igure 2 displays the results for experiments with combinatorial auctions with goods and increasing number of bids. ach data point represents an average over random samples. We observe that the best-first N/ search algorithms (O, O+PVO) outperform their N/ ranch-and-ound counterparts (O, O+PVO), especially when the number of bids increases. When looking at the two best-first algorithms we notice the superiority of O+PVO over O. This demonstrates the power of the dynamic variable selection heuristic which is able in this case to cut the search tree dramatically. In summary, O+PVO is the best performing algorithm and, on some of the hardest instances, it outperforms its competitors with up to one order of magnitude. onclusion In this paper we introduced a best-first N/ search algorithm which extends the classic O algorithm and traverses a context-minimal N/ search graph for solving optimization tasks in graphical models. We also proposed a best-first search algorithm that explores an N/ search tree, rather than a graph, and incorporates dynamic 4 http://groups.yahoo.com/group/lp solve/ cknowledgments This work was supported by the NS grant IIS-42854. References ensana,.; Lemaitre, M.; and Verfaillie, G. 999. arth observation satellite management. onstraints 4(3):293 299. hakrabati, P.; Ghose, S.; charya,.; and de Sarkar, S. 989. Heuristic search in restricted memory. In rtificial Intelligence 3(4):97 22. de Givry, S.; Heras,.; Larrosa, J.; and Zytnicki, M. 25. xistential arc consistency: getting closer to full arc consistency in weighted csps. In IJI. echter, R., and Mateescu, R. 26. nd/or search spaces for graphical models. rtificial Intelligence. echter, R., and Pearl, J. 985. Generalized best-first search strategies and the optimality of a*. In Journal of M 32(3):55 536. echter, R. 23. onstraint Processing. MIT Press. ishelson, M., and Geiger,. 22. xact genetic linkage computations for general pedigrees. ioinformatics. reuder,., and Quinn, M. 985. Taking advantage of stable sets of variables in constraint satisfaction problems. In IJI 76 78. Leyton-rown, K.; Pearson, M.; and Shoham, Y. 2. Towards a universal test suite for combinatorial auction algorithms. In M lectronic ommerce 66 76. Marinescu, R., and echter, R. 25. nd/or branch-andbound for graphical models. In IJI 224 229. Marinescu, R., and echter, R. 26a. nd/or branchand-bound search for pure / integer linear programming problems. In PI 52 66. Marinescu, R., and echter, R. 26b. ynamic orderings for and/or branch-and-bound search in graphical models. In I 38 42. Marinescu, R., and echter, R. 26c. Memory intensive branch-and-bound search for graphical models. In I. Nemhauser, G., and Wolsey, L. 988. Integer and combinatorial optimization. Wiley. Nillson, K. 98. Principles of rtificial Intelligence. Tioga. Pearl, J. 988. Probabilistic Reasoning in Intelligent Systems. Morgan-Kaufmann. 76