An Improved Lower Bound for Bayesian Network Structure Learning

Size: px
Start display at page:

Download "An Improved Lower Bound for Bayesian Network Structure Learning"

Transcription

1 An Improved Lower Bound for Bayesian Network Structure Learning Xiannian Fan and Changhe Yuan Graduate Center and Queens College City University of New York 365 Fifth Avenue, New York Abstract Several heuristic search algorithms such as A* and breadth-first branch and bound have been developed for learning Bayesian network structures that optimize a scoring function. These algorithms rely on a lower bound function called static k-cycle conflict heuristic in guiding the search to explore the most promising search spaces. The heuristic takes as input a partition of the random variables of a data set; the importance of the partition opens up opportunities for further research. This work introduces a new partition method based on information extracted from the potentially optimal parent sets (POPS) of the variables. Empirical results show that the new partition can significantly improve the efficiency and scalability of heuristic search-based structure learning algorithms. Introduction This paper considers the problem of learning an optimal Bayesian network structure for given data and scoring function. Several heuristic search algorithms have been developed for solving this problem by formulating the problem as a shortest path problem (Yuan and Malone 2013), e.g., A* (Yuan, Malone, and Wu 2011), breadth-first branch and bound (BFBnB) (Malone et al. 2011), and anytime Window A* (Malone and Yuan 2013). All of these algorithms need a lower bound function in estimating the quality of a search path so that they can prune paths that are guaranteed to lead to suboptimal solutions and focus on exploring the most promising search spaces. A lower bound function called static k-cycle conflict heuristic (Yuan and Malone 2012) has demonstrated excellent performance when used by the search algorithms to solve many benchmark data sets. It takes as input a partition of the random variables of a data set, and computes the heuristic value of a path by enforcing the acyclicity constraint between the variables within each group and relaxing the acyclicity between different groups. The partition has a significant impact on the tightness of the heuristic. According to Yuan and Malone (2012), a good grouping method should reduce the number of directed cycles between different groups and enforce acyclicity as much as possible. Copyright c 2015, Association for the Advancement of Artificial Intelligence ( All rights reserved. Existing approaches (Fan, Yuan, and Malone 2014) achieve this objective only indirectly; they use independence tests to create an undirected graph, and then use graph partition algorithms to create a partition by minimizing the correlation between the variables in different groups. This work introduces a more direct partition approach. Since the goal is to reduce the cycles between different groups of the partition, we should put variables that have more cycles between them in the same group. We will show that the cycles introduced in the static k-cycle conflict heuristic are originated from the potentially optimal parent sets (POPS) of the variables. We therefore identify groups of variables with more cycles based on information extracted from all or some of the POPS. Empirical results show that the new partition method can significantly improve the efficiency and scalability of heuristic search-based structure learning algorithms. Background This section provides an overview of Bayesian network structure learning, the shortest-path formulation, and the k- cycle conflict heuristic. Bayesian Network Structure Learning A Bayesian network (BN) consists of a directed acyclic graph (DAG) in which the vertices correspond to a set of random variables V = {X 1,..., X n } and a set of conditional probability distributions P (X i P A i ), where all parents of X i are referred to as P A i. The joint probability over all variables factorizes as the product of the conditional probability distributions. We consider the problem of learning a network structure from a discrete dataset D = {D 1,..., D N }, where D i is an instantiation of all the variables in V. A scoring function s measures the goodness of fit of a network structure to D (Heckerman 1998). The goal is to find a structure which optimizes the score. We only require that the scoring function is decomposable (Heckerman 1998); that is, the score of a network s(n) = i s i(p A i ). The s i (P A i ) values are often called local scores. Many commonly used scoring functions, such as MDL (Lam and Bacchus 1994) and BDe (Buntine 1991; Heckerman, Geiger, and Chickering 1995), are decomposable.

2 their appearance, so the search graph is also called an order graph. Each variable selects optimal parents from the variables that precede it, so combining the optimal parent sets yields an optimal structure for that ordering. The shortest path gives the global optimal structure. Figure 1: An order graph for four variables. While the local scores are defined for all 2 n 1 possible parent sets for each variable, this number is greatly reduced by pruning parent sets that are provably never optimal (de Campos and Ji 2011). The scores remaining after pruning are called as potentially optimal parent sets (POPS). The POPS are given as input to the learning problem. We denote the set of POPS for X i as P i.the Bayesian network structure learning problem (BNSL) can be defined as follows. The BNSL Problem INPUT: A set V = {X 1,..., X n } of variables and a set of POPS P i for each X i. TASK: Find a DAG N such that n N arg min s i (P A i ), N i=1 where P A i is the parent set of X i in N and P A i P i. Shortest Path Formulation The above structure learning problem was formulated as a shortest-path problem in (Yuan, Malone, and Wu 2011; Yuan and Malone 2013). Figure 1 shows the implicit search graph for four variables. The top-most node with the empty variable set is the start node, and the bottom-most node with the complete set is the goal node. An arc from U to U {X i } in the graph represents generating a successor node by adding a new variable X i as a leaf to an existing subnetwork of variables U; the cost of the arc is equal to the score of the optimal parent set for X i out of U, which is computed by considering all subsets of the variables in P A U, P A P i, i.e., cost(u U {X i }) = BestScore(X i, U) (1) = min s i (P A i ). P A i U,P A i P i (2) In this search graph, each path from the start to the goal corresponds to an ordering of the variables in the order of Static K-Cycle Conflict Heuristic The following simple heuristic function was introduced in (Yuan, Malone, and Wu 2011) for computing lower bounds for A* search. Definition 1. Let U be a node in the order graph, its heuristic value is h(u) = BestScore(X, V\{X}). (3) X V\U The above heuristic function allows each remaining variable to choose optimal parents from all of the other variables. Therefore it completely relaxes the acyclicity constraint of Bayesian networks in the estimation. The heuristic was proven to be admissible, meaning it never overestimates the future distance (Yuan, Malone, and Wu 2011). Admissible heuristics guarantee the optimality. However, because of the complete relaxation of the acyclicity constraint, the simple heuristic may generate loose lower bounds. In (Yuan and Malone 2012), an improved heuristic function called static k-cycle conflict heuristic was proposed by reducing the amount of relaxation. The idea is to partition the variables V into multiple groups V i (typically two), i.e. V = i V i, and enforce acyclicity within each group while still allowing cycles between the groups. For the partition, we need to compute a pattern database for each group V i. For this particular problem, a pattern database for group V i is basically a full order graph containing all subsets of V i. We will use a backward breadth first search to create the graph layer by layer starting from the node V i. The cost for any reverse arc from U {X} to U in this order graph will be BestScore(X, ( j i V j) U). We then enumerate all subsets of each group V i as the patterns, which can be done by a reverse breadth-first search in an order graph containing only V i (Yuan and Malone 2012). The patterns from different groups are guaranteed to be mutually exclusive, so we simply pick out the maximum-size pattern for each group that is a subset of V \ U and add them together as the lower bound. Figure 2 shows two pattern databases for a 8-variable problem, as well as the procedure of calculating the heuristic value of pattern {X 2, X 3, X 5, X 7 }. The tightness of the static k-cycle conflict heuristic depends highly on the partition being used. The technique can avoid directed cycles for the patterns within the same group, but cannot avoid cycles between different groups. In the example shown in Figure 2, X 3 selects parents {X 1, X 5 } as parents (subset of {X 1, X 4 } V 2 ) and X 5 selects {X 3, X 6 } as parents, then one cycle appears between {X 3 } and {X 5 }. A New Partition Method for Improved Lower Bounds Previously, Fan et al. proposed two methods called Parent Grouping (PG) and Family Grouping (FG) (Fan, Yuan, and

3 Figure 2: Two pattern databases for a 8 variables problem. 8 variables are partitioned into two groups, V 1 = {X 1, X 2, X 3, X 4 } and V 2 = {X 5, X 6, X 7, X 8 }. (Left ) The pattern database for group V 1 ; bold arrows show the path corresponding to the score P 1 for pattern {X 2, X 3 }, where P 1 = BestScore(X 3, {X 1, X 2, X 4 } V 2 ) + BestScore(X 2, {X 1, X 4 } V 2 ). (Right ) The pattern database for group V 2 ; bold arrows show the path corresponding to the score P 2 for pattern {X 5, X 7 }, where P 2 = BestScore(X 7, {X 5, X 6, X 8 } V 1 ) + BestScore(X 5, {X 6, X 8 } V 1 ). The heuristic value for pattern {X 2, X 3, X 5, X 7 } is P 1 + P 2. Malone 2014) to partition the variables based on correlation between the variables. There are three main steps in these two methods. First, they create an undirected skeleton graph. PG uses the best POPS of each variable to create the skeleton, while FG uses the Min-Max Parent Child (MMPC) algorithm (Tsamardinos, Brown, and Aliferis 2006) to get the skeleton. Second, they use the independence tests in MMPC to estimate the weights for the edges of the skeleton. Third, A graph partition algorithm called METIS (Karypis and Kumar 1998) is used to partition the skeleton into two balanced subgraphs by minimizing the total weights of the edges between the subgraphs. Intuitively, since the edge weights measure the correlation between the variables, putting variables that are less correlated into different groups should reduce cycles between the groups. However, these approaches fail to consider the complex relation between a variable and another set of variables (parent set), and are heuristic in nature. This work introduces a new approach to obtaining a partition by directly reducing cycles between its groups. Ancestral Constraints We first use a concrete example to motivate the development of our method. It relies on the fact that all of the potential parent-child relations are contained in the potentially optimal parent sets (POPS) of the variables. Useful information can be extracted from the parent-child relations to find good partition strategies. Table 1 shows the POPS of eight variables in the example. Based on these POPS, we can extract the parent-child relations as follows. The set of all potential parents P i for X i can be collected by taking the union of all of its POPS. For example, X 1 can select parents from {X 2, X 5 }, and X 2 can only select X 1 as its parent. We then use all of the P i s to create a directed graph, which was called parent relation graph in (Fan, Yuan, and Malone 2014). var. POPS X 1 {X 2 } {X 5 } X 2 {X 1 } X 3 {X 1, X 5 } {X 1, X 2 } {X 2, X 4 } {X 1 } X 4 {X 3 } {X 6 } {X 7 } X 5 {X 1, X 3 } {X 3 } X 6 {X 2, X 7 } {X 7 } X 7 {X 8 } {X 6, X 4 } X 8 {X 6 } {X 7 } Table 1: The POPS for eight variables. The ith row shows P i. The POPS for each variable are sorted according to their scores. (We consider the minimization version of BNSL, so the best POPS are those with the lowest scores.) We can use the ancestral relations of the parent relation graph to partition the variables into sets. In particular if the variables in one set can be ancestors of the variables in another set, but not vice versa, there must be no cycles between the two sets. We can put the two sets of variables into different groups of the static k-cycle conflict heuristic in order to reduce cycles. We identify such ancestral relations by extracting strongly connected components (SCCs) from the parent relation graph. The SCCs of the parent relation graph form a DAG called component graph (Cormen et al. 2001); each component c i corresponds to a SCC scc i from the parent relation graph (which in turn corresponds to a set of variables in the Bayesian network). The component graph includes a directed edge from node c i to c j if the parent relation graph includes an edge from a variable X i scc i to X j scc j. The component graph provides some obvious ancestral relation constraints: if c j is a descendent of c i in the component graph, variables in scc j cannot be ancestors of variables in scc i. We call such relations ances-

4 Figure 3: The parent relation graphs created from (a) all of the POPS, (b)top-1 POPS, and (c) top-2 POPS. Each shaded box represents a strongly connected component (SCC) and a node in a corresponding component graph. tral constraints. We can obtain a tight static k-cycle conflict heuristic by putting the different SCCs into different groups of the heuristic as there would not be any cycles between the groups. As a matter of fact, the SCCs represent independent learning problems that can be solved separately (Fan, Malone, and Yuan 2014). There are two possible pitfalls of the above approach. One is that it assumes there exist multiple SCCs. The component graph may only has a single SCC, especially when we use all of the POPS to create the parent relation graph. Figure 3(a) shows the parent relation graph created from the POPS in Table 1. The component graph includes one SCC containing all variables. Therefore, all variables can be ancestors of each other; we have no way of easily dividing the variables into groups. Second, even though there exist multiple SCCs, the largest one may be too large for the approach to be feasible. The reason is that we are performing an exhaustive search when building a pattern database for a group of variables; the search is only feasible for fewer than 30 variables within one group in our testing environment. We will address these issues in the following sections. Ancestral Constraints from Top-K POPS To avoid creating a dense parent relation graph that has either only a single SCC or a SCC that is too large, we propose to use only the top K POPS of each variable to create the graph. Fewer POPS reduce the total number of candidate parents for each variable, resulting in fewer arcs in the parent relation graph. Top K POPS allow us to focus on the most important parent-child relations so that the SCCs of the relation graph still allow the heuristic to remove the most critical cycles. Figure 3(b) shows the parent relation graph created from only the top 1 POPS of each variable. The component graph now has 4 SCCs with no cycles between them. Of course we are ignoring the ancestral constraints outside of the top-1 POPS. For example, after recovering the parent relation X 5 X 1, there will be cycles between SCCs {X 1, X 2 } and {X 3, X 5 } as well as between X 1 and X 5. The parent relation graph created from the top-1 POPS is overly sparse; 4 SCCs are generated as a result. Fortunately, there is a happy medium to explore between using all POPS and using only top-1 POPS; we propose to use top-k POPS of each variable in creating the parent relation graph. Intuitively, including more POPS will introduce more cycles in the parent relation graph. More cycles in the parent relation graph would allow us to remove more cycles in the heuristic by enforcing acyclicity between the variables within each SCC, hence resulting in tighter heuristic values. Figure 3(c) shows the graph created from the top-2 POPS of all variables. Now there are only two SCCs, {X 1, X 2, X 3, X 5 } and {X 4, X 6, X 7, X 8 }. By using each SCC as a group in the static k-cycle conflict heuristic, we can remove the cycles within each SCC. For example, X 5 and X 1 are in the same SCC, the cycle between them can be removed by enforcing acyclicity between them in the heuristic. Moreover, there are no cycles between the SCCs, so putting them into different groups of the heuristic can reduce the possibility of cycles between the groups. Note that we cannot completely avoid cycles between the groups because we only consider top-2 POPS. Finally, since a partition with two groups typically works best for the static k-cycle conflict heuristic, there is no need to increase K to consider more POPS in this case. Components Grouping It was shown in (Yuan and Malone 2012; Fan, Yuan, and Malone 2014) that it is best to divide the variables into only 2 groups in static pattern databases. The reason is that we enforce acyclicity within each group, and allow cycles between groups; using only two groups allows acyclicity to be enforced within the largest possible groups, hence less probability for cycles between the groups. The parent relation graph created from top-k POPS may have more than 2 SCCs, regardless of what K is. We then have a grouping problem. Suppose there are l SCCs in the parent relation graph, denoted by scc 1, scc 2,..., scc l, the grouping problem is to divide these SCCs into two groups, V 1 and V 2. We let γ be the number of variables in the largest pattern database we can build in our testing environment. If the largest SCC is already larger than γ, we cannot succeed in building the heuristic, not mentioning solving the learning problem. In this case we fall back on the parent grouping (PG) method (Fan, Yuan, and Malone 2014), that is, we ignore the direction of the edges in the parent relation graph and use the graph as the skeleton, assign the graph weights, and partition the skeleton to give the two grouping. On the other hand, if the largest SCC is smaller than γ, we should combine it with some other SCCs to get a group as large as possible as long as it is still smaller than γ. Moreover, we want to combine nearby SCCs into a group because variables in nearby SCCs are more likely to have cycles between them outside of the top-k POPS. We therefore propose the following Prim algorithm-like method for dividing the SCCs into two groups. First, we create the parent relation graph and its component graph based on top-k POPS. Initialize V 1 to contain the largest SCC in the graph. Second, we perform the following iteratively until V 1 hits the threshold γ. For all of the SCCs in V 1, we find their neighboring SCCs that are

5 not yet in V 1. Then we select the largest one out of these SCCs to add to V 1 subject to the constraint that V 1 γ. Third, all the remaining SCCs form the second group V 2. It is possible to generalize this method to more groups; we only consider two groups in this research. The K is given in the above procedure. Let P be the maximum number of POPS any variable can have. Since we do not know what the optimal K is, we try each K starting from 1 to P until we get only a single SCC or exhaust all POPS for each variable, or the largest SCC exceeds the γ threshold. We finally accept the grouping of the highest K that produces at least two SCCs subject to the γ threshold. We name the new method components Grouping (CG). Related Work POPS for all variables are the input to the score-based exact BNSL algorithms. To the best of our knowledge, Fan et al. (2014) are the first to extract constraints from POPS to improve the performance of heuristic search on the exact BNSL problem. They decomposed the parent graph into a set of strongly connected components (SCCs). Each SCC corresponds to a smaller subproblem which can be solved independently of the others, thus the search space was reduced. They also used the top-k POPS to further reduce the search space to get a sub-optimal solution, but lost the optimality guarantee. We use the top-k POPS to get groupings for the lower bound. We do not lose global optimality no matter which K we use. The Parent Grouping (PG) method in (Fan, Yuan, and Malone 2014) creates a undirected skeleton from the top-1 POPS, assigns the skeleton weights, and partition the graph into two groupings. Our method works directly on the directed graph created from the top-k POPS. Finally, this research focuses on heuristic search-based algorithms for learning optimal Bayesian networks (Yuan and Malone 2013; Yuan, Malone, and Wu 2011; Malone et al. 2011; Malone and Yuan 2013). Other exact algorithms have been developed for solving the same problem based on dynamic programming (Koivisto and Sood 2004; Ott, Imoto, and Miyano 2004; Silander and Myllymaki 2006; Singh and Moore 2005; Malone, Yuan, and Hansen 2011) and integer linear programming (Cussens 2011; Jaakkola et al. 2010). Please refer to (Yuan and Malone 2013) for a comprehensive study comparing the different approaches. Empirical Results We empirically evaluated our new CG lower bound using the A* and BFBnB algorithms on benchmark datasets from UCI machine learning repository and Bayesian Network Repository ( il/repository/). The experiments were performed on an IBM System x3850 X5 with 16 core 2.67GHz Intel Xeon Processors and 512G RAM. Parameter K CG initializes the first group V 1 with the largest SCC; the size of the largest SCC is monotonically increasing in terms of K. A* was used as A* guarantees to expand the minimal number of nodes. We did not limit the size of the largest SCC Running Time k Expanded Nodes k Figure 4: The running time and number of expanded nodes needed by A* to solve Soybeans with different K. as long as at least two SCCs are obtained. In this experiment, Dataset soybeans (36 variables) was tested bacause it was found to be challenging from a previous study (Fan, Malone, and Yuan 2014). FG failed to solve this dataset within the time limit of 2 hours; PG solved it in more than 518 seconds with more than 9 million nodes expanded. For CG, we varied K and measured the running time and the number of expanded nodes. Figure 4 showed the results. Even for K = 1, the running time of CG was around 76 seconds, 7 times faster than PG, and the number of expanded nodes was around 1 million, 9 times fewer than that of PG; when K = 2, the running time was less than 1 second, and the number of expanded nodes was just around 1 thousand, which were negligible compared to PG; from K = 3 on, the running time and number of expanded nodes were further reduced. The size of largest SCC was 20 when K = 1, 26 when K = 2, and 27 when K = 3, 4, 5. K = 6 and above only generated a single SCC. Generally, we notice that the larger the K, the better performance of the CG lower bound, which is quite understandable because we can avoid cycles within a wider range of POPS. That is exactly why we set K to be the largest value such that at least two SCCs are obtained and the size of the largest SCC is smaller than the threshold γ. Results on Benchmark Datasets We compared the components grouping (CG) against the two existing strategies, the family grouping (FG) and parents grouping (PG) (Fan, Yuan, and Malone 2014), using A* and BFBnB on a set of benchmark datasets. We set the threshold γ to be 22 for small datasets with fewer than 35 variables as the heuristic can be built within 5 seconds, and to be 25 for large datasets with 35 or more variables as the heuristic can be built within 60 seconds. The results are shown in Table 2. A* is guaranteed to expand the least number of search

6 Dataset Results Name n N FG (A*) PG (A*) CG (A*) FG (BFBnB) PG (BFBnB) CG (BFBnB) Autos Time (s) Nodes Insurance* 27 1,000 Time (s) Nodes Horse Colic Time (s) Nodes Flag Time (s) Nodes Mildew* 35 1,000 Time (s) OT OT Nodes OT OT Soybean Time (s) OT OT 1, Nodes OT OT Alarm* 37 1,000 Time (s) Nodes Sponge Time (s) OT OT OT OT Nodes OT OT 1.60 OT OT 2.99 Barley* 48 1,000 Time (s) OT OT Nodes OT OT Table 2: The number of expanded nodes (in millions) and running time (in seconds) of the A* and BFBnB algorithm on a set of benchmark datasets with different lower bounds. The lower bounds are family grouping (FG), parents grouping (PG) and components grouping (CG). n is the total number of variables, and N is the number of data points; * indicates the dataset was generated from a repository network using logic sampling; all other datasets are from UCI; OT means out of time (2 hours). nodes (Yuan, Malone, and Wu 2011). The table shows that the benefit of the new grouping method CG on A* is rather obvious. The improvement brought by CG in running time and the number of expanded nodes ranges from three times to over two orders of magnitude on most of datasets. FG is inferior than CG on all of the datasets except insurance. It failed to solve datasets Mildew, Soybean, Sponge and Barley within 2 hours. In comparison, three out of the four datasets turned out to be quite easy to solve by CG; the other one, Soybean, was solved efficiently by CG within 77 seconds (Note γ is set to be 25 in this experiment, so the final results are from K = 1). Upon further investigation, we found that the skeleton of FG found by the MMPC (Tsamardinos, Brown, and Aliferis 2006) method was very sparse and not well connected. MMPC used a parameter p value to control the generation of the skeleton, and the default value p = 0.05 was shown to result in stable performance. We varied the value of p from 0.01 to 0.99 to test the effect for the sparsity of the network. The results showed that the sparsity issue cannot be solved by with a larger value of p. PG is more robust on large datasets than FG. However, CG consistently outperformed PG; the speedup ranged from several times faster (e.g., Autos) to several orders of magnitude faster (e.g., Flag). Moreover, CG solved the Sponge efficiently (within 24 seconds) while PG and FG failed within the time limit (7200 seconds). The only exception is the easy dataset alarm for which all of the grouping methods can solve it efficiently in around 4 seconds. We also evaluated the different grouping methods with BFBnB (Malone et al. 2011) because it can scale to larger datasets by using external memory. Anytime window A* (AWA*) was used to provided an upper bound for pruning since a previous study (Malone and Yuan 2013) has shown that AWA* is effective at finding high quality solutions quickly, so we provided the upper bound by running AWA* for 5 seconds on small dataset (fewer than 35 variables) and 10 seconds for larger datasets. The results of BF- BnB demonstrate rather similar patterns to A*. Therefore, the new grouping method seems to be generally applicable to other search methods. Conclusion This paper investigates a new method for improving the lower bound of heuristic search for Bayesian networks structure learning (Yuan and Malone 2012). The main idea is to create a parent relation graph from the top-k potentially optimal parent sets (POPS) of each variable, and extract better groupings for the k-cycle conflict heuristic. Empirically, we showed that the new lower bound can significantly improve the efficiency and scalability of the heuristic search-based learning methods. As we said earlier, the need for good groupings in the k-cycle conflict heuristic opens up opportunities for future research. Even though our current method has excellent performance, there may be better methods. For example, one possibility is to investigate how to integrate different domain knowledge and the information contained in the potentially optimal parents set. Acknowledgments This work was supported by NSF grants IIS , IIS , and a PSC-CUNY enhancement award

7 References Buntine, W Theory refinement on Bayesian networks. In Proceedings of the seventh conference (1991) on Uncertainty in artificial intelligence, San Francisco, CA, USA: Morgan Kaufmann Publishers Inc. Cormen, T. H.; Stein, C.; Rivest, R. L.; and Leiserson, C. E Introduction to Algorithms. McGraw-Hill Higher Education. Cussens, J Bayesian network learning with cutting planes. In Proceedings of the Proceedings of the Twenty- Seventh Conference Annual Conference on Uncertainty in Artificial Intelligence (UAI-11), Corvallis, Oregon: AUAI Press. de Campos, C. P., and Ji, Q Efficient learning of Bayesian networks using constraints. Journal of Machine Learning Research 12: Fan, X.; Malone, B.; and Yuan, C Finding optimal bayesian network structures with constraints learned from data. In Proceedings of the 30th Annual Conference on Uncertainty in Artificial Intelligence (UAI-14). Fan, X.; Yuan, C.; and Malone, B Tightening bounds for Bayesian network structure learning. In Proceedings of the 28th AAAI Conference on Artificial Intelligence (AAAI- 14). Heckerman, D.; Geiger, D.; and Chickering, D. M Learning Bayesian networks: The combination of knowledge and statistical data. 20: Heckerman, D A tutorial on learning with Bayesian networks. In Holmes, D., and Jain, L., eds., Innovations in Bayesian Networks, volume 156 of Studies in Computational Intelligence. Springer Berlin / Heidelberg Jaakkola, T.; Sontag, D.; Globerson, A.; and Meila, M Learning Bayesian network structure using LP relaxations. In Proceedings of the 13th International Conference on Artificial Intelligence and Statistics (AISTATS). Karypis, G., and Kumar, V A fast and high quality multilevel scheme for partitioning irregular graphs. SIAM Journal on scientific Computing 20(1): Koivisto, M., and Sood, K Exact Bayesian structure discovery in Bayesian networks. Journal of Machine Learning Research Lam, W., and Bacchus, F Learning Bayesian belief networks: An approach based on the MDL principle. Computational Intelligence 10: Malone, B., and Yuan, C Evaluating anytime algorithms for learning optimal bayesian networks. In In Proceedings of the 29th Conference on Uncertainty in Artificial Intelligence (UAI-13). Malone, B.; Yuan, C.; Hansen, E.; and Bridges, S Improving the scalability of optimal Bayesian network learning with external-memory frontier breadth-first branch and bound search. In Proceedings of the Twenty-Seventh Annual Conference on Uncertainty in Artificial Intelligence (UAI- 11), Corvallis, Oregon: AUAI Press. Malone, B.; Yuan, C.; and Hansen, E Memoryefficient dynamic programming for learning optimal Bayesian networks. In Proceedings of the 25th national conference on Artifical intelligence. Ott, S.; Imoto, S.; and Miyano, S Finding optimal models for small gene networks. In Pac. Symp. Biocomput, Silander, T., and Myllymaki, P A simple approach for finding the globally optimal Bayesian network structure. In Proceedings of the 22nd Annual Conference on Uncertainty in Artificial Intelligence (UAI-06). Arlington, Virginia: AUAI Press. Singh, A., and Moore, A Finding optimal Bayesian networks by dynamic programming. Technical report, Carnegie Mellon University. Tsamardinos, I.; Brown, L.; and Aliferis, C The maxmin hill-climbing Bayesian network structure learning algorithm. Machine Learning 65: /s Yuan, C., and Malone, B An improved admissible heuristic for finding optimal bayesian networks. In Proceedings of the Twenty-Eighth Conference on Uncertainty in Artificial Intelligence (UAI-12). AUAI Press. Yuan, C., and Malone, B Learning optimal Bayesian networks: A shortest path perspective. Journal of Artificial Intelligence Research (JAIR) 48: Yuan, C.; Malone, B.; and Wu, X Learning optimal Bayesian networks using A* search. In Proceedings of the 22nd International Joint Conference on Artificial Intelligence.

BAYESIAN NETWORKS STRUCTURE LEARNING

BAYESIAN NETWORKS STRUCTURE LEARNING BAYESIAN NETWORKS STRUCTURE LEARNING Xiannian Fan Uncertainty Reasoning Lab (URL) Department of Computer Science Queens College/City University of New York http://url.cs.qc.cuny.edu 1/52 Overview : Bayesian

More information

An Improved Admissible Heuristic for Learning Optimal Bayesian Networks

An Improved Admissible Heuristic for Learning Optimal Bayesian Networks An Improved Admissible Heuristic for Learning Optimal Bayesian Networks Changhe Yuan 1,3 and Brandon Malone 2,3 1 Queens College/City University of New York, 2 Helsinki Institute for Information Technology,

More information

Learning Optimal Bayesian Networks Using A* Search

Learning Optimal Bayesian Networks Using A* Search Proceedings of the Twenty-Second International Joint Conference on Artificial Intelligence Learning Optimal Bayesian Networks Using A* Search Changhe Yuan, Brandon Malone, and Xiaojian Wu Department of

More information

Scale Up Bayesian Networks Learning Dissertation Proposal

Scale Up Bayesian Networks Learning Dissertation Proposal Scale Up Bayesian Networks Learning Dissertation Proposal Xiannian Fan xnf1203@gmail.com March 2015 Abstract Bayesian networks are widely used graphical models which represent uncertain relations between

More information

Bayesian Networks Structure Learning Second Exam Literature Review

Bayesian Networks Structure Learning Second Exam Literature Review Bayesian Networks Structure Learning Second Exam Literature Review Xiannian Fan xnf1203@gmail.com December 2014 Abstract Bayesian networks are widely used graphical models which represent uncertainty relations

More information

Scale Up Bayesian Network Learning

Scale Up Bayesian Network Learning City University of New York (CUNY) CUNY Academic Works Dissertations, Theses, and Capstone Projects Graduate Center 6-2016 Scale Up Bayesian Network Learning Xiannian Fan Graduate Center, City University

More information

Learning Bayesian networks with ancestral constraints

Learning Bayesian networks with ancestral constraints Learning Bayesian networks with ancestral constraints Eunice Yuh-Jie Chen and Yujia Shen and Arthur Choi and Adnan Darwiche Computer Science Department University of California Los Angeles, CA 90095 {eyjchen,yujias,aychoi,darwiche}@cs.ucla.edu

More information

On Pruning with the MDL Score

On Pruning with the MDL Score ON PRUNING WITH THE MDL SCORE On Pruning with the MDL Score Eunice Yuh-Jie Chen Arthur Choi Adnan Darwiche Computer Science Department University of California, Los Angeles Los Angeles, CA 90095 EYJCHEN@CS.UCLA.EDU

More information

Learning Bounded Tree-width Bayesian Networks using Integer Linear Programming

Learning Bounded Tree-width Bayesian Networks using Integer Linear Programming Learning Bounded Tree-width Bayesian Networks using Integer Linear Programming Pekka Parviainen Hossein Shahrabi Farahani Jens Lagergren University of British Columbia Department of Pathology Vancouver,

More information

Bayesian network model selection using integer programming

Bayesian network model selection using integer programming Bayesian network model selection using integer programming James Cussens Leeds, 2013-10-04 James Cussens IP for BNs Leeds, 2013-10-04 1 / 23 Linear programming The Belgian diet problem Fat Sugar Salt Cost

More information

Survey of contemporary Bayesian Network Structure Learning methods

Survey of contemporary Bayesian Network Structure Learning methods Survey of contemporary Bayesian Network Structure Learning methods Ligon Liu September 2015 Ligon Liu (CUNY) Survey on Bayesian Network Structure Learning (slide 1) September 2015 1 / 38 Bayesian Network

More information

Improved Local Search in Bayesian Networks Structure Learning

Improved Local Search in Bayesian Networks Structure Learning Proceedings of Machine Learning Research vol 73:45-56, 2017 AMBN 2017 Improved Local Search in Bayesian Networks Structure Learning Mauro Scanagatta IDSIA, SUPSI, USI - Lugano, Switzerland Giorgio Corani

More information

A Parallel Algorithm for Exact Structure Learning of Bayesian Networks

A Parallel Algorithm for Exact Structure Learning of Bayesian Networks A Parallel Algorithm for Exact Structure Learning of Bayesian Networks Olga Nikolova, Jaroslaw Zola, and Srinivas Aluru Department of Computer Engineering Iowa State University Ames, IA 0010 {olia,zola,aluru}@iastate.edu

More information

Finding the k-best Equivalence Classes of Bayesian Network Structures for Model Averaging

Finding the k-best Equivalence Classes of Bayesian Network Structures for Model Averaging Proceedings of the Twenty-Eighth AAAI Conference on Artificial Intelligence Finding the k-best Equivalence Classes of Bayesian Network Structures for Model Averaging Yetian Chen and Jin Tian Department

More information

A* Lasso for Learning a Sparse Bayesian Network Structure for Continuous Variables

A* Lasso for Learning a Sparse Bayesian Network Structure for Continuous Variables A* Lasso for Learning a Sparse Bayesian Network Structure for Continuous Variables Jing Xiang Machine Learning Department Carnegie Mellon University Pittsburgh, PA 523 jingx@cs.cmu.edu Seyoung Kim Lane

More information

Enumerating Equivalence Classes of Bayesian Networks using EC Graphs

Enumerating Equivalence Classes of Bayesian Networks using EC Graphs Enumerating Equivalence Classes of Bayesian Networks using EC Graphs Eunice Yuh-Jie Chen and Arthur Choi and Adnan Darwiche Computer Science Department University of California, Los Angeles {eyjchen,aychoi,darwiche}@cs.ucla.edu

More information

Learning Bayesian Network Structure using LP Relaxations

Learning Bayesian Network Structure using LP Relaxations Tommi Jaakkola David Sontag Amir Globerson Marina Meila MIT CSAIL MIT CSAIL Hebrew University University of Washington Abstract We propose to solve the combinatorial problem of finding the highest scoring

More information

Learning Bayesian Networks with Non-Decomposable Scores

Learning Bayesian Networks with Non-Decomposable Scores Learning Bayesian Networks with Non-Decomposable Scores Eunice Yuh-Jie Chen, Arthur Choi, and Adnan Darwiche Computer Science Department University of California, Los Angeles, USA {eyjchen,aychoi,darwiche}@cs.ucla.edu

More information

Constraint Satisfaction Problems

Constraint Satisfaction Problems Constraint Satisfaction Problems CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2013 Soleymani Course material: Artificial Intelligence: A Modern Approach, 3 rd Edition,

More information

Enumerating Equivalence Classes of Bayesian Networks using EC Graphs

Enumerating Equivalence Classes of Bayesian Networks using EC Graphs Enumerating Equivalence Classes of Bayesian Networks using EC Graphs Eunice Yuh-Jie Chen and Arthur Choi and Adnan Darwiche Computer Science Department University of California, Los Angeles {eyjchen,aychoi,darwiche}@cs.ucla.edu

More information

Integer Programming for Bayesian Network Structure Learning

Integer Programming for Bayesian Network Structure Learning Quality Technology & Quantitative Management Vol. 11, No. 1, pp. 99-110, 2014 QTQM ICAQM 2014 Integer Programming for Bayesian Network Structure Learning James Cussens * Department of Computer Science

More information

Advanced Algorithms Class Notes for Monday, October 23, 2012 Min Ye, Mingfu Shao, and Bernard Moret

Advanced Algorithms Class Notes for Monday, October 23, 2012 Min Ye, Mingfu Shao, and Bernard Moret Advanced Algorithms Class Notes for Monday, October 23, 2012 Min Ye, Mingfu Shao, and Bernard Moret Greedy Algorithms (continued) The best known application where the greedy algorithm is optimal is surely

More information

Understanding the effects of search constraints on structure learning

Understanding the effects of search constraints on structure learning Understanding the effects of search constraints on structure learning Michael Hay, Andrew Fast, and David Jensen {mhay,afast,jensen}@cs.umass.edu University of Massachusetts Amherst Computer Science Department

More information

A Well-Behaved Algorithm for Simulating Dependence Structures of Bayesian Networks

A Well-Behaved Algorithm for Simulating Dependence Structures of Bayesian Networks A Well-Behaved Algorithm for Simulating Dependence Structures of Bayesian Networks Yang Xiang and Tristan Miller Department of Computer Science University of Regina Regina, Saskatchewan, Canada S4S 0A2

More information

Graph. Vertex. edge. Directed Graph. Undirected Graph

Graph. Vertex. edge. Directed Graph. Undirected Graph Module : Graphs Dr. Natarajan Meghanathan Professor of Computer Science Jackson State University Jackson, MS E-mail: natarajan.meghanathan@jsums.edu Graph Graph is a data structure that is a collection

More information

Problem Solving & Heuristic Search

Problem Solving & Heuristic Search 190.08 Artificial 2016-Spring Problem Solving & Heuristic Search Byoung-Tak Zhang School of Computer Science and Engineering Seoul National University 190.08 Artificial (2016-Spring) http://www.cs.duke.edu/courses/fall08/cps270/

More information

A Hybrid Recursive Multi-Way Number Partitioning Algorithm

A Hybrid Recursive Multi-Way Number Partitioning Algorithm Proceedings of the Twenty-Second International Joint Conference on Artificial Intelligence A Hybrid Recursive Multi-Way Number Partitioning Algorithm Richard E. Korf Computer Science Department University

More information

An Exact Approach to Learning Probabilistic Relational Model

An Exact Approach to Learning Probabilistic Relational Model JMLR: Workshop and Conference Proceedings vol 52, 171-182, 2016 PGM 2016 An Exact Approach to Learning Probabilistic Relational Model Nourhene Ettouzi LARODEC, ISG Sousse, Tunisia Philippe Leray LINA,

More information

Dependency detection with Bayesian Networks

Dependency detection with Bayesian Networks Dependency detection with Bayesian Networks M V Vikhreva Faculty of Computational Mathematics and Cybernetics, Lomonosov Moscow State University, Leninskie Gory, Moscow, 119991 Supervisor: A G Dyakonov

More information

Sub-Local Constraint-Based Learning of Bayesian Networks Using A Joint Dependence Criterion

Sub-Local Constraint-Based Learning of Bayesian Networks Using A Joint Dependence Criterion Journal of Machine Learning Research 14 (2013) 1563-1603 Submitted 11/10; Revised 9/12; Published 6/13 Sub-Local Constraint-Based Learning of Bayesian Networks Using A Joint Dependence Criterion Rami Mahdi

More information

3 SOLVING PROBLEMS BY SEARCHING

3 SOLVING PROBLEMS BY SEARCHING 48 3 SOLVING PROBLEMS BY SEARCHING A goal-based agent aims at solving problems by performing actions that lead to desirable states Let us first consider the uninformed situation in which the agent is not

More information

Heuristic (Informed) Search

Heuristic (Informed) Search Heuristic (Informed) Search (Where we try to choose smartly) R&N: Chap., Sect..1 3 1 Search Algorithm #2 SEARCH#2 1. INSERT(initial-node,Open-List) 2. Repeat: a. If empty(open-list) then return failure

More information

Learning decomposable models with a bounded clique size

Learning decomposable models with a bounded clique size Learning decomposable models with a bounded clique size Achievements 2014-2016 Aritz Pérez Basque Center for Applied Mathematics Bilbao, March, 2016 Outline 1 Motivation and background 2 The problem 3

More information

Local Search Methods for Learning Bayesian Networks Using a Modified Neighborhood in the Space of DAGs

Local Search Methods for Learning Bayesian Networks Using a Modified Neighborhood in the Space of DAGs Local Search Methods for Learning Bayesian Networks Using a Modified Neighborhood in the Space of DAGs L.M. de Campos 1, J.M. Fernández-Luna 2, and J.M. Puerta 3 1 Dpto. de Ciencias de la Computación e

More information

Graph Coloring via Constraint Programming-based Column Generation

Graph Coloring via Constraint Programming-based Column Generation Graph Coloring via Constraint Programming-based Column Generation Stefano Gualandi Federico Malucelli Dipartimento di Elettronica e Informatica, Politecnico di Milano Viale Ponzio 24/A, 20133, Milan, Italy

More information

DIT411/TIN175, Artificial Intelligence. Peter Ljunglöf. 23 January, 2018

DIT411/TIN175, Artificial Intelligence. Peter Ljunglöf. 23 January, 2018 DIT411/TIN175, Artificial Intelligence Chapters 3 4: More search algorithms CHAPTERS 3 4: MORE SEARCH ALGORITHMS DIT411/TIN175, Artificial Intelligence Peter Ljunglöf 23 January, 2018 1 TABLE OF CONTENTS

More information

Midterm Examination CS540-2: Introduction to Artificial Intelligence

Midterm Examination CS540-2: Introduction to Artificial Intelligence Midterm Examination CS540-2: Introduction to Artificial Intelligence March 15, 2018 LAST NAME: FIRST NAME: Problem Score Max Score 1 12 2 13 3 9 4 11 5 8 6 13 7 9 8 16 9 9 Total 100 Question 1. [12] Search

More information

Efficient memory-bounded search methods

Efficient memory-bounded search methods Efficient memory-bounded search methods Mikhail Simin Arjang Fahim CSCE 580: Artificial Intelligence Fall 2011 Dr. Marco Voltorta Outline of The Presentation Motivations and Objectives Background - BFS

More information

Branch & Bound (B&B) and Constraint Satisfaction Problems (CSPs)

Branch & Bound (B&B) and Constraint Satisfaction Problems (CSPs) Branch & Bound (B&B) and Constraint Satisfaction Problems (CSPs) Alan Mackworth UBC CS 322 CSP 1 January 25, 2013 P&M textbook 3.7.4 & 4.0-4.2 Lecture Overview Recap Branch & Bound Wrap up of search module

More information

Methods and Models for Combinatorial Optimization Exact methods for the Traveling Salesman Problem

Methods and Models for Combinatorial Optimization Exact methods for the Traveling Salesman Problem Methods and Models for Combinatorial Optimization Exact methods for the Traveling Salesman Problem L. De Giovanni M. Di Summa The Traveling Salesman Problem (TSP) is an optimization problem on a directed

More information

CS 188: Artificial Intelligence Fall 2008

CS 188: Artificial Intelligence Fall 2008 CS 188: Artificial Intelligence Fall 2008 Lecture 4: CSPs 9/9/2008 Dan Klein UC Berkeley Many slides over the course adapted from either Stuart Russell or Andrew Moore 1 1 Announcements Grading questions:

More information

Announcements. CS 188: Artificial Intelligence Fall Large Scale: Problems with A* What is Search For? Example: N-Queens

Announcements. CS 188: Artificial Intelligence Fall Large Scale: Problems with A* What is Search For? Example: N-Queens CS 188: Artificial Intelligence Fall 2008 Announcements Grading questions: don t panic, talk to us Newsgroup: check it out Lecture 4: CSPs 9/9/2008 Dan Klein UC Berkeley Many slides over the course adapted

More information

Algorithms for Integer Programming

Algorithms for Integer Programming Algorithms for Integer Programming Laura Galli November 9, 2016 Unlike linear programming problems, integer programming problems are very difficult to solve. In fact, no efficient general algorithm is

More information

Learning Directed Probabilistic Logical Models using Ordering-search

Learning Directed Probabilistic Logical Models using Ordering-search Learning Directed Probabilistic Logical Models using Ordering-search Daan Fierens, Jan Ramon, Maurice Bruynooghe, and Hendrik Blockeel K.U.Leuven, Dept. of Computer Science, Celestijnenlaan 200A, 3001

More information

Integer Programming ISE 418. Lecture 7. Dr. Ted Ralphs

Integer Programming ISE 418. Lecture 7. Dr. Ted Ralphs Integer Programming ISE 418 Lecture 7 Dr. Ted Ralphs ISE 418 Lecture 7 1 Reading for This Lecture Nemhauser and Wolsey Sections II.3.1, II.3.6, II.4.1, II.4.2, II.5.4 Wolsey Chapter 7 CCZ Chapter 1 Constraint

More information

A PACKAGE FOR DEVELOPMENT OF ALGORITHMS FOR GLOBAL OPTIMIZATION 1

A PACKAGE FOR DEVELOPMENT OF ALGORITHMS FOR GLOBAL OPTIMIZATION 1 Mathematical Modelling and Analysis 2005. Pages 185 190 Proceedings of the 10 th International Conference MMA2005&CMAM2, Trakai c 2005 Technika ISBN 9986-05-924-0 A PACKAGE FOR DEVELOPMENT OF ALGORITHMS

More information

Hybrid Feature Selection for Modeling Intrusion Detection Systems

Hybrid Feature Selection for Modeling Intrusion Detection Systems Hybrid Feature Selection for Modeling Intrusion Detection Systems Srilatha Chebrolu, Ajith Abraham and Johnson P Thomas Department of Computer Science, Oklahoma State University, USA ajith.abraham@ieee.org,

More information

Copyright 2007 Pearson Addison-Wesley. All rights reserved. A. Levitin Introduction to the Design & Analysis of Algorithms, 2 nd ed., Ch.

Copyright 2007 Pearson Addison-Wesley. All rights reserved. A. Levitin Introduction to the Design & Analysis of Algorithms, 2 nd ed., Ch. Iterative Improvement Algorithm design technique for solving optimization problems Start with a feasible solution Repeat the following step until no improvement can be found: change the current feasible

More information

Consistency and Set Intersection

Consistency and Set Intersection Consistency and Set Intersection Yuanlin Zhang and Roland H.C. Yap National University of Singapore 3 Science Drive 2, Singapore {zhangyl,ryap}@comp.nus.edu.sg Abstract We propose a new framework to study

More information

10708 Graphical Models: Homework 2

10708 Graphical Models: Homework 2 10708 Graphical Models: Homework 2 Due October 15th, beginning of class October 1, 2008 Instructions: There are six questions on this assignment. Each question has the name of one of the TAs beside it,

More information

Testing Independencies in Bayesian Networks with i-separation

Testing Independencies in Bayesian Networks with i-separation Proceedings of the Twenty-Ninth International Florida Artificial Intelligence Research Society Conference Testing Independencies in Bayesian Networks with i-separation Cory J. Butz butz@cs.uregina.ca University

More information

A SIMPLE APPROXIMATION ALGORITHM FOR NONOVERLAPPING LOCAL ALIGNMENTS (WEIGHTED INDEPENDENT SETS OF AXIS PARALLEL RECTANGLES)

A SIMPLE APPROXIMATION ALGORITHM FOR NONOVERLAPPING LOCAL ALIGNMENTS (WEIGHTED INDEPENDENT SETS OF AXIS PARALLEL RECTANGLES) Chapter 1 A SIMPLE APPROXIMATION ALGORITHM FOR NONOVERLAPPING LOCAL ALIGNMENTS (WEIGHTED INDEPENDENT SETS OF AXIS PARALLEL RECTANGLES) Piotr Berman Department of Computer Science & Engineering Pennsylvania

More information

Compressing Pattern Databases

Compressing Pattern Databases Compressing Pattern Databases Ariel Felner and Ram Meshulam Computer Science Department Bar-Ilan University Ramat-Gan, Israel 92500 Email: felner,meshulr1 @cs.biu.ac.il Robert C. Holte Computing Science

More information

Ordering-Based Search: A Simple and Effective Algorithm for Learning Bayesian Networks

Ordering-Based Search: A Simple and Effective Algorithm for Learning Bayesian Networks Ordering-Based Search: A Simple and Effective Algorithm for Learning Bayesian Networks Marc Teyssier Computer Science Dept. Stanford University Stanford, CA 94305 Daphne Koller Computer Science Dept. Stanford

More information

Constraint Satisfaction Problems (CSPs)

Constraint Satisfaction Problems (CSPs) 1 Hal Daumé III (me@hal3.name) Constraint Satisfaction Problems (CSPs) Hal Daumé III Computer Science University of Maryland me@hal3.name CS 421: Introduction to Artificial Intelligence 7 Feb 2012 Many

More information

Algorithm Design (8) Graph Algorithms 1/2

Algorithm Design (8) Graph Algorithms 1/2 Graph Algorithm Design (8) Graph Algorithms / Graph:, : A finite set of vertices (or nodes) : A finite set of edges (or arcs or branches) each of which connect two vertices Takashi Chikayama School of

More information

CS521 \ Notes for the Final Exam

CS521 \ Notes for the Final Exam CS521 \ Notes for final exam 1 Ariel Stolerman Asymptotic Notations: CS521 \ Notes for the Final Exam Notation Definition Limit Big-O ( ) Small-o ( ) Big- ( ) Small- ( ) Big- ( ) Notes: ( ) ( ) ( ) ( )

More information

Seminar on. A Coarse-Grain Parallel Formulation of Multilevel k-way Graph Partitioning Algorithm

Seminar on. A Coarse-Grain Parallel Formulation of Multilevel k-way Graph Partitioning Algorithm Seminar on A Coarse-Grain Parallel Formulation of Multilevel k-way Graph Partitioning Algorithm Mohammad Iftakher Uddin & Mohammad Mahfuzur Rahman Matrikel Nr: 9003357 Matrikel Nr : 9003358 Masters of

More information

An Effective Upperbound on Treewidth Using Partial Fill-in of Separators

An Effective Upperbound on Treewidth Using Partial Fill-in of Separators An Effective Upperbound on Treewidth Using Partial Fill-in of Separators Boi Faltings Martin Charles Golumbic June 28, 2009 Abstract Partitioning a graph using graph separators, and particularly clique

More information

Informed Search Algorithms

Informed Search Algorithms Informed Search Algorithms CITS3001 Algorithms, Agents and Artificial Intelligence Tim French School of Computer Science and Software Engineering The University of Western Australia 2017, Semester 2 Introduction

More information

Mini-Buckets: A General Scheme for Generating Approximations in Automated Reasoning

Mini-Buckets: A General Scheme for Generating Approximations in Automated Reasoning Mini-Buckets: A General Scheme for Generating Approximations in Automated Reasoning Rina Dechter* Department of Information and Computer Science University of California, Irvine dechter@ics. uci. edu Abstract

More information

Constructions of hamiltonian graphs with bounded degree and diameter O(log n)

Constructions of hamiltonian graphs with bounded degree and diameter O(log n) Constructions of hamiltonian graphs with bounded degree and diameter O(log n) Aleksandar Ilić Faculty of Sciences and Mathematics, University of Niš, Serbia e-mail: aleksandari@gmail.com Dragan Stevanović

More information

Provably Efficient Non-Preemptive Task Scheduling with Cilk

Provably Efficient Non-Preemptive Task Scheduling with Cilk Provably Efficient Non-Preemptive Task Scheduling with Cilk V. -Y. Vee and W.-J. Hsu School of Applied Science, Nanyang Technological University Nanyang Avenue, Singapore 639798. Abstract We consider the

More information

CSE 473: Artificial Intelligence

CSE 473: Artificial Intelligence CSE 473: Artificial Intelligence Constraint Satisfaction Luke Zettlemoyer Multiple slides adapted from Dan Klein, Stuart Russell or Andrew Moore What is Search For? Models of the world: single agent, deterministic

More information

Coping with the Limitations of Algorithm Power Exact Solution Strategies Backtracking Backtracking : A Scenario

Coping with the Limitations of Algorithm Power Exact Solution Strategies Backtracking Backtracking : A Scenario Coping with the Limitations of Algorithm Power Tackling Difficult Combinatorial Problems There are two principal approaches to tackling difficult combinatorial problems (NP-hard problems): Use a strategy

More information

Learning Optimal Bounded Treewidth Bayesian Networks via Maximum Satisfiability

Learning Optimal Bounded Treewidth Bayesian Networks via Maximum Satisfiability Learning Optimal Bounded Treewidth Bayesian Networks via Maximum Satisfiability Jeremias Berg and Matti Järvisalo and Brandon Malone HIIT & Department of Computer Science, University of Helsinki, Finland

More information

Research Article Structural Learning about Directed Acyclic Graphs from Multiple Databases

Research Article Structural Learning about Directed Acyclic Graphs from Multiple Databases Abstract and Applied Analysis Volume 2012, Article ID 579543, 9 pages doi:10.1155/2012/579543 Research Article Structural Learning about Directed Acyclic Graphs from Multiple Databases Qiang Zhao School

More information

What is Search For? CSE 473: Artificial Intelligence. Example: N-Queens. Example: N-Queens. Example: Map-Coloring 4/7/17

What is Search For? CSE 473: Artificial Intelligence. Example: N-Queens. Example: N-Queens. Example: Map-Coloring 4/7/17 CSE 473: Artificial Intelligence Constraint Satisfaction Dieter Fox What is Search For? Models of the world: single agent, deterministic actions, fully observed state, discrete state space Planning: sequences

More information

An efficient approach for finding the MPE in belief networks

An efficient approach for finding the MPE in belief networks 342 Li and D'Ambrosio An efficient approach for finding the MPE in belief networks Zhaoyu Li Department of Computer Science Oregon State University Corvallis, OR 97331 Bruce D'Ambrosio Department of Computer

More information

Searching with Partial Information

Searching with Partial Information Searching with Partial Information Above we (unrealistically) assumed that the environment is fully observable and deterministic Moreover, we assumed that the agent knows what the effects of each action

More information

Closed Non-Derivable Itemsets

Closed Non-Derivable Itemsets Closed Non-Derivable Itemsets Juho Muhonen and Hannu Toivonen Helsinki Institute for Information Technology Basic Research Unit Department of Computer Science University of Helsinki Finland Abstract. Itemset

More information

CS781 Lecture 2 January 13, Graph Traversals, Search, and Ordering

CS781 Lecture 2 January 13, Graph Traversals, Search, and Ordering CS781 Lecture 2 January 13, 2010 Graph Traversals, Search, and Ordering Review of Lecture 1 Notions of Algorithm Scalability Worst-Case and Average-Case Analysis Asymptotic Growth Rates: Big-Oh Prototypical

More information

Lecture 5 Heuristics. Last Time: A* Search

Lecture 5 Heuristics. Last Time: A* Search CSE 473 Lecture 5 Heuristics CSE AI Faculty Last Time: A* Search Use an evaluation function f(n) for node n. f(n) = estimated total cost of path thru n to goal f(n) = g(n) + h(n) g(n) = cost so far to

More information

3 INTEGER LINEAR PROGRAMMING

3 INTEGER LINEAR PROGRAMMING 3 INTEGER LINEAR PROGRAMMING PROBLEM DEFINITION Integer linear programming problem (ILP) of the decision variables x 1,..,x n : (ILP) subject to minimize c x j j n j= 1 a ij x j x j 0 x j integer n j=

More information

Constraint Satisfaction Problems

Constraint Satisfaction Problems Constraint Satisfaction Problems Berlin Chen Department of Computer Science & Information Engineering National Taiwan Normal University References: 1. S. Russell and P. Norvig. Artificial Intelligence:

More information

56:272 Integer Programming & Network Flows Final Examination -- December 14, 1998

56:272 Integer Programming & Network Flows Final Examination -- December 14, 1998 56:272 Integer Programming & Network Flows Final Examination -- December 14, 1998 Part A: Answer any four of the five problems. (15 points each) 1. Transportation problem 2. Integer LP Model Formulation

More information

INTRODUCTION TO HEURISTIC SEARCH

INTRODUCTION TO HEURISTIC SEARCH INTRODUCTION TO HEURISTIC SEARCH What is heuristic search? Given a problem in which we must make a series of decisions, determine the sequence of decisions which provably optimizes some criterion. What

More information

CSE 431/531: Algorithm Analysis and Design (Spring 2018) Greedy Algorithms. Lecturer: Shi Li

CSE 431/531: Algorithm Analysis and Design (Spring 2018) Greedy Algorithms. Lecturer: Shi Li CSE 431/531: Algorithm Analysis and Design (Spring 2018) Greedy Algorithms Lecturer: Shi Li Department of Computer Science and Engineering University at Buffalo Main Goal of Algorithm Design Design fast

More information

Ordering attributes for missing values prediction and data classification

Ordering attributes for missing values prediction and data classification Ordering attributes for missing values prediction and data classification E. R. Hruschka Jr., N. F. F. Ebecken COPPE /Federal University of Rio de Janeiro, Brazil. Abstract This work shows the application

More information

Basic Search Algorithms

Basic Search Algorithms Basic Search Algorithms Tsan-sheng Hsu tshsu@iis.sinica.edu.tw http://www.iis.sinica.edu.tw/~tshsu 1 Abstract The complexities of various search algorithms are considered in terms of time, space, and cost

More information

Lecture 3: Graphs and flows

Lecture 3: Graphs and flows Chapter 3 Lecture 3: Graphs and flows Graphs: a useful combinatorial structure. Definitions: graph, directed and undirected graph, edge as ordered pair, path, cycle, connected graph, strongly connected

More information

Integrating Machine Learning in Parallel Heuristic Search

Integrating Machine Learning in Parallel Heuristic Search From: Proceedings of the Eleventh International FLAIRS Conference. Copyright 1998, AAAI (www.aaai.org). All rights reserved. Integrating Machine Learning in Parallel Heuristic Search R. Craig Varnell Stephen

More information

Figure : Example Precedence Graph

Figure : Example Precedence Graph CS787: Advanced Algorithms Topic: Scheduling with Precedence Constraints Presenter(s): James Jolly, Pratima Kolan 17.5.1 Motivation 17.5.1.1 Objective Consider the problem of scheduling a collection of

More information

AN APPROACH FOR LOAD BALANCING FOR SIMULATION IN HETEROGENEOUS DISTRIBUTED SYSTEMS USING SIMULATION DATA MINING

AN APPROACH FOR LOAD BALANCING FOR SIMULATION IN HETEROGENEOUS DISTRIBUTED SYSTEMS USING SIMULATION DATA MINING AN APPROACH FOR LOAD BALANCING FOR SIMULATION IN HETEROGENEOUS DISTRIBUTED SYSTEMS USING SIMULATION DATA MINING Irina Bernst, Patrick Bouillon, Jörg Frochte *, Christof Kaufmann Dept. of Electrical Engineering

More information

Midterm Examination CS 540-2: Introduction to Artificial Intelligence

Midterm Examination CS 540-2: Introduction to Artificial Intelligence Midterm Examination CS 54-2: Introduction to Artificial Intelligence March 9, 217 LAST NAME: FIRST NAME: Problem Score Max Score 1 15 2 17 3 12 4 6 5 12 6 14 7 15 8 9 Total 1 1 of 1 Question 1. [15] State

More information

Graph and Digraph Glossary

Graph and Digraph Glossary 1 of 15 31.1.2004 14:45 Graph and Digraph Glossary A B C D E F G H I-J K L M N O P-Q R S T U V W-Z Acyclic Graph A graph is acyclic if it contains no cycles. Adjacency Matrix A 0-1 square matrix whose

More information

U.C. Berkeley CS170 : Algorithms, Fall 2013 Midterm 1 Professor: Satish Rao October 10, Midterm 1 Solutions

U.C. Berkeley CS170 : Algorithms, Fall 2013 Midterm 1 Professor: Satish Rao October 10, Midterm 1 Solutions U.C. Berkeley CS170 : Algorithms, Fall 2013 Midterm 1 Professor: Satish Rao October 10, 2013 Midterm 1 Solutions 1 True/False 1. The Mayan base 20 system produces representations of size that is asymptotically

More information

Branch-and-Cut and GRASP with Hybrid Local Search for the Multi-Level Capacitated Minimum Spanning Tree Problem

Branch-and-Cut and GRASP with Hybrid Local Search for the Multi-Level Capacitated Minimum Spanning Tree Problem Branch-and-Cut and GRASP with Hybrid Local Search for the Multi-Level Capacitated Minimum Spanning Tree Problem Eduardo Uchoa Túlio A.M. Toffolo Mauricio C. de Souza Alexandre X. Martins + Departamento

More information

Distributed minimum spanning tree problem

Distributed minimum spanning tree problem Distributed minimum spanning tree problem Juho-Kustaa Kangas 24th November 2012 Abstract Given a connected weighted undirected graph, the minimum spanning tree problem asks for a spanning subtree with

More information

4 INFORMED SEARCH AND EXPLORATION. 4.1 Heuristic Search Strategies

4 INFORMED SEARCH AND EXPLORATION. 4.1 Heuristic Search Strategies 55 4 INFORMED SEARCH AND EXPLORATION We now consider informed search that uses problem-specific knowledge beyond the definition of the problem itself This information helps to find solutions more efficiently

More information

Dr. Mustafa Jarrar. Chapter 4 Informed Searching. Sina Institute, University of Birzeit

Dr. Mustafa Jarrar. Chapter 4 Informed Searching. Sina Institute, University of Birzeit Lecture Notes, Advanced Artificial Intelligence (SCOM7341) Sina Institute, University of Birzeit 2 nd Semester, 2012 Advanced Artificial Intelligence (SCOM7341) Chapter 4 Informed Searching Dr. Mustafa

More information

Constraint Satisfaction

Constraint Satisfaction Constraint Satisfaction Philipp Koehn 1 October 2015 Outline 1 Constraint satisfaction problems (CSP) examples Backtracking search for CSPs Problem structure and problem decomposition Local search for

More information

Constraint Satisfaction Problems. Chapter 6

Constraint Satisfaction Problems. Chapter 6 Constraint Satisfaction Problems Chapter 6 Constraint Satisfaction Problems A constraint satisfaction problem consists of three components, X, D, and C: X is a set of variables, {X 1,..., X n }. D is a

More information

Network Routing Protocol using Genetic Algorithms

Network Routing Protocol using Genetic Algorithms International Journal of Electrical & Computer Sciences IJECS-IJENS Vol:0 No:02 40 Network Routing Protocol using Genetic Algorithms Gihan Nagib and Wahied G. Ali Abstract This paper aims to develop a

More information

Optimal Sequential Multi-Way Number Partitioning

Optimal Sequential Multi-Way Number Partitioning Optimal Sequential Multi-Way Number Partitioning Richard E. Korf, Ethan L. Schreiber, and Michael D. Moffitt Computer Science Department University of California, Los Angeles Los Angeles, CA 90095 IBM

More information

Artificial Intelligence

Artificial Intelligence Artificial Intelligence Informed Search and Exploration Chapter 4 (4.1 4.2) A General Search algorithm: Chapter 3: Search Strategies Task : Find a sequence of actions leading from the initial state to

More information

Lecture 2 - Graph Theory Fundamentals - Reachability and Exploration 1

Lecture 2 - Graph Theory Fundamentals - Reachability and Exploration 1 CME 305: Discrete Mathematics and Algorithms Instructor: Professor Aaron Sidford (sidford@stanford.edu) January 11, 2018 Lecture 2 - Graph Theory Fundamentals - Reachability and Exploration 1 In this lecture

More information

Constraint Satisfaction Problems

Constraint Satisfaction Problems Constraint Satisfaction Problems Tuomas Sandholm Carnegie Mellon University Computer Science Department [Read Chapter 6 of Russell & Norvig] Constraint satisfaction problems (CSPs) Standard search problem:

More information

CSCI 5454 Ramdomized Min Cut

CSCI 5454 Ramdomized Min Cut CSCI 5454 Ramdomized Min Cut Sean Wiese, Ramya Nair April 8, 013 1 Randomized Minimum Cut A classic problem in computer science is finding the minimum cut of an undirected graph. If we are presented with

More information

Example: Map coloring

Example: Map coloring Today s s lecture Local Search Lecture 7: Search - 6 Heuristic Repair CSP and 3-SAT Solving CSPs using Systematic Search. Victor Lesser CMPSCI 683 Fall 2004 The relationship between problem structure and

More information