Compiling relational Bayesian networks for exact inference

Size: px
Start display at page:

Download "Compiling relational Bayesian networks for exact inference"

Transcription

1 International Journal of Approximate Reasoning 42 (2006) Compiling relational Bayesian networks for exact inference Mark Chavira a, *, Adnan Darwiche a, Manfred Jaeger b a Computer Science Department, UCLA, Los Angeles, CA 90095, United States b Institut for Datalogi, Aalborg Universitet, Fredrik Bajers Vej 7 E, DK-9220 Aalborg Ø, Denmark Available online 15 November 2005 Abstract We describe in this paper a system for exact inference with relational Bayesian networks as defined in the publicly available Primula tool. The system is based on compiling propositional instances of relational Bayesian networks into arithmetic circuits and then performing online inference by evaluating and differentiating these circuits in time linear in their size. We report on experimental results showing successful compilation and efficient inference on relational Bayesian networks, whose Primula-generated propositional instances have thousands of variables, and whose jointrees have clusters with hundreds of variables. Ó 2005 Elsevier Inc. All rights reserved. Keywords: Exact inference; Relational models; Bayesian networks 1. Introduction Relational probabilistic models extend Bayesian network models by representing objects, their attributes, and their relations with other objects. The standard approach for inference with a relational model is based on the generation of a propositional instance of the model in the form of a classical Bayesian network, and then applying classical algorithms, such as jointree [1], to compute answers to queries. * Corresponding author. addresses: chavira@cs.ucla.edu (M. Chavira), darwiche@cs.ucla.edu (A. Darwiche), jaeger@cs.aau.dk (M. Jaeger) X/$ - see front matter Ó 2005 Elsevier Inc. All rights reserved. doi: /j.ijar

2 The propositional instance of a relational model includes one Boolean random variable for each ground relational atom. For example, if we have n domain objects o 1,...,o n, and a binary relation R(Æ,Æ), we generate a propositional variable for each instance of the relation: R(o 1,o 1 ),R(o 1,o 2 ),...,R(o n,o n ). The first task in making Bayesian networks over these random variables tractable for inference is to ensure that the size of the Bayesian network representation does not show exponential growth in the number n of domain objects (as can easily happen due to nodes whose in-degree grows as a function of n). This can often be achieved by decomposing nodes with high in-degree into suitable, sparsely connected sub-networks using a number of new, auxiliary nodes. This approach is systematically employed in the Primula system. Even when a reasonably compact Bayesian network representation (i.e., polynomial in the number of objects) has been constructed for a propositional instance, this model will often be inaccessible to standard algorithms for exact inference, because its global structure does not lead to tractable jointrees. Even though the constructed networks may lack the global structure that would make them accessible to standard inference techniques, they may very well exhibit abundant local structure in the form of determinism. The objective of this paper is to describe a system for inference with propositional instances of relational models which can exploit this local structure, allowing us to reason very efficiently with some relational models whose propositional instances may look quite formidable at first. Specifically, we employ the approach proposed by [2] to compile propositional instances of relational models into arithmetic circuits, and then perform online inference by evaluating and differentiating the compiled circuits in time linear in their size. As our experimental results illustrate, this approach can efficiently handle some relational models whose Primula-generated propositional instances are quite massive. 1 We note here that the inference approach of [2] is applicable to any Bayesian network, but is especially effective on networks with local structure, including determinism. Hence, one of the main points of this paper is to illustrate the extent of local structure available in propositional instances of relational models, and the effectiveness in exploiting this local structure by the approach proposed in [2]. This paper is structured as follows. We start in Section 2 with a review of relational models in general and the specific formalization used in this paper. We then discuss in Section 3 the Primula system, which implements this formalization together with a method for generating propositional instances in the form of Bayesian networks. Section 4 is then dedicated to our proposed approach for compiling relational models. We provide experimental results in Section 5, and finally close with some concluding remarks in Section Relational models M. Chavira et al. / Internat. J. Approx. Reason. 42 (2006) A Bayesian network is a compact representation of a probability distribution and has two parts: a directed acyclic graph and a set of conditional probability tables (CPTs). Each node in the graph represents a random variable, which we assume to be discrete in this paper. Each variable X has associated with it a CPT, which specifies the conditional probabilities Pr(xju), where u is a configuration of the parents U of X in the network. 1 Some may recall the technique of zero-compression which can be used to exploit determinism in the jointree framework [3]. This technique, however, requires that one perform inference on the original jointree before it is zero-compressed, making almost all of our datasets inaccessible to this method. For a more detailed relationship to jointree inference, the reader is referred to [4].

3 6 M. Chavira et al. / Internat. J. Approx. Reason. 42 (2006) 4 20 A Bayesian network over a set of variables specifies a unique probability distribution over these variables. Probabilistic queries with respect to a Bayesian network are to be interpreted as queries with respect to the probability table the network specifies. The main goal of algorithms for Bayesian networks is to answer such queries without having to construct the table explicitly, since the tableõs size is exponential in the number of network variables. Fig. 1 depicts a simple Bayesian network with two of its CPTs. Relational or first-order probabilistic models extend propositional modeling supported by Bayesian networks by allowing one to represent objects explicitly, and to define relations over these objects. Most of the early work on such generic models, which has been subsumed under the title knowledge-based model construction (see e.g. [5]), combines elements of logic-programming with Bayesian networks. Today one can distinguish several distinct representation paradigms for relational and first-order models: (inductive) logicprogramming based approaches [6 8], network fragments [9], frame-based representations [10,11], and probabilistic predicate logic formulas [12]. We review relational models with an example An example Consider the well-known example depicted in Fig. 2(a), in which Holmes becomes alarmed if he receives a call from his neighbor Watson. Watson will likely call if an alarm has sounded at HolmesÕ residence, which is more likely if a burglary occurs. However, Watson is a prankster, so Holmes may receive a call even if the alarm does not sound. We can model this example with a Bayesian network as shown in Fig. 2(b). A query might be the probability that there is a burglary given that Holmes is alarmed. We could also consider similar scenarios. Holmes might have multiple neighbors (only some of whom are pranksters) and become alarmed if any of them calls. There might be multiple individuals who can receive calls, each with distinct neighbors. Or it might be that individuals share neighbors and individuals who receive calls can also make them. For each of these scenarios, we can construct a distinct Bayesian network. Moreover, we can imagine needing to deal with many of these situations, and hence needing to construct many different networks. Each of the situations described represents a combination of various themes, such as the theme of an alarm compelling a neighbor to call or an individual becoming alarmed when some neighbor calls. Relational models address domains involving themes by separating the model construction process into two phases. We first describe a set of general Fig. 1. A Bayesian net with two of its CPTs.

4 M. Chavira et al. / Internat. J. Approx. Reason. 42 (2006) Holmes Phone Watson Alarm Burglar Watson (prankster) (a) Neighbor Holmes is Alarmed Watson calls Holmes Alarm Burglary Holmes (b) (c) Fig. 2. (a) A simple alarm scenario, (b) the corresponding Bayesian network, and (c) a graph depicting the particulars of the situation, as opposed to what is common to all alarm situations. rules that apply to all situations. For example, in the alarm domain described, we need four rules: (1) At a given residence, the probability of burglary is (2) A particular alarm sounds with probability 0.95 if a burglary occurs at the corresponding residence, and with probability 0.01 otherwise. (3) If an alarm sounds at an individualõs residence, then each of the individualõs neighbors will call with probability 0.9; otherwise, if the neighbor is a prankster, then the neighbor will call with probability 0.05; otherwise, the neighbor will not call. (4) An individual is alarmed if one or more neighbors call. We highlight here that whether an individual is alarmed depends on the number of the individualõs neighbors, which makes this domain difficult represent with a template-based language. Once we have specified what is common to all situations, in order to specify a particular situation, we only need specify a small amount of additional information. In the alarm example, that information consists of which individuals are involved (other than burglars), who are neighbors of whom, and who are pranksters. We specify a graph where nodes represent individuals, edges capture the neighbor relationship, and each node is marked if the corresponding individual is a pranktser. Fig. 2(c) depicts the graph corresponding to the situation in Fig. 2(a). One of the main advantages of using a relational model is that a relational model describes a situation involving themes succinctly. This advantage often makes constructing a relational model much easier and less error-prone than constructing a Bayesian network. For example, it is not uncommon for a relational model with a dozen or so general rules to correspond to a Bayesian network that involves hundreds of thousands of CPT parameters. Another advantage is that much of the work performed in constructing a relational model can be directly re-used in describing variations of the model, whereas creating another Bayesian network can involve much more work.

5 8 M. Chavira et al. / Internat. J. Approx. Reason. 42 (2006) Relational Bayesian networks We use in this paper the language of relational Bayesian networks [12] to represent relational models, as implemented in the Primula system available at ~jaeger/primula. The formal semantics of the language is based on Random relational structure models (RRSMs), which we define next. Definition 1. Given (1) a set of relational symbols S, called predefined relations; (2) a set of relational symbols R, called probabilistic relations; and (3) a finite set D, called the domain; we define an S D -structure to be an interpretation of relations S over domain D, that is, a function which maps every ground atom s(d) (s 2 S, d D) to either true or false. We also define a random relational structure model (RRSM) as a partial function which takes an S D -structure as input, and returns a probability distribution over all R D -structures as output. Intuitively, members of domain D represent objects, and members of S and R represent relations that can hold on these objects. These relations can be unary in which case they are called attributes. A user would typically define the relations in S (by providing an S D -structure), and then use an RRSM to induce a probability distribution over the possible definitions of relations in R (R D -structures). We note here that S D -structures correspond to skeleton structures in [11]. For the alarm example above, the set D of objects is the set of individuals. The set of predefined relations S contains a unary relation, prankster, in addition to a binary relation neighbor. There are four probabilistic relations in R for this domain. The first is calls(v,w): whether v calls w in order to warn w that his alarm went off. We also have another probabilistic relation alarmed(v): whether v has been alarmed (called by at least one neighbor). A third is the relation alarm(v): whether võs alarm went off. The last probabilistic relation is burglary(v): whether võs home has been burglarized. The RRSM is the set of four generic rules described previously. We now describe four RRSMs used in our experiments. These models have been implemented in Primula, which provides a syntax for specifying RRSM Random blocks This model describes the random placement of blocks (obstacles) on the locations of a map. The input structures consist of a particular gridmap and a set of blocks. This is represented using a set of predefined relations S ={location,block, leftof, belowof} where location and block are attributes that partition the domain into the two types of objects, and leftof and belowof are binary relations that determine the spatial relationship among locations. Fig. 3 shows an input S D -structure. One of the probabilistic relations in R for this model is the binary relation blocks(b, l) which represents the random placement of a block b on some location l. Another is connected(l 1,l 2 ) between pairs of locations which describes whether, after placement of the blocks, there is an unblocked path between l 1 and l 2. A probabilistic query might be the probability that there is an unblocked path between two locations l 1 and l 2, given the observed locations of some blocks (but uncertainty about the placement of the remaining ones). We experiment with different versions of this relational model, blockmap-l b, where l is the number of locations and b, the number of blocks.

6 M. Chavira et al. / Internat. J. Approx. Reason. 42 (2006) Blocks B1 B2 1 Locations 2 3 belowof 4 leftof 5 Fig. 3. Input S D -structure Mastermind In the game of Mastermind, Player 1 arranges a hidden sequence of colored pegs. Player 2 guesses the exact sequence of colors by arranging guessed sequences of colored pegs. To each guessed sequence, Player 1 responds by stating how many pegs in the guess match pegs in his hidden sequence both in color and position (white feedback), and how many pegs in the guess match pegs in the hidden sequence only in color (black feedback). Player 2 wins if he guesses the hidden sequence within a certain number of rounds. The game can be represented as an RRSM where the domain D consists of objects of types peg, color, and round specified by corresponding unary relations in S, as well as binary relations peg-ord and round-ord in S that impose orders on the peg and round objects, respectively. The probabilistic relations R in the model represent the game configurations after a number of rounds: true-color(p,c) represents that c is the color of the hidden peg p; guessed-color(p,c,r) represents that in round r color c was placed in position p in the guess. Similarly, the arrangement of the feedback pegs can be encoded. A query might be the most probable color configuration of the hidden pegs, given the observed query and feedback pegs. We experiment with different versions of this model, mastermind-c g p, where c is the number of colors, g is the number of guesses, and p is the number of pegs Students and professors This domain was used by [13] to investigate methods for approximate inference for relational models. We have two types of objects in this model: students and professors and two corresponding attributes in the set S. Professors have two probabilistic attributes in R: fame(yes/no) and funding_level (high/low). Students have one probabilistic attribute in R: success(yes/no). Students and professors are related via the binary probabilistic relation advisor(s, p) in R. According to the model, students use the softmax rule, and choose advisor i with funding level y i with probability e y i = Pk ey k. With the funding level discretized into two categories high and low, this reduces to choosing any given rich (poor) professor with probability z h /(Kz h + Lz l )(z l /(Kz h + Lz l )), where K is the number of rich professors, L is the number of poor professors, and z h, z l are the (exponentials of) the funding levels of rich, respectively poor, professors. The probability of success of a student is defined conditional on the funding level. A query for this model can be the probabilities for a professorõs funding level, given the success of his students. Inference in this model becomes hard very quickly with increasing numbers of professors and students in the

7 10 M. Chavira et al. / Internat. J. Approx. Reason. 42 (2006) 4 20 domain [13]. We will experiment with different versions of this relational model, studentsp s, where p is the number of professors and s is the number of students Friends and smokers This domain was introduced in [14]. It involves a number of individuals, with relations in R, such as smokes(v), which indicates whether a person smokes, cancer(v), which indicates whether a person has cancer, and friends(u,v), which indicates who are friends of whom. There are no relations in S for this model. The probabilistic model over R is defined by assigning weights to logical constraints, such as friends(u, v) ^ smokes(u)! smokes(v). A query for this model might be the probability that a person has cancer given information about others who have cancer. The Primula encoding of this model utilizes auxiliary probabilistic relations corresponding to the logical constraints. In ground instances of the model these auxiliary variables manifest themselves as variables in the Bayesian network, on which evidence should be asserted to indicate that they are always true. We experiment with different versions of this relational model, fr&sm-n, where n is the number of people in the domain. 3. The Primula system The RRSM is an abstract semantics of probabilistic relational models. For a practical system, one needs a specific syntax for specifying an RRSM. Primula allows users to encode RRSMs using the language of relational Bayesian networks [12], and outputs the distribution on R D -structures in the form of a standard Bayesian network Specifying RRSMs using Primula We now provide an example of specifying an RRSM using Primula. Consider again the alarm example from Section 2.1 and recall that for this example, the domain is the set of individuals, the set of predefined relations is S ={prankster(v), neighbor(v, w)}, and the set of probabilistic relations is R ={calls(v, w), alarm(v), alarmed(v), burglary(v)}. The probability of calls(v, w) is defined conditional on the predefined neighbor and prankster relations (it is 0 if v and w are not neighbors), and on the probabilistic alarm(v) relation: whether the alarm of v went off. This RRSM is specified in Primula as given in Fig. 4, which provides the probability distribution on probabilistic relations using probability formulas. These formulas can be seen either as probabilistic analogues of predicate logic formulas, or as expressions in a functional programming language. A probability formula defines both the dependency Fig. 4. Specifying an RRSM using Primula.

8 M. Chavira et al. / Internat. J. Approx. Reason. 42 (2006) structure between ground probabilistic atoms (which depends on the predefined relations in the input structure), and the exact conditional probabilities, given the truth values of parent atoms. The specification of the RRSM provides some intuition for why a logic-based approach might work well when applied to Primula generated networks. In addition to certain numbers, we also see in this specification a number of logical constructs. For example, each of the occurrences of (x : y, z) is essentially an application of an if then else, and the noisy-or construct is essentially an existential quantification, which can be converted into a disjunction over a set of auxiliary variables. The utilization of these logical constructs is quite common in relational models From relational to propositional networks To instantiate a generic relational model in Primula, one must provide a definition of an input S D -structure. For the RRSM defined in Fig. 4, one must define the set of individuals in domain D, and then one must define which of these individuals are pranksters (by defining the attribute prankster), and who are neighbors of whom (by defining the relation neighbor). Primula provides a GUI for this purpose, but one can also supply a file-based definition of the domain and corresponding S relations. Fig. 5 presents what one of these files might look like. This file defines the domain to be D ={Holmes,Watson,Gibbon} and specifies that Gibbon is a prankster, that Holmes is a neighbor of Watson and Gibbon and that Watson and Gibbon are neighbors of Holmes. Given the above inputs, the distribution over probabilistic relations can be represented, as described in Section 1, using a standard Bayesian network with a node for each ground probabilistic atom. Our example also illustrates how the in-degree of a node can grow as a function of the number of domain objects: the node alarmed(holmes), for instance, depends on calls(w,holmes) for all of HolmesÕs neighbors w (of which there might be arbitrarily many). The Primula system employs the general method described in [15] to decompose the dependency of a node on multiple parents. This method consists of an iterative algorithm that takes the probability formula defining the distribution of a node, decomposes it into its top-level subformulas by introducing one new auxiliary node for each of these subformulas and defines the probability of the original node conditional only on the new auxiliary nodes. This method can be applied to any relational Bayesian network that only contains multi-linear combination functions (including noisy-or and mean), and yields a Bayesian network where the number of parents is bounded by three for all nodes. Even when one succeeds in constructing a standard Bayesian network of a manageable representation size, inference in this network may be computationally very hard. It is a long-standing open problem in first-order and relational modeling whether one might not design inference techniques that avoid these complexities of inference in the ground propositional instances by performing inference directly on the level of the relational Fig. 5. Specifying an S D structure using Primula.

9 12 M. Chavira et al. / Internat. J. Approx. Reason. 42 (2006) 4 20 representation, perhaps employing techniques of first-order logical inference. Complexity results derived in [16] show that one cannot hope for a better worst-case performance with such inference techniques. This still leaves the possibility that they could often lead to substantial gains in practice. Recent work has described high-level inference techniques that aim at achieving such gains in average-case performance [17,18]. The potential advantage of this and similar techniques seems to be restricted, however, to relational models where individual model instances are given by relatively unstructured input structures, i.e., input structures containing large numbers of indistinguishable objects. The potential of high-level inference techniques lies in their ability to deal with such sets of objects without explicitly naming each object individually. However, in the type of relational models we are considering here, the input structures consist of mostly unique objects (in Section 2.2.1, for instance, the block objects are indistinguishable, but all location objects have unique properties defined by the belowof and leftof relations). We can identify an input structure with the complete ground propositional theory that defines it (for the structure of Fig. 3 this would be the theory blockðb1þ^:locationðb1þ^^leftofð2; 3Þ^^ :belowofð5; 5ÞÞ, and, informally, characterize highly structured input structures as those for which this propositional theory admits no simple first-order abstraction. When a relational model instance, now, is given by an input structure that cannot be succinctly encoded in an abstract, first-order style representation, chances are very small that probabilistic inference for this model instance can gain much efficiency by operating on a nonpropositional level. It thus appears that at least for a fairly large class of interesting models more advantages might be gained by optimizing inference techniques for ground propositional models, than by non-propositional inference techniques. Table 1 depicts the relational models with which we experimented, together with the size of corresponding propositional Bayesian networks generated by Primula. The table also reports the size of the largest cluster for the jointree we constructed for these networks. Obviously, most of these networks are inaccessible to mainstream, structure-based algorithms for exact inference. Yet, we will show later that all of these particular models can be handled efficiently using the compilation approach we propose in this paper. 4. Compiling relational models We describe in this section the approach we use to perform exact inference on propositional instances of relational models, which is based on compiling Bayesian networks into arithmetic circuits [2]. Inference can then be performed using a simple two-pass procedure in which the circuit is evaluated and differentiated given evidence Bayesian networks as polynomials The compilation approach we adopt is based on viewing each Bayesian network as a very large polynomial (multi-linear function in particular), which may be compactly represented using an arithmetic circuit. The function itself contains two types of variables. For each value x of each variable X in the network, we have a variable k x called an evidence indicator. For each instantiation x, u of each variable X and its parents U in the network, we have a variable h xju called a network parameter. The multi-linear function has a

10 M. Chavira et al. / Internat. J. Approx. Reason. 42 (2006) Table 1 Relational Bayesian networks, their corresponding propositional instances, and the sizes of their CNF encodings Relational model Mastermind Bayesian network CNF encoding Arithmetic circuit AC time JT Vars CPT Max Vars Clauses Nodes Edges Inf Comp Inf Params Clst (s) (min) (s) Count Log c r p , , , , , , , , ,228 1,523, , ,490 1,293,323 4,315, , ,351 4,859, , ,355 19,457, , ,453 1,359,391 55,417, Students p s , , , , , , ,099 95, , , , , , , , ,092 2,531, , ,734 38,889 1,319,834 5,236, , ,353 4,586,368 16,936, , ,693 64,325 9,922,233 36,450, Blockmap l b , , , ,083 10,147 56, , ,318 11, , , ,529 17, , , , ,525 58,094 29, , , , ,709 62,299 33,011 1,798, , , ,877 66,443 47,475 7,643, , , , ,164 69, , , , , ,570 75,299 6,989, , , , , ,602 40,172, , , , ,526 96,424 1,103, , , , , ,980 11,707, , , , ,056 44,136 76,649, fr&sm n ,404 1, , ,655 3,686 4, , ,986 42,302 7,689 8, , ,976 77,864 13,919 15, , , ,257 22,824 25, (continued on next page)

11 14 M. Chavira et al. / Internat. J. Approx. Reason. 42 (2006) 4 20 Table 1 (continued) Relational model Bayesian network CNF encoding Arithmetic circuit AC time JT Vars CPT Max Vars Clauses Nodes Edges Inf Comp Inf Params Clst (s) (min) (s) Count Log , , ,397 34,877 38, , , ,200 50,651 55, , , ,582 70,541 76, , , ,342 78,203 84, term for each instantiation of the network variables, which is constructed by multiplying all evidence indicators and network parameters that are consistent with that instantiation. For example, the multi-linear function of the network in Fig. 1 has eight terms corresponding to the eight instantiations of variables A, B, C: f ¼ k a k b k c h a h bja h cja þ k a k b k c h a h bja h cja þþk a k b k c h a h bja h cja. Given this multi-linear function f, we can answer standard queries with respect to its corresponding Bayesian network by simply evaluating and differentiating this function; see [2] for details. The ability to compute answers to probabilistic queries directly from the derivatives of f is interesting semantically, but one must realize that the size of function f is exponential in the number of network variables. Yet, one may be able to factor this function and represent it more compactly using an arithmetic circuit. An arithmetic circuit is a rooted DAG, in which each leaf represents a variable or constant and each internal node represents the product or sum of its children; see Fig. 6. If we can represent the network polynomial efficiently using an arithmetic circuit, then inference can be done in time linear in the size of such circuits, since the (first) partial derivatives of an arithmetic circuit can all be computed simultaneously in time linear in the circuit size [2] Compiling the network polynomial into an arithmetic circuit We now turn to the approach for compiling/factoring network polynomials into arithmetic circuits, which is based on reducing the factoring problem to one of logical reasoning [19]. This approach is based on three conceptual steps, as shown in Fig. 6. First, the network polynomial is encoded using a propositional theory. Next, the propositional theory is factored by converting it to a special logical form. Finally, an arithmetic circuit is extracted from the factored propositional theory. 2 Step 1: Encoding a multi-linear function using a propositional theory. The purpose of this step is to specify the network polynomial using a propositional theory. To illustrate how a multi-linear function can be specified using a propositional theory, consider the following function f = ac + abc + c over real-valued variables a, b, c. The basic idea is to specify this multi-linear function using a propositional theory that has exactly three models, where each model encodes one of the terms in the function. Specifically, suppose we have the 2 A similar approach has been recently proposed in [20], which calls for encoding Bayesian networks into CNFs, and reducing probabilistic inference to weighted model counting on the generated CNFs. The approach is similar in two senses. First, the weighted model counting algorithm applied in [20] is powerful enough to factor the CNF as suggested by Step 2 below see [21]. Second, the factored logical form we generate from the CNF in Step 2 is tractable enough to allow weighted model counting in time linear in the form size [22,23].

12 M. Chavira et al. / Internat. J. Approx. Reason. 42 (2006) Fig. 6. Factoring multi-linear functions into arithmetic circuits. Boolean variables V a, V b, V c. Then the propositional theory D f ¼ðV a _:V b Þ^V c encodes the multi-linear function f as follows: Model V a V b V c Encoded term r 1 true false true ac r 2 true true true abc r 3 false false true c That is, model r encodes term t since r(v j )=true precisely when term t contains the realvalued variable j. This method of specifying network polynomials allows one to easily capture local structure; that is, to declare certain information about values of polynomial variables. For example, if we know that parameter a = 0, then we can exclude all terms that contain a by conjoining :V a with our encoding. Step 2: Factoring the propositional encoding. If we view the conversion of a network polynomial into an arithmetic circuit as a factoring process, then the purpose of this second step is to accomplish a similar task but at the logical level. Instead of starting with a polynomial (set of terms), we start with a propositional theory (set of models). And instead of building an arithmetic circuit, we build a Boolean circuit that satisfies certain properties. Specifically, the circuit must be in negation normal form (NNF): a rooted DAG where leaves are labeled with literals, and where internal nodes are labeled with conjunctions or disjunctions; see Fig. 6. The NNF must satisfy three properties: (1) conjuncts cannot share variables (decomposability), (2) disjuncts must be logically exclusive (determinism), and (3) disjuncts must be over the same variables (smoothness). The NNF in Fig. 6 satisfies the above properties, and encodes the multi-linear function shown in the same figure. In our experimental results, we use a second generation compiler for converting CNFs to NNFs that are decomposable, deterministic and smooth (smooth d-dnnf) [24]. Step 3: Extracting an arithmetic circuit. The purpose of this last step is to extract an arithmetic circuit for the polynomial encoded by an NNF. If D f is an NNF that encodes a network polynomial f, and if D f is a smooth d-dnnf, then an arithmetic circuit for the

13 16 M. Chavira et al. / Internat. J. Approx. Reason. 42 (2006) 4 20 polynomial f can be obtained easily. First, replace and nodes in D f by multiplications; then replace or nodes by additions; and finally, replace each leaf node labeled with V x by x and each node labeled with :V x by 1. The resulting arithmetic circuit is then guaranteed to correspond to polynomial f [19]. Fig. 6 depicts an NNF and its corresponding arithmetic circuit. Note that the generated arithmetic circuit is no larger than the NNF. Hence, if we attempt to minimize the size of NNF, we are also attempting to minimize the size of generated arithmetic circuit Encoding Primula s networks The encoding step described above is semantic; that is, it describes the theory D f which encodes a multi-linear function by describing its models. As mentioned earlier, the Primula system generates propositional instances of relational models in the form of classical Bayesian networks. We now turn to the question of how to syntactically represent in CNF the multi-linear function of a network so generated. We start with the baseline encoding defined in [19], which applies to any Bayesian network. The CNF has one Boolean variable I k for each indicator variable k, and one Boolean variable P h for each parameter variable h. CNF clauses fall into three sets. First, for each network variable X with domain x 1,x 2,...,x n, we have: Indicator clauses : I kx1 _ I kx2 I kxn :I kxi _:I kxj for i < j For example, variable B from Fig. 1 generates the following clauses: I kb _ I kb ; :I kb _:I kb ð1þ These clauses ensure that exactly one indicator variable for B appears in every term of the multi-linear function. The second two sets of clauses correspond to network parameters. In particular, for each parameter h xnjx 1 ;x 2 ;...;x n 1, we have: IP clause : I kx1 ^ I kx2 ^^I kxn ) P hxnjx1 ;x 2 ;...;x n 1 PI clauses : P hxnjx1 ;x 2 ;...;x n 1 ) I kxi for each i For example, parameter h bja in Fig. 1 generates the following clauses: I ka ^ I kb ) P hbja ; P hbja ) I ka ; P hbja ) I kb ð2þ These clauses ensure that h bja appears in a term iff the k a and k b appear. The encoding as discussed does not capture information about parameter values (local structure). However, it is quite easy to encode information about determinism within this encoding. Consider again Fig. 1 and the parameter h bja = 0, which generates the clauses in Eq. 2. Given that this parameter is known to be 0, all multi-linear terms that contain this parameter must vanish. Therefore, we can suppress the generation of a Boolean variable for this parameter, and then replace the above clauses by the single clause: :I ka _:I kb. This clause has the effect of eliminating all CNF models which correspond to vanishing terms, those containing the parameter h bja.

14 M. Chavira et al. / Internat. J. Approx. Reason. 42 (2006) To this basic encoding we apply some optimizations: Primula generated networks contain only binary variables. Therefore, instead of using one propositional variable for each evidence indicator k x, which would be needed in general, we use one propositional variable I X for each Bayesian network variable X, where the positive literal I X represents indicator k x, and the negative literal :I X represents indicator k x. Not only does this cut the number of indicator variables by half, but it also relieves the need for indicator clauses. For example, without the enhancement, variable B in Fig. 1 generates Boolean variables I kb and I kb and the two clauses in Eq. (1). With the optimization, B generates only a single Boolean variable I B and no clauses. This optimization requires a corresponding modification to the decoding step as indicated below. Another enhancement results from the observation that the Boolean indicators and parameters corresponding to the same state of a network root variable are logically equivalent, making it possible to delete the parameter variables and the corresponding IP and PI clauses, which establish the equivalence. The Boolean indicator thus represents both an indicator and a parameter. For example, without the enhancement, parameter h a in Fig. 1 generates one Boolean variable P ha and two clauses, I A ) P ha and P ha ) I A. With the enhancement, the variable and clauses are omitted. This optimization requires a corresponding modification to the decoding step as indicated below. Variables and clauses generated by parameters equal to 1 are redundant and therefore omitted. Applying these enhancements allows us to create the CNF as follows. For each network variable X, we create propositional variable I X.IfX is not a root, then we perform three more steps. (1) For each network parameter h xju not equal to 0 or 1, create a propositional variable P hxju. (2) For each parameter h xju1 ;u 2...;u n equal to 0, create clause :L U 1 _:L U 2 _..._:L U n _:L X, where L U i is a literal over variable I U i whose sign is the same as u i, and similarly for L X with respect to x. (3) For each parameter h xju1 ;u 2 ;...;u n not equal to 0 and not equal to 1, create clauses, L U 1 ^ L U 2 ^^L U n ^ L x ) P ; P hxju1 ;...;un h ) L xju1 ;...;un U 1 ; P ) L hxju1 ;...;un U 2 ;...; P ) L hxju1 ;...;un U n ; P ) L hxju1 ;...;un X, where L U i and L X are as defined earlier. As an example, the CPT for variable B in Fig. 1 generates the following clauses: First CPT row: :I A _:I B. Third CPT row: :I A ^ I B ) P hbja ; P hbja ):I A ; P hbja ) I B. Fourth CPT row: :I A ^:I B ) P hbja ; P hbja ):I A ; P hbja ):I B. Because Primula generates networks with binary variables and nodes with at most three parents, this encoding leads to a CNF whose size is linear in the number of network variables. Table 1 depicts the size of CNF encodings for the relational models with which we experimented. The special encoding used above calls for a slightly different decoding scheme for transforming a smooth d-dnnf into an arithmetic circuit. Specifically, if X is not a root, then literals I X and :I X are replaced with evidence indicators k x and k x, respectively. If X is a root, then literals I X and :I X are replaced with k x h x and k x h x, respectively. Moreover,

15 18 M. Chavira et al. / Internat. J. Approx. Reason. 42 (2006) 4 20 literals P hx;u and :P hx;u are replaced by h x;u and 1, respectively. Finally, conjunctions and disjunctions are replaced by multiplications and additions. We close this section by pointing the reader to [25], which discusses more recent and sophisticated encodings to handle Bayesian networks with context-specific-independence [26], multi-valued variables, large CPTs, and lesser amounts of determinism. 5. Experimental results We ran our experiments on a 1.6 GHz Pentium M with 2 GB of RAM using a system available for download at Table 1 lists for each relational model a number of instances, and for each instance a number of measurements. First is the size and connectivity of the Bayesian network that Primula generated. Primula generates networks in formats acceptable by general purpose tools such as Hugin and Netica, but exact inference in these tools cannot handle most of these networks. Next is the number of variables and clauses in the CNF encodings. Clauses have at most five literals since the networks have at most three parents per node. Table 1 shows additional findings. First, the table shows the size of the compiled arithmetic circuit in terms of both number of nodes and edges (count and log base 2). We also show the time it takes to evaluate and differentiate the circuit, averaged over 31 different randomly generated evidence sets. By evaluating and differentiating the circuit, one obtains marginals over all network families, in addition to other probabilities discussed in [2]. The main points to observe are the efficiency of online inference on compiled circuits and the size of these circuits compared to the size and connectivity of the Bayesian networks. Table 1 also shows the time for jointree propagation using the SamIam inference engine ( on instances whose cluster size was manageable. One can see the big difference between online inference using the compiled AC and corresponding jointrees. Table 1 finally shows the compile time to generate the arithmetic circuits. The compile times range from less than a minute to about 60 min for the largest model. Yet the time for online inference ranges from milliseconds to about 13 s for these models. This clearly shows the benefit of offline compilation in this case, whose time can be amortized over online queries. Friends and smokers produces networks with particularly high connectivity. We mentioned previously that logical constraints in this model give rise to grounded Bayesian networks with evidence that applies to all queries. One might hope that classical pruning techniques such as deleting leaf nodes not part of the query or evidence [27] and deleting edges exiting evidence nodes [28] might reduce the connectivity of these networks, making them accessible to classical inference algorithms. This possibility is not realized though since all of the evidence occur on leaf nodes. However, we can use the method of [29] to place this evidence into the CNF encoding and compile with the evidence. In particular, if we know that network variable A corresponds to a logical constraint that must be true, then we simply add a unit clause k a to the CNF encoding. In fact, injecting these unit clauses into the CNF encoding prior to compilation has a critical effect on both compilation time and AC size, as most of these networks could not be compiled otherwise.

16 M. Chavira et al. / Internat. J. Approx. Reason. 42 (2006) Conclusion We described in this paper an inference system for relational Bayesian networks as defined by Primula. The proposed inference approach is based on compiling propositional instances of these models into arithmetic circuits. The approach exploits determinism in relational models, allowing us to reason efficiently with some relational models whose Primula-generated propositional instances contain thousands of variables, and whose jointrees contain hundreds of variables. The described system appears to significantly expand the scale of Primula-based relational models that can be handled efficiently by exact inference algorithms. It is also equally applicable and effective to any Bayesian network that exhibits similar properties (e.g., determinism), regardless of whether it is synthesized from a relational model. Acknowledgment This work has been partially supported by NSF grant IIS and MURI grant N References [1] F.V. Jensen, S. Lauritzen, K. Olesen, Bayesian updating in recursive graphical models by local computation, Computational Statistics Quarterly 4 (1990) [2] A. Darwiche, A differential approach to inference in Bayesian networks, Journal of the ACM 50 (3) (2003) [3] F. Jensen, S.K. Andersen, Approximations in Bayesian belief universes for knowledge based systems, in: Proceedings of the Conference on Uncertainty in Artificial Intelligence (UAI), Cambridge, MA, 1990, pp [4] J. Park, A. Darwiche, A differential semantics for jointree algorithms, Artificial Intelligence 156 (2004) [5] J.S. Breese, R.P. Goldman, M.P. Wellman, Introduction to the special section on knowledge-based construction of probabilistic decision models, IEEE Transactions on Systems, Man, and Cybernetics 24 (11) (1994) [6] T. Sato, A statistical learning method for logic programs with distribution semantics, in: Proceedings of the International Conference on Logic Programming (ICLP), 1995, pp [7] S. Muggleton, Stochastic logic programs, in: L. de Raedt (Ed.), Advances in Inductive Logic Programming, IOS Press, 1996, pp [8] K. Kersting, L. de Raedt, Towards combining inductive logic programming and Bayesian networks, in: Proceedings of the International Conference on Inductive Logic Programming (ILP), Springer Lecture Notes in AI 2157, [9] K.B. Laskey, S.M. Mahoney, Network fragments: representing knowledge for constructing probabilistic models, in: Proceedings of the Conference on Uncertainty in Artificial Intelligence (UAI), San Francisco, CA, 1997, pp [10] D. Koller, A. Pfeffer, Probabilistic frame-based systems, in: Proceedings of the National Conference on Artificial Intelligence (AAAI), 1998, pp [11] N. Friedman, L. Getoor, D. Koller, A. Pfeffer, Learning probabilistic relational models, in: Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI), [12] M. Jaeger, Relational Bayesian networks, in: D. Geiger, P.P. Shenoy (Eds.), Proceedings of the Conference of Uncertainty in Artificial Intelligence (UAI), Providence, USA, 1997, pp [13] H. Pasula, S. Russell, Approximate inference for first-order probabilistic languages, in: Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI), 2001, pp [14] M. Richardson, P. Domingos, Markov logic networks, Special issue of the Machine Learning Journal on Statistical Relational Learning and Multi-Relational Data Mining, in press.

17 20 M. Chavira et al. / Internat. J. Approx. Reason. 42 (2006) 4 20 [15] M. Jaeger, Complex probabilistic modeling with recursive relational Bayesian networks, Annals of Mathematics and Artificial Intelligence 32 (2001) [16] M. Jaeger, On the complexity of inference about probabilistic relational models, Artificial Intelligence 117 (2000) [17] D. Poole, First-order probabilistic inference, in: Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI), [18] R. de Salvo Braz, E. Amir, D. Roth, Lifted first-order probabilistic inference, in: Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI), 2005, pp [19] A. Darwiche, A logical approach to factoring belief networks, in: Proceedings of the International Conference on Principles of Knowledge Representation and Reasoning (KR), 2002, pp [20] T. Sang, P. Beame, H. Kautz, Solving Bayesian networks by weighted model countingproceedings of the National Conference on Artificial Intelligence (AAAI), vol. 1, AAAI Press, 2005, pp [21] J. Huang, A. Darwiche, DPLL with a trace: from sat to knowledge compilation, in: Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI), 2005, pp [22] A. Darwiche, P. Marquis, A knowledge compilation map, Journal of Artificial Intelligence Research 17 (2002) [23] A. Darwiche, P. Marquis, Compiling propositional weighted bases, Artificial Intelligence 157 (1 2) (2004) [24] A. Darwiche, New advances in compiling CNF to decomposable negational normal form, in: Proceedings of the European Conference on Artificial Intelligence (ECAI), 2004, pp [25] M. Chavira, A. Darwiche, Compiling Bayesian networks with local structure, in: Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI), 2005, pp [26] C. Boutilier, N. Friedman, M. Goldszmidt, D. Koller, Context-specific independence in Bayesian networks, in: Proceedings of the Conference on Uncertainty in Artificial Intelligence (UAI), 1996, pp [27] R.D. Shachter, Evaluating influence diagrams, Operations Research 34 (6) (1986) [28] S. Ross, Evidence absorption and propagation through evidence reversals, in: Proceedings of the Conference on Uncertainty in Artificial Intelligence (UAI), Elsevier Science Publishing Company, Inc., New York, NY, [29] M. Chavira, D. Allen, A. Darwiche, Exploiting evidence in probabilistic inference, in: Proceedings of the Conference on Uncertainty in Artificial Intelligence (UAI), 2005, pp

Compiling Relational Bayesian Networks for Exact Inference

Compiling Relational Bayesian Networks for Exact Inference Compiling Relational Bayesian Networks for Exact Inference Mark Chavira, Adnan Darwiche Computer Science Department, UCLA, Los Angeles, CA 90095 Manfred Jaeger Institut for Datalogi, Aalborg Universitet,

More information

Compiling Probabilistic Graphical Models using Sentential Decision Diagrams

Compiling Probabilistic Graphical Models using Sentential Decision Diagrams Compiling Probabilistic Graphical Models using Sentential Decision Diagrams Arthur Choi, Doga Kisa, and Adnan Darwiche University of California, Los Angeles, California 90095, USA {aychoi,doga,darwiche}@cs.ucla.edu

More information

Modeling and Reasoning with Bayesian Networks. Adnan Darwiche University of California Los Angeles, CA

Modeling and Reasoning with Bayesian Networks. Adnan Darwiche University of California Los Angeles, CA Modeling and Reasoning with Bayesian Networks Adnan Darwiche University of California Los Angeles, CA darwiche@cs.ucla.edu June 24, 2008 Contents Preface 1 1 Introduction 1 1.1 Automated Reasoning........................

More information

Learning Directed Probabilistic Logical Models using Ordering-search

Learning Directed Probabilistic Logical Models using Ordering-search Learning Directed Probabilistic Logical Models using Ordering-search Daan Fierens, Jan Ramon, Maurice Bruynooghe, and Hendrik Blockeel K.U.Leuven, Dept. of Computer Science, Celestijnenlaan 200A, 3001

More information

New Compilation Languages Based on Structured Decomposability

New Compilation Languages Based on Structured Decomposability Proceedings of the Twenty-Third AAAI Conference on Artificial Intelligence (2008) New Compilation Languages Based on Structured Decomposability Knot Pipatsrisawat and Adnan Darwiche Computer Science Department

More information

AIC-PRAiSE User Guide

AIC-PRAiSE User Guide AIC-PRAiSE User Guide Rodrigo de Salvo Braz Artificial Intelligence Center SRI International June 2, 2017 S 1 Contents 1 Introduction 3 2 Installing PRAiSE 3 3 Using PRAiSE 3 3.1 Higher-Order Graphical

More information

A Compiler for Deterministic, Decomposable Negation Normal Form

A Compiler for Deterministic, Decomposable Negation Normal Form From: AAAI-02 Proceedings. Copyright 2002, AAAI (www.aaai.org). All rights reserved. A Compiler for Deterministic, Decomposable Negation Normal Form Adnan Darwiche Computer Science Department University

More information

Markov Logic: Representation

Markov Logic: Representation Markov Logic: Representation Overview Statistical relational learning Markov logic Basic inference Basic learning Statistical Relational Learning Goals: Combine (subsets of) logic and probability into

More information

Compiling Bayesian Networks by Symbolic Probability Calculation Based on Zero-suppressed BDDs

Compiling Bayesian Networks by Symbolic Probability Calculation Based on Zero-suppressed BDDs Compiling Bayesian Networks by Symbolic Probability Calculation Based on Zero-suppressed BDDs Shin-ichi Minato Div. of Computer Science Hokkaido University Sapporo 6 814, Japan Ken Satoh National Institute

More information

Utilizing Device Behavior in Structure-Based Diagnosis

Utilizing Device Behavior in Structure-Based Diagnosis Utilizing Device Behavior in Structure-Based Diagnosis Adnan Darwiche Cognitive Systems Laboratory Department of Computer Science University of California Los Angeles, CA 90024 darwiche @cs. ucla. edu

More information

The Encoding Complexity of Network Coding

The Encoding Complexity of Network Coding The Encoding Complexity of Network Coding Michael Langberg Alexander Sprintson Jehoshua Bruck California Institute of Technology Email: mikel,spalex,bruck @caltech.edu Abstract In the multicast network

More information

Binary Decision Diagrams

Binary Decision Diagrams Logic and roof Hilary 2016 James Worrell Binary Decision Diagrams A propositional formula is determined up to logical equivalence by its truth table. If the formula has n variables then its truth table

More information

Conditional PSDDs: Modeling and Learning with Modular Knowledge

Conditional PSDDs: Modeling and Learning with Modular Knowledge Conditional PSDDs: Modeling and Learning with Modular Knowledge Yujia Shen and Arthur Choi and Adnan Darwiche Computer Science Department University of California, Los Angeles {yujias,aychoi,darwiche}@csuclaedu

More information

Probabilistic Partial Evaluation: Exploiting rule structure in probabilistic inference

Probabilistic Partial Evaluation: Exploiting rule structure in probabilistic inference Probabilistic Partial Evaluation: Exploiting rule structure in probabilistic inference David Poole Department of Computer Science University of British Columbia 2366 Main Mall, Vancouver, B.C., Canada

More information

Knowledge Compilation Properties of Tree-of-BDDs

Knowledge Compilation Properties of Tree-of-BDDs Knowledge Compilation Properties of Tree-of-BDDs Sathiamoorthy Subbarayan IT University of Copenhagen, Denmark sathi@itu.dk Lucas Bordeaux and Youssef Hamadi Microsoft Research, Cambridge, UK lucasb,youssefh@microsoft.com

More information

Av. Prof. Mello Moraes, 2231, , São Paulo, SP - Brazil

Av. Prof. Mello Moraes, 2231, , São Paulo, SP - Brazil " Generalizing Variable Elimination in Bayesian Networks FABIO GAGLIARDI COZMAN Escola Politécnica, University of São Paulo Av Prof Mello Moraes, 31, 05508-900, São Paulo, SP - Brazil fgcozman@uspbr Abstract

More information

Boolean Representations and Combinatorial Equivalence

Boolean Representations and Combinatorial Equivalence Chapter 2 Boolean Representations and Combinatorial Equivalence This chapter introduces different representations of Boolean functions. It then discusses the applications of these representations for proving

More information

Horn Formulae. CS124 Course Notes 8 Spring 2018

Horn Formulae. CS124 Course Notes 8 Spring 2018 CS124 Course Notes 8 Spring 2018 In today s lecture we will be looking a bit more closely at the Greedy approach to designing algorithms. As we will see, sometimes it works, and sometimes even when it

More information

SDD Advanced-User Manual Version 1.1

SDD Advanced-User Manual Version 1.1 SDD Advanced-User Manual Version 1.1 Arthur Choi and Adnan Darwiche Automated Reasoning Group Computer Science Department University of California, Los Angeles Email: sdd@cs.ucla.edu Download: http://reasoning.cs.ucla.edu/sdd

More information

Abstract. 2 Background 2.1 Belief Networks. 1 Introduction

Abstract. 2 Background 2.1 Belief Networks. 1 Introduction Probabilistic Partial Evaluation: Exploiting rule structure in probabilistic inference* David Poole Department of Computer Science University of British Columbia 2366 Main Mall, Vancouver, B.C., Canada

More information

Relax, Compensate and then Recover

Relax, Compensate and then Recover Relax, Compensate and then Recover rthur Choi and dnan Darwiche Computer Science Department, University of California, Los ngeles, US {aychoi,darwiche}@cs.ucla.edu bstract. We present in this paper a framework

More information

Planning and Reinforcement Learning through Approximate Inference and Aggregate Simulation

Planning and Reinforcement Learning through Approximate Inference and Aggregate Simulation Planning and Reinforcement Learning through Approximate Inference and Aggregate Simulation Hao Cui Department of Computer Science Tufts University Medford, MA 02155, USA hao.cui@tufts.edu Roni Khardon

More information

Massively Parallel Seesaw Search for MAX-SAT

Massively Parallel Seesaw Search for MAX-SAT Massively Parallel Seesaw Search for MAX-SAT Harshad Paradkar Rochester Institute of Technology hp7212@rit.edu Prof. Alan Kaminsky (Advisor) Rochester Institute of Technology ark@cs.rit.edu Abstract The

More information

New Worst-Case Upper Bound for #2-SAT and #3-SAT with the Number of Clauses as the Parameter

New Worst-Case Upper Bound for #2-SAT and #3-SAT with the Number of Clauses as the Parameter Proceedings of the Twenty-Fourth AAAI Conference on Artificial Intelligence (AAAI-10) New Worst-Case Upper Bound for #2-SAT and #3-SAT with the Number of Clauses as the Parameter Junping Zhou 1,2, Minghao

More information

The Resolution Algorithm

The Resolution Algorithm The Resolution Algorithm Introduction In this lecture we introduce the Resolution algorithm for solving instances of the NP-complete CNF- SAT decision problem. Although the algorithm does not run in polynomial

More information

Integrating Probabilistic Reasoning with Constraint Satisfaction

Integrating Probabilistic Reasoning with Constraint Satisfaction Integrating Probabilistic Reasoning with Constraint Satisfaction IJCAI Tutorial #7 Instructor: Eric I. Hsu July 17, 2011 http://www.cs.toronto.edu/~eihsu/tutorial7 Getting Started Discursive Remarks. Organizational

More information

A Knowledge Compilation Technique for ALC Tboxes

A Knowledge Compilation Technique for ALC Tboxes A Knowledge Compilation Technique for ALC Tboxes Ulrich Furbach and Heiko Günther and Claudia Obermaier University of Koblenz Abstract Knowledge compilation is a common technique for propositional logic

More information

Dynamic J ointrees. Figure 1: Belief networks and respective jointrees.

Dynamic J ointrees. Figure 1: Belief networks and respective jointrees. 97 Dynamic J ointrees Adnan Darwiche Department of Mathematics American University of Beirut PO Box 11-236 Beirut, Lebanon darwiche@aub. edu.lb Abstract It is well known that one can ignore parts of a

More information

New Advances in Inference by Recursive Conditioning

New Advances in Inference by Recursive Conditioning New Advances in Inference by Recursive Conditioning David Allen and Adnan Darwiche Computer Science Department University of California Los Angeles, CA 90095 {dlallen,darwiche}@cs.ucla.edu Abstract Recursive

More information

Escola Politécnica, University of São Paulo Av. Prof. Mello Moraes, 2231, , São Paulo, SP - Brazil

Escola Politécnica, University of São Paulo Av. Prof. Mello Moraes, 2231, , São Paulo, SP - Brazil Generalizing Variable Elimination in Bayesian Networks FABIO GAGLIARDI COZMAN Escola Politécnica, University of São Paulo Av. Prof. Mello Moraes, 2231, 05508-900, São Paulo, SP - Brazil fgcozman@usp.br

More information

Matthew Richardson Microsoft Research One Microsoft Way Redmond, WA

Matthew Richardson Microsoft Research One Microsoft Way Redmond, WA Speeding up Inference in Statistical Relational Learning by Clustering Similar Query Literals Lilyana Mihalkova 1 Department of Computer Sciences The University of Texas at Austin Austin, TX 78712 lilyanam@cs.utexas.edu

More information

Exploiting Local and Repeated Structure in Dynamic Bayesian Networks

Exploiting Local and Repeated Structure in Dynamic Bayesian Networks Exploiting Local and Repeated Structure in Dynamic Bayesian Networks Jonas Vlasselaer, Wannes Meert, Guy Van den Broeck, Luc De Raedt Departement of Computer Science, Katholieke Universiteit Leuven Celestijnenlaan

More information

Deductive Methods, Bounded Model Checking

Deductive Methods, Bounded Model Checking Deductive Methods, Bounded Model Checking http://d3s.mff.cuni.cz Pavel Parízek CHARLES UNIVERSITY IN PRAGUE faculty of mathematics and physics Deductive methods Pavel Parízek Deductive Methods, Bounded

More information

Core Membership Computation for Succinct Representations of Coalitional Games

Core Membership Computation for Succinct Representations of Coalitional Games Core Membership Computation for Succinct Representations of Coalitional Games Xi Alice Gao May 11, 2009 Abstract In this paper, I compare and contrast two formal results on the computational complexity

More information

PROBABILISTIC GRAPHICAL MODELS SPECIFIED BY PROBABILISTIC LOGIC PROGRAMS: SEMANTICS AND COMPLEXITY

PROBABILISTIC GRAPHICAL MODELS SPECIFIED BY PROBABILISTIC LOGIC PROGRAMS: SEMANTICS AND COMPLEXITY PROBABILISTIC GRAPHICAL MODELS SPECIFIED BY PROBABILISTIC LOGIC PROGRAMS: SEMANTICS AND COMPLEXITY Fabio G. Cozman and Denis D. Mauá Universidade de São Paulo, Brazil September 8, 2016 1 / 23 Many languages

More information

Ch9: Exact Inference: Variable Elimination. Shimi Salant, Barak Sternberg

Ch9: Exact Inference: Variable Elimination. Shimi Salant, Barak Sternberg Ch9: Exact Inference: Variable Elimination Shimi Salant Barak Sternberg Part 1 Reminder introduction (1/3) We saw two ways to represent (finite discrete) distributions via graphical data structures: Bayesian

More information

Structured Bayesian Networks: From Inference to Learning with Routes

Structured Bayesian Networks: From Inference to Learning with Routes Structured Bayesian Networks: From Inference to Learning with Routes Yujia Shen and Anchal Goyanka and Adnan Darwiche and Arthur Choi Computer Science Department University of California, Los Angeles {yujias,anchal,darwiche,aychoi}@csuclaedu

More information

A Framework for Securing Databases from Intrusion Threats

A Framework for Securing Databases from Intrusion Threats A Framework for Securing Databases from Intrusion Threats R. Prince Jeyaseelan James Department of Computer Applications, Valliammai Engineering College Affiliated to Anna University, Chennai, India Email:

More information

Multi Domain Logic and its Applications to SAT

Multi Domain Logic and its Applications to SAT Multi Domain Logic and its Applications to SAT Tudor Jebelean RISC Linz, Austria Tudor.Jebelean@risc.uni-linz.ac.at Gábor Kusper Eszterházy Károly College gkusper@aries.ektf.hu Abstract We describe a new

More information

Uncertain Data Models

Uncertain Data Models Uncertain Data Models Christoph Koch EPFL Dan Olteanu University of Oxford SYNOMYMS data models for incomplete information, probabilistic data models, representation systems DEFINITION An uncertain data

More information

Chapter 3: Propositional Languages

Chapter 3: Propositional Languages Chapter 3: Propositional Languages We define here a general notion of a propositional language. We show how to obtain, as specific cases, various languages for propositional classical logic and some non-classical

More information

COMP4418 Knowledge Representation and Reasoning

COMP4418 Knowledge Representation and Reasoning COMP4418 Knowledge Representation and Reasoning Week 3 Practical Reasoning David Rajaratnam Click to edit Present s Name Practical Reasoning - My Interests Cognitive Robotics. Connect high level cognition

More information

Data Analytics and Boolean Algebras

Data Analytics and Boolean Algebras Data Analytics and Boolean Algebras Hans van Thiel November 28, 2012 c Muitovar 2012 KvK Amsterdam 34350608 Passeerdersstraat 76 1016 XZ Amsterdam The Netherlands T: + 31 20 6247137 E: hthiel@muitovar.com

More information

A New Approach For Convert Multiply-Connected Trees in Bayesian networks

A New Approach For Convert Multiply-Connected Trees in Bayesian networks A New Approach For Convert Multiply-Connected Trees in Bayesian networks 1 Hussein Baloochian, Alireza khantimoory, 2 Saeed Balochian 1 Islamic Azad university branch of zanjan 2 Islamic Azad university

More information

Mini-Buckets: A General Scheme for Generating Approximations in Automated Reasoning

Mini-Buckets: A General Scheme for Generating Approximations in Automated Reasoning Mini-Buckets: A General Scheme for Generating Approximations in Automated Reasoning Rina Dechter* Department of Information and Computer Science University of California, Irvine dechter@ics. uci. edu Abstract

More information

Small Formulas for Large Programs: On-line Constraint Simplification In Scalable Static Analysis

Small Formulas for Large Programs: On-line Constraint Simplification In Scalable Static Analysis Small Formulas for Large Programs: On-line Constraint Simplification In Scalable Static Analysis Isil Dillig, Thomas Dillig, Alex Aiken Stanford University Scalability and Formula Size Many program analysis

More information

International Journal of Approximate Reasoning

International Journal of Approximate Reasoning International Journal of Approximate Reasoning 52 (2) 49 62 Contents lists available at ScienceDirect International Journal of Approximate Reasoning journal homepage: www.elsevier.com/locate/ijar Approximate

More information

Decision Procedures. An Algorithmic Point of View. Decision Procedures for Propositional Logic. D. Kroening O. Strichman.

Decision Procedures. An Algorithmic Point of View. Decision Procedures for Propositional Logic. D. Kroening O. Strichman. Decision Procedures An Algorithmic Point of View Decision Procedures for Propositional Logic D. Kroening O. Strichman ETH/Technion Version 1.0, 2007 Part I Decision Procedures for Propositional Logic Outline

More information

Discrete Optimization. Lecture Notes 2

Discrete Optimization. Lecture Notes 2 Discrete Optimization. Lecture Notes 2 Disjunctive Constraints Defining variables and formulating linear constraints can be straightforward or more sophisticated, depending on the problem structure. The

More information

ABSTRACT 1. INTRODUCTION

ABSTRACT 1. INTRODUCTION ABSTRACT A Framework for Multi-Agent Multimedia Indexing Bernard Merialdo Multimedia Communications Department Institut Eurecom BP 193, 06904 Sophia-Antipolis, France merialdo@eurecom.fr March 31st, 1995

More information

Some Hardness Proofs

Some Hardness Proofs Some Hardness Proofs Magnus Lie Hetland January 2011 This is a very brief overview of some well-known hard (NP Hard and NP complete) problems, and the main ideas behind their hardness proofs. The document

More information

PROPOSITIONAL LOGIC (2)

PROPOSITIONAL LOGIC (2) PROPOSITIONAL LOGIC (2) based on Huth & Ruan Logic in Computer Science: Modelling and Reasoning about Systems Cambridge University Press, 2004 Russell & Norvig Artificial Intelligence: A Modern Approach

More information

To prove something about all Boolean expressions, we will need the following induction principle: Axiom 7.1 (Induction over Boolean expressions):

To prove something about all Boolean expressions, we will need the following induction principle: Axiom 7.1 (Induction over Boolean expressions): CS 70 Discrete Mathematics for CS Fall 2003 Wagner Lecture 7 This lecture returns to the topic of propositional logic. Whereas in Lecture 1 we studied this topic as a way of understanding proper reasoning

More information

1 Definition of Reduction

1 Definition of Reduction 1 Definition of Reduction Problem A is reducible, or more technically Turing reducible, to problem B, denoted A B if there a main program M to solve problem A that lacks only a procedure to solve problem

More information

Hybrid Feature Selection for Modeling Intrusion Detection Systems

Hybrid Feature Selection for Modeling Intrusion Detection Systems Hybrid Feature Selection for Modeling Intrusion Detection Systems Srilatha Chebrolu, Ajith Abraham and Johnson P Thomas Department of Computer Science, Oklahoma State University, USA ajith.abraham@ieee.org,

More information

Multi-relational Decision Tree Induction

Multi-relational Decision Tree Induction Multi-relational Decision Tree Induction Arno J. Knobbe 1,2, Arno Siebes 2, Daniël van der Wallen 1 1 Syllogic B.V., Hoefseweg 1, 3821 AE, Amersfoort, The Netherlands, {a.knobbe, d.van.der.wallen}@syllogic.com

More information

Answer Sets and the Language of Answer Set Programming. Vladimir Lifschitz

Answer Sets and the Language of Answer Set Programming. Vladimir Lifschitz Answer Sets and the Language of Answer Set Programming Vladimir Lifschitz Answer set programming is a declarative programming paradigm based on the answer set semantics of logic programs. This introductory

More information

Exam Topics. Search in Discrete State Spaces. What is intelligence? Adversarial Search. Which Algorithm? 6/1/2012

Exam Topics. Search in Discrete State Spaces. What is intelligence? Adversarial Search. Which Algorithm? 6/1/2012 Exam Topics Artificial Intelligence Recap & Expectation Maximization CSE 473 Dan Weld BFS, DFS, UCS, A* (tree and graph) Completeness and Optimality Heuristics: admissibility and consistency CSPs Constraint

More information

Memory-Efficient Inference in Relational Domains

Memory-Efficient Inference in Relational Domains Memory-Efficient Inference in Relational Domains Parag Singla Pedro Domingos Department of Computer Science and Engineering University of Washington Seattle, WA 98195-2350, U.S.A. {parag, pedrod}@cs.washington.edu

More information

Clone: Solving Weighted Max-SAT in a Reduced Search Space

Clone: Solving Weighted Max-SAT in a Reduced Search Space Clone: Solving Weighted Max-SAT in a Reduced Search Space Knot Pipatsrisawat and Adnan Darwiche {thammakn,darwiche}@cs.ucla.edu Computer Science Department University of California, Los Angeles Abstract.

More information

USING QBF SOLVERS TO SOLVE GAMES AND PUZZLES. Zhihe Shen. Advisor: Howard Straubing

USING QBF SOLVERS TO SOLVE GAMES AND PUZZLES. Zhihe Shen. Advisor: Howard Straubing Boston College Computer Science Senior Thesis USING QBF SOLVERS TO SOLVE GAMES AND PUZZLES Zhihe Shen Advisor: Howard Straubing Abstract There are multiple types of games, such as board games and card

More information

Tractable Probabilistic Knowledge Bases with Existence Uncertainty

Tractable Probabilistic Knowledge Bases with Existence Uncertainty Tractable Probabilistic Knowledge Bases with Existence Uncertainty W. Austin Webb and Pedro Domingos {webb,pedrod@cs.washington.edu Department of Computer Science and Engineering, University of Washington,

More information

Loopy Belief Propagation

Loopy Belief Propagation Loopy Belief Propagation Research Exam Kristin Branson September 29, 2003 Loopy Belief Propagation p.1/73 Problem Formalization Reasoning about any real-world problem requires assumptions about the structure

More information

Normal Forms for Boolean Expressions

Normal Forms for Boolean Expressions Normal Forms for Boolean Expressions A NORMAL FORM defines a class expressions s.t. a. Satisfy certain structural properties b. Are usually universal: able to express every boolean function 1. Disjunctive

More information

Example: Map coloring

Example: Map coloring Today s s lecture Local Search Lecture 7: Search - 6 Heuristic Repair CSP and 3-SAT Solving CSPs using Systematic Search. Victor Lesser CMPSCI 683 Fall 2004 The relationship between problem structure and

More information

8.1 Polynomial-Time Reductions

8.1 Polynomial-Time Reductions 8.1 Polynomial-Time Reductions Classify Problems According to Computational Requirements Q. Which problems will we be able to solve in practice? A working definition. Those with polynomial-time algorithms.

More information

Pbmodels Software to Compute Stable Models by Pseudoboolean Solvers

Pbmodels Software to Compute Stable Models by Pseudoboolean Solvers Pbmodels Software to Compute Stable Models by Pseudoboolean Solvers Lengning Liu and Mirosław Truszczyński Department of Computer Science, University of Kentucky, Lexington, KY 40506-0046, USA Abstract.

More information

On Resolution Proofs for Combinational Equivalence Checking

On Resolution Proofs for Combinational Equivalence Checking On Resolution Proofs for Combinational Equivalence Checking Satrajit Chatterjee Alan Mishchenko Robert Brayton Department of EECS U. C. Berkeley {satrajit, alanmi, brayton}@eecs.berkeley.edu Andreas Kuehlmann

More information

Lecture 1 Contracts : Principles of Imperative Computation (Fall 2018) Frank Pfenning

Lecture 1 Contracts : Principles of Imperative Computation (Fall 2018) Frank Pfenning Lecture 1 Contracts 15-122: Principles of Imperative Computation (Fall 2018) Frank Pfenning In these notes we review contracts, which we use to collectively denote function contracts, loop invariants,

More information

A Brief Introduction to Bayesian Networks AIMA CIS 391 Intro to Artificial Intelligence

A Brief Introduction to Bayesian Networks AIMA CIS 391 Intro to Artificial Intelligence A Brief Introduction to Bayesian Networks AIMA 14.1-14.3 CIS 391 Intro to Artificial Intelligence (LDA slides from Lyle Ungar from slides by Jonathan Huang (jch1@cs.cmu.edu)) Bayesian networks A simple,

More information

A Roadmap to an Enhanced Graph Based Data mining Approach for Multi-Relational Data mining

A Roadmap to an Enhanced Graph Based Data mining Approach for Multi-Relational Data mining A Roadmap to an Enhanced Graph Based Data mining Approach for Multi-Relational Data mining D.Kavinya 1 Student, Department of CSE, K.S.Rangasamy College of Technology, Tiruchengode, Tamil Nadu, India 1

More information

Finite Model Generation for Isabelle/HOL Using a SAT Solver

Finite Model Generation for Isabelle/HOL Using a SAT Solver Finite Model Generation for / Using a SAT Solver Tjark Weber webertj@in.tum.de Technische Universität München Winterhütte, März 2004 Finite Model Generation for / p.1/21 is a generic proof assistant: Highly

More information

Chapter 8. NP-complete problems

Chapter 8. NP-complete problems Chapter 8. NP-complete problems Search problems E cient algorithms We have developed algorithms for I I I I I finding shortest paths in graphs, minimum spanning trees in graphs, matchings in bipartite

More information

Reals 1. Floating-point numbers and their properties. Pitfalls of numeric computation. Horner's method. Bisection. Newton's method.

Reals 1. Floating-point numbers and their properties. Pitfalls of numeric computation. Horner's method. Bisection. Newton's method. Reals 1 13 Reals Floating-point numbers and their properties. Pitfalls of numeric computation. Horner's method. Bisection. Newton's method. 13.1 Floating-point numbers Real numbers, those declared to be

More information

Choice Logic Programs and Nash Equilibria in Strategic Games

Choice Logic Programs and Nash Equilibria in Strategic Games Choice Logic Programs and Nash Equilibria in Strategic Games Marina De Vos and Dirk Vermeir Dept. of Computer Science Free University of Brussels, VUB Pleinlaan 2, Brussels 1050, Belgium Tel: +32 2 6293308

More information

10708 Graphical Models: Homework 2

10708 Graphical Models: Homework 2 10708 Graphical Models: Homework 2 Due October 15th, beginning of class October 1, 2008 Instructions: There are six questions on this assignment. Each question has the name of one of the TAs beside it,

More information

Validating Plans with Durative Actions via Integrating Boolean and Numerical Constraints

Validating Plans with Durative Actions via Integrating Boolean and Numerical Constraints Validating Plans with Durative Actions via Integrating Boolean and Numerical Constraints Roman Barták Charles University in Prague, Faculty of Mathematics and Physics Institute for Theoretical Computer

More information

Performing Incremental Bayesian Inference by Dynamic Model Counting

Performing Incremental Bayesian Inference by Dynamic Model Counting Performing Incremental Bayesian Inference by Dynamic Model Counting Wei Li and Peter van Beek and Pascal Poupart School of Computer Science University of Waterloo Waterloo, Ontario N2L 3G1, Canada {w22li,

More information

Declarative programming. Logic programming is a declarative style of programming.

Declarative programming. Logic programming is a declarative style of programming. Declarative programming Logic programming is a declarative style of programming. Declarative programming Logic programming is a declarative style of programming. The programmer says what they want to compute,

More information

Full CNF Encoding: The Counting Constraints Case

Full CNF Encoding: The Counting Constraints Case Full CNF Encoding: The Counting Constraints Case Olivier Bailleux 1 and Yacine Boufkhad 2 1 LERSIA, Université de Bourgogne Avenue Alain Savary, BP 47870 21078 Dijon Cedex olivier.bailleux@u-bourgogne.fr

More information

NP-Completeness of 3SAT, 1-IN-3SAT and MAX 2SAT

NP-Completeness of 3SAT, 1-IN-3SAT and MAX 2SAT NP-Completeness of 3SAT, 1-IN-3SAT and MAX 2SAT 3SAT The 3SAT problem is the following. INSTANCE : Given a boolean expression E in conjunctive normal form (CNF) that is the conjunction of clauses, each

More information

Approximate Discrete Probability Distribution Representation using a Multi-Resolution Binary Tree

Approximate Discrete Probability Distribution Representation using a Multi-Resolution Binary Tree Approximate Discrete Probability Distribution Representation using a Multi-Resolution Binary Tree David Bellot and Pierre Bessière GravirIMAG CNRS and INRIA Rhône-Alpes Zirst - 6 avenue de l Europe - Montbonnot

More information

Local Two-Level And-Inverter Graph Minimization without Blowup

Local Two-Level And-Inverter Graph Minimization without Blowup Local Two-Level And-Inverter Graph Minimization without Blowup Robert Brummayer and Armin Biere Institute for Formal Models and Verification Johannes Kepler University Linz, Austria {robert.brummayer,

More information

Part I Logic programming paradigm

Part I Logic programming paradigm Part I Logic programming paradigm 1 Logic programming and pure Prolog 1.1 Introduction 3 1.2 Syntax 4 1.3 The meaning of a program 7 1.4 Computing with equations 9 1.5 Prolog: the first steps 15 1.6 Two

More information

Discrete mathematics

Discrete mathematics Discrete mathematics Petr Kovář petr.kovar@vsb.cz VŠB Technical University of Ostrava DiM 470-2301/02, Winter term 2018/2019 About this file This file is meant to be a guideline for the lecturer. Many

More information

COMP 410 Lecture 1. Kyle Dewey

COMP 410 Lecture 1. Kyle Dewey COMP 410 Lecture 1 Kyle Dewey About Me I research automated testing techniques and their intersection with CS education My dissertation used logic programming extensively This is my second semester at

More information

Safe Stratified Datalog With Integer Order Does not Have Syntax

Safe Stratified Datalog With Integer Order Does not Have Syntax Safe Stratified Datalog With Integer Order Does not Have Syntax Alexei P. Stolboushkin Department of Mathematics UCLA Los Angeles, CA 90024-1555 aps@math.ucla.edu Michael A. Taitslin Department of Computer

More information

CS 188: Artificial Intelligence Spring Today

CS 188: Artificial Intelligence Spring Today CS 188: Artificial Intelligence Spring 2006 Lecture 7: CSPs II 2/7/2006 Dan Klein UC Berkeley Many slides from either Stuart Russell or Andrew Moore Today More CSPs Applications Tree Algorithms Cutset

More information

Bayesian Logic Networks (Extended Version )

Bayesian Logic Networks (Extended Version ) Bayesian Logic Networks (Extended Version ) Technical Report IAS-2009-03 Dominik Jain, Stefan Waldherr and Michael Beetz Intelligent Autonomous Systems Group, Technische Universität München Boltzmannstr.

More information

SAT solver of Howe & King as a logic program

SAT solver of Howe & King as a logic program SAT solver of Howe & King as a logic program W lodzimierz Drabent June 6, 2011 Howe and King [HK11b, HK11a] presented a SAT solver which is an elegant and concise Prolog program of 22 lines. It is not

More information

Today. CS 188: Artificial Intelligence Fall Example: Boolean Satisfiability. Reminder: CSPs. Example: 3-SAT. CSPs: Queries.

Today. CS 188: Artificial Intelligence Fall Example: Boolean Satisfiability. Reminder: CSPs. Example: 3-SAT. CSPs: Queries. CS 188: Artificial Intelligence Fall 2007 Lecture 5: CSPs II 9/11/2007 More CSPs Applications Tree Algorithms Cutset Conditioning Today Dan Klein UC Berkeley Many slides over the course adapted from either

More information

Lecture 1 Contracts. 1 A Mysterious Program : Principles of Imperative Computation (Spring 2018) Frank Pfenning

Lecture 1 Contracts. 1 A Mysterious Program : Principles of Imperative Computation (Spring 2018) Frank Pfenning Lecture 1 Contracts 15-122: Principles of Imperative Computation (Spring 2018) Frank Pfenning In these notes we review contracts, which we use to collectively denote function contracts, loop invariants,

More information

2386 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 6, JUNE 2006

2386 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 6, JUNE 2006 2386 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 6, JUNE 2006 The Encoding Complexity of Network Coding Michael Langberg, Member, IEEE, Alexander Sprintson, Member, IEEE, and Jehoshua Bruck,

More information

Note that in this definition, n + m denotes the syntactic expression with three symbols n, +, and m, not to the number that is the sum of n and m.

Note that in this definition, n + m denotes the syntactic expression with three symbols n, +, and m, not to the number that is the sum of n and m. CS 6110 S18 Lecture 8 Structural Operational Semantics and IMP Today we introduce a very simple imperative language, IMP, along with two systems of rules for evaluation called small-step and big-step semantics.

More information

Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms for Inference Fall 2014

Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms for Inference Fall 2014 Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms for Inference Fall 2014 1 Course Overview This course is about performing inference in complex

More information

5 The Control Structure Diagram (CSD)

5 The Control Structure Diagram (CSD) 5 The Control Structure Diagram (CSD) The Control Structure Diagram (CSD) is an algorithmic level diagram intended to improve the comprehensibility of source code by clearly depicting control constructs,

More information

SAT-CNF Is N P-complete

SAT-CNF Is N P-complete SAT-CNF Is N P-complete Rod Howell Kansas State University November 9, 2000 The purpose of this paper is to give a detailed presentation of an N P- completeness proof using the definition of N P given

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Overview of Part Two Probabilistic Graphical Models Part Two: Inference and Learning Christopher M. Bishop Exact inference and the junction tree MCMC Variational methods and EM Example General variational

More information

The Satisfiability Problem [HMU06,Chp.10b] Satisfiability (SAT) Problem Cook s Theorem: An NP-Complete Problem Restricted SAT: CSAT, k-sat, 3SAT

The Satisfiability Problem [HMU06,Chp.10b] Satisfiability (SAT) Problem Cook s Theorem: An NP-Complete Problem Restricted SAT: CSAT, k-sat, 3SAT The Satisfiability Problem [HMU06,Chp.10b] Satisfiability (SAT) Problem Cook s Theorem: An NP-Complete Problem Restricted SAT: CSAT, k-sat, 3SAT 1 Satisfiability (SAT) Problem 2 Boolean Expressions Boolean,

More information

Learning Tractable Probabilistic Models Pedro Domingos

Learning Tractable Probabilistic Models Pedro Domingos Learning Tractable Probabilistic Models Pedro Domingos Dept. Computer Science & Eng. University of Washington 1 Outline Motivation Probabilistic models Standard tractable models The sum-product theorem

More information