Inference in Higher Order MRF-MAP Problems with Small and Large Cliques

Size: px
Start display at page:

Download "Inference in Higher Order MRF-MAP Problems with Small and Large Cliques"

Transcription

1 Inference in Higher Order MRF-MAP Problems with Small and Large Cliques Ishant Shanu 1 Chetan Arora 1 S.N. Maheshwari 2 1 IIIT Delhi 2 IIT Delhi Abstract Higher Order MRF-MAP formulation has been a popular technique for solving many problems in computer vision. Inference in a general MRF-MAP problem is NP Hard, but can be performed in polynomial time for the special case when potential functions are submodular. Two popular combinatorial approaches for solving such formulations are flow based and polyhedral approaches. Flow based approaches work well with small cliques and in that mode can handle problems with millions of variables. Polyhedral approaches can handle large cliques but in small numbers. We show in this paper that the variables in these seemingly disparate techniques can be mapped to each other. This allows us to combine the two styles in a joint framework exploiting the strength of both of them. Using the proposed joint framework, we are able to perform tractable inference in MRF-MAP problems with millions of variables and a mix of small and large cliques, a formulation which can not be solved by either of the two styles individually. We show applicability of this hybrid framework on object segmentation problem as an example of a situation where quality of results is significantly better than systems which are based only on the use of small or large cliques. 1. Introduction Many problems in computer vision can be posed as labeling problems [33, 16]. Modeling them as MRF and finding the labeling configuration with maximum a posteriori probability converts them to the following discrete optimization problem: l V = arg min l V p V f p (l p ) + c C f c (l c ). (1) Here, a pixel is denoted by p and V is the set of all pixels. A clique is a set of pixels conditionally dependent upon each other and is denoted by c, and l c is the set of labels assigned to the pixels in clique c. C denotes the set of all cliques. f p (l p ), called unary energy, assigns the cost for labeling a particular pixel p as l p, whereas f c (l c ), called clique potential, is the cost of assigning labeling configuration l c to clique c. The inference problem, popularly known as MRF- MAP, is to find the labeling, l V which minimizes the right hand side of Eq. (1). Many computer vision problems like image restoration, segmentation of videos and images, super resolution, texture synthesis, stereo matching to object detection, can be modelled as MRF-MAP inference problems [33, 16]. MRF-MAP is computationally a hard problem even in the restricted setting of binary labels and pairwise cliques. However, its applicability to modeling computer vision problems is so well established that variety of approaches have been explored to solve the problem in the restricted setting mentioned above. These approaches have ranged from message passing [24], dual decomposition [26], semidefinite programming relaxation [34], and graph cut [1, 25]. Since higher order clique potentials are more expressive compared to pairwise and lead to improved visual quality in the inferred output, the above mentioned techniques have been generalized to handle such problems also. The inference problem, however, becomes significantly more complicated. Apart from generalization of message passing and dual decomposition [23, 31], reducing the higher order problem to an equivalent pseudo Boolean form and solving it approximately using the QPBO algorithm is the other notable approach to handle such generalizations [8, 11, 15]. It should be noted that all the methods mentioned above provide only approximate solutions. Therefore, developing efficient inference techniques for some special subclass of clique potentials of interest has become an area of active research [18, 19, 20, 29]. One such special class of clique potentials whose importance to solving MRF MAP inference problems in computer vision is increasingly getting recognized is based on the notion of submodularity (formally defined in the next section). Algorithms that minimize submodular functions (SFM) in polynomial time have now been known for over two decades [9, 30]. These algorithms work directly with the submodular polyhedron (defined formally in the next section), but have been found to be unusable as their run time is bounded by very high degree polynomials. There have been attempts to adapt/improvise various SFM techniques 1

2 for MRF-MAP inference problem [14, 17, 32]. A significant breakthrough is Generic Cuts (GC) [2, 3], a flow based algorithm that solves the inference problem for sum of submodular potentials in low order polynomial time for fixed clique sizes. GC has been shown to run efficiently even when the number of cliques runs into hundreds of thousands but does not scale well on problems with cliques of size larger than 16. The problem of minimizing sum of submodular functions is beginning to attract attention of researchers [22, 31]. Kolmogorov [22] has reported a polynomial time submodular flow [6] based algorithm using capacity scaling. No results exist, however, to assess it s viability on practical vision problems. Recently [31] has reported a submodular function minimization algorithm, SOSMNP (Sum of Submodular Min Norm Point), based on Minimum Norm Point (MNP) algorithm [9]. SOSMNP works on realistic vision problems (millions of variables) with few but very large clique sizes. This paper arises out of a realization that a typical vision problem involving millions of variables may require a model in which there are a very large number of small cliques (of sizes in the range 2 to 16) and a relatively few but large cliques (sizes of the order of 1000). One typical example could be when smaller cliques enforce regularization while larger cliques penalize deviation from the region/patch level statistics (e.g. a prior on the expected object size). Such problems are not solvable practically either by flow based algorithms like [2] or submodular polyhedron based algorithm like [31]. The first cannot handle large cliques and the second is applicable only when the number of cliques is small. What is required is a framework which simultaneously runs GC, a flow based algorithm on the small cliques, and SOSMNP, a submodular polyhedron based min norm style of algorithm on the large cliques. We show that this is indeed possible. The specific contributions of this paper are as follows: 1. Mapping: We show that variables in the, seemingly disparate, flow based and submodular polyhedron based approaches can be mapped to one other. 2. Fusion: The mapping allows us to propose a hybrid framework where we can adopt different styles to solve different sub-problems within the overall inference problem. 3. Efficiency: Using the proposed framework inference problems involving large number of variables with a mix of small and large cliques can be solved efficiently. 4. Bridge: The proposed framework is general and can act as a model for combining other algorithms chosen for applicability to other vision problems. 2. Background Central to this paper is the notion of submodularity. A set function, f : S V R is called submodular if: f(a) + f(b) f(a B) + f(a B), (2) where A, B V. Convex polyhedron based SFM algorithms associate a base polyhedra, B(f), with each submodular function f: B(f) = {x x(s) f(s), S V; x(v) = f(v)}, where x R V and x(u) = u U x(u). Any point x B(f) is called a base vector or simply a base. For any base vector x, we denote x (U) = u U min {0, x(u)}. It is easy to see that, x (V) x(y ) f(y ) holds for any x B(f) and any subset Y V. An important result in SFM theory is the Min Max Theorem [12] which states that for a submodular function f : S V R with f( ) = 0: max {x (V) x B(f)} = min {f(s) S V}. (3) Strongly polynomial convex polyhedron based SFM algorithms actually solve the dual problem of maximizing x (V) [13, 28, 30]. The Min Norm Point algorithm of Fuzishige et al. [9] on the other hand reduces the problem of finding a base vector, x, with maximum x (V) to finding a base vector x with minimum norm: i.e., x = min x B(f) x 2. While no polynomial bound is known for the algorithm, experimental studies have shown that it runs much faster than the other available algorithms [14, 27, 31] MRF-MAP and SFM Relationship Eq. (1), with label set {a, b} and the clique potential f c, can be looked upon as a set function over the set c, such that f c (l c ) defines the cost of a subset of pixels in c given a particular label (we use label a as a convention here) in the labeling l c. Further, it is possible to consider p f p(l p ) in Eq. (1) as a modular function 1 defined over a clique spanning all pixels of the image. With this, the MRF-MAP problem of Eq. (1) can be equivalently written as: l V = arg min l V f c (l c ). (4) c C The above MRF-MAP formulation can be looked upon as minimizing a sum of submodular functions. Shanu et al. [31] have suggested a block coordinate descent framework to implement the Min Norm Point algorithm when cliques are large. 1 Modular functions are special type of submodular function when Eq. (2) is satisfied with equality. Unlike submodular functions, modular functions can be minimized trivially by finding the minimum of each element independently.

3 2.2. Sum of Submodular Functions and Block Coordinate Descent With each f c, as given in Eq. (4), one can associate a base polyhedron such that: B(f c ) := { y c R c y c (U) f c (U), U c; (5) } y c (c) = f c (c). (6) Shanu et al. [31] have given the following results to relate a base vector, x, of the function f and a set of base vectors y c of a f c : Lemma 2.1. Let x(s) = c y c(c S) where each y c belongs to base polyhedra B(f c ). Then the vector x belongs to base polyhedron B(f). Lemma 2.2. Let x be a vector belonging to the base polyhedron B(f). Then, x can be expressed as the sum: x(s) = c y c(s c), where each y c belongs to the submodular polyhedron B(f c ) i.e., y c B(f c ) c. The above results were used to minimize x 2 using a block coordinate descent approach, where each block represents a base vector y c as defined above (c.f. [31]). Note that a base vector y c is of dimension c (clique size), whereas a base x is of dimension V (number of pixels in an image). Since c V, minimizing the norm of y c over it s submodular polyhedron B(f c ) is much more efficient than minimizing the norm of x by just applying the MNP algorithm. It should be noted that the above scheme is equally applicable with any other polyhedral based algorithm to minimize sum of submodular functions [9, 13, 28, 30] Generic Cuts (GC) An important result from [2] states that an arbitrary submodular function, f, defined over a set V, can always be factorized into a sum of modular function, f m, and a submodular function, f, such that f ( ) = f (V) = 0 and f (S) 0, S V. In GC a given function (Eq. (1)) is reparameterized to such a form, f = f m + f c. GC then formulates the reparameterized problem as an integer program, relaxes it to a linear program, and maximizes its Lagrangian dual given below: max p V U p, subject to U p f m (l) + V c,p,l p V, l {a, b}, and c:p c V c,p,l p c f c(l c ), c C, l c {a, b} c. (7) p c For all submodular functions, without loss of generality, all V c,p,b can be set to 0, thereby reducing Eq. (7) to: p:l p c=a V c,p,a f c(l c ), c C, l c {a, b} c. (8) The algorithm then creates a gadget based flow graph (a gadget for each clique) with a node corresponding to each pixel and two special nodes s and t for source and sink respectively. The capacities of the edges from s and t to pixels are set using modular function f m, whereas f c(l c ) are used to restrict sum of flows in the edges (measured by the value of variables V c,p,a ) of a gadget [2]. 3. Proposed Framework From hereon, we will use a slightly modified notation for flow variables in GC. We will denote the set of variables V c,p,a as a vector v c, with each variable V c,p,a denoted by v c (p) (a being implicit in the notation, since all the variables corresponding to b always remain 0 in GC). Recall that a base vector is denoted by y c and each element in the base vector is denoted as y c (p). Further, we will assume that each submodular function f c is of the form such that f c ( ) = f c (c) = 0 and f(s) 0, S c. As shown in [2], the assumption is not restrictive as every submodular function can be reparameterized to such a form. The following results establish the correspondence between v c and y c. Lemma 3.1. Any vector v c derived from a valid flow in the GC flow graph is a vector in the base polyhedron of f c. Proof. Equations (5) and (6) give the conditions for any vector to be in the base polyhedron. By virtue of Eq. (8), it is easy to see that the flow vector satisfies Eq. (5). Further, by the property of flow conservation in the GC flow graph p c v c(p) = 0 = f c (c) (c.f. [2]), thus satisfying Eq. (6) as well. Lemma 3.1 serves to indicate that equations (5) and (8) are essentially the same, and a flow vector also satisfies the conditions of a base vector. This allows us to directly map a flow vector v c to y c. But, the converse is not true, since, a general base vector may not satisfy the flow conservation constraint at each vertex in the flow graph. In flow minus out flow, i.e. excess at the vertex, may be negative or positive. However, this is easy to fix as follows. If the excess is negative then an edge directed from the source s is to the vertex with negative excess is added with flow equal to the absolute value of the excess. If the excess is positive then an edge from the positive excess vertex to the sink t is added with flow equal to the excess. We have, in effect, shown the following:

4 Lemma 3.2. For every base vector y c there exists a GC flow graph with valid flow obtained through one to one mapping of y c to the flow vector v c in the gadget corresponding to clique c in the GC flow graph. 4. Hybrid Algorithms Lemmas 3.1 and 3.2 enable us to put in place a block coordinate descent strategy consistent with the framework of [31]. We make a coordinate block corresponding to sum of large clique potentials to be solved by a polyhedral method maximizing x (V). Another coordinate block is made (corresponding to sum of smaller cliques) to be solved by GC by maximizing flow. At each iteration we choose a coordinate block and solve it s optimization problem using the chosen method. Consistent with the coordinate blocks, Eq. (4) can be rewritten as l V = arg min f c (l c ) + f c (l c ). (9) l V c L c S Here L and S are the set of large and small cliques respectively such that C = L S. If x is a base vector corresponding to the submodular function as given in Eq. (9), then using Lemma 2.2 it can also be written as: x = c L y c (l c ) + c S y c (l c ), or (10) x = x l + x s, (11) where y c (l c ), c L and y c (l c ), c S are the base vectors corresponding to large and small cliques respectively. Vectors, x l and x s denote the factors of x due to large and small cliques respectively. A hybrid algorithm achieves the objective of maximizing x (V) by alternately carrying out a block coordinate descent iterations on the blocks corresponding to L and S. While maximizing x l (V) we keep x s (and its associated y s) as constant. Any base polyhedron based SFM algorithm which works on the principle of maximizing x (V) can be used for the maximizing step. Note that since there will be multiple number of large cliques it is quite likely that the sum of submodular function algorithm used may itself be based on block coordinate descent scheme. We use GC to maximize x s (V) and when we do that, we keep x l constant, in effect treating them as unary costs in GC flow graph. As explained in Section 2.3 the role of unary costs is in setting the terminal edge capacities in the GC graph. The flow graph is created as per the construction given in Section 3. It may be noted that during multiple such block coordinate ascent iterations GC edge capacities remain constant and the minimum cut to be found changes due to changes in the terminal edge capacities. In principle, it is possible to exploit such structure by initializing the flow in an iteration from the flow in the last iteration, fixing the overflows and then send more flow if possible. This kind of structure have been exploited by Kohli and Torr [21] using graph cuts. Our current implementation, however, does not exploit this. 5. Hybrid Algorithm Based On MNP and GC It must be pointed out that the above framework is not applicable as written if the base polyhedron based algorithm is the Min Norm Point Algorithm, the algorithm of choice for working with large cliques [27, 31]. We give below the reasoning in detail and the changes required to use MNP to handle large cliques Problems in Combining GC with MNP Before we delve into the problem in combing MNP and GC, it is important to understand the differences between polyhedral methods which maximize x (V) and MNP which minimizes x 2. Polyhedral methods that maximize x (V) can be considered to move from one base vector to another in which the value of x (V) increases. The extent of increase is a function of the next base vector that the algorithm chooses. For example, in Schrijver s algorithm [30] this increase is effected by carrying out what is known as block exchange. Given a base vector x and two elements say p and q in V, in a block exchange, a δ is computed to generate a new base vector x, which differs from x only in that the value of x (p) is larger than x(p) by δ and the value of x (q) is smaller than x(q) by δ. In many polyhedral based methods the value of x (V) is maximized through appropriately chosen block exchanges. Let S be a set that minimizes the given submodular function. It can be shown that S can also be obtained by including all the elements with negative value in a base vector x that maximizes x (V) along with some other elements of x, with zero value, chosen based upon some reachability criterion that is particular to specific SFM algorithms. Note that if p and q are in S a block exchange may be possible which increases the value of x(p) and decreases the value of x(q) without making x(p) positive. In this case neither the value of x (V) changes nor the optimal set S changes, but new optimal base vector has been computed. By picking two elements, which either both belong to S or it s complement, one can, therefore, generate infinitely many optimal base vectors that leave the optimal set S unchanged. It can be easily shown that when such block exchanges also leave the relative order of the values of elements unchanged the L 2 norm value of the resultant base vector reduces. The discussion not only indicates that polyhedral methods using block exchanges as the basic operation to maximize x (V) may not minimize L 2 norm, but also motivates a strategy to do so using GC, as we show later in the next section.

5 Note that, apart from the edge emanating from the source and the edge incident at the sink, the edges in a flow augmentation path in GC consists of a sequence of pairs of edges each pair belonging to a gadget through which the path passes. During flow augmentation flow increases in one edge of such a pair and decreases by the same amount in the other resulting in change in the flow vetor value at just two vertices in the gadget involved. Lemma 3.2 implies that this results in change in the value of the corresponding base vector at just two vertices. In effect flow augmentation in GC, at the macro level, involves performing of a series of block exchange operations (corresponding to the gadgets through which the augmenting path passes), and results in maximized x (V) at termination. The source of problem faced in combining GC and SOSMNP in a common block descent framework is, the additional requirement, that output of GC should also minimize the L 2 norm. When GC flow augmentation stops, the set of pixels gets partitioned into a (S, T ) cut set: namely those which are reachable from the source s form the set S, and the remaining nodes form the set T. It can be easily shown that while the total flow in the edges across the cutset is the value of the maximum flow and also the value of the optimal solution to the sum of submodular function minimization problem, the values of the elements in S in the corresponding final base vector are all negative or zero and their sum is equal to the max flow in the GC flow graph Outputting a Min L 2 Norm Solution in GC As shown in the previous section, GC can be interpreted as a block exchange algorithm. Therefore, the reachability property from nodes s and t to verticices in sets S and T in the max flow (S, T ) cutset in the GC flowgraph are the same reachability properties as those by the attributes in the sets S and T outputted by any other block exchange based algorithm, say Schrijver s. Therefore, the counterpart of block exchange among nodes in set S in Schrijver s algorithm that results in lowering of the L 2 norm in GC is nothing but the result of sending flow from s to some node p in S. The following lemma can be shown to hold. Lemma 5.1. The base vector x corresponding to max flow flow in GC with (S, T ) cut set does not satisfy the L 2 norm as long as there exist augmenting paths 1. from s to any vertex p in S such that the value of x(q), q being the first vertex on that path after s, is less than x(p), and 2. to t from any vertex p in T such that the value of x(q), q being the first vertex on that path after t, is larger than x(p). We propose the following scheme for moving towards minimum L 2 norm solution in GC. Let (S, T ) be the min cut in the flow graph outputted by GC. We will explain below the process for the set S. The process for set T will be identical and consistent with condition 2 of Lemma 5.1 and effectively be just the mirror image of process explained below. Let x denote the base vector corresponding to a GC block coordinate descent iteration. Let x (S) denote the sum of the negative elements (corresponding to pixels) in set S. It is easy to see that the ideal redistribution of values in vector x such that the x (S) remains same, but the norm x(s) 2 achieves minimum is when all non zero elements in set S have the value x (S)/ S. To move towards the desired objective, we create a new flow graph containing all the nodes in S and two auxiliary nodes s and t. We create an edge from source s to a vertex, p, whenever x(p) < x (S)/ S. The capacity of the edge is set to x (S) S x(p). From the rest of the negative valued vertices in S we direct edges to sink t. Capacity of an edge from such a node q to t is set equal to x(q) x+ (S) S. All the other edges between two pixel nodes of the set S are the edges of the original GC graph with their capacities set equal to the residual capacities they had when GC s previous iteration terminated. We now solve the max flow problem in the flow subgraph so created. If all the edges out of s are saturated then we have effectively shown that there exists an optimal solution to the original problem in which all the elements on the S side have the same value, which the optimal min norm solution should have. Otherwise, we have discovered a new (S, T ) cut in the subproblem the base vector corresponding to which can be moved towards the min L 2 norm recursively by solving two max flow problems (for the new S and T sets) as explained above on increasingly smaller flow graphs. Note that the this method of moving towards the min L 2 norm solution, consistent with Lemma 5.1, has the merit that the number of additional max flow problems that need to be solved are not only finite but are bounded by the number of nodes in the original GC graph Proposed Algorithm We formalize the proposed hybrid strategy for minimizing the sum of submodular functions. As it is common in SFM literature, we assume that f c ( ) = 0 for large clique potentials. For small clique potentials, as is done in GC, we assume that the function has been appropriately reparameterized such that f c ( ) = f c (c) = 0 and the value of clique potential for all other configurations is non-negative. No such reparameterization is done for large clique potential. Values of modular function, f m, can be arbitrary (positive/negative/zero), but we reparameterize it in our algorithm such that f m (l p = a) = f m (l p = a) f m (l p = b) and f m (l p = b) = 0. This causes a decrease of constant, p f m(l p = b), in all the function values, which we add

6 Procedure 1 HybridMNGC: Algorithm for minimizing a sum of submodular function with mixed sized cliques Input: f = f m + f s + f l. Output: x = y m + x s + x l = arg min x 2 subject to x B(f). 1: Set y m = f m and y c = 0, c S. 2: c L initialize y c from a random base vector. 3: while Norm of x decreases do 4: Call SOSMNP with y m + x s as unary cost and f l as the clique potential; 5: MinNormGC(V, f s, y m + x l, 0); 6: end while Procedure 2 MinNormGC(V, f s, U, b) Input: V: Set of Nodes/Elements for which norm is to computed. Input: U: Function specifying unary/modular cost for each pixel in V. Input: f s : Clique potential for the small cliques. Input: b: Base value above which we add edge from s. # We create the GC gadget graph using f s and assume that the flow graph once created is available globally during the recursive calls. Only the edges from s and to t change. 1: for p V do 2: if U(p) > b then 3: Add s to p edge with capacity U(p) b; 4: else 5: Add p to t edge with capacity b U(p); 6: end if 7: end for 8: Run GC; Find minimum cut and corresponding S and T sets; 9: if S > 1 then 10: Denote by U S, the residual edge capacity between source s and pixels p S; 11: Denote average of U S by b S ; 12: MinNormGC(S, f s, U S, b S ); 13: end if 14: if T > 1 then 15: Denote by U T, the residual edge capacity between pixels p T and sink node t; 16: Denote average of U T by b T ; 17: MinNormGC(T, f s, U T, b T ); 18: end if back in the returned value of the minimum. We solve the minimization problem in the dual domain by finding a vector with minimum norm in the base polyhedron of sum of f m + f l + f s. We make use of the block coordinate descent framework of [31] and maintain a base vector, x, of the overall function, as the sum of base vectors y m, x s and x l, corresponding to modular, small and large clique potential functions. We initialize the vector y m as the value of function f m and keep it constant throughout the algorithm. We initialize the y c corresponding to small cliques to be zero, which is equivalent to initializing the GC flow graph with flow of zero. For large cliques, we start with an arbitrary base as the starting y c. The various iterations of the algorithm follow the scheme as detailed out in Section 4. The above detailing of the initialization process takes into account modular function costs which were subsumed earlier. Algorithms 1 and 2 give the details of proposed algorithm in pseudocode format Convergence Overall algorithm implements block coordinate descent over two blocks. One consists of all the large cliques, and the other contains all the small cliques. Each iteration minimizes the norm of the solution vector x corresponding to the block chosen. Convergence of the over all algorithm follows from the in principle convergence of Min Norm algorithm [10], polynomial time convergence of GC [2], and convergence of the block coordinate descent based min norm algorithm [14, 31]. The additional work that is happening is transformation of GC output in every run of GC block to satisfy L 2 norm. That can take as little as O(log n) GC iterations in the best case to O(n) iterations in the worst case, where n is the number of pixels. Experiments reported in the next section indicate that HybridMNGC is significantly faster than SOSMNP which, experiments reported earlier [31] indicate, is the algorithm of choice for polyhedral based algorithms working with large cliques. 6. Experiments We have conducted all the experiments on a regular workstation with Intel Core i7 CPU, 8 GB of RAM and running Windows 10 OS. We have implemented the proposed algorithm in C++. We have taken implementations of SOSMNP and GC from the authors website and modified them for our purpose. All timings reported are in seconds. The focus of the experiments is to establish the efficacy of the hybrid scheme both in terms of the efficiency and quality of the results outputted. The experiments focussing on efficiency have been on synthetic problems of small sizes. The results and discussion are in Section 6.1. Quality comparison has been done on the pixelwise object segmentation problem where the images for the demonstration are taken from Pascal VOC dataset [7]. Outputs of HybridMNGC are compared with that of SegNet [4] (chosen to represent state of the art methods based on deep neural networks). GC runs with cliques of size two, and SOSMNPs run with cliques of size approximately In experiments which focus on efficiency, as in [31], we consider edge based clique potentials based on counting the

7 Big Clique Size =100, No. of cliques =25 Small Clique Size HybridMNGC SOSMNP E E E E+06 Dual Primal Table 1. Clique potentials are Count Based. Small cliques span the image in sliding window style. All running time are in seconds. Big Clique Size =100, No. of cliques =25 Small Clique Size HybridMNGC SOSMNP Table 2. Similar comparison as in Table 1 with the proviso that clique potentials are edge based. All running time are in seconds. No. of Pixels 2500 #No. of Big Cliques HybridMNGC SOSMNP Table 3. Count based big cliques with size=100, pairwise small cliques span whole image. All running time are in seconds. No. of Big Cliques = 50 #No. of Pixels HybridMNGC SOSMNP Table 4. Time as a function of Image size. Count based big cliques with size=100, pairwise small cliques span whole image. All running time are in seconds. square root of the number of edges, and count based potential which involve counting the product of number of ones and zeros in the clique. The unary potentials are assigned randomly. In experiments on real images from Pascal VOC dataset we have used count based potentials and unary costs are based on the score received from SegNet for the image under consideration Performance Comparison Tables 1 and 2 compare the performance of the two algorithms on images of size (50 50) as small clique size is changed. The unary potentials, seeding of the large cliques, as well as the number of small cliques is the same. Note that HybridMNGC takes significantly less time than SOSMNP at all points of comparison. Observe that run time for SOSMNP does not show any monotonic pattern. Since at each iteration both algorithms work with optimal 0.00E E E E+06 Figure 1. Primal dual curve for a sample run of our algorithm subsolutions this fluctuation is possibly due to different ways in which values get transmitted among large number of small cliques under SOSMNP. In comparison HybridM- NGC, which is an order of magnitude faster than SOSMNP, exhibits typical GC trends where performance decreases with increase in clique size. Similar trend is observed with both count as well as edge based potentials. Tables 3 and 4 give the results of changes in time taken when both the number of big cliques and the image size is increased. The time for both the algorithms increase, but HybridMNGC continues to outperform the baseline SOSMNP Comparing Object Segmentation Quality We show some indicative results in Figures 2 and 3 for pixel wise object segmentation. Each image has a single object consistent with our binary labeling formulation. In the images in Figure 3 Gaussian noise with zero mean and intensity dependent variance has been added. We use Segnet [4], output in two ways. First as the basis for comparison of deep network based object segmentation, and second as a source for generating singleton confidence scores for independent pixel labels. The other points of comparison of the output of HybridMNGC are GC as an example of an algorithm for MRF MAP optimization using small clique models, and SOSMNP for an algorithm which can handle large cliques. The pairwise cliques span the image in 4-connected neighborhood style. Big cliques are region grown as in [32]. The number of big cliques and their size varies from image to image. Suffice to say that the number of big cliques range from 300 to 500. The average big clique size is 1000 with the max being The images themselves are pixels in size. Since both the images and the big clique sizes are too large for HybridMNGC and SOSMNP to terminate with optimal solutions in acceptable times, these algorithms are run in ɛ-approximate mode where ɛ is defined as in [5]. We run these algorithms with ɛ as 100. This condition seems to be arrived, on the average, after about iterations. Exper-

8 Original Image Ground-truth Segnet GC SOSMNP HybridMNGC Figure 2. Input image with no noise. Original Image Ground-truth Segnet GC SOSMNP HybridMNGC Figure 3. Input images with Gaussian noise. imental evidence indicates Figure 1 that the primal labeling stabilizes much before the termination condition is reached. Figure 3 has the comparative output images. Note the significant improvement in HybridMNGC output in comparison to that of GC and SOSMNP. Figure 2 contains an example of object segmentation with no added noise. Note that Segnet output deteriorates greatly with added noise. Output of HybridMNGC seems to be much more resilient to noise. The object segmentation experiments have been done with the objective of establishing the usefulness of working simultaneously with both higher order and small cliques. What should their sizes be, how should they be seeded in the image, how should the potentials be chosen for an object segmentation system based on the ideas outlined here is a subject for future research. 7. Conclusions Our objective in this paper has been to establish the efficacy of combining algorithms like Generic Cuts [2] and SOSMNP [31] which have complimentary strengths and weaknesses. The resultant algorithm, HybridMNGC, not only is shown to be more efficient, but has the potential of providing very high quality solutions to computer vision problems. Not only does it open up possibilities of developing new algorithms around HybridMNGC, it shows the way in which seemingly different techniques can be combined under the block coordinate descent framework. Solving problems in parallel by suitably defining blocks is an obvious possibility and future research direction for us. Acknowledgement: Chetan Arora has been supported by Infosys Center for AI, Visversaraya Young Faculty Research Fellowship, and EMR Funding by DST-SERB, Government of India. Ishant Shanu has been supported by Visvesaraya PhD Fellowship and a travel grant by Google.

9 References [1] C. Arora, S. Banerjee, P. Kalra, and S. Maheshwari. An efficient graph cut algorithm for computer vision problems. In ECCV, pages Springer, [2] C. Arora, S. Banerjee, P. Kalra, and S. Maheshwari. Generalized flows for optimal inference in higher order MRF-MAP. TPAMI, 37(7): , , 3, 6, 8 [3] C. Arora and S. Maheshwari. Multi label generic cuts: Optimal inference in multi label multi clique MRF-MAP problems. In CVPR, pages , [4] V. Badrinarayanan, A. Kendall, and R. Cipolla. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. arxiv preprint arxiv: , , 7 [5] D. Chakrabarty, P. Jain, and P. Kothari. Provable submodular minimization using wolfe s algorithm. In NIPS, pages , [6] J. Edmonds and R. Giles. A min-max relation for submodular functions on graphs. Annals of Discrete Mathematics, 1: , [7] M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn, and A. Zisserman. The PASCAL Visual Object Classes Challenge 2012 (VOC2012) Results. 6 [8] A. Fix, A. Gruber, E. Boros, and R. Zabih. A graph cut algorithm for higher-order markov random fields. In ICCV, pages , [9] S. Fujishige, T. Hayashi, and S. Isotani. The minimum-normpoint algorithm applied to submodular function minimization and linear programming , 2, 3 [10] S. Fujishige, T. Hayashi, and S. Isotani. The minimum-normpoint algorithm applied to submodular function minimization and linear programming [11] H. Ishikawa. Higher-order clique reduction without auxiliary variables. In CVPR, pages , [12] S. Iwata. Submodular function minimization. Mathematical Programming, 112(1):45, [13] S. Iwata, L. Fleischer, and S. Fujishige. A combinatorial strongly polynomial algorithm for minimizing submodular functions. JACM, 48(4): , , 3 [14] S. Jegelka, F. Bach, and S. Sra. Reflection methods for userfriendly submodular optimization. In NIPS, pages , , 6 [15] F. Kahl and P. Strandmark. Generalized roof duality for pseudo-boolean optimization. In ICCV, pages , [16] J. H. Kappes, B. Andres, F. A. Hamprecht, C. Schnörr, S. Nowozin, D. Batra, S. Kim, B. X. Kausler, T. Kröger, J. Lellmann, N. Komodakis, B. Savchynskyy, and C. Rother. A comparative study of modern inference techniques for structured discrete energy minimization problems. IJCV, 115(2): , [17] D. Khandelwal, K. Bhatia, C. Arora, and P. Singla. Lazy generic cuts. CVIU, 143:80 91, [18] P. Kohli. Graph cuts for minimizing robust higher order potentials. 1 [19] P. Kohli, M. P. Kumar, and P. H. Torr. P3 & beyond: Solving energies with higher order cliques. In CVPR, pages 1 8. IEEE, [20] P. Kohli, P. H. Torr, et al. Robust higher order potentials for enforcing label consistency. IJCV, 82(3): , [21] P. Kohli and P. H. S. Torr. Dynamic graph cuts for efficient inference in markov random fields. TPAMI, 29(12): , Dec [22] V. Kolmogorov. Minimizing a sum of submodular functions. Discrete Applied Mathematics, 160(15): , [23] V. Kolmogorov. A new look at reweighted message passing. TPAMI, 37(5): , [24] V. Kolmogorov and M. Wainwright. On the optimality of tree-reweighted max-product message-passing. arxiv preprint arxiv: , [25] V. Kolmogorov and R. Zabin. What energy functions can be minimized via graph cuts? TPAMI, 26(2): , [26] N. Komodakis, N. Paragios, and G. Tziritas. Mrf energy minimization and beyond via dual decomposition. TPAMI, 33(3): , [27] S. McCormick. Submodular function minimization , 4 [28] J. B. Orlin. A faster strongly polynomial time algorithm for submodular function minimization. Mathematical Programming, 118(2): , , 3 [29] C. Rother, P. Kohli, W. Feng, and J. Jia. Minimizing sparse higher order energy functions of discrete variables. In CVPR, pages , [30] A. Schrijver. A combinatorial algorithm minimizing submodular functions in strongly polynomial time. Journal of Combinatorial Theory, Series B, 80(2): , , 2, 3, 4 [31] I. Shanu, C. Arora, and P. Singla. Min norm point algorithm for higher order mrf-map inference. In CVPR, pages , , 2, 3, 4, 6, 8 [32] P. Stobbe and A. Krause. Efficient minimization of decomposable submodular functions. In NIPS, pages , , 7 [33] R. Szeliski, R. Zabih, D. Scharstein, O. Veksler, V. Kolmogorov, A. Agarwala, M. Tappen, and C. Rother. A comparative study of energy minimization methods for Markov random fields with smoothness-based priors. TPAMI, 30(6): , June [34] P. Wang, C. Shen, A. van den Hengel, and P. H. Torr. Efficient semidefinite branch-and-cut for MAP-MRF inference. IJCV, 117(3): ,

Learning Higher Order Potentials For MRFs

Learning Higher Order Potentials For MRFs Learning Higher Order Potentials For MRFs Dinesh Khandelwal IIT Delhi Parag Singla IIT Delhi Chetan Arora IIIT Delhi Abstract Higher order MRF-MAP formulation has been shown to improve solutions in many

More information

Quadratic Pseudo-Boolean Optimization(QPBO): Theory and Applications At-a-Glance

Quadratic Pseudo-Boolean Optimization(QPBO): Theory and Applications At-a-Glance Quadratic Pseudo-Boolean Optimization(QPBO): Theory and Applications At-a-Glance Presented By: Ahmad Al-Kabbany Under the Supervision of: Prof.Eric Dubois 12 June 2012 Outline Introduction The limitations

More information

A discriminative view of MRF pre-processing algorithms supplementary material

A discriminative view of MRF pre-processing algorithms supplementary material A discriminative view of MRF pre-processing algorithms supplementary material Chen Wang 1,2 Charles Herrmann 2 Ramin Zabih 1,2 1 Google Research 2 Cornell University {chenwang, cih, rdz}@cs.cornell.edu

More information

Alina Ene. From Minimum Cut to Submodular Minimization Leveraging the Decomposable Structure. Boston University

Alina Ene. From Minimum Cut to Submodular Minimization Leveraging the Decomposable Structure. Boston University From Minimum Cut to Submodular Minimization Leveraging the Decomposable Structure Alina Ene Boston University Joint work with Huy Nguyen (Northeastern University) Laszlo Vegh (London School of Economics)

More information

Dynamic Planar-Cuts: Efficient Computation of Min-Marginals for Outer-Planar Models

Dynamic Planar-Cuts: Efficient Computation of Min-Marginals for Outer-Planar Models Dynamic Planar-Cuts: Efficient Computation of Min-Marginals for Outer-Planar Models Dhruv Batra 1 1 Carnegie Mellon University www.ece.cmu.edu/ dbatra Tsuhan Chen 2,1 2 Cornell University tsuhan@ece.cornell.edu

More information

Iteratively Reweighted Graph Cut for Multi-label MRFs with Non-convex Priors

Iteratively Reweighted Graph Cut for Multi-label MRFs with Non-convex Priors Iteratively Reweighted Graph Cut for Multi-label MRFs with Non-convex Priors Thalaiyasingam Ajanthan, Richard Hartley, Mathieu Salzmann, and Hongdong Li Australian National University & NICTA Canberra,

More information

CS599: Convex and Combinatorial Optimization Fall 2013 Lecture 14: Combinatorial Problems as Linear Programs I. Instructor: Shaddin Dughmi

CS599: Convex and Combinatorial Optimization Fall 2013 Lecture 14: Combinatorial Problems as Linear Programs I. Instructor: Shaddin Dughmi CS599: Convex and Combinatorial Optimization Fall 2013 Lecture 14: Combinatorial Problems as Linear Programs I Instructor: Shaddin Dughmi Announcements Posted solutions to HW1 Today: Combinatorial problems

More information

Reduce, Reuse & Recycle: Efficiently Solving Multi-Label MRFs

Reduce, Reuse & Recycle: Efficiently Solving Multi-Label MRFs Reduce, Reuse & Recycle: Efficiently Solving Multi-Label MRFs Karteek Alahari 1 Pushmeet Kohli 2 Philip H. S. Torr 1 1 Oxford Brookes University 2 Microsoft Research, Cambridge Abstract In this paper,

More information

ESSP: An Efficient Approach to Minimizing Dense and Nonsubmodular Energy Functions

ESSP: An Efficient Approach to Minimizing Dense and Nonsubmodular Energy Functions 1 ESSP: An Efficient Approach to Minimizing Dense and Nonsubmodular Functions Wei Feng, Member, IEEE, Jiaya Jia, Senior Member, IEEE, and Zhi-Qiang Liu, arxiv:145.4583v1 [cs.cv] 19 May 214 Abstract Many

More information

CAP5415-Computer Vision Lecture 13-Image/Video Segmentation Part II. Dr. Ulas Bagci

CAP5415-Computer Vision Lecture 13-Image/Video Segmentation Part II. Dr. Ulas Bagci CAP-Computer Vision Lecture -Image/Video Segmentation Part II Dr. Ulas Bagci bagci@ucf.edu Labeling & Segmentation Labeling is a common way for modeling various computer vision problems (e.g. optical flow,

More information

Efficient Parallel Optimization for Potts Energy with Hierarchical Fusion

Efficient Parallel Optimization for Potts Energy with Hierarchical Fusion Efficient Parallel Optimization for Potts Energy with Hierarchical Fusion Olga Veksler University of Western Ontario London, Canada olga@csd.uwo.ca Abstract Potts frequently occurs in computer vision applications.

More information

A Principled Deep Random Field Model for Image Segmentation

A Principled Deep Random Field Model for Image Segmentation A Principled Deep Random Field Model for Image Segmentation Pushmeet Kohli Microsoft Research Cambridge, UK Anton Osokin Moscow State University Moscow, Russia Stefanie Jegelka UC Berkeley Berkeley, CA,

More information

Big and Tall: Large Margin Learning with High Order Losses

Big and Tall: Large Margin Learning with High Order Losses Big and Tall: Large Margin Learning with High Order Losses Daniel Tarlow University of Toronto dtarlow@cs.toronto.edu Richard Zemel University of Toronto zemel@cs.toronto.edu Abstract Graphical models

More information

IMPROVED FINE STRUCTURE MODELING VIA GUIDED STOCHASTIC CLIQUE FORMATION IN FULLY CONNECTED CONDITIONAL RANDOM FIELDS

IMPROVED FINE STRUCTURE MODELING VIA GUIDED STOCHASTIC CLIQUE FORMATION IN FULLY CONNECTED CONDITIONAL RANDOM FIELDS IMPROVED FINE STRUCTURE MODELING VIA GUIDED STOCHASTIC CLIQUE FORMATION IN FULLY CONNECTED CONDITIONAL RANDOM FIELDS M. J. Shafiee, A. G. Chung, A. Wong, and P. Fieguth Vision & Image Processing Lab, Systems

More information

Maximum flows and minimal cuts. Filip Malmberg

Maximum flows and minimal cuts. Filip Malmberg Maximum flows and minimal cuts Filip Malmberg MSF cuts I A Minimum Spanning Forest with respect to a set of seed-points divides a graph into a number of connected components. Thus, it defines a cut on

More information

MULTI-REGION SEGMENTATION

MULTI-REGION SEGMENTATION MULTI-REGION SEGMENTATION USING GRAPH-CUTS Johannes Ulén Abstract This project deals with multi-region segmenation using graph-cuts and is mainly based on a paper by Delong and Boykov [1]. The difference

More information

Graphs and Network Flows IE411. Lecture 21. Dr. Ted Ralphs

Graphs and Network Flows IE411. Lecture 21. Dr. Ted Ralphs Graphs and Network Flows IE411 Lecture 21 Dr. Ted Ralphs IE411 Lecture 21 1 Combinatorial Optimization and Network Flows In general, most combinatorial optimization and integer programming problems are

More information

Loopy Belief Propagation

Loopy Belief Propagation Loopy Belief Propagation Research Exam Kristin Branson September 29, 2003 Loopy Belief Propagation p.1/73 Problem Formalization Reasoning about any real-world problem requires assumptions about the structure

More information

3 No-Wait Job Shops with Variable Processing Times

3 No-Wait Job Shops with Variable Processing Times 3 No-Wait Job Shops with Variable Processing Times In this chapter we assume that, on top of the classical no-wait job shop setting, we are given a set of processing times for each operation. We may select

More information

Models for grids. Computer vision: models, learning and inference. Multi label Denoising. Binary Denoising. Denoising Goal.

Models for grids. Computer vision: models, learning and inference. Multi label Denoising. Binary Denoising. Denoising Goal. Models for grids Computer vision: models, learning and inference Chapter 9 Graphical Models Consider models where one unknown world state at each pixel in the image takes the form of a grid. Loops in the

More information

GRMA: Generalized Range Move Algorithms for the Efficient Optimization of MRFs

GRMA: Generalized Range Move Algorithms for the Efficient Optimization of MRFs International Journal of Computer Vision manuscript No. (will be inserted by the editor) GRMA: Generalized Range Move Algorithms for the Efficient Optimization of MRFs Kangwei Liu Junge Zhang Peipei Yang

More information

NP-Hardness. We start by defining types of problem, and then move on to defining the polynomial-time reductions.

NP-Hardness. We start by defining types of problem, and then move on to defining the polynomial-time reductions. CS 787: Advanced Algorithms NP-Hardness Instructor: Dieter van Melkebeek We review the concept of polynomial-time reductions, define various classes of problems including NP-complete, and show that 3-SAT

More information

On Covering a Graph Optimally with Induced Subgraphs

On Covering a Graph Optimally with Induced Subgraphs On Covering a Graph Optimally with Induced Subgraphs Shripad Thite April 1, 006 Abstract We consider the problem of covering a graph with a given number of induced subgraphs so that the maximum number

More information

CS6670: Computer Vision

CS6670: Computer Vision CS6670: Computer Vision Noah Snavely Lecture 19: Graph Cuts source S sink T Readings Szeliski, Chapter 11.2 11.5 Stereo results with window search problems in areas of uniform texture Window-based matching

More information

Learning CRFs using Graph Cuts

Learning CRFs using Graph Cuts Appears in European Conference on Computer Vision (ECCV) 2008 Learning CRFs using Graph Cuts Martin Szummer 1, Pushmeet Kohli 1, and Derek Hoiem 2 1 Microsoft Research, Cambridge CB3 0FB, United Kingdom

More information

CS 598CSC: Approximation Algorithms Lecture date: March 2, 2011 Instructor: Chandra Chekuri

CS 598CSC: Approximation Algorithms Lecture date: March 2, 2011 Instructor: Chandra Chekuri CS 598CSC: Approximation Algorithms Lecture date: March, 011 Instructor: Chandra Chekuri Scribe: CC Local search is a powerful and widely used heuristic method (with various extensions). In this lecture

More information

The Lazy Flipper: Efficient Depth-limited Exhaustive Search in Discrete Graphical Models

The Lazy Flipper: Efficient Depth-limited Exhaustive Search in Discrete Graphical Models The Lazy Flipper: Efficient Depth-limited Exhaustive Search in Discrete Graphical Models Bjoern Andres, Jörg H. Kappes, Thorsten Beier, Ullrich Köthe and Fred A. Hamprecht HCI, University of Heidelberg,

More information

A MRF Shape Prior for Facade Parsing with Occlusions Supplementary Material

A MRF Shape Prior for Facade Parsing with Occlusions Supplementary Material A MRF Shape Prior for Facade Parsing with Occlusions Supplementary Material Mateusz Koziński, Raghudeep Gadde, Sergey Zagoruyko, Guillaume Obozinski and Renaud Marlet Université Paris-Est, LIGM (UMR CNRS

More information

Parameterized graph separation problems

Parameterized graph separation problems Parameterized graph separation problems Dániel Marx Department of Computer Science and Information Theory, Budapest University of Technology and Economics Budapest, H-1521, Hungary, dmarx@cs.bme.hu Abstract.

More information

1. Lecture notes on bipartite matching February 4th,

1. Lecture notes on bipartite matching February 4th, 1. Lecture notes on bipartite matching February 4th, 2015 6 1.1.1 Hall s Theorem Hall s theorem gives a necessary and sufficient condition for a bipartite graph to have a matching which saturates (or matches)

More information

Integer Programming Theory

Integer Programming Theory Integer Programming Theory Laura Galli October 24, 2016 In the following we assume all functions are linear, hence we often drop the term linear. In discrete optimization, we seek to find a solution x

More information

Approximation Algorithms

Approximation Algorithms Approximation Algorithms Prof. Tapio Elomaa tapio.elomaa@tut.fi Course Basics A 4 credit unit course Part of Theoretical Computer Science courses at the Laboratory of Mathematics There will be 4 hours

More information

1 Linear programming relaxation

1 Linear programming relaxation Cornell University, Fall 2010 CS 6820: Algorithms Lecture notes: Primal-dual min-cost bipartite matching August 27 30 1 Linear programming relaxation Recall that in the bipartite minimum-cost perfect matching

More information

THE preceding chapters were all devoted to the analysis of images and signals which

THE preceding chapters were all devoted to the analysis of images and signals which Chapter 5 Segmentation of Color, Texture, and Orientation Images THE preceding chapters were all devoted to the analysis of images and signals which take values in IR. It is often necessary, however, to

More information

Energy Minimization Under Constraints on Label Counts

Energy Minimization Under Constraints on Label Counts Energy Minimization Under Constraints on Label Counts Yongsub Lim 1, Kyomin Jung 1, and Pushmeet Kohli 2 1 Korea Advanced Istitute of Science and Technology, Daejeon, Korea yongsub@kaist.ac.kr, kyomin@kaist.edu

More information

6. Lecture notes on matroid intersection

6. Lecture notes on matroid intersection Massachusetts Institute of Technology 18.453: Combinatorial Optimization Michel X. Goemans May 2, 2017 6. Lecture notes on matroid intersection One nice feature about matroids is that a simple greedy algorithm

More information

Chapter 8 of Bishop's Book: Graphical Models

Chapter 8 of Bishop's Book: Graphical Models Chapter 8 of Bishop's Book: Graphical Models Review of Probability Probability density over possible values of x Used to find probability of x falling in some range For continuous variables, the probability

More information

CS 5540 Spring 2013 Assignment 3, v1.0 Due: Apr. 24th 11:59PM

CS 5540 Spring 2013 Assignment 3, v1.0 Due: Apr. 24th 11:59PM 1 Introduction In this programming project, we are going to do a simple image segmentation task. Given a grayscale image with a bright object against a dark background and we are going to do a binary decision

More information

Topic: Local Search: Max-Cut, Facility Location Date: 2/13/2007

Topic: Local Search: Max-Cut, Facility Location Date: 2/13/2007 CS880: Approximations Algorithms Scribe: Chi Man Liu Lecturer: Shuchi Chawla Topic: Local Search: Max-Cut, Facility Location Date: 2/3/2007 In previous lectures we saw how dynamic programming could be

More information

Approximate MRF Inference Using Bounded Treewidth Subgraphs

Approximate MRF Inference Using Bounded Treewidth Subgraphs Approximate MRF Inference Using Bounded Treewidth Subgraphs Alexander Fix 1, Joyce Chen 1, Endre Boros 2, and Ramin Zabih 1 1 Cornell University, Computer Science Department, Ithaca, New York {afix,yuhsin,rdz}@cs.cornell.edu

More information

Paths, Flowers and Vertex Cover

Paths, Flowers and Vertex Cover Paths, Flowers and Vertex Cover Venkatesh Raman M. S. Ramanujan Saket Saurabh Abstract It is well known that in a bipartite (and more generally in a König) graph, the size of the minimum vertex cover is

More information

Lecture and notes by: Nate Chenette, Brent Myers, Hari Prasad November 8, Property Testing

Lecture and notes by: Nate Chenette, Brent Myers, Hari Prasad November 8, Property Testing Property Testing 1 Introduction Broadly, property testing is the study of the following class of problems: Given the ability to perform (local) queries concerning a particular object (e.g., a function,

More information

Markov Random Fields and Gibbs Sampling for Image Denoising

Markov Random Fields and Gibbs Sampling for Image Denoising Markov Random Fields and Gibbs Sampling for Image Denoising Chang Yue Electrical Engineering Stanford University changyue@stanfoed.edu Abstract This project applies Gibbs Sampling based on different Markov

More information

Discrete Optimization Methods in Computer Vision CSE 6389 Slides by: Boykov Modified and Presented by: Mostafa Parchami Basic overview of graph cuts

Discrete Optimization Methods in Computer Vision CSE 6389 Slides by: Boykov Modified and Presented by: Mostafa Parchami Basic overview of graph cuts Discrete Optimization Methods in Computer Vision CSE 6389 Slides by: Boykov Modified and Presented by: Mostafa Parchami Basic overview of graph cuts [Yuri Boykov, Olga Veksler, Ramin Zabih, Fast Approximation

More information

The Encoding Complexity of Network Coding

The Encoding Complexity of Network Coding The Encoding Complexity of Network Coding Michael Langberg Alexander Sprintson Jehoshua Bruck California Institute of Technology Email: mikel,spalex,bruck @caltech.edu Abstract In the multicast network

More information

Segmentation with non-linear constraints on appearance, complexity, and geometry

Segmentation with non-linear constraints on appearance, complexity, and geometry IPAM February 2013 Western Univesity Segmentation with non-linear constraints on appearance, complexity, and geometry Yuri Boykov Andrew Delong Lena Gorelick Hossam Isack Anton Osokin Frank Schmidt Olga

More information

Integer Programming ISE 418. Lecture 7. Dr. Ted Ralphs

Integer Programming ISE 418. Lecture 7. Dr. Ted Ralphs Integer Programming ISE 418 Lecture 7 Dr. Ted Ralphs ISE 418 Lecture 7 1 Reading for This Lecture Nemhauser and Wolsey Sections II.3.1, II.3.6, II.4.1, II.4.2, II.5.4 Wolsey Chapter 7 CCZ Chapter 1 Constraint

More information

A Tutorial Introduction to Belief Propagation

A Tutorial Introduction to Belief Propagation A Tutorial Introduction to Belief Propagation James Coughlan August 2009 Table of Contents Introduction p. 3 MRFs, graphical models, factor graphs 5 BP 11 messages 16 belief 22 sum-product vs. max-product

More information

Using Combinatorial Optimization within Max-Product Belief Propagation

Using Combinatorial Optimization within Max-Product Belief Propagation Using Combinatorial Optimization within Max-Product Belief Propagation John Duchi Daniel Tarlow Gal Elidan Daphne Koller Department of Computer Science Stanford University Stanford, CA 94305-9010 {jduchi,dtarlow,galel,koller}@cs.stanford.edu

More information

Vertex 3-colorability of claw-free graphs

Vertex 3-colorability of claw-free graphs Algorithmic Operations Research Vol.2 (27) 5 2 Vertex 3-colorability of claw-free graphs Marcin Kamiński a Vadim Lozin a a RUTCOR - Rutgers University Center for Operations Research, 64 Bartholomew Road,

More information

Planar Cycle Covering Graphs

Planar Cycle Covering Graphs Planar Cycle Covering Graphs Julian Yarkony, Alexander T. Ihler, Charless C. Fowlkes Department of Computer Science University of California, Irvine {jyarkony,ihler,fowlkes}@ics.uci.edu Abstract We describe

More information

Chapter 5 Graph Algorithms Algorithm Theory WS 2012/13 Fabian Kuhn

Chapter 5 Graph Algorithms Algorithm Theory WS 2012/13 Fabian Kuhn Chapter 5 Graph Algorithms Algorithm Theory WS 2012/13 Fabian Kuhn Graphs Extremely important concept in computer science Graph, : node (or vertex) set : edge set Simple graph: no self loops, no multiple

More information

Analysis: TextonBoost and Semantic Texton Forests. Daniel Munoz Februrary 9, 2009

Analysis: TextonBoost and Semantic Texton Forests. Daniel Munoz Februrary 9, 2009 Analysis: TextonBoost and Semantic Texton Forests Daniel Munoz 16-721 Februrary 9, 2009 Papers [shotton-eccv-06] J. Shotton, J. Winn, C. Rother, A. Criminisi, TextonBoost: Joint Appearance, Shape and Context

More information

The Structure of Bull-Free Perfect Graphs

The Structure of Bull-Free Perfect Graphs The Structure of Bull-Free Perfect Graphs Maria Chudnovsky and Irena Penev Columbia University, New York, NY 10027 USA May 18, 2012 Abstract The bull is a graph consisting of a triangle and two vertex-disjoint

More information

Measuring Uncertainty in Graph Cut Solutions - Efficiently Computing Min-marginal Energies using Dynamic Graph Cuts

Measuring Uncertainty in Graph Cut Solutions - Efficiently Computing Min-marginal Energies using Dynamic Graph Cuts Measuring Uncertainty in Graph Cut Solutions - Efficiently Computing Min-marginal Energies using Dynamic Graph Cuts Pushmeet Kohli Philip H.S. Torr Department of Computing, Oxford Brookes University, Oxford.

More information

Statistical and Learning Techniques in Computer Vision Lecture 1: Markov Random Fields Jens Rittscher and Chuck Stewart

Statistical and Learning Techniques in Computer Vision Lecture 1: Markov Random Fields Jens Rittscher and Chuck Stewart Statistical and Learning Techniques in Computer Vision Lecture 1: Markov Random Fields Jens Rittscher and Chuck Stewart 1 Motivation Up to now we have considered distributions of a single random variable

More information

Comparison of Graph Cuts with Belief Propagation for Stereo, using Identical MRF Parameters

Comparison of Graph Cuts with Belief Propagation for Stereo, using Identical MRF Parameters Comparison of Graph Cuts with Belief Propagation for Stereo, using Identical MRF Parameters Marshall F. Tappen William T. Freeman Computer Science and Artificial Intelligence Laboratory Massachusetts Institute

More information

Random projection for non-gaussian mixture models

Random projection for non-gaussian mixture models Random projection for non-gaussian mixture models Győző Gidófalvi Department of Computer Science and Engineering University of California, San Diego La Jolla, CA 92037 gyozo@cs.ucsd.edu Abstract Recently,

More information

Joint Entity Resolution

Joint Entity Resolution Joint Entity Resolution Steven Euijong Whang, Hector Garcia-Molina Computer Science Department, Stanford University 353 Serra Mall, Stanford, CA 94305, USA {swhang, hector}@cs.stanford.edu No Institute

More information

CSE 417 Network Flows (pt 3) Modeling with Min Cuts

CSE 417 Network Flows (pt 3) Modeling with Min Cuts CSE 417 Network Flows (pt 3) Modeling with Min Cuts Reminders > HW6 is due on Friday start early bug fixed on line 33 of OptimalLineup.java: > change true to false Review of last two lectures > Defined

More information

Markov/Conditional Random Fields, Graph Cut, and applications in Computer Vision

Markov/Conditional Random Fields, Graph Cut, and applications in Computer Vision Markov/Conditional Random Fields, Graph Cut, and applications in Computer Vision Fuxin Li Slides and materials from Le Song, Tucker Hermans, Pawan Kumar, Carsten Rother, Peter Orchard, and others Recap:

More information

Linear Programming Duality and Algorithms

Linear Programming Duality and Algorithms COMPSCI 330: Design and Analysis of Algorithms 4/5/2016 and 4/7/2016 Linear Programming Duality and Algorithms Lecturer: Debmalya Panigrahi Scribe: Tianqi Song 1 Overview In this lecture, we will cover

More information

Object Detection with Partial Occlusion Based on a Deformable Parts-Based Model

Object Detection with Partial Occlusion Based on a Deformable Parts-Based Model Object Detection with Partial Occlusion Based on a Deformable Parts-Based Model Johnson Hsieh (johnsonhsieh@gmail.com), Alexander Chia (alexchia@stanford.edu) Abstract -- Object occlusion presents a major

More information

Fast Approximate Energy Minimization via Graph Cuts

Fast Approximate Energy Minimization via Graph Cuts IEEE Transactions on PAMI, vol. 23, no. 11, pp. 1222-1239 p.1 Fast Approximate Energy Minimization via Graph Cuts Yuri Boykov, Olga Veksler and Ramin Zabih Abstract Many tasks in computer vision involve

More information

Variable Grouping for Energy Minimization

Variable Grouping for Energy Minimization Variable Grouping for Energy Minimization Taesup Kim, Sebastian Nowozin, Pushmeet Kohli, Chang D. Yoo Dept. of Electrical Engineering, KAIST, Daejeon Korea Microsoft Research Cambridge, Cambridge UK kim.ts@kaist.ac.kr,

More information

Curvature Prior for MRF-based Segmentation and Shape Inpainting

Curvature Prior for MRF-based Segmentation and Shape Inpainting Curvature Prior for MRF-based Segmentation and Shape Inpainting Alexander Shekhovtsov, Pushmeet Kohli +, Carsten Rother + Center for Machine Perception, Czech Technical University, Microsoft Research Cambridge,

More information

Theorem 2.9: nearest addition algorithm

Theorem 2.9: nearest addition algorithm There are severe limits on our ability to compute near-optimal tours It is NP-complete to decide whether a given undirected =(,)has a Hamiltonian cycle An approximation algorithm for the TSP can be used

More information

DISCRETE CONVEX ANALYSIS

DISCRETE CONVEX ANALYSIS DISCRETE CONVEX ANALYSIS o KAZUO MUROTA University of Tokyo; PRESTO, JST Tokyo, Japan Society for Industrial and Applied Mathematics Philadelphia List of Figures Notation Preface xi xiii xxi 1 Introduction

More information

Higher-order models in Computer Vision

Higher-order models in Computer Vision Chapter 1 Higher-order models in Computer Vision PUSHMEET KOHLI Machine Learning and Perception Microsoft Research Cambridge, UK Email: pkohli@microsoft.com CARSTEN ROTHER Machine Learning and Perception

More information

CRFs for Image Classification

CRFs for Image Classification CRFs for Image Classification Devi Parikh and Dhruv Batra Carnegie Mellon University Pittsburgh, PA 15213 {dparikh,dbatra}@ece.cmu.edu Abstract We use Conditional Random Fields (CRFs) to classify regions

More information

Image Segmentation with a Bounding Box Prior Victor Lempitsky, Pushmeet Kohli, Carsten Rother, Toby Sharp Microsoft Research Cambridge

Image Segmentation with a Bounding Box Prior Victor Lempitsky, Pushmeet Kohli, Carsten Rother, Toby Sharp Microsoft Research Cambridge Image Segmentation with a Bounding Box Prior Victor Lempitsky, Pushmeet Kohli, Carsten Rother, Toby Sharp Microsoft Research Cambridge Dylan Rhodes and Jasper Lin 1 Presentation Overview Segmentation problem

More information

Dr. Ulas Bagci

Dr. Ulas Bagci CAP- Computer Vision Lecture - Image Segmenta;on as an Op;miza;on Problem Dr. Ulas Bagci bagci@ucf.edu Reminders Oct Guest Lecture: SVM by Dr. Gong Oct 8 Guest Lecture: Camera Models by Dr. Shah PA# October

More information

Lecture notes on the simplex method September We will present an algorithm to solve linear programs of the form. maximize.

Lecture notes on the simplex method September We will present an algorithm to solve linear programs of the form. maximize. Cornell University, Fall 2017 CS 6820: Algorithms Lecture notes on the simplex method September 2017 1 The Simplex Method We will present an algorithm to solve linear programs of the form maximize subject

More information

Supplementary Material: Adaptive and Move Making Auxiliary Cuts for Binary Pairwise Energies

Supplementary Material: Adaptive and Move Making Auxiliary Cuts for Binary Pairwise Energies Supplementary Material: Adaptive and Move Making Auxiliary Cuts for Binary Pairwise Energies Lena Gorelick Yuri Boykov Olga Veksler Computer Science Department University of Western Ontario Abstract Below

More information

Mathematical and Algorithmic Foundations Linear Programming and Matchings

Mathematical and Algorithmic Foundations Linear Programming and Matchings Adavnced Algorithms Lectures Mathematical and Algorithmic Foundations Linear Programming and Matchings Paul G. Spirakis Department of Computer Science University of Patras and Liverpool Paul G. Spirakis

More information

On the Space-Time Trade-off in Solving Constraint Satisfaction Problems*

On the Space-Time Trade-off in Solving Constraint Satisfaction Problems* Appeared in Proc of the 14th Int l Joint Conf on Artificial Intelligence, 558-56, 1995 On the Space-Time Trade-off in Solving Constraint Satisfaction Problems* Roberto J Bayardo Jr and Daniel P Miranker

More information

Supplementary Material: Unconstrained Salient Object Detection via Proposal Subset Optimization

Supplementary Material: Unconstrained Salient Object Detection via Proposal Subset Optimization Supplementary Material: Unconstrained Salient Object via Proposal Subset Optimization 1. Proof of the Submodularity According to Eqns. 10-12 in our paper, the objective function of the proposed optimization

More information

Notes on Minimum Cuts and Modular Functions

Notes on Minimum Cuts and Modular Functions Notes on Minimum Cuts and Modular Functions 1 Introduction The following are my notes on Cunningham s paper [1]. Given a submodular function f and a set S, submodular minimisation is the problem of finding

More information

Beyond Pairwise Energies: Efficient Optimization for Higher-order MRFs

Beyond Pairwise Energies: Efficient Optimization for Higher-order MRFs Beyond Pairwise Energies: Efficient Optimization for Higher-order MRFs Nikos Komodakis Computer Science Department, University of Crete komod@csd.uoc.gr Nikos Paragios Ecole Centrale de Paris/INRIA Saclay

More information

Faster parameterized algorithms for Minimum Fill-In

Faster parameterized algorithms for Minimum Fill-In Faster parameterized algorithms for Minimum Fill-In Hans L. Bodlaender Pinar Heggernes Yngve Villanger Technical Report UU-CS-2008-042 December 2008 Department of Information and Computing Sciences Utrecht

More information

Learning Low-order Models for Enforcing High-order Statistics

Learning Low-order Models for Enforcing High-order Statistics Patrick Pletscher ETH Zurich Zurich, Switzerland Pushmeet Kohli Microsoft Research Cambridge, UK Abstract Models such as pairwise conditional random fields (CRFs) are extremely popular in computer vision

More information

Discrete Optimization of Ray Potentials for Semantic 3D Reconstruction

Discrete Optimization of Ray Potentials for Semantic 3D Reconstruction Discrete Optimization of Ray Potentials for Semantic 3D Reconstruction Marc Pollefeys Joined work with Nikolay Savinov, Christian Haene, Lubor Ladicky 2 Comparison to Volumetric Fusion Higher-order ray

More information

Energy Minimization for Segmentation in Computer Vision

Energy Minimization for Segmentation in Computer Vision S * = arg S min E(S) Energy Minimization for Segmentation in Computer Vision Meng Tang, Dmitrii Marin, Ismail Ben Ayed, Yuri Boykov Outline Clustering/segmentation methods K-means, GrabCut, Normalized

More information

Combinatorial optimization and its applications in image Processing. Filip Malmberg

Combinatorial optimization and its applications in image Processing. Filip Malmberg Combinatorial optimization and its applications in image Processing Filip Malmberg Part 1: Optimization in image processing Optimization in image processing Many image processing problems can be formulated

More information

Exact Algorithms Lecture 7: FPT Hardness and the ETH

Exact Algorithms Lecture 7: FPT Hardness and the ETH Exact Algorithms Lecture 7: FPT Hardness and the ETH February 12, 2016 Lecturer: Michael Lampis 1 Reminder: FPT algorithms Definition 1. A parameterized problem is a function from (χ, k) {0, 1} N to {0,

More information

University Alexandru Ioan Cuza of Iaşi Faculty of Computer Science. A simple push-relabel algorithm for sub-modular flows

University Alexandru Ioan Cuza of Iaşi Faculty of Computer Science. A simple push-relabel algorithm for sub-modular flows University Alexandru Ioan Cuza of Iaşi Faculty of Computer Science T E C H N I C A L R E P O R T A simple push-relabel algorithm for sub-modular flows Cristian Frasinaru, Emanuel F. Olariu TR 13-02, October

More information

FOUR EDGE-INDEPENDENT SPANNING TREES 1

FOUR EDGE-INDEPENDENT SPANNING TREES 1 FOUR EDGE-INDEPENDENT SPANNING TREES 1 Alexander Hoyer and Robin Thomas School of Mathematics Georgia Institute of Technology Atlanta, Georgia 30332-0160, USA ABSTRACT We prove an ear-decomposition theorem

More information

Estimating Human Pose in Images. Navraj Singh December 11, 2009

Estimating Human Pose in Images. Navraj Singh December 11, 2009 Estimating Human Pose in Images Navraj Singh December 11, 2009 Introduction This project attempts to improve the performance of an existing method of estimating the pose of humans in still images. Tasks

More information

FMA901F: Machine Learning Lecture 6: Graphical Models. Cristian Sminchisescu

FMA901F: Machine Learning Lecture 6: Graphical Models. Cristian Sminchisescu FMA901F: Machine Learning Lecture 6: Graphical Models Cristian Sminchisescu Graphical Models Provide a simple way to visualize the structure of a probabilistic model and can be used to design and motivate

More information

A CSP Search Algorithm with Reduced Branching Factor

A CSP Search Algorithm with Reduced Branching Factor A CSP Search Algorithm with Reduced Branching Factor Igor Razgon and Amnon Meisels Department of Computer Science, Ben-Gurion University of the Negev, Beer-Sheva, 84-105, Israel {irazgon,am}@cs.bgu.ac.il

More information

Data-driven Depth Inference from a Single Still Image

Data-driven Depth Inference from a Single Still Image Data-driven Depth Inference from a Single Still Image Kyunghee Kim Computer Science Department Stanford University kyunghee.kim@stanford.edu Abstract Given an indoor image, how to recover its depth information

More information

Figure : Example Precedence Graph

Figure : Example Precedence Graph CS787: Advanced Algorithms Topic: Scheduling with Precedence Constraints Presenter(s): James Jolly, Pratima Kolan 17.5.1 Motivation 17.5.1.1 Objective Consider the problem of scheduling a collection of

More information

1. Lecture notes on bipartite matching

1. Lecture notes on bipartite matching Massachusetts Institute of Technology 18.453: Combinatorial Optimization Michel X. Goemans February 5, 2017 1. Lecture notes on bipartite matching Matching problems are among the fundamental problems in

More information

An ILP Solver for Multi-label MRFs with Connectivity Constraints

An ILP Solver for Multi-label MRFs with Connectivity Constraints An ILP Solver for Multi-label MRFs with Connectivity Constraints arxiv:1712.06020v2 [cs.cv] 20 Mar 2018 Ruobing Shen 1, Eric Kendinibilir 2, Ismail Ben Ayed 3, Andrea Lodi 4, Andrea Tramontani 5, and Gerhard

More information

Outline: Finish uncapacitated simplex method Negative cost cycle algorithm The max-flow problem Max-flow min-cut theorem

Outline: Finish uncapacitated simplex method Negative cost cycle algorithm The max-flow problem Max-flow min-cut theorem Outline: Finish uncapacitated simplex method Negative cost cycle algorithm The max-flow problem Max-flow min-cut theorem Uncapacitated Networks: Basic primal and dual solutions Flow conservation constraints

More information

Algorithm and Complexity of Disjointed Connected Dominating Set Problem on Trees

Algorithm and Complexity of Disjointed Connected Dominating Set Problem on Trees Algorithm and Complexity of Disjointed Connected Dominating Set Problem on Trees Wei Wang joint with Zishen Yang, Xianliang Liu School of Mathematics and Statistics, Xi an Jiaotong University Dec 20, 2016

More information

An Introduction to LP Relaxations for MAP Inference

An Introduction to LP Relaxations for MAP Inference An Introduction to LP Relaxations for MAP Inference Adrian Weller MLSALT4 Lecture Feb 27, 2017 With thanks to David Sontag (NYU) for use of some of his slides and illustrations For more information, see

More information

Dual-Based Approximation Algorithms for Cut-Based Network Connectivity Problems

Dual-Based Approximation Algorithms for Cut-Based Network Connectivity Problems Dual-Based Approximation Algorithms for Cut-Based Network Connectivity Problems Benjamin Grimmer bdg79@cornell.edu arxiv:1508.05567v2 [cs.ds] 20 Jul 2017 Abstract We consider a variety of NP-Complete network

More information

Spectral Clustering and Community Detection in Labeled Graphs

Spectral Clustering and Community Detection in Labeled Graphs Spectral Clustering and Community Detection in Labeled Graphs Brandon Fain, Stavros Sintos, Nisarg Raval Machine Learning (CompSci 571D / STA 561D) December 7, 2015 {btfain, nisarg, ssintos} at cs.duke.edu

More information

The Min Norm Point on a Convex Polytope

The Min Norm Point on a Convex Polytope The Min Norm Point on a Convex Polytope Jamie Haddock January 13, 2018 Graduate Group in Applied Mathematics UC Davis joint with Jesús De Loera and Luis Rademacher https://arxiv.org/abs/1710.02608 1 Minimum

More information