Filter Flow. Steven M. Seitz University of Washington. Simon Baker. Abstract. 1. Introduction. at Microsoft Research.

Size: px
Start display at page:

Download "Filter Flow. Steven M. Seitz University of Washington. Simon Baker. Abstract. 1. Introduction. at Microsoft Research."

Transcription

1 Filter Flow Steven M. Seitz University of Washington Simon Baker Microsoft Research Abstract The filter flow problem is to compute a space-variant linear filter that transforms one image into another. This framework encompasses a broad range of transformations including stereo, optical flow, lighting changes, blur, and combinations of these effects. Parametric models such as affine motion, vignetting, and radial distortion can also be modeled within the same framework. All such transformations are modeled by selecting a number of constraints and objectives on the filter entries from a catalog which we enumerate. Most of the constraints are linear, leading to globally optimal solutions (via linear programming) for affine transformations, depth-from-defocus, and other problems. Adding a (non-convex) compactness objective enables solutions for optical flow with illumination changes, spacevariant defocus, and higher-order smoothness. 1. Introduction How do two photos of the same scene differ? The camera or objects in the scene may have moved, producing optical flow. The lighting, shading, or shadows may have changed, transforming the pixel intensities. The camera or camera parameters may be different, modifying the amount of (space-varying) defocus, and altering exposure, zoom, color balance, vignetting, etc. In this paper we introduce a framework called filter flow for modeling a broad class of image transformations, encompassing all of the aforementioned effects. The key idea is to model image transformations as a space-variant linear filter. Like a convolution, each pixel in the first image is combined with its neighborhood to produce the pixel at the same coordinates in the second image. The difference is that the filter kernel (matrix of weights) may vary from pixel to pixel. As shown in Figure 1, this model supports motion (optical flow), intensity changes, blur, and combinations of these effects. In addition to being very general, filter flow is attractive because the transformation is linear. Hence, the in- This work was conducted while the first author was a Consulting Researcher at Microsoft Research. F1: move up 1 F: move + darken.5 input image F3: move + blur filter flow output image Figure 1. Filter flow generalizes optical flow by applying a spacevariant filter at each pixel, allowing very general transformations. verse problem of computing a transformation (i.e., perpixel kernels) from the two images is also linear (e.g., via pseudo-inverse). The inverse problem, however, is massively underconstrained there are many different transformations that produce a given image pair. Computing the correct transformation therefore requires identifying additional constraints on the filters that enable a unique solution without sacrificing too much expressive power. This paper presents a catalog of constraints and objectives that can be placed on the filters to solve a wide variety of computer vision problems ranging from image alignment to shape from defocus. A particular algorithm is designed by selecting a subset of these constraints and objectives. Most of the constraints and objectives are linear. Any combination of the linear constraints and objectives can be solved to obtain the globally optimal solution using Linear Programming (LP). When non-linear constraints are added, the problem can be solved using a short sequence of LPs. After describing the framework, we present five case studies. In the first we present a linear algorithm for computing a global affine transformation. In the second we introduce a stereo algorithm that integrates higher order smoothness priors such as nd order smoothness [] and pixelwise affine smoothness. In the third we show how filter flow is used to perform stereo between a pair of grey-scale images with substantial illumination variation. By imposing MRF smoothness on both the disparities and the illumination variation, we are able to recover estimates of both. In the fourth case study we show how to compute optical flow using the filter flow framework. We present results

2 on the Middlebury Flow benchmark [1] and show our algorithm to be reasonably competitive with the state of the art in terms of flow accuracy. In terms of interpolation accuracy our algorithm scores second overall, at the time of submission. Adding an illumination variation model significantly improves performance on benchmark images that contain moving shadows and other photometric effects. In the final case study we consider the problem of estimating space-variant image defocus from a pair of images. When the images are aligned, our formulation yields a globally optimal algorithm to compute the defocus kernels (whose widths correlate to depth). We then show how to model both motion and defocus, resulting in an alignment algorithm that is robust to space-varying defocus. Our algorithm is an important step toward the practical use of depth-from-defocus in less controlled settings.. Space-variant Linear Filters We consider the class of transformations that can be written as an N N matrix T that operates on an image I 1 represented as a vector of N pixels to yield an image I : I = TI 1. (1) Each row of T defines a linear combination of pixels in I 1 that produces a single pixel at the appropriate position in I. Another way to write Eq. (1) is in terms of filtering operations rather than a single large matrix multiplication. Specifically, denote an image I(u, v) where u = (u, v) denotes pixel location. At each pixel, we define a filter T u ij that operates on a region of the image centered at pixel u: I (u, v) = T u iji 1 (u + i, v + j). () Note that this is a double summation, and the domains i [ k 1, k ] and j [ k 3, k 4 ] are omitted to simplify notation. Eq. () defines a space-variant linear filter. Spacevariant linear filters, well-known in the shape-from-defocus literature (e.g., [4,1]), can represent a broad class of transformations including optical flow, lighting changes, blur, and their combinations (see Figure 1)..1. Factored Filters Rather than treat the transformations atomically, it is useful to separate the geometric and photometric components. We thus factor the transformations into a composition of a space-variant motion with a space-variant filtering step: each pixel in I 1 is filtered with its neighborhood and then moved to the appropriate location in I. Formally, T is factored into a product of a motion M and a kernel K: T = MK. (3) In the case of a pure motion, K is the identity matrix. Similarly, if there is no motion, M is the identity. The transformation I = MKI 1 can be rewritten: M 1 I = KI 1. (4) Note that M 1 encodes the motion from I to I 1 instead of from I 1 to I. The advantage of Eq. (4) is that it yields equations that are linear with respect to the unknowns of M 1 and K. We therefore solve for inverse motion, and redefine M to be motion from I to I 1, yielding: MI = KI 1. (5) As with T, we can alternatively define K and M as filters Kij u and M ij u and rewrite Eq. (5) as: M u iji (u + i, v + j) = K u iji 1 (u + i, v + j). (6) Note that the filter kernels Kij u and M ij u need not be the same size; e.g., we may want to allow large motions but limit the amount of blur. Eq. (6) can then be converted into a linear (L1) Data Objective: MijI u (u + i, v + j) KijI u 1 (u + i, v + j) 1 (DO) This equation is hugely underconstrained. To solve for the motion and kernel filters, additional constraints are need. In the next two sections we describe a catalog of possibilities. 3. Catalog of Motion Models We first consider constraints on the motion filter. M u ij Representing Motion We can represent (integer-precision) optical flow in a motion filter Mij u by the constraint that, for each pixel u, exactly one element of Mij u is 1 and the rest are 0. We can state this constraint alternatively using the following three constraints on Mij u. The first is Non-Negativity: M u ij 0 i, j. (POS-M) (We use the name (POS-M) to refer to this non-negativity constraint) The second is Sum-To-One: Mij u = 1. (SUM1-M) The third is Compactness: [i j] T M u = 0 (CPSTRICT-M) M u ij

3 where is the L norm and the centroid is defined: M u = [i j] T M u ij. (11) For each pixel u in I 1, note that the centroid M u encodes u s displacement in I. I.e., the centroid represents optical flow. Eqs. (CPSTRICT-M) and (POS-M) require M u ij to be zero everywhere except the centroid, and Eq. (SUM1-M) requires the filter to be 1 at the centroid. The compactness constraint (CPSTRICT-M) is too strict for our purposes in that it does not allow subpixel motions. It is also relatively difficult to enforce in an optimization framework. Hence, we relax Eq. (CPSTRICT-M) with a Compactness Objective that we seek to minimize: M u ij [i j] T M u. (CP-M) Eq. (CP-M) imposes compactness in a soft manner, the weight of which can be regulated, and can be optimized iteratively as described in Section 5. We also consider an even softer, Soft Compactness Objective: [ ( max 0, [i j] T M )] u c (SCP-M) M u ij where c is a small constant (c = 1 in our experiments.) Eq. (SCP-M) is softer in the sense that it does not affect any pixels which are less than L distance c from the centroid M u, useful for accurately modeling sub-pixel motions. Applying Eqs. (POS-M), (SUM1-M) and (CP-M) constrain M u to approximate a pure flow field. Solving for a unique motion, however, requires adding constraints or objectives on the form of the motion, as described next. 3.. Global Motion Models 3..1 Global translations The motion must be a global translation if we constrain the centroid of every filter to be the same as the first one using the Global Translation Constraints: M u = M 0 u. (XLATE-M) The combination of (XLATE-M), (POS-M), (SUM1-M) leads to a linear algorithm to estimate a global translation. 3.. Global affine motion To constrain the motion to be a global affine transformation, we introduce the unknown affine parameters as a single 3 matrix A. The Global Affine Constraints are then: A[u v 1] T = M u u. (GLOBA-M) (a) Centroid Smoothness (c) Affine Smoothness (b) nd Order Smoothness (d) Filter Smoothness Figure. Comparison of MRF smoothness terms. nd order smoothness produces smoother disparities compared to centroid smoothness, but over-smooths discontinuities. Affine smoothness yields both smooth disparities and relatively clean discontinuities. Filter smoothness is not quite as smooth as affine, but does best at discontinuities, due to the implicit truncation. These were computed using the filter flow stereo approach presented in Section 6.. The affine approach is easily generalized to polynomial spatial transformations (e.g., radial lens distortion), where the affine matrix is replaced by a K matrix P of unknown coefficients, and [u v 1] is replaced with the vector of all (a priori known) K monomial terms u a v b Local Motion (MRF) Models Local motion models (i.e., spatial regularization or smoothness) are also straightforward to enforce. In particular, we support Markov Random Field (MRF) smoothness terms that are popular in optical flow and stereo. Here we consider only linear (L1) MRF objectives. Note that while this restriction eliminates useful models like error truncation [19], it also trivially enables second-order smoothness and other higher order linear relations Centroid smoothness A standard smoothness condition used in optical flow and stereo is that the motion/disparity of a pixel should be similar to that of its neighbors. The L1 form of this objective in filter flow is the Centroid Smoothness Objective: M u M u 1 (MRF1-M) u u N(u) where N(u) is the set of neighbors of a pixel u (our experiments employ 4-connectivity). A known limitation of Eq. (MRF1-M), is that it tends to produce piece-wise constant flow fields (see Figure (a)). In the context of stereo, it

4 has a fronto-parallel bias, which penalizes surfaces that are not parallel to the image plane nd order smoothness Penalizing the nd derivative of flow rather than first derivative avoids this fronto-parallel bias []. In filter flow, the nd Order Smoothness Objective is defined as follows: M u M u + M u 1 (MRF-M) u,u,u where the sum is over all sets of three consecutive horizontal or three consecutive vertical pixels. Results are shown in Figure (b). Eq. (MRF-M) is much harder to optimize in the context of graph cuts because it introduces trinary terms (cliques of size 3), and authors have therefore proposed more complex solution methods such as QPBO and fusion moves []. Observe, however, that (MRF-M) is a linear (L1) objective and is therefore straightforward to optimize within the Filter Flow framework Affine smoothness While second-order smoothness produces very smooth flow fields, we have observed that it significantly over-smooths discontinuities (see Figure (b)). We obtain better results using affine smoothness [14], in which an explicit sixparameter affine transformation is estimated at each pixel, and the affine parameters are regularized within a linear (L1) MRF. I.e., the affine transformations may be different at each pixel, but encouraged to vary smoothly. A key advantage of this approach is that it leads to a pairwise (rather than trinary) objective, and is therefore better behaved around discontinuities (see Figure (c)). We introduce a per-pixel 3 matrix A u and introduce Local Affine Constraints: A u [u v 1] T = M u u. (LOCA-M) We also introduce an Affine Smoothness Objective: A u A u 1. (MRFA-M) u u N(u) Note that affine smoothness would be very difficult to solve using existing state-of-the-art MRF solvers as it depends on having hard constraints and continuous variables for A Filter smoothness Another option is to encourage neighboring pixels to have similar filters, rather than similar flow fields. This is accomplished via the following Filter Smoothness Objective: M u M u 1 (MRFF-M) u u N(u) Eq. (MRFF-M) approximates the Potts model [3], becoming exact when M u is binary with a single 1. (If the 1s are in different places, MRFF-M is always 1.) Filter smoothness has a fronto-parallel bias, but the error truncation of the Potts model better preserves boundaries (see Figure (d)). Note, however, that Eq. (MRFF-M) also has more objective terms and is significantly (around a factor of 10 on the results in Figure ) more expensive to optimize compared to the above alternatives Edge/Segment weighted smoothness All of the above local smoothness objectives weight each edge in a neighborhood equally. This uniform weighting can be generalized to a gradient- or segment-based [19] weighting. We use a segment-based weighting in our optical flow results. See Section 6.4 for the details. All the other experiments use uniform weighting. 4. Catalog of Kernel Models We now consider the kernel filter. This list is not comprehensive, but includes models used in our case studies, and a few others that give a broader sense of what is possible. Since motion is already encoded in M u, we can enforce the constraint that the kernel filter has zero centroid using the Zero-Centroid Constraints: [i j] T Kij u = 0 u. (0CENT-K) In some scenarios (e.g., when we know there is no sharpening ), we can also enforce Non-Negativity on the kernel: K u ij 0 i, j. (POS-K) A Sum-To-One constraint may also be appropriate in cases (e.g., for a pure blur): Kij u = 1. (SUM1-K) Eqs. (0CENT-K), (POS-K), and/or sum-to-one (SUM1-K) are insufficient to constrain the problem sufficiently to obtain a unique solution. Solving for a unique set of filters generally requires adding constraints or objectives on the spatial variation of the filters, as described next Global kernel models A space-invariant kernel, i.e., convolution, is enforced by constraining all filters to be the same. We call this constraint the Convolution Constraint: K u = K 0 u. (CONV-K) More generally, we can constrain the variation of each entry of the kernel to satisfy a global parametric form. For

5 example, the vignetting model from [6] (assuming vignette origin at [0 0]) is expressed as a Vignette Constraint: V[ u 6 u 4 u ] T + 1 = ij K u ij u (VIG-K) where V is a global vector of coefficients. Eq. (VIG-K) is linear in all unknowns (V and K). 4.. MRF kernel smoothness For many scenes, we may assume that the kernel varies smoothly throughout most of the image, i.e., neighboring pixels have similar kernels. Accordingly, we define the L1 MRF Kernel Smoothness Objective: Kij u K u. (MRF-K) 1 u N(u) 4.3. Parameterizing Symmetric Kernels Many filters exhibit bilateral or rotational symmetry. Although it is possible to formulate constraints on a kernel filter Kij u to enforce symmetry, it is more efficient to reparameterize the filter to eliminate the redundant degrees of freedom. For rotational symmetry, we represent a 1D radial slice of the filter as a vector ki u. We then express each element of the D filter Kij u as a linear combination of the 1D filter entries ki u. More specifically, Ku ij = ku d, where d = i + j (sub-pixel values of d require interpolating the vector k). Hence, we can represent a K K filter with approximately K/ parameters. Note that rotationally symmetric filters automatically have zero centroid. 5. Optimization Sections 3 and 4 enumerate a catalog of constraints and objectives. Almost all of these expressions are either linear, or are L1-norms of linear expressions which can converted to a set of linear expressions through the introduction of a dummy variable [11]. The only exceptions are compactness (CP-M) and soft compactness (SCP-M). In the absence of these nonlinear objectives, any (linear) combination of the constraints and objectives is linear, convex, and can be optimized using Linear Programming (LP). In our experiments we used both MOSEK s interior point solver and an internal Microsoft solver also based on an interior point method. We found both algorithms to perform similarly, and both performed significantly better than the builtin Matlab LP solver, linsolve. When either compactness or soft compactness is added, we solve a sequence of linear programs, each obtained by linearizing the non-linear term(s) about the previous solution. Empirically, this procedure converges after 3 4 iterations. We initialize the algorithm by omitting the nonlinear ij objective(s) and solving the resulting LP. The derivative of the compactness term (CP-M) with respect to M u ij is: D u ij = a,b M u ab [i j] T M u [ i ( a c,d cm cd u ) ( + j b )] c,d dm cd u. Note that if sum to one (SUM1-M) is enforced, the expression in the second line above reduces to 0. Dij u is computed using the estimate of Mij u in the previous iteration. The Linearized Compactness Objective: (CPL-M) M u ijd u ij is then added to the LP and solved. A similar derivation can be performed for soft compactness (SCP-M). Note that in the case of (CP-M), as opposed to (SCP-M), the procedure is a Concave-Convex Program which guarantees convergence to a local minimum or saddle point [3]. For moderately large motion filters the LPs can become extremely large and no longer fit in memory. For an N N image with an m m motion filter and a k k kernel filter, there are O((k + m )N ) unknowns. In addition, a reasonable initialization is required when using either of the non-convex compactness objectives (CP-M) or (CP- M). We therefore employ a coarse-to-fine strategy, via a pyramid. Each level of a traditional Image Pyramid is constructed by blurring and then downsampling the previous level. A second variant, which we call a Filter Pyramid, also blurs the image, but instead of downsampling, we double the sampling interval of a filter, e.g., replacing a filtering operation such as Eq. () with: TijI u 1 (u + l i, v + l j). (8) at level l in the pyramid. The advantage of using a Filter Pyramid is that we estimatxke the filter at each pixel in the original image. When the results are propagated down the pyramid there is therefore less blurring over edges. We use pyramids as needed based on problem complexity. No pyramids are used for affine alignment (Section 6.1) and higher order constraints on stereo (Section 6.). For stereo with illumination change (Section 6.3) and defocus with motion (Section 6.5) we only use filter pyramids. For optical flow (Section 6.4) we use a -level image pyramid. For each level of the image pyramid we use a 3-level filter pyramid. 6. Case Studies 6.1. Affine Alignment: A Convex Formulation To compute a global affine motion, we impose nonnegativity (POS-M), sum to one (SUM1-M), and the global

6 affine constraint (GLOBA-M) on the motion filter. The kernel filter is set to be the identity. The only objective is the data term (DO). Hence, we optimize: arg min (DO) s.t. (POS-M), (SUM1-M), (GLOBA-M), a single LP involving only linear constraints, objectives, and L1 norms. To confirm our LP solution is correct we synthetically warped a pixel image with an affine warp. The motions ranged from 6.1 to 5.5 pixels in x and 7.5 to 7.0 pixels in y. We used a pixel filter. The resulting LP has 8.7M unknowns (including about 100K dummy variables) and was solved in approximately 165 minutes. All timing results are computed on a x Quad Core.5Ghz Xeon. No pyramid was used. The solution is essentially perfect, up to interpolation errors, with a maximum error across the entire image of 0.08 pixels (at one of the corners, the errors in the middle are far less.) Our approach is unique in that it provides a globally optimal solution for affine motion, via linear programming. In contrast, other known techniques are nonlinear even for linear motion models. E.g., Bergen et al. [] linearize the image in each iteration. Feature-based methods are linear, but assume known feature correspondence. (a) input image (c) true depth (e) disparity with illum. (b) input, different illumination (d) true shading change (f) estimated shading change 6.. Higher Order Priors on Stereo For dense stereo, we replace the global affine model with a local smoothness model, and restrict each motion filter to a 1D row vector whose width is the disparity search range. We found that adding the compactness objective (CP-M) improves results, particularly for larger images and disparity search ranges. For centroid smoothness we optimize: arg min (DO) + α (MRF1-M) + β (CP-M) (9) s.t. (POS-M), (SUM1-M) (30) where α = 1.0 and β = 4.0 are weights that regulate the importance of the three objectives. For nd order smoothness we simply replace (MRF1-M) with (MRF-M). For affine smoothness, we replace (MRF1-M) with (MRFA- M) and add the local affine constraints (LOCA-M). For filter smoothness we use (MRFF-M). Figure compares centroid smoothness, nd order smoothness, affine smoothness, and filter smoothness on the half size (17 191) Venus images from [19] Stereo with Illumination Change Illumination changes between images are multiplicative and can be modeled as a per-pixel intensity scale, represented as a 1 1 kernel K u in filter flow. To add the kernel filter to stereo we just add the kernel smoothness constraint γ(mrf-k) to Eq. (9) which encourages the intensity scales to vary smoothly. In our experiments we used (g) disparity without illum. Figure 3. Stereo with illumination changes. (d) shows the shading changes (ratio of intensities) between images (a) and (b), as a result of moving the light source. Without illumination compensation, stereo performs poorly (g). Adding an MRF over shading changes greatly improves results (e), and the estimated shading (f) closely match (a regularized version of) the ground truth (d). α = 5.0, β = 1.0, and γ = 160. The results, shown in Figure 3, are computed on 1/6th size size (13 185) Rocks1 images from [?]. See Figure 3 for a description and this paper s supplemental material for more examples. Prior work has addressed illumination changes in stereo by use of color invariants [8] or robust matching metrics (see [?] for an excellent overview). While these techniques can be very effective, they don t exploit the property that shading changes tend to be smooth with isolated discontinuities (object and shadow borders). Negahdaripour [13] therefore proposed to regularize intensity changes in a flow algorithm; however, his approach assumed constant parameters over a local window. Our MRF formulation globally regularizes both illumination and depth, and also produces a regularized shading change image. See Figure 3(f) for the computed shading change Optical Flow Optical flow is formulated in the same manner as stereo, but with D motion filters. We found that soft compactness (SCP-M) produced slight more accurate subpixel flow es-

7 timates than compactness (CP-M). We also obtained better results on the Middlebury Flow benchmark [1] by adding a 1 1 intensity-scale kernel (same scale applied to all three color channels) to model illumination changes (as in Section 6.3). In particular, intensity-scaling better handled moving shadows and other photometric effects that occur on some images in the Middlebury test set [1] (see the supplemental material for numerical results and comparison images). Note that other authors have integrated illumination modeling in optical flow (e.g., [7, 13, 0]) but did not regularize these effects in an MRF. We optimize: (DO) + α (SCP-M) + β (MRFA-M) + γ (MRF-K) s.t. (POS-M), (SUM1-M), (LOCA-M). We set α = 1.0, γ = 160, and β = 3.5 but also used segment based smoothness weighting where we multiply β by.5 within segments and up to.5 between segments depending on the average color difference; if the average colors are identical, β is multiplied by.5. We used a meanshift-based over-segmentation algorithm, and a -level image pyramid, with a 3-level filter pyramid at each level of the image pyramid. For the Urban sequence (which has one of the largest motions) we used a 8 3 pixel filter at the top level. The largest LPs at the lowest level in the image pyramid have approximately 9.M unknowns (including around 4.6M dummy variables). The running time with the kernel filter is around hrs, and 9hrs without. The results with the kernel filter are available on the Middlebury benchmark webpage [1]. In terms of flow accuracy, our algorithm is reasonably competitive with the state-ofthe-art. Our results are the amongst the best to date on the difficult Urban sequence. In terms of interpolation accuracy, our algorithm is within the top two algorithms at the time of submission. Our algorithm appears to be the only one that correctly renders the orange ball in the Backyard sequence. Please see supplemental results for more details Defocus Given a pair of images taken at different aperture or focus settings, we wish to estimate both the motion and the per-pixel defocus kernel. When the images are pre-aligned our framework provides a convex MRF formulation for the defocus problem. We set the motion filter to be the identity, use a radial parameterization of K u (see Section 4.3), and apply MRF kernel smoothness (MRF-K): arg min (DO) + α (MRF-K) s.t. (SUM1-K) (31) where α = 7.0. We note that while several linear solutions to defocus have been proposed [4,5,15,1], they generally assume that the scene is locally-planar so that defocus is constant over a local window. Our MRF formulation avoids this assumption, and allows global regularization. Input (one of two) Kernel Filter Width Figure 4. Defocus result on an image pair from [1] using our linear MRF-based formulation, showing estimated kernel width (W ) for each pixel. Light values (small W ) correspond to sharpening kernels, where dark values (large W ) are blur kernels. We note that [1, 17] also proposed MRF frameworks for shape-from-defocus, but do not ensure a global optimum. As our approach is non-parametric, it does not provide a direct measurement of depth (see [4] for a method to calibrate depth). For the purposes of visualization, we define the squared width of the kernel as: W (Kij u ) = ij ( i + j )Ku ij based on the fact that for a Gaussian blur G(σ) this quantity is proportional to σ. Because kernel width is related to depth, it provides a compelling depth visualization. One result using our linear algorithm is shown in Figure 4. This result is significantly cleaner than prior published depth from defocus results see supplemental material for more results and comparisons. Almost all work in depth-from-defocus assumes that the images are pre-aligned (a notable exception is [18]). Our linear formulation can be extended to include motion simply by adding terms from the motion catalog to Eq. (31) This allows us to reliably align sequences with substantial amounts of blur. In particular we conducted experiments to solve for a global affine transform to model the zoom between two images captured with different focus settings by adding the global affine constraint (GLOBA-M) to Eq. (31). Figure 5 shows results. The input consists of one image focused on the front part of the carving, and the other on the back. The input images were captured with a 10MPixel digitial still camera in JPG and downsampled by a factor of 10 using the Lanczos algorithm. Figure 5(c) includes the affine motion, which is roughly a zoom. Due to presence of some blur in the motion filters and subpixel interpolation in the kernel filter, we obtained better results by first using defocus-invariant affine estimation (Eq. (31) and (GLOBA-M)) to warp one image and then use our convex solution in Eq. (31). Our radially parameterized 5 5 kernel filter has 3 unknowns at distance 0, 1, and. Pixels with distance > are coded to be zero. While we do not recover metric depth (although this may be possible [4]), The W plot (Figure 5(b)) qualitatively captures scene depth as the object recedes to the right. Note that in parts of the image, the recovered filter is a blur, and in other parts the filter is a sharpening filter with negative entries and negative W (Figure 5(d)).

8 (a) Input image (1 of ) (c) flow (color-coded) (b) Filter width (bright = large) blur kernel sharpen kernel (d) computed kernels Figure 5. Two images (a) at different focus settings yield an affine flow field (c) and symmetric per-pixel kernels whose widths correlate to scene depth (b). The kernel profiles (d) correctly encode a blur to the left and a sharpen to the right. 7. Conclusions This paper introduced filter flow, a framework for modeling a broad class of image transformations via space-variant filters. A major contribution of the filter flow framework, beyond its generality, is that it introduces new ways of attacking traditional motion and deconvolution problems. For example, reparameterizing optical flow from a single pair of continuous unknowns (f x, f y ) to an m m motion filter leads to an optimization problem that may be solved via linear programming. While there is prior work on relaxing discrete MRFs using continuous variables [10, 16], recasting the problem into filters yields new capabilities. Filter flow has a number of advantages over prior optical flow methods, such as enabling 1) linear but non-submodular objectives (e.g., nd order smoothness), ) global constraints (e.g., affine transformations), and 3) spatially-varying intensity changes and blur. Finally, we show how it is possible to derive globally optimal (convex) solutions for interesting classes of D motion and defocus problems. A major limitation is speed most results require several hours to compute. Clearly more research is needed to improve run-time. Incorporating nonlinear metrics (e.g., truncated linear error) and constraints is another important topic of future work. And finally, it would be interesting to cast several other problems within the filter flow framework, including parametric registration and fitting problems, active appearance models, and motion deconvolution. Acknowledgement: we thank Sameer Agarwal for invaluable guidance on both formulation and optimization issues. References [1] S. Baker, D. Scharstein, J. Lewis, S. Roth, M. Black, and R. Szeliski. A database and evaluation methodology for optical flow. In Proc. ICCV, 007. [] J. Bergen, P. Anandan, K. Hanna, and R. Hingorani. Hierarchical model-based motion estimation. In Proc. ECCV, pages 37 5, 199. [3] Y. Boykov, O. Veksler, and R. Zabih. Fast approximate energy minimization via graph cuts. PAMI, 3:001, 001. [4] J. Ens and P. Lawrence. An investigation of methods for determining depth from focus. PAMI, 15():97 108, [5] P. Favaro and S. Soatto. Learning shape from defocus. In Proc. ECCV, pages , 00. [6] D. B. Goldman and J.-H. Chen. Vignette and exposure calibration and compensation. In CVPR, pages , 005. [7] H. Haussecker and D. Fleet. Computing optical flow with physical models of brightness variation. PAMI, 3(6): , Jun 001. [8] Y. S. Heo, K. M. Lee, and S. U. Lee. Illumination and camera invariant stereo matching. In Proc. CVPR, 008. [9] H. Hirschmller and D. Scharstein. Evaluation of stereo matching costs on images with radiometric differences. PAMI, 99(1), [10] M. Kumar, V. Kolmogorov, and P. Torr. An analysis of convex relaxations for MAP estimation. In NIPS, 007. [11] MOSEK ApS. The mosek optimization toolbox for matlab manual. 0/ tools/doc/pdf/toolbox.pdf, Section , Page 37. [1] V. P. Namboodiri, S. Chaudhuri, and S. Hadap. Regularized depth from defocus. In ICIP, pages , 008. [13] S. Negahdaripour. Revised definition of optical flow: Integration of radiometric and geometric cues for dynamic scene analysis. PAMI, 0(9): , [14] T. Nir, A. Bruckstein, and R. Kimmel. Over-parameterized variational optical flow. IJCV, 76():05 16, 008. [15] A. P. Pentland. A new sense for depth of field. PAMI, 9(4):53 531, [16] T. Pock, T. Schoenemann, G. Graber, H. Bischof, and D. Cremers. A convex formulation of continuous multi-label problems. In Proc. ECCV, pages , 008. [17] A. N. Rajagopalan and S. Chaudhuri. Optimal recovery of depth from defocused images using an mrf model. In Proc. ICCV, pages , [18] A. N. Rajagopalan and U. Mudenagudi. Depth estimation and image restoration using defocused stereo pairs. PAMI, 6(11): , 004. [19] D. Scharstein and R. Szeliski. A taxonomy and evaluation of dense two-frame stereo correspondence algorithms. IJCV, 47(1-3):7 4, 00. [0] D. Sun, S. Roth, J. P. Lewis, and M. J. Black. Learning optical flow. In Proc. ECCV, pages 83 97, 008. [1] M. Watanabe and S. Nayar. Rational Filters for Passive Depth from Defocus. IJCV, 7(3):03 5, [] O. Woodford, P. Torr, I. Reid, and A. Fitzgibbon. Global stereo reconstruction under second order smoothness priors. In Proc. CVPR, 008. [3] A. L. Yuille and A. Rangarajan. The concave-convex procedure. Neural Comput., 15(4): , 003.

Filter Flow: Supplemental Material

Filter Flow: Supplemental Material Filter Flow: Supplemental Material Steven M. Seitz University of Washington Simon Baker Microsoft Research We include larger images and a number of additional results obtained using Filter Flow [5]. 1

More information

EE795: Computer Vision and Intelligent Systems

EE795: Computer Vision and Intelligent Systems EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 14 130307 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Stereo Dense Motion Estimation Translational

More information

Super Resolution Using Graph-cut

Super Resolution Using Graph-cut Super Resolution Using Graph-cut Uma Mudenagudi, Ram Singla, Prem Kalra, and Subhashis Banerjee Department of Computer Science and Engineering Indian Institute of Technology Delhi Hauz Khas, New Delhi,

More information

CS6670: Computer Vision

CS6670: Computer Vision CS6670: Computer Vision Noah Snavely Lecture 19: Optical flow http://en.wikipedia.org/wiki/barberpole_illusion Readings Szeliski, Chapter 8.4-8.5 Announcements Project 2b due Tuesday, Nov 2 Please sign

More information

EE795: Computer Vision and Intelligent Systems

EE795: Computer Vision and Intelligent Systems EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 11 140311 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Motion Analysis Motivation Differential Motion Optical

More information

Segmentation Based Stereo. Michael Bleyer LVA Stereo Vision

Segmentation Based Stereo. Michael Bleyer LVA Stereo Vision Segmentation Based Stereo Michael Bleyer LVA Stereo Vision What happened last time? Once again, we have looked at our energy function: E ( D) = m( p, dp) + p I < p, q > We have investigated the matching

More information

What have we leaned so far?

What have we leaned so far? What have we leaned so far? Camera structure Eye structure Project 1: High Dynamic Range Imaging What have we learned so far? Image Filtering Image Warping Camera Projection Model Project 2: Panoramic

More information

Learning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009

Learning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009 Learning and Inferring Depth from Monocular Images Jiyan Pan April 1, 2009 Traditional ways of inferring depth Binocular disparity Structure from motion Defocus Given a single monocular image, how to infer

More information

Stereo Matching.

Stereo Matching. Stereo Matching Stereo Vision [1] Reduction of Searching by Epipolar Constraint [1] Photometric Constraint [1] Same world point has same intensity in both images. True for Lambertian surfaces A Lambertian

More information

Notes 9: Optical Flow

Notes 9: Optical Flow Course 049064: Variational Methods in Image Processing Notes 9: Optical Flow Guy Gilboa 1 Basic Model 1.1 Background Optical flow is a fundamental problem in computer vision. The general goal is to find

More information

Video Alignment. Literature Survey. Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin

Video Alignment. Literature Survey. Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin Literature Survey Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin Omer Shakil Abstract This literature survey compares various methods

More information

Photometric Stereo with Auto-Radiometric Calibration

Photometric Stereo with Auto-Radiometric Calibration Photometric Stereo with Auto-Radiometric Calibration Wiennat Mongkulmann Takahiro Okabe Yoichi Sato Institute of Industrial Science, The University of Tokyo {wiennat,takahiro,ysato} @iis.u-tokyo.ac.jp

More information

Lucas-Kanade Without Iterative Warping

Lucas-Kanade Without Iterative Warping 3 LucasKanade Without Iterative Warping Alex RavAcha School of Computer Science and Engineering The Hebrew University of Jerusalem 91904 Jerusalem, Israel EMail: alexis@cs.huji.ac.il Abstract A significant

More information

Markov Networks in Computer Vision

Markov Networks in Computer Vision Markov Networks in Computer Vision Sargur Srihari srihari@cedar.buffalo.edu 1 Markov Networks for Computer Vision Some applications: 1. Image segmentation 2. Removal of blur/noise 3. Stereo reconstruction

More information

Efficient Belief Propagation for Early Vision

Efficient Belief Propagation for Early Vision Efficient Belief Propagation for Early Vision Pedro F. Felzenszwalb and Daniel P. Huttenlocher Department of Computer Science Cornell University {pff,dph}@cs.cornell.edu Abstract Markov random field models

More information

Markov Networks in Computer Vision. Sargur Srihari

Markov Networks in Computer Vision. Sargur Srihari Markov Networks in Computer Vision Sargur srihari@cedar.buffalo.edu 1 Markov Networks for Computer Vision Important application area for MNs 1. Image segmentation 2. Removal of blur/noise 3. Stereo reconstruction

More information

Video Alignment. Final Report. Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin

Video Alignment. Final Report. Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin Final Report Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin Omer Shakil Abstract This report describes a method to align two videos.

More information

Hierarchical Belief Propagation To Reduce Search Space Using CUDA for Stereo and Motion Estimation

Hierarchical Belief Propagation To Reduce Search Space Using CUDA for Stereo and Motion Estimation Hierarchical Belief Propagation To Reduce Search Space Using CUDA for Stereo and Motion Estimation Scott Grauer-Gray and Chandra Kambhamettu University of Delaware Newark, DE 19716 {grauerg, chandra}@cis.udel.edu

More information

A Patch Prior for Dense 3D Reconstruction in Man-Made Environments

A Patch Prior for Dense 3D Reconstruction in Man-Made Environments A Patch Prior for Dense 3D Reconstruction in Man-Made Environments Christian Häne 1, Christopher Zach 2, Bernhard Zeisl 1, Marc Pollefeys 1 1 ETH Zürich 2 MSR Cambridge October 14, 2012 A Patch Prior for

More information

Motion. 1 Introduction. 2 Optical Flow. Sohaib A Khan. 2.1 Brightness Constancy Equation

Motion. 1 Introduction. 2 Optical Flow. Sohaib A Khan. 2.1 Brightness Constancy Equation Motion Sohaib A Khan 1 Introduction So far, we have dealing with single images of a static scene taken by a fixed camera. Here we will deal with sequence of images taken at different time intervals. Motion

More information

Data Term. Michael Bleyer LVA Stereo Vision

Data Term. Michael Bleyer LVA Stereo Vision Data Term Michael Bleyer LVA Stereo Vision What happened last time? We have looked at our energy function: E ( D) = m( p, dp) + p I < p, q > N s( p, q) We have learned about an optimization algorithm that

More information

Introduction à la vision artificielle X

Introduction à la vision artificielle X Introduction à la vision artificielle X Jean Ponce Email: ponce@di.ens.fr Web: http://www.di.ens.fr/~ponce Planches après les cours sur : http://www.di.ens.fr/~ponce/introvis/lect10.pptx http://www.di.ens.fr/~ponce/introvis/lect10.pdf

More information

Ninio, J. and Stevens, K. A. (2000) Variations on the Hermann grid: an extinction illusion. Perception, 29,

Ninio, J. and Stevens, K. A. (2000) Variations on the Hermann grid: an extinction illusion. Perception, 29, Ninio, J. and Stevens, K. A. (2000) Variations on the Hermann grid: an extinction illusion. Perception, 29, 1209-1217. CS 4495 Computer Vision A. Bobick Sparse to Dense Correspodence Building Rome in

More information

Continuous and Discrete Optimization Methods in Computer Vision. Daniel Cremers Department of Computer Science University of Bonn

Continuous and Discrete Optimization Methods in Computer Vision. Daniel Cremers Department of Computer Science University of Bonn Continuous and Discrete Optimization Methods in Computer Vision Daniel Cremers Department of Computer Science University of Bonn Oxford, August 16 2007 Segmentation by Energy Minimization Given an image,

More information

Feature Tracking and Optical Flow

Feature Tracking and Optical Flow Feature Tracking and Optical Flow Prof. D. Stricker Doz. G. Bleser Many slides adapted from James Hays, Derek Hoeim, Lana Lazebnik, Silvio Saverse, who 1 in turn adapted slides from Steve Seitz, Rick Szeliski,

More information

Capturing, Modeling, Rendering 3D Structures

Capturing, Modeling, Rendering 3D Structures Computer Vision Approach Capturing, Modeling, Rendering 3D Structures Calculate pixel correspondences and extract geometry Not robust Difficult to acquire illumination effects, e.g. specular highlights

More information

Computer Vision I - Filtering and Feature detection

Computer Vision I - Filtering and Feature detection Computer Vision I - Filtering and Feature detection Carsten Rother 30/10/2015 Computer Vision I: Basics of Image Processing Roadmap: Basics of Digital Image Processing Computer Vision I: Basics of Image

More information

Motion Estimation. There are three main types (or applications) of motion estimation:

Motion Estimation. There are three main types (or applications) of motion estimation: Members: D91922016 朱威達 R93922010 林聖凱 R93922044 謝俊瑋 Motion Estimation There are three main types (or applications) of motion estimation: Parametric motion (image alignment) The main idea of parametric motion

More information

Matching. Compare region of image to region of image. Today, simplest kind of matching. Intensities similar.

Matching. Compare region of image to region of image. Today, simplest kind of matching. Intensities similar. Matching Compare region of image to region of image. We talked about this for stereo. Important for motion. Epipolar constraint unknown. But motion small. Recognition Find object in image. Recognize object.

More information

Discrete Optimization Methods in Computer Vision CSE 6389 Slides by: Boykov Modified and Presented by: Mostafa Parchami Basic overview of graph cuts

Discrete Optimization Methods in Computer Vision CSE 6389 Slides by: Boykov Modified and Presented by: Mostafa Parchami Basic overview of graph cuts Discrete Optimization Methods in Computer Vision CSE 6389 Slides by: Boykov Modified and Presented by: Mostafa Parchami Basic overview of graph cuts [Yuri Boykov, Olga Veksler, Ramin Zabih, Fast Approximation

More information

Global Flow Estimation. Lecture 9

Global Flow Estimation. Lecture 9 Motion Models Image Transformations to relate two images 3D Rigid motion Perspective & Orthographic Transformation Planar Scene Assumption Transformations Translation Rotation Rigid Affine Homography Pseudo

More information

Multi-Label Moves for Multi-Label Energies

Multi-Label Moves for Multi-Label Energies Multi-Label Moves for Multi-Label Energies Olga Veksler University of Western Ontario some work is joint with Olivier Juan, Xiaoqing Liu, Yu Liu Outline Review multi-label optimization with graph cuts

More information

Dense 3D Reconstruction. Christiano Gava

Dense 3D Reconstruction. Christiano Gava Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Today: dense 3D reconstruction The matching problem

More information

CS 4495 Computer Vision Motion and Optic Flow

CS 4495 Computer Vision Motion and Optic Flow CS 4495 Computer Vision Aaron Bobick School of Interactive Computing Administrivia PS4 is out, due Sunday Oct 27 th. All relevant lectures posted Details about Problem Set: You may *not* use built in Harris

More information

USING ONE GRAPH-CUT TO FUSE MULTIPLE CANDIDATE MAPS IN DEPTH ESTIMATION

USING ONE GRAPH-CUT TO FUSE MULTIPLE CANDIDATE MAPS IN DEPTH ESTIMATION USING ONE GRAPH-CUT TO FUSE MULTIPLE CANDIDATE MAPS IN DEPTH ESTIMATION F. Pitié Sigmedia Group, Trinity College Dublin, fpitie@mee.tcd.ie Abstract Graph-cut techniques for depth and disparity estimations

More information

Public Library, Stereoscopic Looking Room, Chicago, by Phillips, 1923

Public Library, Stereoscopic Looking Room, Chicago, by Phillips, 1923 Public Library, Stereoscopic Looking Room, Chicago, by Phillips, 1923 Teesta suspension bridge-darjeeling, India Mark Twain at Pool Table", no date, UCR Museum of Photography Woman getting eye exam during

More information

Ping Tan. Simon Fraser University

Ping Tan. Simon Fraser University Ping Tan Simon Fraser University Photos vs. Videos (live photos) A good photo tells a story Stories are better told in videos Videos in the Mobile Era (mobile & share) More videos are captured by mobile

More information

Motion Estimation using Block Overlap Minimization

Motion Estimation using Block Overlap Minimization Motion Estimation using Block Overlap Minimization Michael Santoro, Ghassan AlRegib, Yucel Altunbasak School of Electrical and Computer Engineering, Georgia Institute of Technology Atlanta, GA 30332 USA

More information

Particle Tracking. For Bulk Material Handling Systems Using DEM Models. By: Jordan Pease

Particle Tracking. For Bulk Material Handling Systems Using DEM Models. By: Jordan Pease Particle Tracking For Bulk Material Handling Systems Using DEM Models By: Jordan Pease Introduction Motivation for project Particle Tracking Application to DEM models Experimental Results Future Work References

More information

Lecture 10: Multi view geometry

Lecture 10: Multi view geometry Lecture 10: Multi view geometry Professor Fei Fei Li Stanford Vision Lab 1 What we will learn today? Stereo vision Correspondence problem (Problem Set 2 (Q3)) Active stereo vision systems Structure from

More information

Part II: Modeling Aspects

Part II: Modeling Aspects Yosemite test sequence Illumination changes Motion discontinuities Variational Optical Flow Estimation Part II: Modeling Aspects Discontinuity Di ti it preserving i smoothness th tterms Robust data terms

More information

Digital Image Processing. Prof. P. K. Biswas. Department of Electronic & Electrical Communication Engineering

Digital Image Processing. Prof. P. K. Biswas. Department of Electronic & Electrical Communication Engineering Digital Image Processing Prof. P. K. Biswas Department of Electronic & Electrical Communication Engineering Indian Institute of Technology, Kharagpur Lecture - 21 Image Enhancement Frequency Domain Processing

More information

Dense 3D Reconstruction. Christiano Gava

Dense 3D Reconstruction. Christiano Gava Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Wide baseline matching (SIFT) Today: dense 3D reconstruction

More information

Segmentation with non-linear constraints on appearance, complexity, and geometry

Segmentation with non-linear constraints on appearance, complexity, and geometry IPAM February 2013 Western Univesity Segmentation with non-linear constraints on appearance, complexity, and geometry Yuri Boykov Andrew Delong Lena Gorelick Hossam Isack Anton Osokin Frank Schmidt Olga

More information

Feature Tracking and Optical Flow

Feature Tracking and Optical Flow Feature Tracking and Optical Flow Prof. D. Stricker Doz. G. Bleser Many slides adapted from James Hays, Derek Hoeim, Lana Lazebnik, Silvio Saverse, who in turn adapted slides from Steve Seitz, Rick Szeliski,

More information

SURVEY OF LOCAL AND GLOBAL OPTICAL FLOW WITH COARSE TO FINE METHOD

SURVEY OF LOCAL AND GLOBAL OPTICAL FLOW WITH COARSE TO FINE METHOD SURVEY OF LOCAL AND GLOBAL OPTICAL FLOW WITH COARSE TO FINE METHOD M.E-II, Department of Computer Engineering, PICT, Pune ABSTRACT: Optical flow as an image processing technique finds its applications

More information

CHAPTER 3 DISPARITY AND DEPTH MAP COMPUTATION

CHAPTER 3 DISPARITY AND DEPTH MAP COMPUTATION CHAPTER 3 DISPARITY AND DEPTH MAP COMPUTATION In this chapter we will discuss the process of disparity computation. It plays an important role in our caricature system because all 3D coordinates of nodes

More information

Edge and corner detection

Edge and corner detection Edge and corner detection Prof. Stricker Doz. G. Bleser Computer Vision: Object and People Tracking Goals Where is the information in an image? How is an object characterized? How can I find measurements

More information

EECS 442 Computer vision. Stereo systems. Stereo vision Rectification Correspondence problem Active stereo vision systems

EECS 442 Computer vision. Stereo systems. Stereo vision Rectification Correspondence problem Active stereo vision systems EECS 442 Computer vision Stereo systems Stereo vision Rectification Correspondence problem Active stereo vision systems Reading: [HZ] Chapter: 11 [FP] Chapter: 11 Stereo vision P p p O 1 O 2 Goal: estimate

More information

Video Mosaics for Virtual Environments, R. Szeliski. Review by: Christopher Rasmussen

Video Mosaics for Virtual Environments, R. Szeliski. Review by: Christopher Rasmussen Video Mosaics for Virtual Environments, R. Szeliski Review by: Christopher Rasmussen September 19, 2002 Announcements Homework due by midnight Next homework will be assigned Tuesday, due following Tuesday.

More information

Global Flow Estimation. Lecture 9

Global Flow Estimation. Lecture 9 Global Flow Estimation Lecture 9 Global Motion Estimate motion using all pixels in the image. Parametric flow gives an equation, which describes optical flow for each pixel. Affine Projective Global motion

More information

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching Stereo Matching Fundamental matrix Let p be a point in left image, p in right image l l Epipolar relation p maps to epipolar line l p maps to epipolar line l p p Epipolar mapping described by a 3x3 matrix

More information

Image Based Reconstruction II

Image Based Reconstruction II Image Based Reconstruction II Qixing Huang Feb. 2 th 2017 Slide Credit: Yasutaka Furukawa Image-Based Geometry Reconstruction Pipeline Last Lecture: Multi-View SFM Multi-View SFM This Lecture: Multi-View

More information

A New Approach to 3D Shape Recovery of Local Planar Surface Patches from Shift-Variant Blurred Images

A New Approach to 3D Shape Recovery of Local Planar Surface Patches from Shift-Variant Blurred Images A New Approach to 3D Shape Recovery of Local Planar Surface Patches from Shift-Variant Blurred Images Xue Tu, Murali Subbarao, Youn-Sik Kang {tuxue, murali, yskang}@ecesunysbedu, Dept of Electrical and

More information

Stereo and Epipolar geometry

Stereo and Epipolar geometry Previously Image Primitives (feature points, lines, contours) Today: Stereo and Epipolar geometry How to match primitives between two (multiple) views) Goals: 3D reconstruction, recognition Jana Kosecka

More information

Mathematics in Image Processing

Mathematics in Image Processing Mathematics in Image Processing Michal Šorel Department of Image Processing Institute of Information Theory and Automation (ÚTIA) Academy of Sciences of the Czech Republic http://zoi.utia.cas.cz/ Mathematics

More information

Fundamental matrix. Let p be a point in left image, p in right image. Epipolar relation. Epipolar mapping described by a 3x3 matrix F

Fundamental matrix. Let p be a point in left image, p in right image. Epipolar relation. Epipolar mapping described by a 3x3 matrix F Fundamental matrix Let p be a point in left image, p in right image l l Epipolar relation p maps to epipolar line l p maps to epipolar line l p p Epipolar mapping described by a 3x3 matrix F Fundamental

More information

EE795: Computer Vision and Intelligent Systems

EE795: Computer Vision and Intelligent Systems EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 WRI C225 Lecture 04 130131 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Histogram Equalization Image Filtering Linear

More information

Broad field that includes low-level operations as well as complex high-level algorithms

Broad field that includes low-level operations as well as complex high-level algorithms Image processing About Broad field that includes low-level operations as well as complex high-level algorithms Low-level image processing Computer vision Computational photography Several procedures and

More information

Translation Symmetry Detection: A Repetitive Pattern Analysis Approach

Translation Symmetry Detection: A Repetitive Pattern Analysis Approach 2013 IEEE Conference on Computer Vision and Pattern Recognition Workshops Translation Symmetry Detection: A Repetitive Pattern Analysis Approach Yunliang Cai and George Baciu GAMA Lab, Department of Computing

More information

Supplementary Material: Adaptive and Move Making Auxiliary Cuts for Binary Pairwise Energies

Supplementary Material: Adaptive and Move Making Auxiliary Cuts for Binary Pairwise Energies Supplementary Material: Adaptive and Move Making Auxiliary Cuts for Binary Pairwise Energies Lena Gorelick Yuri Boykov Olga Veksler Computer Science Department University of Western Ontario Abstract Below

More information

CS 664 Structure and Motion. Daniel Huttenlocher

CS 664 Structure and Motion. Daniel Huttenlocher CS 664 Structure and Motion Daniel Huttenlocher Determining 3D Structure Consider set of 3D points X j seen by set of cameras with projection matrices P i Given only image coordinates x ij of each point

More information

Stereo Correspondence with Occlusions using Graph Cuts

Stereo Correspondence with Occlusions using Graph Cuts Stereo Correspondence with Occlusions using Graph Cuts EE368 Final Project Matt Stevens mslf@stanford.edu Zuozhen Liu zliu2@stanford.edu I. INTRODUCTION AND MOTIVATION Given two stereo images of a scene,

More information

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching Stereo Matching Fundamental matrix Let p be a point in left image, p in right image l l Epipolar relation p maps to epipolar line l p maps to epipolar line l p p Epipolar mapping described by a 3x3 matrix

More information

Multi-View 3D-Reconstruction

Multi-View 3D-Reconstruction Multi-View 3D-Reconstruction Cedric Cagniart Computer Aided Medical Procedures (CAMP) Technische Universität München, Germany 1 Problem Statement Given several calibrated views of an object... can we automatically

More information

Geometric Reconstruction Dense reconstruction of scene geometry

Geometric Reconstruction Dense reconstruction of scene geometry Lecture 5. Dense Reconstruction and Tracking with Real-Time Applications Part 2: Geometric Reconstruction Dr Richard Newcombe and Dr Steven Lovegrove Slide content developed from: [Newcombe, Dense Visual

More information

Models for grids. Computer vision: models, learning and inference. Multi label Denoising. Binary Denoising. Denoising Goal.

Models for grids. Computer vision: models, learning and inference. Multi label Denoising. Binary Denoising. Denoising Goal. Models for grids Computer vision: models, learning and inference Chapter 9 Graphical Models Consider models where one unknown world state at each pixel in the image takes the form of a grid. Loops in the

More information

Topics to be Covered in the Rest of the Semester. CSci 4968 and 6270 Computational Vision Lecture 15 Overview of Remainder of the Semester

Topics to be Covered in the Rest of the Semester. CSci 4968 and 6270 Computational Vision Lecture 15 Overview of Remainder of the Semester Topics to be Covered in the Rest of the Semester CSci 4968 and 6270 Computational Vision Lecture 15 Overview of Remainder of the Semester Charles Stewart Department of Computer Science Rensselaer Polytechnic

More information

Accurate and Dense Wide-Baseline Stereo Matching Using SW-POC

Accurate and Dense Wide-Baseline Stereo Matching Using SW-POC Accurate and Dense Wide-Baseline Stereo Matching Using SW-POC Shuji Sakai, Koichi Ito, Takafumi Aoki Graduate School of Information Sciences, Tohoku University, Sendai, 980 8579, Japan Email: sakai@aoki.ecei.tohoku.ac.jp

More information

Image Coding with Active Appearance Models

Image Coding with Active Appearance Models Image Coding with Active Appearance Models Simon Baker, Iain Matthews, and Jeff Schneider CMU-RI-TR-03-13 The Robotics Institute Carnegie Mellon University Abstract Image coding is the task of representing

More information

COMPUTER VISION > OPTICAL FLOW UTRECHT UNIVERSITY RONALD POPPE

COMPUTER VISION > OPTICAL FLOW UTRECHT UNIVERSITY RONALD POPPE COMPUTER VISION 2017-2018 > OPTICAL FLOW UTRECHT UNIVERSITY RONALD POPPE OUTLINE Optical flow Lucas-Kanade Horn-Schunck Applications of optical flow Optical flow tracking Histograms of oriented flow Assignment

More information

Graph Cuts vs. Level Sets. part I Basics of Graph Cuts

Graph Cuts vs. Level Sets. part I Basics of Graph Cuts ECCV 2006 tutorial on Graph Cuts vs. Level Sets part I Basics of Graph Cuts Yuri Boykov University of Western Ontario Daniel Cremers University of Bonn Vladimir Kolmogorov University College London Graph

More information

Peripheral drift illusion

Peripheral drift illusion Peripheral drift illusion Does it work on other animals? Computer Vision Motion and Optical Flow Many slides adapted from J. Hays, S. Seitz, R. Szeliski, M. Pollefeys, K. Grauman and others Video A video

More information

Local Image Registration: An Adaptive Filtering Framework

Local Image Registration: An Adaptive Filtering Framework Local Image Registration: An Adaptive Filtering Framework Gulcin Caner a,a.murattekalp a,b, Gaurav Sharma a and Wendi Heinzelman a a Electrical and Computer Engineering Dept.,University of Rochester, Rochester,

More information

There are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few...

There are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few... STEREO VISION The slides are from several sources through James Hays (Brown); Srinivasa Narasimhan (CMU); Silvio Savarese (U. of Michigan); Bill Freeman and Antonio Torralba (MIT), including their own

More information

The Lucas & Kanade Algorithm

The Lucas & Kanade Algorithm The Lucas & Kanade Algorithm Instructor - Simon Lucey 16-423 - Designing Computer Vision Apps Today Registration, Registration, Registration. Linearizing Registration. Lucas & Kanade Algorithm. 3 Biggest

More information

Segmentation and Tracking of Partial Planar Templates

Segmentation and Tracking of Partial Planar Templates Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract

More information

Optical Flow-Based Motion Estimation. Thanks to Steve Seitz, Simon Baker, Takeo Kanade, and anyone else who helped develop these slides.

Optical Flow-Based Motion Estimation. Thanks to Steve Seitz, Simon Baker, Takeo Kanade, and anyone else who helped develop these slides. Optical Flow-Based Motion Estimation Thanks to Steve Seitz, Simon Baker, Takeo Kanade, and anyone else who helped develop these slides. 1 Why estimate motion? We live in a 4-D world Wide applications Object

More information

Algorithms for Markov Random Fields in Computer Vision

Algorithms for Markov Random Fields in Computer Vision Algorithms for Markov Random Fields in Computer Vision Dan Huttenlocher November, 2003 (Joint work with Pedro Felzenszwalb) Random Field Broadly applicable stochastic model Collection of n sites S Hidden

More information

Computer Vision 2. SS 18 Dr. Benjamin Guthier Professur für Bildverarbeitung. Computer Vision 2 Dr. Benjamin Guthier

Computer Vision 2. SS 18 Dr. Benjamin Guthier Professur für Bildverarbeitung. Computer Vision 2 Dr. Benjamin Guthier Computer Vision 2 SS 18 Dr. Benjamin Guthier Professur für Bildverarbeitung Computer Vision 2 Dr. Benjamin Guthier 1. IMAGE PROCESSING Computer Vision 2 Dr. Benjamin Guthier Content of this Chapter Non-linear

More information

Computer Vision II Lecture 4

Computer Vision II Lecture 4 Computer Vision II Lecture 4 Color based Tracking 29.04.2014 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Course Outline Single-Object Tracking Background modeling

More information

EXAM SOLUTIONS. Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006,

EXAM SOLUTIONS. Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006, School of Computer Science and Communication, KTH Danica Kragic EXAM SOLUTIONS Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006, 14.00 19.00 Grade table 0-25 U 26-35 3 36-45

More information

Final Exam Study Guide

Final Exam Study Guide Final Exam Study Guide Exam Window: 28th April, 12:00am EST to 30th April, 11:59pm EST Description As indicated in class the goal of the exam is to encourage you to review the material from the course.

More information

Stereo II CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz

Stereo II CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz Stereo II CSE 576 Ali Farhadi Several slides from Larry Zitnick and Steve Seitz Camera parameters A camera is described by several parameters Translation T of the optical center from the origin of world

More information

Motion Estimation (II) Ce Liu Microsoft Research New England

Motion Estimation (II) Ce Liu Microsoft Research New England Motion Estimation (II) Ce Liu celiu@microsoft.com Microsoft Research New England Last time Motion perception Motion representation Parametric motion: Lucas-Kanade T I x du dv = I x I T x I y I x T I y

More information

Multi-view Stereo. Ivo Boyadzhiev CS7670: September 13, 2011

Multi-view Stereo. Ivo Boyadzhiev CS7670: September 13, 2011 Multi-view Stereo Ivo Boyadzhiev CS7670: September 13, 2011 What is stereo vision? Generic problem formulation: given several images of the same object or scene, compute a representation of its 3D shape

More information

3D Computer Vision. Dense 3D Reconstruction II. Prof. Didier Stricker. Christiano Gava

3D Computer Vision. Dense 3D Reconstruction II. Prof. Didier Stricker. Christiano Gava 3D Computer Vision Dense 3D Reconstruction II Prof. Didier Stricker Christiano Gava Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de

More information

Fast Approximate Energy Minimization via Graph Cuts

Fast Approximate Energy Minimization via Graph Cuts IEEE Transactions on PAMI, vol. 23, no. 11, pp. 1222-1239 p.1 Fast Approximate Energy Minimization via Graph Cuts Yuri Boykov, Olga Veksler and Ramin Zabih Abstract Many tasks in computer vision involve

More information

Multi-view stereo. Many slides adapted from S. Seitz

Multi-view stereo. Many slides adapted from S. Seitz Multi-view stereo Many slides adapted from S. Seitz Beyond two-view stereo The third eye can be used for verification Multiple-baseline stereo Pick a reference image, and slide the corresponding window

More information

Quadratic Pseudo-Boolean Optimization(QPBO): Theory and Applications At-a-Glance

Quadratic Pseudo-Boolean Optimization(QPBO): Theory and Applications At-a-Glance Quadratic Pseudo-Boolean Optimization(QPBO): Theory and Applications At-a-Glance Presented By: Ahmad Al-Kabbany Under the Supervision of: Prof.Eric Dubois 12 June 2012 Outline Introduction The limitations

More information

Discrete Optimization of Ray Potentials for Semantic 3D Reconstruction

Discrete Optimization of Ray Potentials for Semantic 3D Reconstruction Discrete Optimization of Ray Potentials for Semantic 3D Reconstruction Marc Pollefeys Joined work with Nikolay Savinov, Christian Haene, Lubor Ladicky 2 Comparison to Volumetric Fusion Higher-order ray

More information

Lecture 16: Computer Vision

Lecture 16: Computer Vision CS4442/9542b: Artificial Intelligence II Prof. Olga Veksler Lecture 16: Computer Vision Motion Slides are from Steve Seitz (UW), David Jacobs (UMD) Outline Motion Estimation Motion Field Optical Flow Field

More information

Lecture 16: Computer Vision

Lecture 16: Computer Vision CS442/542b: Artificial ntelligence Prof. Olga Veksler Lecture 16: Computer Vision Motion Slides are from Steve Seitz (UW), David Jacobs (UMD) Outline Motion Estimation Motion Field Optical Flow Field Methods

More information

Supervised texture detection in images

Supervised texture detection in images Supervised texture detection in images Branislav Mičušík and Allan Hanbury Pattern Recognition and Image Processing Group, Institute of Computer Aided Automation, Vienna University of Technology Favoritenstraße

More information

Minimizing Noise and Bias in 3D DIC. Correlated Solutions, Inc.

Minimizing Noise and Bias in 3D DIC. Correlated Solutions, Inc. Minimizing Noise and Bias in 3D DIC Correlated Solutions, Inc. Overview Overview of Noise and Bias Digital Image Correlation Background/Tracking Function Minimizing Noise Focus Contrast/Lighting Glare

More information

Optical Flow Estimation with CUDA. Mikhail Smirnov

Optical Flow Estimation with CUDA. Mikhail Smirnov Optical Flow Estimation with CUDA Mikhail Smirnov msmirnov@nvidia.com Document Change History Version Date Responsible Reason for Change Mikhail Smirnov Initial release Abstract Optical flow is the apparent

More information

Perceptual Grouping from Motion Cues Using Tensor Voting

Perceptual Grouping from Motion Cues Using Tensor Voting Perceptual Grouping from Motion Cues Using Tensor Voting 1. Research Team Project Leader: Graduate Students: Prof. Gérard Medioni, Computer Science Mircea Nicolescu, Changki Min 2. Statement of Project

More information

Supplementary Material : Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision

Supplementary Material : Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision Supplementary Material : Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision Due to space limitation in the main paper, we present additional experimental results in this supplementary

More information

BIL Computer Vision Apr 16, 2014

BIL Computer Vision Apr 16, 2014 BIL 719 - Computer Vision Apr 16, 2014 Binocular Stereo (cont d.), Structure from Motion Aykut Erdem Dept. of Computer Engineering Hacettepe University Slide credit: S. Lazebnik Basic stereo matching algorithm

More information

Optical flow and tracking

Optical flow and tracking EECS 442 Computer vision Optical flow and tracking Intro Optical flow and feature tracking Lucas-Kanade algorithm Motion segmentation Segments of this lectures are courtesy of Profs S. Lazebnik S. Seitz,

More information