Prior-based Segmentation by Projective Registration and Level Sets

Similar documents
Prior-based Segmentation and Shape Registration in the Presence of Perspective Distortion

Geodesic Active Contours with Combined Shape and Appearance Priors

Kernel Density Estimation and Intrinsic Alignment for Knowledge-driven Segmentation: Teaching Level Sets to Walk

Automated Segmentation Using a Fast Implementation of the Chan-Vese Models

Multiphase Dynamic Labeling for Variational Recognition-Driven Image Segmentation

Multiregion level set tracking with transformation invariant shape priors

Implicit Active Shape Models for 3D Segmentation in MR Imaging

Multiple Motion and Occlusion Segmentation with a Multiphase Level Set Method

Region Based Image Segmentation using a Modified Mumford-Shah Algorithm

An Active Contour Model without Edges

URL: /ICPR

Interactive Image Segmentation Using Level Sets and Dempster-Shafer Theory of Evidence

College of Engineering, Trivandrum.

On Symmetry, Perspectivity and Level-set based Segmentation

Implicit Active Contours Driven by Local Binary Fitting Energy

Converting Level Set Gradients to Shape Gradients

Active Geodesics: Region-based Active Contour Segmentation with a Global Edge-based Constraint

Research Article Local- and Global-Statistics-Based Active Contour Model for Image Segmentation

METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS

Towards Segmentation Based on a Shape Prior Manifold

Continuous and Discrete Optimization Methods in Computer Vision. Daniel Cremers Department of Computer Science University of Bonn

Robust Automatic Tracking of Skin-Colored Objects with Level Set based Occlusion Handling

Stereo Matching for Calibrated Cameras without Correspondence

Extract Object Boundaries in Noisy Images using Level Set. Literature Survey

On-line, Incremental Learning of a Robust Active Shape Model

Notes 9: Optical Flow

Yunyun Yang, Chunming Li, Chiu-Yen Kao and Stanley Osher. Speaker: Chiu-Yen Kao (Math Department, The Ohio State University) BIRS, Banff, Canada

A Survey of Image Segmentation Based On Multi Region Level Set Method

A Probabilistic Level Set Formulation for Interactive Organ Segmentation

Level Set Evolution without Reinitilization

NSCT BASED LOCAL ENHANCEMENT FOR ACTIVE CONTOUR BASED IMAGE SEGMENTATION APPLICATION

The Level Set Method. Lecture Notes, MIT J / 2.097J / 6.339J Numerical Methods for Partial Differential Equations

Segmentation Using Active Contour Model and Level Set Method Applied to Medical Images

IMA Preprint Series # 2235

Variational Space-Time Motion Segmentation

Structure from Motion. Prof. Marco Marcon

Active Contour-Based Visual Tracking by Integrating Colors, Shapes, and Motions Using Level Sets

Universität des Saarlandes. Fachrichtung 6.1 Mathematik

Method of Background Subtraction for Medical Image Segmentation

Level-set MCMC Curve Sampling and Geometric Conditional Simulation

Dr. Ulas Bagci

Object and Motion Recognition using Plane Plus Parallax Displacement of Conics

SCIENCE & TECHNOLOGY

Automatic Feature Extraction of Pose-measuring System Based on Geometric Invariants

Euclidean Reconstruction Independent on Camera Intrinsic Parameters

Shape Priors for Level Set Representations

Midterm Exam Solutions

Submitted by Wesley Snyder, Ph.D. Department of Electrical and Computer Engineering. North Carolina State University. February 29 th, 2004.

ISSN: X International Journal of Advanced Research in Electronics and Communication Engineering (IJARECE) Volume 6, Issue 8, August 2017

3D Computer Vision. Dense 3D Reconstruction II. Prof. Didier Stricker. Christiano Gava

Technion - Computer Science Department - Tehnical Report CIS

Active contours for multi-region image segmentation with a single level set function

Tracking with Active Contours Using Dynamically Updated Shape Information

Shape Prior Segmentation of Multiple Objects with Graph Cuts

Accurate Image Registration from Local Phase Information

Unstructured Mesh Generation for Implicit Moving Geometries and Level Set Applications

Shape Analysis. Erkut Erdem. BIL717, April 2012!

Camera Calibration Using Line Correspondences

STATISTICS AND ANALYSIS OF SHAPE

Level Set Segmentation using Statistical Shape Priors

Curve Sampling and Geometric Conditional Simulation

Curvature guided surface registration using level sets

Mathematics of a Multiple Omni-Directional System

weighted minimal surface model for surface reconstruction from scattered points, curves, and/or pieces of surfaces.

A 3D shape constraint on video

CALIBRATION BETWEEN DEPTH AND COLOR SENSORS FOR COMMODITY DEPTH CAMERAS. Cha Zhang and Zhengyou Zhang

SELF-ORGANIZING APPROACH TO LEARN A LEVEL-SET FUNCTION FOR OBJECT SEGMENTATION IN COMPLEX BACKGROUND ENVIRONMENTS

Hierarchical Segmentation of Thin Structures in Volumetric Medical Images

Metric Rectification for Perspective Images of Planes

Isophote-Based Interpolation

A Level Set Based Predictor-Corrector Algorithm for Vessel Segmentation

Computer Vision. Coordinates. Prof. Flávio Cardeal DECOM / CEFET- MG.

Edge and local feature detection - 2. Importance of edge detection in computer vision

MULTIPHASE LEVEL SET EVOLUTION WITH APPLICATIONS TO AUTOMATIC GENERATIONAL TRACKING OF CELL DIVISION OF ESCHERICHIA COLI. A Thesis.

Camera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration

A Note on Computation of a Ray Bending Path

How to Compute the Pose of an Object without a Direct View?

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing

Synchronized Ego-Motion Recovery of Two Face-to-Face Cameras

Automatic Logo Detection and Removal

A General Expression of the Fundamental Matrix for Both Perspective and Affine Cameras

Multiple View Geometry in computer vision

METR Robotics Tutorial 2 Week 2: Homogeneous Coordinates

Flexible Calibration of a Portable Structured Light System through Surface Plane

Catadioptric camera model with conic mirror

Subpixel Corner Detection Using Spatial Moment 1)

Machine vision. Summary # 11: Stereo vision and epipolar geometry. u l = λx. v l = λy

CS 664 Slides #9 Multi-Camera Geometry. Prof. Dan Huttenlocher Fall 2003

3-D D Euclidean Space - Vectors

ACTIVE CONTOURS BASED OBJECT DETECTION & EXTRACTION USING WSPF PARAMETER: A NEW LEVEL SET METHOD

arxiv: v1 [cs.cv] 28 Sep 2018

Image denoising using TV-Stokes equation with an orientation-matching minimization

Visual Recognition: Image Formation

Isophote-Based Interpolation

Stereo Vision. MAN-522 Computer Vision

A Comparison of Shape Matching Methods for Contour Based Pose Estimation

Feature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies

Split Bregman Method for Minimization of Region-Scalable Fitting Energy for Image Segmentation

Edge Detection and Active Contours

Transcription:

Prior-based Segmentation by Projective Registration and Level Sets Tammy Riklin-Raviv Faculty of Engineering Tel-Aviv University, Israel tammy@eng.tau.ac.il Nahum Kiryati Faculty of Engineering Tel-Aviv University, Israel nk@eng.tau.ac.il Nir Sochen School of Mathematics Tel-Aviv University, Israel sochen@post.tau.ac.il Abstract Object detection and segmentation can be facilitated by the availability of a reference object. However, accounting for possible transformations between the different object views, as part of the segmentation process, remains a challenge. Recent works address this problem by using comprehensive training data. Other approaches are applicable only to limited object classes or can only accommodate similarity transformations. We suggest a novel variational approach to prior-based segmentation, which accounts for planar projective transformation, using a single reference object. The prior shape is registered concurrently with the segmentation process, without point correspondence. The algorithm detects the object of interest and correctly extracts its boundaries. The homography between the two object views is accurately recovered as well. Extending the Chan-Vese level set framework, we propose a region-based segmentation functional that includes explicit representation of the projective homography between the prior shape and the shape to segment. The formulation is derived from two-view geometry. Segmentation of a variety of objects is demonstrated and the recovered transformation is verified. 1. Introduction The coupled tasks of segmentation and object detection are essential for the extraction of semantic content from images. Prior knowledge on the shape of interest can significantly facilitate these processes, particularly when the object boundaries are not well defined. Variational methods solve segmentation problems by means of energy minimization, integrating top-down and bottom-up information, see [1] and references therein. Specifically, the level-set framework [14] for curve evolution, has been successfully applied to extract complex object boundaries allowing an automatic change in the topology. Recent approaches incorporate a representation of a reference shape within the energy functional. Thus, the recovered object boundary should resemble the expected contour, in addition to being constrained by length, smoothness and compatibility with the image features. The main difficulties in the integration of shape information into the segmentation process is the determination of a suitable shape representation and the adjustment of similarity measure between shape representations that can accommodate the possible group of transformations. The statistical approach [3, 4, 9, 11, 16, 18] addresses this challenge by using a comprehensive training set to account for shape deformations. It characterizes the probability distribution of the shapes and then measures the similarity between the evolving object boundary (or level-set function) and representatives of the training data. The performance of these methods depends on the size and coverage of the training set. Several studies [5, 6, 16, 18] accommodate similarity transformations, however, besides [15] that handles perspectivity, none of the existing methods explicitly accounts for projective transformations between the prior shape and the shape of interest. We suggest a novel variational approach to prior-based segmentation, that accounts for homographies, using a single reference object. Segmentation process is carried out concurrently with the registration of the prior shape to the shape of interest. Neither point correspondence nor direct methods [10] are used. The outcome of the algorithm includes the detection of the object of interest and a correct extraction of its boundaries. The planar projective transformation between the two object views is accurately recovered as well. This is accomplished by a novel extension of the Chan-Vese [2] level-set framework. The proposed region-based functional for segmentation includes an explicit expression of the projective homography between the prior shape and the shape to segment, which is derived from two-view geometry. It is supported by the introduction of a shape-similarity measure that admits general transformations, beyond similarity, into the functional. Segmentation of a variety of objects is demonstrated and the recovered transformation is verified.

2. Level Set Framework 2.1. The Chan-Vese two-phase model In the level-set formulation for curve evolution [14], a segmenting boundary C in an image plane Ω R 2, is the zero level-set of a 3D function φ, C = {x Ω φ(x) = 0}. Chan and Vese [2], inspired by the segmentation model of Mumford and Shah [13], suggested to segment an input image f: Ω R by minimizing an energy functional based on a piecewise constant approximation of the image. They used the Heaviside function of the evolving level-set function H(φ) as an indicator for object and background regions in the image. Thus, they have turned the segmentation framework to be a region-based, establishing a cost functional with well defined integration limits: [ E CV = (f u+ ) 2 H(φ) + (f u ) 2 (1 H(φ)) Ω + ν H(φ) ] dx, (1) where ν is a real and positive weight parameter. The scalars u + and u denote the average gray level values of the input image in the regions indicated by φ 0 and by φ < 0 respectively. They are alternately updated with the evolution of the level-set function φ and take the form: u + = f(x)h(φ)dx H(φ)dx u = f(x) (1 H(φ)) dx (1 H(φ))dx (2) The gradient descent equation for the evolution of φ is derived using the Euler-Lagrange equations for the functional (1): φ t = δ(φ) [ ( ) ] φ ν div (f u + ) 2 + (f u ) 2. φ (3) Practically, a smooth approximation of the Heaviside function H ɛ, rather than a step function, is used [2]: H ɛ (φ) = 1 2 (1 + 2 π arctan(φ )) (4) ɛ δ ɛ (φ) = dh ɛ(φ) dφ 2.2. Prior shape model = 1 π ɛ ɛ 2 + φ 2. (5) The energy functional (1) can be extended by adding a prior shape term, e.g. [5, 15]: E(φ, u +, u ) = E CV (φ, u +, u ) + µe shape (φ) µ 0. (6) One of the contributions of this work is the similarity measure between the prior and the evolving shape. The shapeterm incorporated in the energy functional measures the non-overlapping areas between the prior shape and the zero level-set of φ. Let φ : R 2 {0, 1} be a labeling function of the shape in the prior image. Note that H ɛ ( φ) φ as ɛ 0. Thus, the shape term takes the form: ( φ(x))) 2 E shape (φ) = H ɛ (φ(x)) H ɛ ( dx (7) Ω Note, that we do not force the evolving level-set function φ to resemble φ. Instead, we demand similarity of the regions within the respective contours. Since φ is not constrained to a distance function, the proposed similarity measure can accommodate planar projective transformations between the prior contour C and the evolving segmenting contour C. The following evolution equation is obtained by minimizing (7) with respect to φ: δe shape δφ = φ ( t = δ ɛ(φ) H ɛ (φ) H ɛ ( φ) ) Hence, at each time step, φ will be changed in regions where there are inconsistencies between the object-background areas indicated by H ɛ (φ) and H ɛ ( φ). The spatial evolution in φ is controlled by ɛ, according to (4) and (5). The shape-term (7) is adequate when the prior and segmented shapes are aligned. Otherwise, the representation of the prior shape-term should incorporate the projective transformation, as is detailed in section 4. However, a few geometric concepts should be reviewed first. 3. Projectivity 3.1. Planar projective homography To draw the geometric relation between two corresponding shape contours, we review the concept of planar projective homography. The identification of geometric projectivity with algebraic homography is supported by the theorems of Desargues [17]. Planar projective homography (projectivity) is a mapping H : P 2 P 2 such that the points p i are collinear if and only if H(p i ) are collinear (projectivity preserves lines) [8, 17]. The relation between corresponding views of points on a world plane Π in a 3D space, as is illustrated in Fig. 1, can be modeled by a planar homography induced by the plane. An explicit expression for the induced homorgraphy can be derived as follows. Consider two views p and p of a world point P Π, in two camera frames f and f respectively. We will denote their corresponding homogeneous coordinates by x, x and X. Let M = K[ I 0 ] and M = K [ R t ] be the first and the second camera projection matrices (respectively), where R and t are the relative rotation and translation between the cameras and K and K are the respective internal calibration matrices. Thus, x = K[ I 0 ]X and x = K [ R t ]X. (8)

H = h 11 h 12 h 13 h 21 h 22 h 23 h 31 h 32 h 33 R 3 3 (11) and let x = (x, y, 1) and x = (x, y, 1) be the homogeneous representations of the points x π, x π such that x = Hx. The eight unknowns of H (the ratios of its nine entries) can be recovered by solving at least four pairs of equations of the form: x = h 11x + h 12 y + h 13 h 31 x + h 32 y + h 33, y = h 21x + h 22 y + h 23 h 31 x + h 32 y + h 33(12) Figure 1. The homography induced by a plane. The ray that intersects the first image plane at a point p meets the world plane Π at a point P. The point P is projected to a point p in the second image plane. The mapping from p to p is the homography, denoted by H, induced by the plane Π. Illustrated after [8]. In this research, the corresponding points p and p are on corresponding planar contours. Let n be the unit normal vector to the plane Π, and let d > 0 denote the distance of Π from the optical center of the first camera. The linear transformation from x to x can be expressed as x = K ( R + 1 d tnt ) K 1 x = Hx. (9) Assuming calibrated cameras, we set K = K = I. Thus, the planar homography matrix takes the form [7, 8, 12]: H = R + 1 d tnt. (10) The matrix H is determined by the translation and rotation between the two cameras {R, t}, and by the structure parameters {n, d} of the world plane Π. Note that only the ratio t/d can be recovered from H. We now show how the structure of the homography can be used to recover the relation between the camera frames and thus between the respective shape contours. 3.2. Implicit recovery of the homography A nonsingular homography between two planes π,π is determined (up to a scale factor) by four pairs of corresponding points x π, x π with no collinear triplet of points in either plane. Consider the homography matrix Classic approaches recover H by solving an overdetermined set of equations as in (12). The rela tive position of the image planes (R, t), and the world plane orientation and pose (n, d) are recovered, by decomposition methods from the known homography matrix (see [7, 8, 12] and references therein). Our novel approach calculates the homography directly in its explicit representation (10). No point correspondence is needed. Instead, we use the calculus of variations to recover the relative motion between the two camera frames, given two corresponding views of the shape of interest. Practically, since the recovery of the homography is done simultaneously with the process of object detection and segmentation, only the reference view is known in advance. The shape to segment is extracted by the segmentation process concurrently with the registration of the prior to it. 3.3. Explicit recovery of the homography We now use the explicit formulation of the homography (10) to reformulate equation (12). We denote by γ, β and α the Euler rotation angles around the Z,Y and X axes respectively, in this order of rotation. They determine the relative orientation between the first and the second camera coordinate systems. The rotation matrix R R 3 3 operating on a vector (x, y, z) T takes the form: R = c β c γ c β s γ s β s α s β c γ c α s γ s α s β s γ + c α c γ s α c β c α s β c γ + s α s γ c α s β s γ s α s γ c α c β (13) where s α is sin(α) and c α is cos(α). The relative position between the two camera frames is determined by t = (t x, t y, t z ). Consider first the special case in which the world plane and the first image plane coincide, and d is set to 1. In this case, the normal to the world plane n is (0, 0, 1). Substituting d and n in (10), we obtain the entries of the homography matrix H: h 13 = R 13 + t x, h 23 = R 23 + t y, h 33 = R 33 + t z and h ij = R ij otherwise. Generally, the world plane is not perpendicular to the optical axis of the first camera: n (0, 0, 1). As before, we

represent n using the coordinate system of the first camera. The unit vector n is obtained by rotating the vector (0, 0, 1) by an angle ξ around the y-axis, then by an angle ψ around the x-axis. Hence, n = ( sin ξ, sin ψ cos ξ, cos ψ cos ξ). Substituting R, t and n in (10), we obtain the components of the homography matrix H: h 11 = cos β cos γ tx d sin ξ h 12 = cos β sin γ + tx d sin ψ cos ξ h 13 = sin β + tx d cos ψ cos ξ h 21 = sin α sin β cos γ cos α sin γ ty d sin ξ h 22 = sin α sin β sin γ + cos α cos γ + ty d sin ψ cos ξ h 23 = sin α cos β + ty d cos ψ cos ξ h 31 = cos α sin β cos γ + sin α sin γ tz d sin ξ h 32 = cos α sin β sin γ sin α sin γ + tz d sin ψ cos ξ h 33 = cos α cos β + tz d cos ψ cos ξ (14) Hereafter, ξ and ψ are referred to as structure parameters. 4. The cost functional 4.1. Representation of the prior shape The shape-term in the energy functional (6) is now extended to account for projective transformations. The evolution of the level-set function, given the prior contour and an estimate of the transformation parameters, is considered in this section. Consider a prior image f : Ω R + with labeled object and background regions Ω + and Ω respectively. The prior shape is represented by a function φ R 3 : φ(x, y) = { 1 f(x, y) Ω+ 0 f(x, y) Ω (15) Let T p : ( φ(x, y), x, y) ( φ (x, y ), x, y ) be a transformation of a point (x, y) and its φ value to a (projectively) equivalent point (x, y ), with φ (x, y ) = φ(x, y). Expressions for x and y are obtained by substituting the explicit entries of the homography matrix (14) in (12). For example, when ξ = ψ = 0 and d = 1, the expressions are: x = c β c γ x+c β s γ y s β +t x (c α s β c γ +s α s γ ) x+(c α s β s γ s α c γ ) y+c α c β +t z (16) y = (s αs β c γ c α s γ ) x+(s α s β s γ +c α c γ ) y+s α c β +t y (c α s β c γ +s α s γ ) x+(c α s β s γ s α c γ ) y+c α c β +t z (17) The evolution of the level-set function φ is determined by an energy shape term that measures the dissimilarity between its zero level-set and the zero level-set of the projectively transformed φ. Formally, E(φ) = (f u + ) 2 H ɛ (φ)+(f u ) 2 (1 H ɛ (φ)) Ω + ν H ɛ (φ) +µ [ H ɛ (φ) H ɛ (T p ( φ))] 2 dx (18) and the gradient descent equation for the evaluation of φ is [ ( ) φ = δ ɛ (φ) (f u ) 2 (f u + ) 2 φɛ +ν div t φ ɛ ( + µ H ɛ (φ) H ɛ (T p ( φ) )]. (19) 4.2. Recovery of the transformation In order to solve (19), one has to evaluate φ simultaneously with the recovery of the transformation T p of the prior level-set function φ. The transformation parameters (α, β, γ, t x /d, t y /d, t z /d, ψ and ξ) are evaluated via the gradient equations obtained by minimizing the energy functional (18) with respect to each of them. The general gradient descent equation for each of the transformation parameters (denoted here by η) is of the form: η t =2µ δ(t p ( φ)) ( H ɛ (φ) H ɛ (T p ( φ)) ) Tp ( φ) dx Ω η (20) where, T p ( φ) = T p( φ) ( x x η x x η + x y ) y η + T p( φ) ( y x y x η + y y ) y (21) η The values of T p ( φ)/ x and T p ( φ)/ y are derived numerically using the finite difference method. The expressions for x / η and y / η for each of the transformation parameters η are evaluated by differentiating the explicit expressions for x and y (generated by substituting equations (14) in (12))with respect to η. We use the implicit function theorem to derive the equations for x/ x, x/ y, y/ x, y/ y. 4.3. Algorithm We summarize the proposed algorithm assuming the following setup. The input is two images f and f of the same object, taken with the same camera, but under different viewing conditions. Contour points of the object to segment should be approximately co-planar. The segmentation of the reference object in f is known and used to construct the prior shape representation φ as in (15) 1. Choose an initial level-set function φ. Its zero level-set should form an initial contour within the image. 2. Set initial values (e.g. zero) for the transformation parameters α, β, γ, t x, t y, t z, ξ and ψ. Set d = 1. 3. Compute the average gray levels of the estimated object and background pixels, u + and u, using (2).

4. Transform the prior shape representation φ, by applying a coordinate transformation (substituting (14) in (12) ) with the currently estimated parameters. 5. Update φ using the gradient descent equation (19). 6. Update the transformation parameters α, β, γ, t x, t y and t z, ψ and ξ using the derivatives of (18) with respect to each, according to (20). 7. Repeat steps 3-6 until convergence. 5. Experimental Results To demonstrate our algorithm, we present segmentation results on various images. Relative transformation parameters between the known shape and the shape to segment have been evaluated and verified. In all experiments we set: dt = 0.1, ɛ = 1, µ = 1.5 and d = 1. The contributions of the first three terms in the gradient descent equation of φ (19) are normalized to [ 1, 1], thus ν is set to 1. Consider the synthetic reference image in Fig. 2a, that contains several components of different sizes and gray levels. The prior shape has been obtained by thresholding. The image to segment is shown in Fig. 2b. The initial segmenting contour is drawn on the image. Note, that this image is a noisy 1, transformed and corrupted version of the image in Fig. 2a. Successful segmentation result is shown in Fig. 2c (red contour). The recovered transformation parameters are validated by comparison to the actual transformation parameters in Table 1. The unregistered prior contour is drawn on the image to segment in Fig. 2d, to demonstrate the misalignment between the two images, Fig. 2a-b. Fig. 2e shows segmentation without a prior shape. We next consider two views, Fig. 3a-b, of a real object, a toy elephant. The initial segmenting contour is drawn on the image to segment in Fig. 3a. The prior shape has been constructed from the reference image, Fig. 3b. The desired segmentation result is shown (in red) in Fig. 3c. The contour precisely tracks the elephant profile, recovering the missing part of its trunk. The segmenting contour in Fig. 3c is the zero level-set of the final level-set function (φ) shown in Fig. 3d. Fig. 3e demonstrates the significant transformation between the two elephant views. It shows the prior contour of the reference object drawn on the image to segment. Since the ground truth is unavailable, we compared the homography matrix generated from the recovered transformation parameters, with that obtained with point correspondence method, using 13 pairs of corresponding points on the object contour, that were manually selected (Table 2). Visual verification can be seen in Fig. 3f in which the projectively transformed prior shape is shown with the final segmenting contour (same as in Fig. 3c). 1 Zero-mean Gaussian noise with STD=0.14 In the next example, Fig. 4a, the bottom-right part of the bottle opener assimilates the background. Nevertheless, the opener is extracted precisely, as shown in Fig. 4c, using the prior shape generated from the reference image shown in Fig. 4b. In Fig. 4d, we visually verify the recovered transformation by showing the projectively transformed prior shape with the final segmenting contour. Fig. 4e demonstrates the fairly significant projective transformation between the image to segment and the prior contour shape (red). When no prior is used, as seen in Fig. 4f, the segmenting contour traces (undesirably) the gradients within the object. Finally, Fig. 5a-b are two views of Europe from a satellite, at night and day time respectively. The images were adapted from NASA public domain: earthobservatory.nasa.gov. Successful segmentation is demonstrated in Fig. 5c. The coast line of the continent and the islands nearby are precisely extracted, despite the neither constant nor smooth object gray levels. Note the significant transformation, Fig. 5e, between the prior and the image to segment. The accuracy of the recovered transformation is visually verified in Fig. 5d. Segmentation when no prior is used 2, is shown for comparison in Fig. 5f. The dark areas of the eastern Adriatic shore and Scandinavia are not well segmented. 6. Summary This paper presents a novel variational approach for object segmentation in the presence of noise, specular reflections and occlusions, given a reference image from a different view. Segmentation process is carried out concurrently with a registration of the prior instance of the object to the image to segment. The outcome of the algorithm includes the detection of the object of interest and a correct extraction of its boundaries. The planar projective transformation between the two object views is accurately recovered as well. The successful segmentation results and the reliable estimation of the transformation parameters, suggest this method as a promising tool for a wide range of applications. These include perspective-invariant search in image databases, registration and structure recovery in stereo imaging, and object tracking in video sequences. Acknowledgments This research was supported by MUSCLE: Multimedia Understanding through Semantics, Computation and Learning, a European Network of Excellence funded by the EC 6th Framework IST Programme. It was also supported by the Israel Science Foundation. 2 Similar segmentation result, without prior, appears in [2], Fig.11

(a) (b) (c) (d) (e) Figure 2. (a) Reference image: the object is composed of several components in different sizes and gray levels. Prior shape was obtained by thresholding. (b) Noisy, transformed and corrupted version of the image in (a). The initial contour (red) is drawn on the image. (c) Correct segmentation. The final contour (red) is drawn on the corrupted, transformed and noisy image. (d) The unregistered prior shape contour (red) drawn on the input image. This demonstrates the misalignment between the reference (prior) image and the image to segment. (e) Segmentation result (red) when no prior shape is used. (a) (b) (c) (d) (e) (f) Figure 3. (a) Input image (a toy elephant).the initial contour (red) used for the segmentation process is drawn on the image. (b) Reference (prior) image. (c) Successful segmentation: the final contour (red) on the projectively transformed elephant. (d) Final step of the evolving level-set function φ for a correct segmentation. (e) The final contour as in (c), drawn on the projectively transformed prior shape according to the recovered homography. This shows the compatibility between the calculated and actual transformation parameters. (f) The unregistered prior shape contour (red) drawn on the input image. This demonstrates the misalignment between the reference (prior) image and the image to segment.

(a) (b) (c) (d) (e) (f) Figure 4. (a) Input image (bottle opener). The initial contour (red) used for the segmentation process is drawn on the image. (b) Reference (prior) image. (c) Successful segmentation: the final contour (red) on the projectively transformed opener. The segmentation accurately completes the missing contour part despite the specular reflection. (d) The final contour as in (c), drawn on the projectively transformed prior shape according to the recovered homography. This shows the compatibility between the calculated and actual transformation parameters. (e) The unregistered prior shape contour (red) drawn on the input image. (f) Segmentation result without a prior. (a) (b) (c) (d) (e) (f) Figure 5. (a) Input image: Europe at night from a satellite. The initial contour (red) is drawn on the image. (b) Prior image. Fig.a-b were adapted from earthobservatory.nasa.gov. (c) Successful segmentation: the final contour (red) accurately traces the coast lines, despite the shaded parts of the continent. (d) Verification of the estimated transformation parameters: The final contour as in (c), drawn on the projectively transformed prior shape according to the recovered homography. (e) The unregistered prior shape contour (red) drawn on the input image. (f) Segmentation result without a prior. The dark areas of the eastern Adriatic shore and Scandinavia are not well segmented.

Transformation parameters α β γ t x t y t z ψ ξ Proposed algorithm 0.069 0 0.075 0 7.4 0-10 9.6-0.1 0.086 0 0.14 0 Ground truth 0.075 0 0.075 0 7.5 0-10 10-0.1 0.075 0 0.15 0 Table 1. Comparison of the recovered and true transformation parameters, for the example in Fig. 2. Homography matrix entries h 11 h 12 h 13 h 21 h 22 h 23 h 31 h 32 Proposed algorithm 0.92 0.27 2.8-0.30 0.70 33.35-0.0003 0.0018 Point correspondence 0.91 0.28 3.2-0.31 0.71 34.05-0.0003 0.0019 Table 2. Comparison of the homography matrix entries obtained using the proposed algorithm and via manually selected corresponding point pairs, for the example shown in Fig. 3. References [1] G. Aubert and P. Kornprobst. Mathematical Problems in Image Processing: Partial Differential Equations and the Calculus of Variations. Springer, 2002. [2] T.F. Chan and L.A. Vese. Active contours without edges. IEEE Trans. Image Processing, 10(2):266 277, Feb. 2001. [3] Y. Chen, H.D. Tagare, S. Thiruvenkadam, F. Huang, D. Wilson, K.S. Gopinath, R.W. Briggs, and E.A. Geiser. Using prior shapes in geometric active contours in a variational framework. IJCV, 50(3):315 328, Dec. 2002. [4] D. Cremers, T. Kohlberger, and C. Schnorr. Shape statistics in kernel space for variational image segmentation. Pattern Recognition, 36(9):1929 1943, Sep. 2003. [5] D. Cremers, N. Sochen, and C. Schnorr. Multiphase dynamic labeling for variational recognition-driven image segmentation. In ECCV, volume IV, pages 74 86, 2004. [6] A. Duci, A. Yezzi, S. Mitter, and S. Soatto. Region matching with missing parts. In ECCV, volume 3, pages 48 64, May 2002. [7] O. Faugeras, Q.T. Luong, and T. Papadopoulo. The Geometry of Multiple Images. MIT Press, 2001. [8] R. I. Hartley and A. Zisserman. Multiple View Geometry in Computer Vision. Cambridge University Press, 2nd edition, 2003. [9] X. Huang, Z. Li, and D. Metaxas. Learning coupled prior-shape and appearance models for segmentation. In MICCAI, volume I, pages 60 69, 2004. [10] M. Irani and P. Anandan. All about direct methods. In W. Triggs, A. Zisserman, and R. Szeliski, editors, Vision Algorithms: Theory and Practice. Springer- Verlag, 1999. [11] M.E. Leventon, W.E.L. Grimson, and O. Faugeras. Statistical shape influence in geodesic active contours. In CVPR, volume I, pages 316 323, 2000. [12] Y. Ma, S. Soatto, J. Košecká, and S. S. Sastry. An Invitation to 3-D Vision. Springer-Verlag, 2003. [13] D. Mumford and J. Shah. Optimal approximations by piecewise smooth functions and associated variational problems. Communications on Pure and Applied Mathematics, 42:577 684, 1989. [14] S. Osher and J.A. Sethian. Fronts propagating with curvature-dependent speed: Algorithms based on Hamilton-Jacobi formulations. Journal of Computational Physics, 79:12 49, 1988. [15] T. Riklin-Raviv, N. Kiryati, and N. Sochen. Unlevelsets: Geometry and prior-based segmentation. In ECCV, volume 4, pages 50 61, 2004. [16] M. Rousson and N. Paragios. Shape priors for level set representation. In ECCV, pages 78 92, 2002. [17] C. E. Springer. Geometry and Analysis of Projective Spaces. Freeman, 1964. [18] A. Tsai, A. Yezzi, Jr., W.M. Wells, III, C. Tempany, D. Tucker, A. Fan, W.E.L. Grimson, and A.S. Willsky. A shape-based approach to the segmentation of medical imagery using level sets. IEEE Transactions on Medical Imaging, 22(2):137 154, Feb. 2003.