Graphcut-based Interactive Segmentation using Colour and Depth cues

Size: px
Start display at page:

Download "Graphcut-based Interactive Segmentation using Colour and Depth cues"

Transcription

1 Graphcut-based Interactive Segmentation using Colour and Depth cues Hu He University of Queensland,Australia David McKinnon Queensland University of Technology,Australia Michael Warren University of Queensland,Australia Abstract Segmentation of novel or dynamic objects in a scene, often referred to as background subtraction or foreground segmentation, is critical for robust high level computer vision applications such as object tracking, object classification and recognition. However, automatic realtime segmentation for robotics still poses challenges including global illumination changes, shadows, inter-reflections, colour similarity of foreground to background, and cluttered backgrounds. This paper introduces depth cues provided by structure from motion (SFM) for interactive segmentation to alleviate some of these challenges. In this paper, two prevailing interactive segmentation algorithms are compared; Lazysnapping [Li et al., 2004] and Grabcut [Rother et al., 2004], both based on graphcut optimisation [Boykov and Jolly, 2001]. The algorithms are extended to include depth cues rather than colour only as in the original papers. Results show interactive segmentation based on colour and depth cues enhances the performance of segmentation with a lower error with respect to ground truth. 1 Introduction Object recognition is an integral component for building robots capable of interacting in human environments. However, real-time object recognition in real scenes remains one of the most challenging problems in computer vision. Many issues must be overcome such as tracking, robustly identifying objects in complex cluttered backgrounds, viewpoint changes, etc [Schulz et al., 2001; Batavia and Singh, 2001; Salembier et al., 1997]. Furthermore, computational complexity must be reduced to a minimum for mobile robotic applications as object information can rapidly become obsolete in a dynamic world. This paper moves towards robust object recognition using reliable object segmentation based on graph Ben Upcroft University of Queensland, Australia ben.upcroft@uq.edu.au cut optimisation which combines both colour and depth cues. Here, we only consider interactive segmentation requiring input from a human operator. However, we propose that depth combined with colour cues can provide rich information enabling reliable automatic segmentation and eventual object recognition for mobile robotics. This paper is motivated by the Lazysnapping [Li et al., 2004] and Grabcut [Rother et al., 2004] algorithms which are based on graphcut optimization [Boykov and Jolly, 2001]. Both algorithms require user input for initialisation. For Lazysnapping [Li et al., 2004], users need to draw strokes on the original colour image indicating foreground and background (Figure 3). Pixels in the strokes are collected to model energy terms in an energy function framework. The graphcut algorithm is then employed to minimise the energy function resulting in the segmented foreground pixels stored as a label vector L. Grabcut [Rother et al., 2004] initialisation requires users to draw a rectangle around the objects of interest. The pixels internal and external to the rectangular boundary are used to model the energy terms. The graphcut algorithm is run iteratively until the resulting label vector L remains constant. The common attribute for both algorithms is the use of colour as the only cue to model the energy terms. Thus both algorithms are prone to failure when foreground and background colours are similar. In this paper, we use both colour and depth cues to model the energy terms, while using graphcuts to minimise the new energy function. Depth cues are acquired using 3D reconstruction from structure from motion, an increasingly common vision technology in mobile robotics. If there are considerable depth differences between foreground and background even though they have similar colour, segmentation should be improved. Our experimental results show segmentation based on both colour and depth cues outperform those relying purely on colour based on an error evaluation with respect to ground truth.

2 Figure 1: Statue dataset. The colour image and the corresponding depth image. White pixel indicates closer range, while black one indicates more distant range. 2 Colour-Based Segmentation Currently, graphcut based interactive segmentation uses region (colour or intensity similarity) [Agarwala and Dontcheva, 2004; Barrett and Cheney, 2002; Reese and Barrett, 2002] and/or boundary (colour or intensity contrast) [Gleicher, 1995; Mortensen and Barrett, 1995; 1999] information to construct an objective function, often referred to as an energy function. Hard constraints are then introduced by a user, e.g., by selecting foreground and background pixels. Soft constraints refer to the inherent properties, e.g., assumption on elsewhere smoothness or piecewise smoothness across pixels. For completeness we briefly summarise the graphcut optimisation algorithm as follows. The original image is represented as the corresponding graph G =< V, E > which is defined as a set of nodes (V) and a set of unordered edges (E) that connect these nodes. The nodes relate to the pixels in the original image, while the edges relate to the relationship between the adjacent pixels. Let L = (L 1,, L i,, L V ) be the binary vector whose elements L i specify assignments to pixels i in V. Each L i can be either background (L i = 0) or foreground (L i = 1). Here the optimised vector L defines the final segmentation. The energy function used in this paper is similar to the Gibbs energy function described in [Geman Figure 2: Camel dataset. The colour image and the corresponding depth image. White pixel indicates closer range, while black one indicates most distant range. and Geman, 1984]: E(L) = i V E R (L i ) + λ (i,j) E E B (L i, L j ) (1) where E R (L i ) is the region-based energy encoding the cost when the label of the node i is L i, and E B (L i, L j ) is the boundary-based energy, representing the cost when the label of adjacent nodes (pixels) i and j are L i and L j respectively, and λ [0, 1] indicates the relative importance of the region-based energy versus the boundarybased energy. Note that when we make a negative logarithm to both sides of the basic Bayesian formula (2), we can get Eq. (3) as follows: P (L D) P (D L)P (L) (2) ln(p (L D)) ln(p (D L)) + ( ln(p (L))) (3) where L denotes the label for all pixels, D encodes the observed data, i.e., the pixels with pre-defined labels, and P (L D) is a conditional probability. Comparing (1) and (3), the term E R (L i ) in the energy function can be seen as the likelihood probability, while the term E B (L i, L j ) can be seen as the prior probability. So we also refer E R (L i ) and E B (L i, L j ) as the likelihood energy and prior energy respectively.

3 Figure 3: Lazysnapping GUI. Red strokes select the foreground seed, and blue strokes select the background seed. Using the colour or intensity similarity and contrast, we can obtain the weights between each node in the graph corresponding to the original image. Then the combinatorial optimisation algorithm max-flow/min-cut [Boykov and Kolmogorov, 2004] is employed to minimise the energy function. This results in nodes grouped into different classes (e.g., source and sink in binary graph theory) which is equivalent to assigning the corresponding pixels into different classes (e.g., foreground and background in bi-layer segmentation). 3 Depth and Colour-Based Segmentation In contrast to the above formulation, we use both colour and depth cues to construct the energy function (1), and can be formed as follows: " # X X d c E(L) = θ ER (Li ) + (1 θ) ER (Li ) i V i V +λ θ X (i,j) E c EB (Li, Lj ) + (1 θ) X d EB (Li, Lj ) (i,j) E (4) where θ denotes the relative importance among the colour and depth terms, the superscript c, d of the energy term encode the colour term and depth term respectively,and the other variables have the same meaning with that in (1). The method to construct the depth term is as in the framework (4), and the tuning parameters except θ are set to be the same value described in [Li et al., 2004; Rother et al., 2004]. 3.1 Colour Data and Depth Data Modelling In the implementation of Lazysnapping, we select some pixels as foreground and others as background using pen Figure 4: GrabCut GUI. The rectangle is drawn by the user, while the internal thin red line indicates the graph cut between foreground and background generated by Grabcut. strokes (Figure 3). The corresponding pixels in the depth image are also selected. This results in two sets; foreground set F and background set B. All the known labelled pixels in F and B are used to construct the colour and depth energy terms respectively. With respect to the region-based energy term, we use the distance between each unlabelled pixel and the centroid of the selected foreground and background pixels precomputed using K-means [Kwatra et al., 2003]. Specifically, the selected pixels of foreground and background are used to construct the models of foreground and background by K-means based on colour distribution. While for the boundary-based energy term, we use the gradient or contrast between two adjacent pixels. In the implementation of Grabcut, we use a rectangle to select the internal and external pixels of the colour image denoting the uncertain set U (used to model the foreground) and background set B respectively (Figure 4). The corresponding pixels in the depth image are chosen as well. Here we construct a Gaussian Mixture Model (GMMs) based on the selected set to represent the region-based energy term, while the boundary-based energy term is still constructed using the gradient or contrast between two adjacent pixels. Further details on user input and modelling are shown in [Li et al., 2004; Rother et al., 2004]. 3.2 Segmentation using Lazysnapping and GrabCut Image resolution highly affects computational efficiency. To improve efficiency, we use the Watershed algorithm [Vincent and Soille, 1991] to divide the entire image into several new patches which still locate boundaries well and preserve small differences between each patch. This process is referred to as pre-segmentation in [Li et al., 2004].

4 (a) Weight θ = 0 (b) Weight θ = 0.9 (c) Weight θ = 1 Figure 5: Lazysnapping results with varying weights between the colour and depth term. Note that as θ increases which means the weight of colour cue increases, the segmentation performance gets better due to the depth ambiguity. Lazysnapping runs the graphcut algorithm once to acquire the final segmentation result, while Grabcut runs iteratively until the final segmentation result remains constant. 4 Results This section presents the results from an implementation of the Lazysnapping and Grabcut algorithms using depth and colour cues on two datasets of a statue and a toy camel. Modifications to freely available code provided by Gupta were used to produce the following results [Gupta, 2005]. Each dataset consists of colour images and corresponding depth map images. The depth map for the colour images is obtained from SFM. To a pair of colour images, the 3D position of a dense set of visual features was estimated using SFM algorithms [Ullman, 1979; Zhang et al., 2006]. Here we use GPU to extract the SURF [Bay et al., 2006] features and then match them to make the features trackable across multi-views. RANSAC [Fischler and Bolles, 1981] will be used to get rid of outliers, i.e., mismatched feature pairs. Finally,depth map will be generated by triangulation and interpolation over multi-views. The tuning parameter θ in (4) is varied from 0 to 1. If θ = 0, segmentation depends only on depth information, while θ = 1, forces a dependence on colour information only - identical to the original Lazysnapping and Grabcut algorithms. Both colour and depth cues contribute to the segmentation when θ (0, 1). The images used in this paper are shown in Figure 1 and Figure 2. In Figure 1, there is a high contrast in depth information which can complement the drawback of colour only. While in Figure 2, the dataset has similar depth information, i.e., the objects lie in an approximate plane, so the colour information should be more useful in this scenario. 4.1 Lazysnapping based on Colour and Depth Here the pixels intersecting with the red stroke are foreground set F, and the ones intersecting with the blue stroke are background set B. Figure 3 illustrates the GUI for Lazysnapping. The final segmentation result is shown in Figure GrabCut based on Colour and Depth With Grabcut, we use a red rectangle to divide the pixels into background set B (external to the rectangle) and the uncertain set U (internal to the rectangle). The Grabcut GUI is shown in Figure 4. Figure 6 illustrates the final segmentation result with different weights for the colour term. Note that Grabcut algorithm is an iterative energy minimisation. The energy should decrease with the number of iterations, which is shown in Figure 7. Figure 7: The energy E for the segmentation converges over 13 iterations..

5 (a) Weight θ = 0 (b) Weight θ = 0.55 (c) Weight θ = 1 Figure 6: GrabCut results with varying weights between the colour and depth term. Note that as θ increases which means the weight of colour cue increases, the segmentation performance gets worse due to the colour ambiguity. 4.3 Evaluation To evaluate the segmentation performance, we manually segmented both images to form a ground truth (Figure 8). Results were then compared using two methods of evaluation; 1) the L2 distance between the segmentation result from the proposed method and ground truth segmentation, referred to as ε1 in (5), 2) the number of pixels mislabelled as compared to the ground truth as a ratio with the total pixels in the original image, denoted by ε2 in (6). ε1 = P P G 2 (5) ε2 = N error N total (6) where P is the intensity of pixels in the segmentation derived by the proposed method, while P G is the intensity of pixels in the ground truth segmentation. In equation (6), N error is the number of misclassified pixels comparing to the ground truth and N total is the number of total pixels in original image. 4.4 Result Discussion In Figure 7, the energy converges after approximate 10 iterations. Particularly, it decreases dramatically at the beginning couple of iterations. It indicates Grabcut algorithm still converges quickly even after introducing the depth cue. Figure 9 indicates the best results are obtained when the weight θ is set to 0.55 for the statue dataset using Grabcut algorithm. For the camel dataset using Lazysnapping algorithm, the weight θ is set to 0.9 meaning colour information is more important than depth information in this case. The quantitative evaluations show that jointly using colour and depth cues in general, achieves better accuracy than using colour alone. Furthermore, the performance of segmentation can be refined by determining the weight θ based on the discriminative capabilities of the colour and depth cues. By further investigating Figure 9 we can find that the colour information is quite ambiguous for the statue segmentation. The depth information seems to be a reliable cue and demonstrates good performance, especially when the weight for depth term reaches But the accuracy of segmentation on statue slightly decreases as the weight for depth term continuously increases. For this case, the discriminative capability of depth cue achieves a maximum at 0.55, and the increased importance of the depth information causes the background objects to be mislabelled as foreground. However, depth information is more ambiguous for the camel dataset as all the small objects lie on the table, and the depth distributions of the toy camel and the books are quite similar. In this case, increasing the weight of colour cue refines the segmentation results. 5 Conclusion and Future Work We have introduced the depth cue into the basic energy function framework for interactive segmentation in a static image. As mentioned before, better segmentation results will be obtained based on colour and depth cues rather than using only one. With an appropriate weight θ, the proposed method outperforms Lazysnapping and Grabcut methods based only on colour cue. In the future, we would like to explore several directions in which the accuracy of segmentation can be further improved. Develop a novel model to automatically construct the foreground and background data selected by the user Develop an adaptive way to adjust the weight θ Extend the current method to video sequences Use both depth and colour to design an automatic segmentation method with reasonable accuracy for object tracking for mobile robotics [Wang et al., 2006; Zhao et al., 2008]

6 (a) Groundtruth for camel segmentation (b) Groundtruth for statue segmentation Figure 8: Groundtruth dataset. 1.6 x 106 The L2 distance between the segmentation result and groundtruth 0.03 The ratio of mislabling pixels verse the total pixels The L2 distance for all pixels ratio of mislabeling(wrong segmentation) weight for depth term in the energy function(%) (a) ɛ1 over different depth weights θ for statue dataset weight for depth term in the energy function(%) (b) ɛ2 over different depth weights θ for statue dataset 4.6 x 106 The L2 distance between the segmentation result and groundtruth 0.02 The ratio of mislabling pixels verse the total pixels The L2 distance for all pixels ratio of mislabeling(wrong segmentation) weight for depth term in the energy function(%) (c) ɛ1 over different depth weights θ for camel dataset weight for depth term in the energy function(%) (d) ɛ2 over different depth weights θ for camel dataset Figure 9: Segmentation error verse varying weights θ between colour and depth. (a,b) shows error for the statue segmentation, while (c,d) indicates error for the camel segmentation. As can be seen error decreases with the weight of depth term increasing to some extent for the statue dataset, however, error increases for the camel dataset.

7 Acknowledgement This research is supported by CRCMining and CSC scholarship. The authors acknowledge helpful discussions with Mohit Gupta. References [Agarwala and Dontcheva, 2004] A. Agarwala and M. et.al Dontcheva. Interactive digital photomontage. ACM Transactions on Graphics (TOG), 23(3): , [Barrett and Cheney, 2002] W.A. Barrett and A.S. Cheney. Object-based image editing. ACM Transactions on Graphics, 21(3): , [Batavia and Singh, 2001] P.H. Batavia and S. Singh. Obstacle detection using adaptive color segmentation and color stereo homography. In IEEE International Conference on Robotics and Automation, volume 1, pages Citeseer, [Bay et al., 2006] H. Bay, T. Tuytelaars, and L. Van Gool. Surf: Speeded up robust features. Computer Vision ECCV 2006, pages , [Boykov and Jolly, 2001] Y. Boykov and M.P. Jolly. Interactive graph cuts for optimal boundary and region segmentation of objects in ND images. In International Conference on Computer Vision, volume 1, pages Citeseer, [Boykov and Kolmogorov, 2004] Y. Boykov and V. Kolmogorov. An experimental comparison of mincut/max-flow algorithms for energy minimization in vision. IEEE Transactions on Pattern Analysis and Machine Intelligence, pages , [Fischler and Bolles, 1981] M.A. Fischler and R.C. Bolles. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Communications of the ACM, 24(6): , [Geman and Geman, 1984] S. Geman and D. Geman. Stochastic relaxation,gibbs distributions, and the Bayesian restoration of images. IEEE trans on pattern analysis and machine inteligence, [Gleicher, 1995] M. Gleicher. Image snapping. In Proceedings of the 22nd annual conference on Computer graphics and interactive techniques, page 190. ACM, [Gupta, 2005] M. Gupta. Interactive segmentation toolbox. In http: // www. cs. cmu. edu/ ~ mohitg/ Research/ segmentation. htm, [Kwatra et al., 2003] V. Kwatra, A. Schodl, I. Essa, G. Turk, and A. Bobick. Graphcut textures: Image and video synthesis using graph cuts. ACM Transactions on Graphics, 22(3): , [Li et al., 2004] Y. Li, J. Sun, C.K. Tang, and H.Y. Shum. Lazy snapping. ACM Transactions on Graphics (TOG), 23(3): , [Mortensen and Barrett, 1995] E N Mortensen and W A Barrett. Intelligent scissors for image composition. In Proceedings of the 22nd annual conference on Computer graphics and interactive techniques, pages ACM, [Mortensen and Barrett, 1999] E N Mortensen and W A Barrett. Toboggan-based intelligent scissors with a four-parameter edgemodel. In Computer Vision and Pattern Recognition, IEEE Computer Society Conference on., volume 2, [Reese and Barrett, 2002] LJ Reese and WA Barrett. Image editing with intelligent paint. In Proceedings of Eurographics, volume 21, pages Citeseer, [Rother et al., 2004] C. Rother, V. Kolmogorov, and A. Blake. Grabcut: Interactive foreground extraction using iterated graph cuts. In ACM SIGGRAPH 2004 Papers, page 314. ACM, [Salembier et al., 1997] P. Salembier, F. Marques, M. Pardas, J.R. Morros, I. Corset, S. Jeannin, L. Bouchard, F. Meyer, and B. Marcotegui. Segmentation-based video coding system allowing the manipulation ofobjects. IEEE Transactions on Circuits and Systems for Video Technology, 7(1):60 74, [Schulz et al., 2001] D. Schulz, W. Burgard, D. Fox, and A.B. Cremers. Tracking multiple moving targets with a mobile robot using particle filters and statistical data association. In IEEE international conference on robotics and automation, Proceedings 2001 ICRA, volume 2, [Ullman, 1979] S. Ullman. The interpretation of structure from motion. Proceedings of the Royal Society of London. Series B. Biological Sciences, 203(1153):405, [Vincent and Soille, 1991] L. Vincent and P. Soille. Watersheds in digital spaces: an efficient algorithm based onimmersion simulations. IEEE transactions on pattern analysis and machine intelligence, 13(6): , [Wang et al., 2006] L. Wang, M. Liao, M. Gong, R. Yang, and D. Nister. High-quality real-time stereo using adaptive cost aggregation and dynamic programming. In 3D Data Processing, Visualization, and Transmission, Third International Symposium on, pages , [Zhang et al., 2006] G. Zhang, J. Jia, W. Xiong, T.T. Wong, P.A. Heng, and H. Bao. Moving object extrac-

8 tion with a hand-held camera. In Proc. International Conference on Computer Vision, pages 1 8. Citeseer, [Zhao et al., 2008] T. Zhao, R. Nevatia, and B. Wu. Segmentation and tracking of multiple humans in crowded environments. IEEE transactions on pattern analysis and machine intelligence, 30(7): , 2008.

Image Segmentation Using Iterated Graph Cuts Based on Multi-scale Smoothing

Image Segmentation Using Iterated Graph Cuts Based on Multi-scale Smoothing Image Segmentation Using Iterated Graph Cuts Based on Multi-scale Smoothing Tomoyuki Nagahashi 1, Hironobu Fujiyoshi 1, and Takeo Kanade 2 1 Dept. of Computer Science, Chubu University. Matsumoto 1200,

More information

Image Segmentation Using Iterated Graph Cuts BasedonMulti-scaleSmoothing

Image Segmentation Using Iterated Graph Cuts BasedonMulti-scaleSmoothing Image Segmentation Using Iterated Graph Cuts BasedonMulti-scaleSmoothing Tomoyuki Nagahashi 1, Hironobu Fujiyoshi 1, and Takeo Kanade 2 1 Dept. of Computer Science, Chubu University. Matsumoto 1200, Kasugai,

More information

Prof. Feng Liu. Spring /17/2017. With slides by F. Durand, Y.Y. Chuang, R. Raskar, and C.

Prof. Feng Liu. Spring /17/2017. With slides by F. Durand, Y.Y. Chuang, R. Raskar, and C. Prof. Feng Liu Spring 2017 http://www.cs.pdx.edu/~fliu/courses/cs510/ 05/17/2017 With slides by F. Durand, Y.Y. Chuang, R. Raskar, and C. Rother Last Time Image segmentation Normalized cut and segmentation

More information

Video Object Cut and Paste

Video Object Cut and Paste Video Object Cut and Paste Yin Li Jian Sun Heung-Yeung Shum Microsoft Research Asia (a) Figure 1: Given a video (a) containing an opaque video object on a complicated background, our system can generate

More information

A Feature Point Matching Based Approach for Video Objects Segmentation

A Feature Point Matching Based Approach for Video Objects Segmentation A Feature Point Matching Based Approach for Video Objects Segmentation Yan Zhang, Zhong Zhou, Wei Wu State Key Laboratory of Virtual Reality Technology and Systems, Beijing, P.R. China School of Computer

More information

Color Me Right Seamless Image Compositing

Color Me Right Seamless Image Compositing Color Me Right Seamless Image Compositing Dong Guo and Terence Sim School of Computing National University of Singapore Singapore, 117417 Abstract. This paper introduces an approach of creating an image

More information

Image Stitching using Watersheds and Graph Cuts

Image Stitching using Watersheds and Graph Cuts Image Stitching using Watersheds and Graph Cuts Patrik Nyman Centre for Mathematical Sciences, Lund University, Sweden patnym@maths.lth.se 1. Introduction Image stitching is commonly used in many different

More information

MRFs and Segmentation with Graph Cuts

MRFs and Segmentation with Graph Cuts 02/24/10 MRFs and Segmentation with Graph Cuts Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem Today s class Finish up EM MRFs w ij i Segmentation with Graph Cuts j EM Algorithm: Recap

More information

intro, applications MRF, labeling... how it can be computed at all? Applications in segmentation: GraphCut, GrabCut, demos

intro, applications MRF, labeling... how it can be computed at all? Applications in segmentation: GraphCut, GrabCut, demos Image as Markov Random Field and Applications 1 Tomáš Svoboda, svoboda@cmp.felk.cvut.cz Czech Technical University in Prague, Center for Machine Perception http://cmp.felk.cvut.cz Talk Outline Last update:

More information

Graph-Based Superpixel Labeling for Enhancement of Online Video Segmentation

Graph-Based Superpixel Labeling for Enhancement of Online Video Segmentation Graph-Based Superpixel Labeling for Enhancement of Online Video Segmentation Alaa E. Abdel-Hakim Electrical Engineering Department Assiut University Assiut, Egypt alaa.aly@eng.au.edu.eg Mostafa Izz Cairo

More information

A Survey of Light Source Detection Methods

A Survey of Light Source Detection Methods A Survey of Light Source Detection Methods Nathan Funk University of Alberta Mini-Project for CMPUT 603 November 30, 2003 Abstract This paper provides an overview of the most prominent techniques for light

More information

Soft Scissors : An Interactive Tool for Realtime High Quality Matting

Soft Scissors : An Interactive Tool for Realtime High Quality Matting Soft Scissors : An Interactive Tool for Realtime High Quality Matting Jue Wang University of Washington Maneesh Agrawala University of California, Berkeley Michael F. Cohen Microsoft Research Figure 1:

More information

Tri-modal Human Body Segmentation

Tri-modal Human Body Segmentation Tri-modal Human Body Segmentation Master of Science Thesis Cristina Palmero Cantariño Advisor: Sergio Escalera Guerrero February 6, 2014 Outline 1 Introduction 2 Tri-modal dataset 3 Proposed baseline 4

More information

QUT Digital Repository: This is the author version published as:

QUT Digital Repository:   This is the author version published as: QUT Digital Repository: http://eprints.qut.edu.au/ This is the author version published as: This is the accepted version of this article. To be published as : This is the author version published as: Chen,

More information

Lazy Snapping. A paper from Siggraph04 by Yin Li, Jian Sun, Chi-KeungTang, Heung-Yeung Shum, Microsoft Research Asia. Presented by Gerhard Röthlin

Lazy Snapping. A paper from Siggraph04 by Yin Li, Jian Sun, Chi-KeungTang, Heung-Yeung Shum, Microsoft Research Asia. Presented by Gerhard Röthlin A paper from Siggraph04 by Yin Li, Jian Sun, Chi-KeungTang, Heung-Yeung Shum, Microsoft Research Asia Presented by Gerhard Röthlin 1 Image Cutout Composing a foreground object with an alternative background

More information

Segmentation and Tracking of Partial Planar Templates

Segmentation and Tracking of Partial Planar Templates Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract

More information

Human Head-Shoulder Segmentation

Human Head-Shoulder Segmentation Human Head-Shoulder Segmentation Hai Xin, Haizhou Ai Computer Science and Technology Tsinghua University Beijing, China ahz@mail.tsinghua.edu.cn Hui Chao, Daniel Tretter Hewlett-Packard Labs 1501 Page

More information

Segmenting Objects in Weakly Labeled Videos

Segmenting Objects in Weakly Labeled Videos Segmenting Objects in Weakly Labeled Videos Mrigank Rochan, Shafin Rahman, Neil D.B. Bruce, Yang Wang Department of Computer Science University of Manitoba Winnipeg, Canada {mrochan, shafin12, bruce, ywang}@cs.umanitoba.ca

More information

IMA Preprint Series # 2153

IMA Preprint Series # 2153 DISTANCECUT: INTERACTIVE REAL-TIME SEGMENTATION AND MATTING OF IMAGES AND VIDEOS By Xue Bai and Guillermo Sapiro IMA Preprint Series # 2153 ( January 2007 ) INSTITUTE FOR MATHEMATICS AND ITS APPLICATIONS

More information

Perception-based Seam-cutting for Image Stitching

Perception-based Seam-cutting for Image Stitching Perception-based Seam-cutting for Image Stitching Nan Li Tianli Liao Chao Wang Received: xxx / Accepted: xxx Abstract Image stitching is still challenging in consumerlevel photography due to imperfect

More information

People Tracking and Segmentation Using Efficient Shape Sequences Matching

People Tracking and Segmentation Using Efficient Shape Sequences Matching People Tracking and Segmentation Using Efficient Shape Sequences Matching Junqiu Wang, Yasushi Yagi, and Yasushi Makihara The Institute of Scientific and Industrial Research, Osaka University 8-1 Mihogaoka,

More information

arxiv: v1 [cs.cv] 28 Sep 2018

arxiv: v1 [cs.cv] 28 Sep 2018 Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,

More information

AUTOMATIC OBJECT EXTRACTION IN SINGLE-CONCEPT VIDEOS. Kuo-Chin Lien and Yu-Chiang Frank Wang

AUTOMATIC OBJECT EXTRACTION IN SINGLE-CONCEPT VIDEOS. Kuo-Chin Lien and Yu-Chiang Frank Wang AUTOMATIC OBJECT EXTRACTION IN SINGLE-CONCEPT VIDEOS Kuo-Chin Lien and Yu-Chiang Frank Wang Research Center for Information Technology Innovation, Academia Sinica, Taipei, Taiwan {iker, ycwang}@citi.sinica.edu.tw

More information

Markov/Conditional Random Fields, Graph Cut, and applications in Computer Vision

Markov/Conditional Random Fields, Graph Cut, and applications in Computer Vision Markov/Conditional Random Fields, Graph Cut, and applications in Computer Vision Fuxin Li Slides and materials from Le Song, Tucker Hermans, Pawan Kumar, Carsten Rother, Peter Orchard, and others Recap:

More information

Using temporal seeding to constrain the disparity search range in stereo matching

Using temporal seeding to constrain the disparity search range in stereo matching Using temporal seeding to constrain the disparity search range in stereo matching Thulani Ndhlovu Mobile Intelligent Autonomous Systems CSIR South Africa Email: tndhlovu@csir.co.za Fred Nicolls Department

More information

Object detection using non-redundant local Binary Patterns

Object detection using non-redundant local Binary Patterns University of Wollongong Research Online Faculty of Informatics - Papers (Archive) Faculty of Engineering and Information Sciences 2010 Object detection using non-redundant local Binary Patterns Duc Thanh

More information

arxiv: v1 [cs.cv] 28 Sep 2018

arxiv: v1 [cs.cv] 28 Sep 2018 Extrinsic camera calibration method and its performance evaluation Jacek Komorowski 1 and Przemyslaw Rokita 2 arxiv:1809.11073v1 [cs.cv] 28 Sep 2018 1 Maria Curie Sklodowska University Lublin, Poland jacek.komorowski@gmail.com

More information

Lecture 10: Multi view geometry

Lecture 10: Multi view geometry Lecture 10: Multi view geometry Professor Fei Fei Li Stanford Vision Lab 1 What we will learn today? Stereo vision Correspondence problem (Problem Set 2 (Q3)) Active stereo vision systems Structure from

More information

Transductive Object Cutout

Transductive Object Cutout Transductive Object Cutout Jingyu Cui 1, Qiong Yang 3, Fang Wen 2, Qiying Wu 4, Changshui Zhang 1, Luc Van Gool 3, and Xiaoou Tang 2 1 Tsinghua University, Beijing, China cui-jy@mails.thu.edu.cn, zcs@mail.thu.edu.cn

More information

Supervised texture detection in images

Supervised texture detection in images Supervised texture detection in images Branislav Mičušík and Allan Hanbury Pattern Recognition and Image Processing Group, Institute of Computer Aided Automation, Vienna University of Technology Favoritenstraße

More information

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching Stereo Matching Fundamental matrix Let p be a point in left image, p in right image l l Epipolar relation p maps to epipolar line l p maps to epipolar line l p p Epipolar mapping described by a 3x3 matrix

More information

Depth Propagation with Key-Frame Considering Movement on the Z-Axis

Depth Propagation with Key-Frame Considering Movement on the Z-Axis , pp.131-135 http://dx.doi.org/10.1457/astl.014.47.31 Depth Propagation with Key-Frame Considering Movement on the Z-Axis Jin Woo Choi 1, Taeg Keun Whangbo 1 Culture Technology Institute, Gachon University,

More information

UNSUPERVISED CO-SEGMENTATION BASED ON A NEW GLOBAL GMM CONSTRAINT IN MRF. Hongkai Yu, Min Xian, and Xiaojun Qi

UNSUPERVISED CO-SEGMENTATION BASED ON A NEW GLOBAL GMM CONSTRAINT IN MRF. Hongkai Yu, Min Xian, and Xiaojun Qi UNSUPERVISED CO-SEGMENTATION BASED ON A NEW GLOBAL GMM CONSTRAINT IN MRF Hongkai Yu, Min Xian, and Xiaojun Qi Department of Computer Science, Utah State University, Logan, UT 84322-4205 hongkai.yu@aggiemail.usu.edu,

More information

FOREGROUND DETECTION ON DEPTH MAPS USING SKELETAL REPRESENTATION OF OBJECT SILHOUETTES

FOREGROUND DETECTION ON DEPTH MAPS USING SKELETAL REPRESENTATION OF OBJECT SILHOUETTES FOREGROUND DETECTION ON DEPTH MAPS USING SKELETAL REPRESENTATION OF OBJECT SILHOUETTES D. Beloborodov a, L. Mestetskiy a a Faculty of Computational Mathematics and Cybernetics, Lomonosov Moscow State University,

More information

III. VERVIEW OF THE METHODS

III. VERVIEW OF THE METHODS An Analytical Study of SIFT and SURF in Image Registration Vivek Kumar Gupta, Kanchan Cecil Department of Electronics & Telecommunication, Jabalpur engineering college, Jabalpur, India comparing the distance

More information

Epipolar Geometry Based On Line Similarity

Epipolar Geometry Based On Line Similarity 2016 23rd International Conference on Pattern Recognition (ICPR) Cancún Center, Cancún, México, December 4-8, 2016 Epipolar Geometry Based On Line Similarity Gil Ben-Artzi Tavi Halperin Michael Werman

More information

Adding a Transparent Object on Image

Adding a Transparent Object on Image Adding a Transparent Object on Image Liliana, Meliana Luwuk, Djoni Haryadi Setiabudi Informatics Department, Petra Christian University, Surabaya, Indonesia lilian@petra.ac.id, m26409027@john.petra.ac.id,

More information

Thin Plate Spline Feature Point Matching for Organ Surfaces in Minimally Invasive Surgery Imaging

Thin Plate Spline Feature Point Matching for Organ Surfaces in Minimally Invasive Surgery Imaging Thin Plate Spline Feature Point Matching for Organ Surfaces in Minimally Invasive Surgery Imaging Bingxiong Lin, Yu Sun and Xiaoning Qian University of South Florida, Tampa, FL., U.S.A. ABSTRACT Robust

More information

Saliency Based Video Object Recognition in Single Concept Video

Saliency Based Video Object Recognition in Single Concept Video Saliency Based Video Object Recognition in Single Concept Video Raihanath A.S 1, Chithra Rani P R 2 1 P G Student, Department of Computer Science, Ilahia College of Engineering & Technology (ICET), Kerala,

More information

[2006] IEEE. Reprinted, with permission, from [Wenjing Jia, Huaifeng Zhang, Xiangjian He, and Qiang Wu, A Comparison on Histogram Based Image

[2006] IEEE. Reprinted, with permission, from [Wenjing Jia, Huaifeng Zhang, Xiangjian He, and Qiang Wu, A Comparison on Histogram Based Image [6] IEEE. Reprinted, with permission, from [Wenjing Jia, Huaifeng Zhang, Xiangjian He, and Qiang Wu, A Comparison on Histogram Based Image Matching Methods, Video and Signal Based Surveillance, 6. AVSS

More information

Visualization 2D-to-3D Photo Rendering for 3D Displays

Visualization 2D-to-3D Photo Rendering for 3D Displays Visualization 2D-to-3D Photo Rendering for 3D Displays Sumit K Chauhan 1, Divyesh R Bajpai 2, Vatsal H Shah 3 1 Information Technology, Birla Vishvakarma mahavidhyalaya,sumitskc51@gmail.com 2 Information

More information

From Orientation to Functional Modeling for Terrestrial and UAV Images

From Orientation to Functional Modeling for Terrestrial and UAV Images From Orientation to Functional Modeling for Terrestrial and UAV Images Helmut Mayer 1 Andreas Kuhn 1, Mario Michelini 1, William Nguatem 1, Martin Drauschke 2 and Heiko Hirschmüller 2 1 Visual Computing,

More information

Image Inpainting and Selective Motion Blur

Image Inpainting and Selective Motion Blur Image Inpainting and Selective Motion Blur Gaurav Verma Dept. of Electrical Engineering, IIT Kanpur 14244, gverma@iitk.ac.in Abstract: This report is presented as a part of the coursework for EE604A, Image

More information

Fundamental Matrices from Moving Objects Using Line Motion Barcodes

Fundamental Matrices from Moving Objects Using Line Motion Barcodes Fundamental Matrices from Moving Objects Using Line Motion Barcodes Yoni Kasten (B), Gil Ben-Artzi, Shmuel Peleg, and Michael Werman School of Computer Science and Engineering, The Hebrew University of

More information

Discrete Optimization Methods in Computer Vision CSE 6389 Slides by: Boykov Modified and Presented by: Mostafa Parchami Basic overview of graph cuts

Discrete Optimization Methods in Computer Vision CSE 6389 Slides by: Boykov Modified and Presented by: Mostafa Parchami Basic overview of graph cuts Discrete Optimization Methods in Computer Vision CSE 6389 Slides by: Boykov Modified and Presented by: Mostafa Parchami Basic overview of graph cuts [Yuri Boykov, Olga Veksler, Ramin Zabih, Fast Approximation

More information

Fast Interactive Image Segmentation by Discriminative Clustering

Fast Interactive Image Segmentation by Discriminative Clustering Fast Interactive Image Segmentation by Discriminative Clustering Dingding Liu Department of Electrical Engineering University of Washington Seattle, WA, USA, 98105 liudd@u.washington.edu Kari Pulli Nokia

More information

arxiv: v4 [cs.cv] 8 Jan 2017

arxiv: v4 [cs.cv] 8 Jan 2017 Epipolar Geometry Based On Line Similarity Gil Ben-Artzi Tavi Halperin Michael Werman Shmuel Peleg School of Computer Science and Engineering The Hebrew University of Jerusalem, Israel arxiv:1604.04848v4

More information

Interactive Shadow Editing from Single Images

Interactive Shadow Editing from Single Images Interactive Shadow Editing from Single Images Han Gong, Darren Cosker Department of Computer Science, University of Bath Abstract. We present a system for interactive shadow editing from single images

More information

Iterated Graph Cuts for Image Segmentation

Iterated Graph Cuts for Image Segmentation Iterated Graph Cuts for Image Segmentation Bo Peng 1, Lei Zhang 1, and Jian Yang 2 1 Department of Computing, The Hong Kong Polytechnic University, Kowloon, Hong Kong, China. 2 School of Computer Science

More information

Image Guided Cost Aggregation for Hierarchical Depth Map Fusion

Image Guided Cost Aggregation for Hierarchical Depth Map Fusion Image Guided Cost Aggregation for Hierarchical Depth Map Fusion Thilo Borgmann and Thomas Sikora Communication Systems Group, Technische Universität Berlin {borgmann, sikora}@nue.tu-berlin.de Keywords:

More information

Announcements. Image Segmentation. From images to objects. Extracting objects. Status reports next Thursday ~5min presentations in class

Announcements. Image Segmentation. From images to objects. Extracting objects. Status reports next Thursday ~5min presentations in class Image Segmentation Announcements Status reports next Thursday ~5min presentations in class Project voting From Sandlot Science Today s Readings Forsyth & Ponce, Chapter 1 (plus lots of optional references

More information

Segmentation. Separate image into coherent regions

Segmentation. Separate image into coherent regions Segmentation II Segmentation Separate image into coherent regions Berkeley segmentation database: http://www.eecs.berkeley.edu/research/projects/cs/vision/grouping/segbench/ Slide by L. Lazebnik Interactive

More information

Active 3D Shape Acquisition Using Smartphones

Active 3D Shape Acquisition Using Smartphones Active 3D Shape Acquisition Using Smartphones Jae Hyun Won won1425@gmail.com Man Hee Lee maninara@gmail.com In Kyu Park pik@inha.ac.kr School of Information and Communication Engineering, Inha University,

More information

Textureless Layers CMU-RI-TR Qifa Ke, Simon Baker, and Takeo Kanade

Textureless Layers CMU-RI-TR Qifa Ke, Simon Baker, and Takeo Kanade Textureless Layers CMU-RI-TR-04-17 Qifa Ke, Simon Baker, and Takeo Kanade The Robotics Institute Carnegie Mellon University 5000 Forbes Avenue Pittsburgh, PA 15213 Abstract Layers are one of the most well

More information

Ensemble of Bayesian Filters for Loop Closure Detection

Ensemble of Bayesian Filters for Loop Closure Detection Ensemble of Bayesian Filters for Loop Closure Detection Mohammad Omar Salameh, Azizi Abdullah, Shahnorbanun Sahran Pattern Recognition Research Group Center for Artificial Intelligence Faculty of Information

More information

FOREGROUND SEGMENTATION BASED ON MULTI-RESOLUTION AND MATTING

FOREGROUND SEGMENTATION BASED ON MULTI-RESOLUTION AND MATTING FOREGROUND SEGMENTATION BASED ON MULTI-RESOLUTION AND MATTING Xintong Yu 1,2, Xiaohan Liu 1,2, Yisong Chen 1 1 Graphics Laboratory, EECS Department, Peking University 2 Beijing University of Posts and

More information

Markov Random Fields and Segmentation with Graph Cuts

Markov Random Fields and Segmentation with Graph Cuts Markov Random Fields and Segmentation with Graph Cuts Computer Vision Jia-Bin Huang, Virginia Tech Many slides from D. Hoiem Administrative stuffs Final project Proposal due Oct 27 (Thursday) HW 4 is out

More information

USER DRIVEN SPARSE POINT-BASED IMAGE SEGMENTATION. Sachin Meena Kannappan Palaniappan Guna Seetharaman

USER DRIVEN SPARSE POINT-BASED IMAGE SEGMENTATION. Sachin Meena Kannappan Palaniappan Guna Seetharaman USER DRIVEN SPARSE POINT-BASED IMAGE SEGMENTATION Sachin Meena Kannappan Palaniappan Guna Seetharaman Department of Computer Science, University of Missouri-Columbia, MO 6511 USA US Naval Research Laboratory,

More information

Automatic Colorization of Grayscale Images

Automatic Colorization of Grayscale Images Automatic Colorization of Grayscale Images Austin Sousa Rasoul Kabirzadeh Patrick Blaes Department of Electrical Engineering, Stanford University 1 Introduction ere exists a wealth of photographic images,

More information

Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision

Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision Zhiyan Zhang 1, Wei Qian 1, Lei Pan 1 & Yanjun Li 1 1 University of Shanghai for Science and Technology, China

More information

VIDEO STABILIZATION WITH L1-L2 OPTIMIZATION. Hui Qu, Li Song

VIDEO STABILIZATION WITH L1-L2 OPTIMIZATION. Hui Qu, Li Song VIDEO STABILIZATION WITH L-L2 OPTIMIZATION Hui Qu, Li Song Institute of Image Communication and Network Engineering, Shanghai Jiao Tong University ABSTRACT Digital videos often suffer from undesirable

More information

A Benchmark for Interactive Image Segmentation Algorithms

A Benchmark for Interactive Image Segmentation Algorithms A Benchmark for Interactive Image Segmentation Algorithms Yibiao Zhao 1,3, Xiaohan Nie 2,3, Yanbiao Duan 2,3, Yaping Huang 1, Siwei Luo 1 1 Beijing Jiaotong University, 2 Beijing Institute of Technology,

More information

Edge tracking for motion segmentation and depth ordering

Edge tracking for motion segmentation and depth ordering Edge tracking for motion segmentation and depth ordering P. Smith, T. Drummond and R. Cipolla Department of Engineering University of Cambridge Cambridge CB2 1PZ,UK {pas1001 twd20 cipolla}@eng.cam.ac.uk

More information

Principal-channels for One-sided Object Cutout

Principal-channels for One-sided Object Cutout Principal-channels for One-sided Object Cutout Lior Gavish, Lior Wolf, Lior Shapira, and Daniel Cohen-Or Tel-Aviv University, Tel-Aviv, Israel Abstract We introduce principal-channels for cutting out objects

More information

CS6670: Computer Vision

CS6670: Computer Vision CS6670: Computer Vision Noah Snavely Lecture 19: Graph Cuts source S sink T Readings Szeliski, Chapter 11.2 11.5 Stereo results with window search problems in areas of uniform texture Window-based matching

More information

Segmentation. Bottom up Segmentation Semantic Segmentation

Segmentation. Bottom up Segmentation Semantic Segmentation Segmentation Bottom up Segmentation Semantic Segmentation Semantic Labeling of Street Scenes Ground Truth Labels 11 classes, almost all occur simultaneously, large changes in viewpoint, scale sky, road,

More information

Animating Characters in Pictures

Animating Characters in Pictures Animating Characters in Pictures Shih-Chiang Dai jeffrey@cmlab.csie.ntu.edu.tw Chun-Tse Hsiao hsiaochm@cmlab.csie.ntu.edu.tw Bing-Yu Chen robin@ntu.edu.tw ABSTRACT Animating pictures is an interesting

More information

Segmentation with non-linear constraints on appearance, complexity, and geometry

Segmentation with non-linear constraints on appearance, complexity, and geometry IPAM February 2013 Western Univesity Segmentation with non-linear constraints on appearance, complexity, and geometry Yuri Boykov Andrew Delong Lena Gorelick Hossam Isack Anton Osokin Frank Schmidt Olga

More information

Probabilistic Facial Feature Extraction Using Joint Distribution of Location and Texture Information

Probabilistic Facial Feature Extraction Using Joint Distribution of Location and Texture Information Probabilistic Facial Feature Extraction Using Joint Distribution of Location and Texture Information Mustafa Berkay Yilmaz, Hakan Erdogan, Mustafa Unel Sabanci University, Faculty of Engineering and Natural

More information

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching Stereo Matching Fundamental matrix Let p be a point in left image, p in right image l l Epipolar relation p maps to epipolar line l p maps to epipolar line l p p Epipolar mapping described by a 3x3 matrix

More information

Chaplin, Modern Times, 1936

Chaplin, Modern Times, 1936 Chaplin, Modern Times, 1936 [A Bucket of Water and a Glass Matte: Special Effects in Modern Times; bonus feature on The Criterion Collection set] Multi-view geometry problems Structure: Given projections

More information

Viewpoint Invariant Features from Single Images Using 3D Geometry

Viewpoint Invariant Features from Single Images Using 3D Geometry Viewpoint Invariant Features from Single Images Using 3D Geometry Yanpeng Cao and John McDonald Department of Computer Science National University of Ireland, Maynooth, Ireland {y.cao,johnmcd}@cs.nuim.ie

More information

Stereo and Epipolar geometry

Stereo and Epipolar geometry Previously Image Primitives (feature points, lines, contours) Today: Stereo and Epipolar geometry How to match primitives between two (multiple) views) Goals: 3D reconstruction, recognition Jana Kosecka

More information

Optimization. Intelligent Scissors (see also Snakes)

Optimization. Intelligent Scissors (see also Snakes) Optimization We can define a cost for possible solutions Number of solutions is large (eg., exponential) Efficient search is needed Global methods: cleverly find best solution without considering all.

More information

What have we leaned so far?

What have we leaned so far? What have we leaned so far? Camera structure Eye structure Project 1: High Dynamic Range Imaging What have we learned so far? Image Filtering Image Warping Camera Projection Model Project 2: Panoramic

More information

Extracting Smooth and Transparent Layers from a Single Image

Extracting Smooth and Transparent Layers from a Single Image Extracting Smooth and Transparent Layers from a Single Image Sai-Kit Yeung Tai-Pang Wu Chi-Keung Tang saikit@ust.hk pang@ust.hk cktang@cse.ust.hk Vision and Graphics Group The Hong Kong University of Science

More information

PAPER Video Segmentation with Motion Smoothness

PAPER Video Segmentation with Motion Smoothness IEICE TRANS. INF. & SYST., VOL.E93 D, NO.4 APRIL 2010 873 PAPER Video Segmentation with Motion Smoothness Chung-Lin WEN a), Nonmember, Bing-Yu CHEN b), and Yoichi SATO c), Members SUMMARY In this paper,

More information

Occlusion Detection of Real Objects using Contour Based Stereo Matching

Occlusion Detection of Real Objects using Contour Based Stereo Matching Occlusion Detection of Real Objects using Contour Based Stereo Matching Kenichi Hayashi, Hirokazu Kato, Shogo Nishida Graduate School of Engineering Science, Osaka University,1-3 Machikaneyama-cho, Toyonaka,

More information

An Integrated System for Digital Restoration of Prehistoric Theran Wall Paintings

An Integrated System for Digital Restoration of Prehistoric Theran Wall Paintings An Integrated System for Digital Restoration of Prehistoric Theran Wall Paintings Nikolaos Karianakis 1 Petros Maragos 2 1 University of California, Los Angeles 2 National Technical University of Athens

More information

VIDEO MATTING USING MOTION EXTENDED GRABCUT

VIDEO MATTING USING MOTION EXTENDED GRABCUT VIDEO MATTING USING MOTION EXTENDED GRABCUT David Corrigan 1, Simon Robinson 2 and Anil Kokaram 1 1 Sigmedia Group, Trinity College Dublin, Ireland. Email: {corrigad, akokaram}@tcd.ie 2 The Foundry, London,

More information

Depth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth

Depth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth Common Classification Tasks Recognition of individual objects/faces Analyze object-specific features (e.g., key points) Train with images from different viewing angles Recognition of object classes Analyze

More information

Separating Objects and Clutter in Indoor Scenes

Separating Objects and Clutter in Indoor Scenes Separating Objects and Clutter in Indoor Scenes Salman H. Khan School of Computer Science & Software Engineering, The University of Western Australia Co-authors: Xuming He, Mohammed Bennamoun, Ferdous

More information

Learning to Grasp Objects: A Novel Approach for Localizing Objects Using Depth Based Segmentation

Learning to Grasp Objects: A Novel Approach for Localizing Objects Using Depth Based Segmentation Learning to Grasp Objects: A Novel Approach for Localizing Objects Using Depth Based Segmentation Deepak Rao, Arda Kara, Serena Yeung (Under the guidance of Quoc V. Le) Stanford University Abstract We

More information

An Approach for Real Time Moving Object Extraction based on Edge Region Determination

An Approach for Real Time Moving Object Extraction based on Edge Region Determination An Approach for Real Time Moving Object Extraction based on Edge Region Determination Sabrina Hoque Tuli Department of Computer Science and Engineering, Chittagong University of Engineering and Technology,

More information

Image Blending and Compositing NASA

Image Blending and Compositing NASA Image Blending and Compositing NASA CS194: Image Manipulation & Computational Photography Alexei Efros, UC Berkeley, Fall 2016 Image Compositing Compositing Procedure 1. Extract Sprites (e.g using Intelligent

More information

Pairwise Threshold for Gaussian Mixture Classification and its Application on Human Tracking Enhancement

Pairwise Threshold for Gaussian Mixture Classification and its Application on Human Tracking Enhancement Pairwise Threshold for Gaussian Mixture Classification and its Application on Human Tracking Enhancement Daegeon Kim Sung Chun Lee Institute for Robotics and Intelligent Systems University of Southern

More information

Image Segmentation Via Iterative Geodesic Averaging

Image Segmentation Via Iterative Geodesic Averaging Image Segmentation Via Iterative Geodesic Averaging Asmaa Hosni, Michael Bleyer and Margrit Gelautz Institute for Software Technology and Interactive Systems, Vienna University of Technology Favoritenstr.

More information

Fast Lighting Independent Background Subtraction

Fast Lighting Independent Background Subtraction Fast Lighting Independent Background Subtraction Yuri Ivanov Aaron Bobick John Liu [yivanov bobick johnliu]@media.mit.edu MIT Media Laboratory February 2, 2001 Abstract This paper describes a new method

More information

Image Resizing Based on Gradient Vector Flow Analysis

Image Resizing Based on Gradient Vector Flow Analysis Image Resizing Based on Gradient Vector Flow Analysis Sebastiano Battiato battiato@dmi.unict.it Giovanni Puglisi puglisi@dmi.unict.it Giovanni Maria Farinella gfarinellao@dmi.unict.it Daniele Ravì rav@dmi.unict.it

More information

Map-Enhanced UAV Image Sequence Registration and Synchronization of Multiple Image Sequences

Map-Enhanced UAV Image Sequence Registration and Synchronization of Multiple Image Sequences Map-Enhanced UAV Image Sequence Registration and Synchronization of Multiple Image Sequences Yuping Lin and Gérard Medioni Computer Science Department, University of Southern California 941 W. 37th Place,

More information

Calibration of a Different Field-of-view Stereo Camera System using an Embedded Checkerboard Pattern

Calibration of a Different Field-of-view Stereo Camera System using an Embedded Checkerboard Pattern Calibration of a Different Field-of-view Stereo Camera System using an Embedded Checkerboard Pattern Pathum Rathnayaka, Seung-Hae Baek and Soon-Yong Park School of Computer Science and Engineering, Kyungpook

More information

Supplementary Materials

Supplementary Materials GONG, COSKER: INTERACTIVE SHADOW REMOVAL AND GROUND TRUTH 1 Supplementary Materials Han Gong http://www.cs.bath.ac.uk/~hg299 Darren Cosker http://www.cs.bath.ac.uk/~dpc Department of Computer Science University

More information

Automatic Shadow Removal by Illuminance in HSV Color Space

Automatic Shadow Removal by Illuminance in HSV Color Space Computer Science and Information Technology 3(3): 70-75, 2015 DOI: 10.13189/csit.2015.030303 http://www.hrpub.org Automatic Shadow Removal by Illuminance in HSV Color Space Wenbo Huang 1, KyoungYeon Kim

More information

Data Term. Michael Bleyer LVA Stereo Vision

Data Term. Michael Bleyer LVA Stereo Vision Data Term Michael Bleyer LVA Stereo Vision What happened last time? We have looked at our energy function: E ( D) = m( p, dp) + p I < p, q > N s( p, q) We have learned about an optimization algorithm that

More information

Interactive Differential Segmentation of the Prostate using Graph-Cuts with a Feature Detector-based Boundary Term

Interactive Differential Segmentation of the Prostate using Graph-Cuts with a Feature Detector-based Boundary Term MOSCHIDIS, GRAHAM: GRAPH-CUTS WITH FEATURE DETECTORS 1 Interactive Differential Segmentation of the Prostate using Graph-Cuts with a Feature Detector-based Boundary Term Emmanouil Moschidis emmanouil.moschidis@postgrad.manchester.ac.uk

More information

A robust stereo prior for human segmentation

A robust stereo prior for human segmentation A robust stereo prior for human segmentation Glenn Sheasby, Julien Valentin, Nigel Crook, Philip Torr Oxford Brookes University Abstract. The emergence of affordable depth cameras has enabled significant

More information

A Method of Annotation Extraction from Paper Documents Using Alignment Based on Local Arrangements of Feature Points

A Method of Annotation Extraction from Paper Documents Using Alignment Based on Local Arrangements of Feature Points A Method of Annotation Extraction from Paper Documents Using Alignment Based on Local Arrangements of Feature Points Tomohiro Nakai, Koichi Kise, Masakazu Iwamura Graduate School of Engineering, Osaka

More information

HYBRID CENTER-SYMMETRIC LOCAL PATTERN FOR DYNAMIC BACKGROUND SUBTRACTION. Gengjian Xue, Li Song, Jun Sun, Meng Wu

HYBRID CENTER-SYMMETRIC LOCAL PATTERN FOR DYNAMIC BACKGROUND SUBTRACTION. Gengjian Xue, Li Song, Jun Sun, Meng Wu HYBRID CENTER-SYMMETRIC LOCAL PATTERN FOR DYNAMIC BACKGROUND SUBTRACTION Gengjian Xue, Li Song, Jun Sun, Meng Wu Institute of Image Communication and Information Processing, Shanghai Jiao Tong University,

More information

Stereo Vision. MAN-522 Computer Vision

Stereo Vision. MAN-522 Computer Vision Stereo Vision MAN-522 Computer Vision What is the goal of stereo vision? The recovery of the 3D structure of a scene using two or more images of the 3D scene, each acquired from a different viewpoint in

More information

Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation

Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation ÖGAI Journal 24/1 11 Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation Michael Bleyer, Margrit Gelautz, Christoph Rhemann Vienna University of Technology

More information