Volumetric Shape Grammars for Image Segmentation and Shape Estimation

Size: px
Start display at page:

Download "Volumetric Shape Grammars for Image Segmentation and Shape Estimation"

Transcription

1 Volumetric Shape Grammars for Image Segmentation and Shape Estimation Elias Mahfoud Dept. of Computer Science University of North Carolina at Charlotte Charlotte, NC Andrew Willis Dept. of Elect. and Comp. Engineering University of North Carolina at Charlotte Charlotte, NC Abstract In this paper, we present a technique that uses 3D shape grammars to segment and estimate rectilinear shapes in non-rectified images. While others have proposed the use of shape grammars for segmentation, i.e., assigning labels to rectangular 2D regions within rectified images, the proposed method is innovative for the following reasons: (1) it uses projections of 3D shapes to define non-rectangular segmentation regions and (2) the approach also estimates the unknown shape parameters of the imaged object. As others have done in the past, we requireuser interaction to learn probability distributions for classifying pixels from the input image to each of the unknown classes. We then use a 3D shape grammar to hypothesize 3D models for the imaged object. A search procedure hypothesizes different 3D models by modifying the shape and pose parameters of the 3D shape grammar. The parameters of the hypothesized model that best fits the input image provide a segmentation of the image into semantic parts and shape estimates for each of the segmented parts. The key difference between the proposed approach and previous approaches is the 3D nature of our shape generation and estimation process which represents an advancement over existing approaches that are currently restricted to 2D representations. We describe the method and show segmentation and estimation results using non-rectified images. I. INTRODUCTION Segmentation, the process of assigning semantic labels to objects within images, has been a core computer vision problem for more than 30 years. Shape grammars provide a way to specify dynamic geometric models that result in structured semantic labeling and segmentation with plausible parameters. This has numerous applications in the games and movies industry, mapping applications such as Google maps and Bing maps, and reconstruction of archaeological sites from images. Several attempts have been made to use shape grammars as models for objects to aid in solving segmentation and estimation problems [1], [2]. The majority of these methods [3], [4], [5], [6] use rectified images of building facades as input, and apply two-dimensional grammars to segment, i.e., divide, the input image into semantically meaningful rectangular regions, e.g., windows, walls, doors, etc. Other methods such as that described in [7] use non-shape grammars to segment non-rectified images, yet such grammars have limited representation capabilities applicable to Manhattanstyle buildings (i.e. buildings with faces aligned to the world axes) and used calibrated oblique-angle images (i.e. images which have their pixels mapped to the actual longitude and latitude coordinates). Our proposed method addresses the constrains of having to use rectified images and limiting the space to Manhattan-style buildings, by using a generic volumetric grammar and utilizing a free camera model. We created PSML (Procedural Shape Modeling Language); a volumetric shape grammar language that generates hierarchies of terminal and non-terminal solid shapes with associated semantic labels. The models generated by this language are volumetric and can be viewed from any angle. Some terminals might be partially or fully occluded by other terminals. Our system is capable of rendering image masks of isolated terminals, taking into account occlusions by other terminals with different labels. The image masks can then be used to select regions from pixel-level probabilities and feed them into an optimization algorithm. Any optimization method can be used on the resulting pixel-probabilities in order to find the best fit between the generated grammar and the input image. The contribution of this paper lies mainly in the introduction of such a modular framework, in which different modules can be easily replaced by more advanced and complicated subsystems. This yields a highly customizable optimization system capable of producing plausible image segmentation from non-rectified images. II. RELATED WORK Procedural modeling is an important topic in computer graphics where they are often used to automatically generate very large and detailed models for buildings and cities. Many different types of procedural methods have been devised. Early examples used L-Systems to automatically generate 3D plant models [8] and simulate their interaction with the environment [9]. Subsequent work developed shape grammars for modeling cities [10] and buildings [11]. Other work used tensor fields instead of grammars for modeling street networks [12], where a tensor refers to a symmetric and traceless matrix used to guide the generated streets. One recent and very popular method for generating buildings using shape grammars is CGA (Computer Generated Architecture) [13]. CGA is capable of generating highly detailed building shells which include windows, doors, and walls. (We are not claiming a shape grammar representation or model contribution in this paper).

2 Figure 1: A diagram demonstrating the high-level modules in the optimization engine. Methods which apply shape grammars to help solve computer vision segmentation problems are relatively new and several approaches have been proposed. Teboul et. al. have published a sequence of articles that use two-dimensional grammars based on CGA to represent rectified facades of urban style buildings. Their first approach used randomized forests to learn a likelihood distribution which expresses the likelihood of each grammar object for each image pixel. An optimization procedure based on random walks uses the shape grammar to hypothesize a segmentation of the image into a collection rectangular tiles, each of which is assigned a grammar object. A score for each hypothesized segmentation is taken as the likelihood of the observed image pixels given segmentation hypothesized by the shape grammar [5]. Later work described in [6] improved the performance of the optimization process by using reinforcement learning, which was shown to achieve better results in less computing time on their test data. However, as mentioned previously, these approaches are limited to using 2D grammars and are constrained to only work on rectified images. Our method does not require the use of rectified images and, as such, can be applied to a wider range of problems. Another approach was proposed by Vanegas [7] which uses 3D grammars to estimate Manhattan-style buildings from aerial images. Although this approach is not limited to rectified images and 2D grammars, it limits the problem to using calibrated oblique-angle images. In addition, the grammar used for representing building changes produces very simple building shapes with no details such as windows and doors. Our proposed method utilizes a more generic grammar capable of generating complex shapes with different terminal labels allowing for generating optimal solutions over all labels. III. A PPROACH Our approach formulates the problem as a Bayesian inference problem where we seek to find the segmentation that maximizes the likelihood of the observed data, i.e., as a maximum likelihood estimation (MLE) problem. The input to the system is a 2D image D, a shape grammar (in PSML format; see III-A for details) having parameters θ g, and a set of view parameters θ v. The output is a segmentation of the image, provided as a collection of pixel labels, L, an estimate of the object shape, θ g, and, if appropriate, an estimate of the view from which the image was generated, θ v. While it Figure 2: Some examples of models generated using PSML. is theoretically possible to estimate all of these parameters in the proposed framework, this article assumes that the view parameters, θ v, are a-priori known and are not estimated for any of the presented results in this article. A flow chart of the estimation process is shown in figure 1. The system is divided into several subsystems that collectively perform the estimation: A supervised learning module that generates pixel-level probabilities A shape grammar which takes in grammar variables,θ g, and produces 3D shapes S(θ g ) A rendering engine that synthesizes mask images from the created shape S(θ g ) An optimization procedure that searches for the collection of parameters that maximize the likelihood of the observed data; (θ g,l ) It is worth noting that each subsystem can be independently replaced by more advanced algorithms to improve the performance or accuracy of the proposed method. However, in this initial version of the system, we chose very simple implementations for each of those modules. Specifically, we manually train the pixel classifier and we use a standard gradient-descent algorithm as the optimization procedure for exploring the space of plausible solutions. More advanced algorithms can be added in the future. A. Model Generation using PSML (Procedural Shape Modeling Language) The purpose of the shape grammar is to provide plausible 3D models for objects within the image. Shape grammars use values of the shape parameters to automatically generate shapes which would otherwise require a very large number of variables to accurately represent (see figure 2). This greatly reduces the number of unknown variables associated with possibly complex shapes and provides a computationally tractable way to estimate possibly complex shapes from sparse measurements, e.g., a single image.

3 The shape grammar parser generates an instance of a 3D model using a value for the shape parameters, θ g, and a shape grammar program. Shape grammar programs are specified in PSML, a volumetric 3D modeling language we developed to create 3D models of real-world objects. The shape grammar program contains a collection of rules which describe how to construct the 3D model and labels for each of the elements within constructed 3D models. These labels are used as the unknown segmentation labels which we seek to assign to the image pixels. PSML derives its syntax from Java and adds syntax and functionality needed for specifying shape grammars. In particular rules blocks may be declared within the source code which specify the methods to automatically construct a shape from a collection of primitive shape components, e.g., pieces of spheres, cylinders, 3D rectangles. Each rule within a rule block is provided and input shape, the predecessor, and replaces the shape with one or more possibly different shapes referred to as successors. Rules have a syntax as shown below: predecessor:condition:[modifiers]{successors}; where condition represents an expression that, if true, allows the rule to be executed and [modifiers] denotes an optional list of modifiers which can modify the successor symbols, e.g., change their color. Algorithm 1 is an example of a shape grammar specified in PSML to represent the shape in figure 3a. The program starts with the main() method at line 24, which creates a root object and selects a properties file. The properties file is what defines the grammar parameters which will be optimized by our system. The BuildingMassModel() function generates the shape using these parameters. The process starts by creating a box with the dimensions of the building (lines 4-5) then it splits the box horizontally into two parts having different widths (lines 6 and 8). The resulting parts are then split vertically into the boxes having different heights (lines 11, 14). The shape parameters of this grammar determine the specific locations for the height and width splits (shown as variables with prefix in Algorithm 1). Different values for these parameters allow the shape grammar to represent simple rectilinear objects, i.e., objects whose surface is well approximated as a collection of three different 3D rectangles. B. Rendering Engine The rendering engine takes 3D shapes generated by the shape grammar, S(θ g ), and values for the view, θ v, and renders the synthesized object. The rendering process decomposes the image into a collection of mask images, one for each semantic label within the shape grammar. These mask images denote, for the provided shape and view parameters, where we expect to observe pixels from each semantic class assuming that the values of both the shape and view parameters are correct. The system utilizes jmonkey 3.0 as the rendering engine to generate the mask images. jmonkey is a modern, open-source 3D engine written in Java. The engine facilitates Algorithm 1 The PSML program used to create figure 3a 1 grammar BuildingMassModel { 2 method BuildingMassModel() { 3 rules { 4 parent::i("box", 6 mainmass::split("x", {@westwidth, scope.sx - 7 westwidth}){westmass, east}; 8 westmass::split("z", {@west.northdepth, 9 scope.sz 10 {westnorthmass, westsouthmass}; 11 westsouthmass::split("y", {@west.southheight, 12 scope.sy 13 {south, space}; 14 westnorthmass::split("y", {@west.northheight, 15 scope.sy 16 {north, space}; 17 south::{j3d.terminal}; 18 north::{j3d.terminal}; 19 space::void(){j3d.terminal}; 20 east::{j3d.terminal}; 21 } 22 } method main() { 25 rules { 26 Axiom::I("box", {1, 1, 1}){model}; 27 model::useattributes("building.properties", 28 "building"){buildingmassmodel()}; 29 } 30 } 31 } the creation of custom scene processors that can be applied to each frame before or after it is rendered. We used this functionality to write a mask scene processor which renders a specified list of terminals in place, occluding them by the nonspecified terminals which are invisible. The result is a bitmap mask showing the silhouette of the visible parts of the selected terminals (see figure 3). (a) (b) (c) (d) Figure 3: Example of bitmap mask rendering. (a) shows three adjacent boxes with different sizes and different labels. (bd) show three different mask images rendered for the three different boxes. Notice how the white silhouettes take into account occlusions by other labels. C. Optimization The optimization process searches the grammar parameter space for values of the shape parameter vector, θ g, that best fit the specified input image, i.e. it searches for parameter values that generate shapes that, when projected using the view and overlaid onto the input image, partition the input image data into collections of pixels which agree with the proposed labeling. As we mentioned earlier, our main contribution lies

4 (a) Building (b) Ceiling (c) Walls (d) Windows Figure 4: A manually-labeled non-rectified image of a synthetic building (image courtesy to in the introduction of such a framework, not in the specific algorithms used in its modules. We chose gradient descent as the search engine in this current implementation as it is simple to implement. On the other hand, this algorithm is not very robust and it is prone to converging to local minima. We believe that replacing this algorithm with a more advanced one will greatly improve performance. Our current implementation involves the following steps: 1) Use a label classifier to compute likelihood distributions for each pixel. The distribution is represented as a collection of images which referred to as likelihood images. One image must be computed for every label in the label set L. 2) Provide a PSML grammar representing the class of objects which are expected to be found in the input image data. 3) Apply the optimization engine to search for the best value of the shape grammar parameters. a) Using the current guess for the shape parameters, θ g, compute image masks for each label and apply them to the corresponding likelihood images to identify those pixels lying in regions associated with a given label. b) Compute the joint likelihood of all the pixel data given the labels that they have been assigned from (a). c) Compute the numerical gradient of the likelihood function as a function of the shape parameters and update the shape parameters to increase the joint likelihood. d) Stop when the likelihood cannot be increased, the corresponding shape parameters provide a shape estimate for the objects in the image and the image masks specify a segmentation of the image for each of the labels. Collectively these steps allow us to estimate the best segmentation and shape model for the objects in the image. D. Optimization Process Details 1) Pixel-Level Probabilities: The first step in the optimization process is to use a supervised-learning algorithm to generate initial approximate pixel probabilities. For each label, we input a gray-scale image of the same size as the original image, where the pixel value at position (x, y) represents the likelihood of the pixel belonging to the corresponding label L i, denoted as P (D(x, y) L i ). The gray-scale likelihood images can be generated using any standard image classifier such as Gaussian Mixture Model (GMM) or randomized forests. In our case and for demonstration purposes, we simply used manually assigned probability images as input, where areas that belong to a specific label are painted with white while the remaining areas remained black as shown in figure 4. 2) Input PSML Grammar: Each shape grammar is written to represent a class of objects, e.g., one grammar might represent various realizations of a coffee mug. Hence, a shape grammar program must be selected as is appropriate to the object(s) contained in the input image. The various shapes that can be realized by changing the shape parameters is referred to as the shape space of the grammar or, more simply, they are thought of as shapes that satisfy the grammar. As mentioned earlier, the labels of shapes listed within the shape grammar program provide the possible values for the label set L and the pixel probability images state how likely each pixel is to have a specific label value, e.g., the Figure 4(c) shows the likelihood image for the label Wall. 3) Generate Mask Images: For each label L i, a mask is rendered using the current shape, S(θ g ), and the current view, θ v. Each mask image, I Li (x, y θ g ), indicates the image pixels which are assumed to have label. This is represented mathematically using the indicator function as defined in equation (1). I Li (x, y θ g ) = { 0 if the pixel at (x, y) Li 1 if the pixel at (x, y) = L i (1) The indicator function is used by the search engine to select regions from pixel-probabilities for each label. The use of image masks allows selecting non-rectangular regions of pixels, which is the main difference from previous methods that used flat rectangles to select pixels. For a specific label L i, a region R Li is the set of pixel positions (x, y) such that the indicator function is not zero, as defined in equation (2). R Li = {(x, y) I Li (x, y θ g ) = 1} (2) 4) Generate an Energy Function: The probability of the region associated with the i th label from the label set is given by equation (3). p(r Li θ g ) = p(x, y L i, θ g ) (3) (x,y) R Li Note that in this equation the labels of (x, y) locations are presumed to be independent. Products of a large number of probabilities such as that in equation (3) tend to generate extremely small numbers which can introduce difficulties to optimization algorithms due to numerical instabilities. For this reason we take the negative log of the probabilities which converts equation (3) into an energy function as shown in equation (4).

5 E(R Li θ g ) = (x,y) R Li log(p(x, y L i, θ g )) (4) The energy of the chosen shape parameters is the sum of region energies for each of the label values shown as equation (5). (a) BG (b) East (c) North (d) South (e) Input (f) Diff E(θ g ) = L i L E(R Li θ g ) (5) To facilitate convergence of the gradient based energy minimization process, we blur the likelihood images. This makes the energy function smoother and facilitates convergence of the minimization. 5) Gradient Descent Optimization: A gradient of a function f(θ g ) where θ g = (θ g1, θ g2,, θ n ) is a vector field that points in the direction of maximum increase in the value of that function. The components of the gradient are the partial derivatives of the function E(θ g ) as seen in equation (6). ( E E(θ g ) =, E ) E,..., (6) θ g1 θ g2 θ gn Gradient descent is a simple optimization algorithm that finds the minimum of a function. Starting at an initial value of the input parameters θ 0 g, the algorithm repeatedly computes the value of the gradient vector and changes the parameters vector according to the value of the gradient vector using equation (7). θ n+1 g = θ n g γ E(θ n g ) (7) where γ is a small number. Theoretically, the value of E(θ n+1 g ) is less than or equal to E(θ n g ). As a result, the value of the function will decrease until it reaches a local minimum. The process stops when E(θ n+1 g ) E(θ n g ) < ɛ. The value of θ n g at that point is the optimization result ˆθ g. For our implementation, we numerically compute the gradient of the parameters vector by introducing small perturbations to the vector components one at a time, and computing the difference in the energy function of the generated image. After the differences have been computed for all the parameters, we multiply the resulting gradient vector by γ and subtract the result from the current parameters vector, then repeat the process until convergence. However, since computing the numerical derivative is not accurate, the actual process does not always converge and the value of E(θ n+1 g ) might actually increase in some cases. From a performance point of view, it is worth noting that in order to compute the energy function, the number of render operations needed is equal to the number of available classes. Furthermore, to compute the gradient, the energy function needs to be computed once for the original parameter values, and then once for every perturbation for each parameter. Let the number of available labels be N, and the number of grammar parameters be M, then the total number of render operations needed per gradient-descent iteration is N M. This (g) Result Figure 5: (a,b,c,d) collectively show the likelihood distribution for the image pixels, P (D(x, y) L i ) where for (a) L i = BG, for (b) L i = East, for (c) L i = North, and for (d) L i = South. Sub-figure (e) shows the input image, (f) shows the difference between the input image and the final segmentation result, and (g) shows the energy function values for each iteration of the minimization process and images along this curve depict intermediate results for the estimation. emphasizes the importance of using a fast rendering engine so that the rendering step does not form a bottleneck. IV. RESULTS We ran our test cases on an Intel Core 2 Quad CPU Q9550 running at 2.83GHz, with 4GB of RAM with Windows 7 64bit. We used synthesized renderings as the input images to make sure that the image object can be represented by the specified grammar. For pixel probabilities, we used manual labeling where the user specifies the areas associated with each label by marking them with a higher gray-scale value than other areas. The user labeling does not have to be accurate, as the process is reasonably tolerant to noise. The user labeling is then blurred to make sure the generated energy function is smooth. The first test case was a 3D shape consisting of three blocks that occlude each other. We used initial values for grammar parameters that were different from the original values used to generate the input image, θ g. We recorded the energy differences at various times as gradient descent approached θ g. The result is shown in figure 5. For the second test case we used a non-rectified facade image. The user input for this case is much more noisy and inaccurate than the previous case, yet the process was able to find parameter values that are reasonably close to θ g. The result is shown in figure 6. In the case of the facade, the gradient descent algorithm was less robust than the previous case and took longer to converge. We can see a spike in the energy difference in figure 6f

6 (a) BG (b) Wall (c) Window (d) Input (e) Final (f) Result Figure 6: (a,b,c) collectively show the likelihood distribution for the image pixels, P (D(x, y) L i ) where for (a) L i = BG, for (b) L i = W all and for (c) L i = W indow. (d,e) show a side-by-side comparison of the input image, (d), and the final segmentation result, (e). (f) shows the energy function values for each iteration of the minimization process and images along this curve depict intermediate results for the estimation. where the algorithm diverged for a while before it started converging again. There are two reasons for this behavior: first, the provided pixel-probabilities are very noisy and inaccurate. The second reason is due to the more complicated behavior of the facade grammar in combination with the numerical computation of the gradient. The windows and floors in this grammar are created by a periodic split operation in the x and y directions. As a result, even small changes of some parameters cause a significant change in the generate facade and subsequently the energy function gradient might become inaccurate. Addressing this problem requires the use of more advanced search algorithm than gradient descent. Table I shows the initial, input, and final values after gradient descent has converged for the facade test case. Facade Param. Name θg θ 0 g ˆθ g Error floor.height window.height window.width window.wallpane facade.width facade.depth facade.height Table I: Comparison between initial values, θ 0 g, input values, θ g, and final optimization values, ˆθ g, and percent error. is not limited to using 2D shapes or rectified images. It is also is capable of segmenting higher details from a single image. The main contribution is the introduction of such modular framework that facilitates the application of newer algorithms to improve segmentation results. In the future, we plan to address performance issues to improve optimization speed. Since the current approach uses non-rectangular silhouettes of labels, it is not possible to apply optimization techniques such as integral images [5]. We plan to use Graphics Processing Unit (GPU) to apply label masks in real time and then support multi-threaded summations to generate the energy function. We expect a significant increase in performance after applying those improvements. We also plan to apply similar principles to allow inferring parameters of occluded parts of 3D shapes without the need for multiple views from different directions. REFERENCES [1] J. Schlecht, K. Barnard, E. Spriggs, and B. Pryor, Inferring grammarbased structure models from 3d microscopy data, in Computer Vision and Pattern Recognition, CVPR 07. IEEE Conference on. IEEE, 2007, pp [2] A. Toshev, P. Mordohai, and B. Taskar, Detecting and parsing architecture at city scale from range data, in Computer Vision and Pattern Recognition (CVPR), 2010 IEEE Conference on. IEEE, 2010, pp [3] S. Lee and R. Nevatia, Extraction and integration of window in a 3d building model from ground view images, in Computer Vision and Pattern Recognition, CVPR Proceedings of the 2004 IEEE Computer Society Conference on, vol. 2. IEEE, 2004, pp. II 113. [4] P. Müller, G. Zeng, P. Wonka, and L. Van Gool, Image-based procedural modeling of facades, in ACM SIGGRAPH 2007 papers, ser. SIGGRAPH 07. New York, NY, USA: ACM, [5] O. Teboul, L. Simon, P. Koutsourakis, and N. Paragios, Segmentation of building facades using procedural shape priors, in Computer Vision and Pattern Recognition (CVPR), 2010 IEEE Conference on, 2010, pp [6] O. Teboul, I. Kokkinos, L. Simon, P. Koutsourakis, and N. Paragios, Shape grammar parsing via reinforcement learning, in Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on, 2011, pp [7] C. A. Vanegas, D. G. Aliaga, and B. Benes, Building reconstruction using manhattan-world grammars, in Computer Vision and Pattern Recognition (CVPR), 2010 IEEE Conference on, 2010, pp [8] P. Prusinkiewicz, A. Lindenmayer, and J. Hanan, The algorithmic beauty of plants, ser. Virtual laboratory. Springer-Verlag, [9] R. Měch and P. Prusinkiewicz, Visual models of plants interacting with their environment, in Proceedings of the 23rd annual conference on Computer graphics and interactive techniques. ACM, 1996, pp [10] Y. Parish and P. Müller, Procedural modeling of cities, in Proceedings of the 28th annual conference on Computer graphics and interactive techniques. ACM, 2001, pp [11] J. Duarte, Customizing mass housing: a discursive grammar for siza s malagueira houses, Ph.D. dissertation, Massachusetts Institute of Technology, [12] G. Chen, G. Esch, P. Wonka, P. Müller, and E. Zhang, Interactive procedural street modeling, in ACM SIGGRAPH 2008 papers, ser. SIGGRAPH 08. New York, NY, USA: ACM, 2008, pp. 103:1 103:10. [13] P. Müller, P. Wonka, S. Haegler, A. Ulmer, and L. Van Gool, Procedural modeling of buildings, in ACM SIGGRAPH 2006 Papers, ser. SIGGRAPH 06. New York, NY, USA: ACM, 2006, pp V. CONCLUSIONS AND FUTURE WORK We presented a system for segmenting non-rectified images using a 3D volumetric shape grammar. The proposed method

Image-Based Buildings and Facades

Image-Based Buildings and Facades Image-Based Buildings and Facades Peter Wonka Associate Professor of Computer Science Arizona State University Daniel G. Aliaga Associate Professor of Computer Science Purdue University Challenge Generate

More information

GRAMMAR SUPPORTED FACADE RECONSTRUCTION FROM MOBILE LIDAR MAPPING

GRAMMAR SUPPORTED FACADE RECONSTRUCTION FROM MOBILE LIDAR MAPPING GRAMMAR SUPPORTED FACADE RECONSTRUCTION FROM MOBILE LIDAR MAPPING Susanne Becker, Norbert Haala Institute for Photogrammetry, University of Stuttgart Geschwister-Scholl-Straße 24D, D-70174 Stuttgart forename.lastname@ifp.uni-stuttgart.de

More information

Automatic generation of 3-d building models from multiple bounded polygons

Automatic generation of 3-d building models from multiple bounded polygons icccbe 2010 Nottingham University Press Proceedings of the International Conference on Computing in Civil and Building Engineering W Tizani (Editor) Automatic generation of 3-d building models from multiple

More information

3D Automatic Building Footprints Generation

3D Automatic Building Footprints Generation 3D Automatic Building Footprints Generation Mohamed Shaarawy Ahmed Kaboudan Shaimaa Toriah Abstract Virtual building modeling became an increasingly urgent need for computer graphics, simulation, games,

More information

Urban Layouts and Road Networks

Urban Layouts and Road Networks Urban Layouts and Road Networks Daniel G. Aliaga Associate Professor of Computer Science Purdue University Challenge Design and model realistic/plausible road networks and urban layouts Road network: a

More information

Grammar Rule Extraction and Transfer in Buildings

Grammar Rule Extraction and Transfer in Buildings Grammar Rule Extraction and Transfer in Buildings Asad Khalid Ismail Lahore University of Management Sciences Sector U, DHA, Lahore 13100004@lums.edu.pk Zuha Agha Lahore University of Management Sciences

More information

Automatic Generation of 3D Building Models with Efficient Solar Photovoltaic Generation

Automatic Generation of 3D Building Models with Efficient Solar Photovoltaic Generation International review for spatial planning and sustainable development, Vol.5 No.1 (2017), 4-14 ISSN: 2187-3666 (online) DOI: http://dx.doi.org/10.14246/irspsd.5.1_4 Copyright@SPSD Press from 2010, SPSD

More information

A Survey of Light Source Detection Methods

A Survey of Light Source Detection Methods A Survey of Light Source Detection Methods Nathan Funk University of Alberta Mini-Project for CMPUT 603 November 30, 2003 Abstract This paper provides an overview of the most prominent techniques for light

More information

GRAMMAR-BASED AUTOMATIC 3D MODEL RECONSTRUCTION FROM TERRESTRIAL LASER SCANNING DATA

GRAMMAR-BASED AUTOMATIC 3D MODEL RECONSTRUCTION FROM TERRESTRIAL LASER SCANNING DATA GRAMMAR-BASED AUTOMATIC 3D MODEL RECONSTRUCTION FROM TERRESTRIAL LASER SCANNING DATA Qian Yu, Petra Helmholz, David Belton and Geoff West Cooperated Research Centre for Spatial Sciences (CRCSI) Department

More information

A Constraint Based System to Populate Procedurally Modeled Cities with Buildings

A Constraint Based System to Populate Procedurally Modeled Cities with Buildings A Constraint Based System to Populate Procedurally Modeled Cities with Buildings Johannes Scharl Supervised by: Daniel Scherzer Institute of Computer Graphics and Algorithms Vienna University of Technology

More information

Procedural Articulation of Form

Procedural Articulation of Form R. Stouffs, P. Janssen, S. Roudavski, B. Tunçer (eds.), Open Systems: Proceedings of the 18th International Conference on Computer-Aided Architectural Design Research in Asia (CAADRIA 2013), 821 829. 2013,

More information

Articulated Pose Estimation with Flexible Mixtures-of-Parts

Articulated Pose Estimation with Flexible Mixtures-of-Parts Articulated Pose Estimation with Flexible Mixtures-of-Parts PRESENTATION: JESSE DAVIS CS 3710 VISUAL RECOGNITION Outline Modeling Special Cases Inferences Learning Experiments Problem and Relevance Problem:

More information

Panoramic Image Stitching

Panoramic Image Stitching Mcgill University Panoramic Image Stitching by Kai Wang Pengbo Li A report submitted in fulfillment for the COMP 558 Final project in the Faculty of Computer Science April 2013 Mcgill University Abstract

More information

Towards Knowledge-Based Extraction of Roads from 1m-resolution Satellite Images

Towards Knowledge-Based Extraction of Roads from 1m-resolution Satellite Images Towards Knowledge-Based Extraction of Roads from 1m-resolution Satellite Images Hae Yeoun Lee* Wonkyu Park** Heung-Kyu Lee* Tak-gon Kim*** * Dept. of Computer Science, Korea Advanced Institute of Science

More information

Silhouette-based Multiple-View Camera Calibration

Silhouette-based Multiple-View Camera Calibration Silhouette-based Multiple-View Camera Calibration Prashant Ramanathan, Eckehard Steinbach, and Bernd Girod Information Systems Laboratory, Electrical Engineering Department, Stanford University Stanford,

More information

Graph-based Modeling of Building Roofs Judith Milde, Claus Brenner Institute of Cartography and Geoinformatics, Leibniz Universität Hannover

Graph-based Modeling of Building Roofs Judith Milde, Claus Brenner Institute of Cartography and Geoinformatics, Leibniz Universität Hannover 12th AGILE International Conference on Geographic Information Science 2009 page 1 of 5 Graph-based Modeling of Building Roofs Judith Milde, Claus Brenner Institute of Cartography and Geoinformatics, Leibniz

More information

DETERMINATION OF FACADE ATTRIBUTES FOR FACADE RECONSTRUCTION

DETERMINATION OF FACADE ATTRIBUTES FOR FACADE RECONSTRUCTION DETERMINATION OF FACADE ATTRIBUTES FOR FACADE RECONSTRUCTION Nora Ripperda Institute of Cartography and Geoinformatics Leibniz University of Hannover nora.ripperda@ikg.uni-hannover.de KEY WORDS: data analysis,

More information

Urban Scene Segmentation, Recognition and Remodeling. Part III. Jinglu Wang 11/24/2016 ACCV 2016 TUTORIAL

Urban Scene Segmentation, Recognition and Remodeling. Part III. Jinglu Wang 11/24/2016 ACCV 2016 TUTORIAL Part III Jinglu Wang Urban Scene Segmentation, Recognition and Remodeling 102 Outline Introduction Related work Approaches Conclusion and future work o o - - ) 11/7/16 103 Introduction Motivation Motivation

More information

Extracting Axially Symmetric Geometry From Limited 3D Range Data

Extracting Axially Symmetric Geometry From Limited 3D Range Data Extracting Axially Symmetric Geometry From Limited 3D Range Data Andrew Willis, Xavier Orriols, Senem Velipasalar, Xavier Binefa, David B. Cooper Brown University, Providence, RI 02912 Computer Vision

More information

Judging Whether Multiple Silhouettes Can Come from the Same Object

Judging Whether Multiple Silhouettes Can Come from the Same Object Judging Whether Multiple Silhouettes Can Come from the Same Object David Jacobs 1, eter Belhumeur 2, and Ian Jermyn 3 1 NEC Research Institute 2 Yale University 3 New York University Abstract. We consider

More information

Semi-Automatic Techniques for Generating BIM Façade Models of Historic Buildings

Semi-Automatic Techniques for Generating BIM Façade Models of Historic Buildings Semi-Automatic Techniques for Generating BIM Façade Models of Historic Buildings C. Dore, M. Murphy School of Surveying & Construction Management Dublin Institute of Technology Bolton Street Campus, Dublin

More information

Learning to Segment Document Images

Learning to Segment Document Images Learning to Segment Document Images K.S. Sesh Kumar, Anoop Namboodiri, and C.V. Jawahar Centre for Visual Information Technology, International Institute of Information Technology, Hyderabad, India Abstract.

More information

Semiautomatic Rule Assist Architecture Modeling

Semiautomatic Rule Assist Architecture Modeling Semiautomatic Rule Assist Architecture Modeling Hua Liu, Hongxin Zhang, and Hujun Bao State Key lab of CAD&CG, Zhejiang University, Hangzhou, China {sun day,zhx,bao}@cad.zju.edu.cn Abstract. This paper

More information

Edge and local feature detection - 2. Importance of edge detection in computer vision

Edge and local feature detection - 2. Importance of edge detection in computer vision Edge and local feature detection Gradient based edge detection Edge detection by function fitting Second derivative edge detectors Edge linking and the construction of the chain graph Edge and local feature

More information

IRIS SEGMENTATION OF NON-IDEAL IMAGES

IRIS SEGMENTATION OF NON-IDEAL IMAGES IRIS SEGMENTATION OF NON-IDEAL IMAGES William S. Weld St. Lawrence University Computer Science Department Canton, NY 13617 Xiaojun Qi, Ph.D Utah State University Computer Science Department Logan, UT 84322

More information

CREATION OF THE TREE MODEL ADAPTING TO ENVIRONMENT

CREATION OF THE TREE MODEL ADAPTING TO ENVIRONMENT CREATION OF THE TREE MODEL ADAPTING TO ENVIRONMENT Ryota Ueno Yoshio Ohno {ryota ohno}@on.ics.keio.ac.jp Graduate School of Science and Technology, Keio University 3-14-1 Hiyoshi, Kohoku-ku Yokohama 223-8522

More information

Interactive Design and Visualization of Urban Spaces using Geometrical and Behavioral Modeling

Interactive Design and Visualization of Urban Spaces using Geometrical and Behavioral Modeling Interactive Design and Visualization of Urban Spaces using Geometrical and Behavioral Modeling Carlos Vanegas 1,4,5 Daniel Aliaga 1 Bedřich Beneš 2 Paul Waddell 3 1 Computer Science, Purdue University,

More information

Real-Time Human Detection using Relational Depth Similarity Features

Real-Time Human Detection using Relational Depth Similarity Features Real-Time Human Detection using Relational Depth Similarity Features Sho Ikemura, Hironobu Fujiyoshi Dept. of Computer Science, Chubu University. Matsumoto 1200, Kasugai, Aichi, 487-8501 Japan. si@vision.cs.chubu.ac.jp,

More information

METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS

METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS M. Lefler, H. Hel-Or Dept. of CS, University of Haifa, Israel Y. Hel-Or School of CS, IDC, Herzliya, Israel ABSTRACT Video analysis often requires

More information

Procedural Modeling. Last Time? Reading for Today. Reading for Today

Procedural Modeling. Last Time? Reading for Today. Reading for Today Last Time? Procedural Modeling Modern Graphics Hardware Cg Programming Language Gouraud Shading vs. Phong Normal Interpolation Bump, Displacement, & Environment Mapping G P R T F P D Reading for Today

More information

L1 - Introduction. Contents. Introduction of CAD/CAM system Components of CAD/CAM systems Basic concepts of graphics programming

L1 - Introduction. Contents. Introduction of CAD/CAM system Components of CAD/CAM systems Basic concepts of graphics programming L1 - Introduction Contents Introduction of CAD/CAM system Components of CAD/CAM systems Basic concepts of graphics programming 1 Definitions Computer-Aided Design (CAD) The technology concerned with the

More information

CS 534: Computer Vision Segmentation and Perceptual Grouping

CS 534: Computer Vision Segmentation and Perceptual Grouping CS 534: Computer Vision Segmentation and Perceptual Grouping Spring 2005 Ahmed Elgammal Dept of Computer Science CS 534 Segmentation - 1 Where are we? Image Formation Human vision Cameras Geometric Camera

More information

Structural and Syntactic Pattern Recognition

Structural and Syntactic Pattern Recognition Structural and Syntactic Pattern Recognition Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Fall 2017 CS 551, Fall 2017 c 2017, Selim Aksoy (Bilkent

More information

Human Upper Body Pose Estimation in Static Images

Human Upper Body Pose Estimation in Static Images 1. Research Team Human Upper Body Pose Estimation in Static Images Project Leader: Graduate Students: Prof. Isaac Cohen, Computer Science Mun Wai Lee 2. Statement of Project Goals This goal of this project

More information

HOG-Based Person Following and Autonomous Returning Using Generated Map by Mobile Robot Equipped with Camera and Laser Range Finder

HOG-Based Person Following and Autonomous Returning Using Generated Map by Mobile Robot Equipped with Camera and Laser Range Finder HOG-Based Person Following and Autonomous Returning Using Generated Map by Mobile Robot Equipped with Camera and Laser Range Finder Masashi Awai, Takahito Shimizu and Toru Kaneko Department of Mechanical

More information

cse 252c Fall 2004 Project Report: A Model of Perpendicular Texture for Determining Surface Geometry

cse 252c Fall 2004 Project Report: A Model of Perpendicular Texture for Determining Surface Geometry cse 252c Fall 2004 Project Report: A Model of Perpendicular Texture for Determining Surface Geometry Steven Scher December 2, 2004 Steven Scher SteveScher@alumni.princeton.edu Abstract Three-dimensional

More information

A Statistical Consistency Check for the Space Carving Algorithm.

A Statistical Consistency Check for the Space Carving Algorithm. A Statistical Consistency Check for the Space Carving Algorithm. A. Broadhurst and R. Cipolla Dept. of Engineering, Univ. of Cambridge, Cambridge, CB2 1PZ aeb29 cipolla @eng.cam.ac.uk Abstract This paper

More information

A New Image Based Ligthing Method: Practical Shadow-Based Light Reconstruction

A New Image Based Ligthing Method: Practical Shadow-Based Light Reconstruction A New Image Based Ligthing Method: Practical Shadow-Based Light Reconstruction Jaemin Lee and Ergun Akleman Visualization Sciences Program Texas A&M University Abstract In this paper we present a practical

More information

Mixture Models and EM

Mixture Models and EM Mixture Models and EM Goal: Introduction to probabilistic mixture models and the expectationmaximization (EM) algorithm. Motivation: simultaneous fitting of multiple model instances unsupervised clustering

More information

HOUGH TRANSFORM CS 6350 C V

HOUGH TRANSFORM CS 6350 C V HOUGH TRANSFORM CS 6350 C V HOUGH TRANSFORM The problem: Given a set of points in 2-D, find if a sub-set of these points, fall on a LINE. Hough Transform One powerful global method for detecting edges

More information

Accurate 3D Face and Body Modeling from a Single Fixed Kinect

Accurate 3D Face and Body Modeling from a Single Fixed Kinect Accurate 3D Face and Body Modeling from a Single Fixed Kinect Ruizhe Wang*, Matthias Hernandez*, Jongmoo Choi, Gérard Medioni Computer Vision Lab, IRIS University of Southern California Abstract In this

More information

Module 1 Lecture Notes 2. Optimization Problem and Model Formulation

Module 1 Lecture Notes 2. Optimization Problem and Model Formulation Optimization Methods: Introduction and Basic concepts 1 Module 1 Lecture Notes 2 Optimization Problem and Model Formulation Introduction In the previous lecture we studied the evolution of optimization

More information

Computer Vision 6 Segmentation by Fitting

Computer Vision 6 Segmentation by Fitting Computer Vision 6 Segmentation by Fitting MAP-I Doctoral Programme Miguel Tavares Coimbra Outline The Hough Transform Fitting Lines Fitting Curves Fitting as a Probabilistic Inference Problem Acknowledgements:

More information

What have we leaned so far?

What have we leaned so far? What have we leaned so far? Camera structure Eye structure Project 1: High Dynamic Range Imaging What have we learned so far? Image Filtering Image Warping Camera Projection Model Project 2: Panoramic

More information

Computational Foundations of Cognitive Science

Computational Foundations of Cognitive Science Computational Foundations of Cognitive Science Lecture 16: Models of Object Recognition Frank Keller School of Informatics University of Edinburgh keller@inf.ed.ac.uk February 23, 2010 Frank Keller Computational

More information

EE795: Computer Vision and Intelligent Systems

EE795: Computer Vision and Intelligent Systems EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 WRI C225 Lecture 02 130124 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Basics Image Formation Image Processing 3 Intelligent

More information

Overview. Related Work Tensor Voting in 2-D Tensor Voting in 3-D Tensor Voting in N-D Application to Vision Problems Stereo Visual Motion

Overview. Related Work Tensor Voting in 2-D Tensor Voting in 3-D Tensor Voting in N-D Application to Vision Problems Stereo Visual Motion Overview Related Work Tensor Voting in 2-D Tensor Voting in 3-D Tensor Voting in N-D Application to Vision Problems Stereo Visual Motion Binary-Space-Partitioned Images 3-D Surface Extraction from Medical

More information

Room Reconstruction from a Single Spherical Image by Higher-order Energy Minimization

Room Reconstruction from a Single Spherical Image by Higher-order Energy Minimization Room Reconstruction from a Single Spherical Image by Higher-order Energy Minimization Kosuke Fukano, Yoshihiko Mochizuki, Satoshi Iizuka, Edgar Simo-Serra, Akihiro Sugimoto, and Hiroshi Ishikawa Waseda

More information

ONE of the most fundamental tasks in computer graphics

ONE of the most fundamental tasks in computer graphics EDIC RESEARCH PROPOSAL 1 Structure-aware Procedural Editing Stefan Lienhard LGG, I&C, EPFL Abstract This research proposal summarizes and reviews three important papers on procedural modelling of buildings,

More information

AUTOMATIC EXTRACTION OF BUILDING ROOFS FROM PICTOMETRY S ORTHOGONAL AND OBLIQUE IMAGES

AUTOMATIC EXTRACTION OF BUILDING ROOFS FROM PICTOMETRY S ORTHOGONAL AND OBLIQUE IMAGES AUTOMATIC EXTRACTION OF BUILDING ROOFS FROM PICTOMETRY S ORTHOGONAL AND OBLIQUE IMAGES Yandong Wang Pictometry International Corp. Suite A, 100 Town Centre Dr., Rochester, NY14623, the United States yandong.wang@pictometry.com

More information

Object Recognition Using Pictorial Structures. Daniel Huttenlocher Computer Science Department. In This Talk. Object recognition in computer vision

Object Recognition Using Pictorial Structures. Daniel Huttenlocher Computer Science Department. In This Talk. Object recognition in computer vision Object Recognition Using Pictorial Structures Daniel Huttenlocher Computer Science Department Joint work with Pedro Felzenszwalb, MIT AI Lab In This Talk Object recognition in computer vision Brief definition

More information

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing Visual servoing vision allows a robotic system to obtain geometrical and qualitative information on the surrounding environment high level control motion planning (look-and-move visual grasping) low level

More information

ROTATION INVARIANT SPARSE CODING AND PCA

ROTATION INVARIANT SPARSE CODING AND PCA ROTATION INVARIANT SPARSE CODING AND PCA NATHAN PFLUEGER, RYAN TIMMONS Abstract. We attempt to encode an image in a fashion that is only weakly dependent on rotation of objects within the image, as an

More information

7 Fractions. Number Sense and Numeration Measurement Geometry and Spatial Sense Patterning and Algebra Data Management and Probability

7 Fractions. Number Sense and Numeration Measurement Geometry and Spatial Sense Patterning and Algebra Data Management and Probability 7 Fractions GRADE 7 FRACTIONS continue to develop proficiency by using fractions in mental strategies and in selecting and justifying use; develop proficiency in adding and subtracting simple fractions;

More information

Simultaneous Appearance Modeling and Segmentation for Matching People under Occlusion

Simultaneous Appearance Modeling and Segmentation for Matching People under Occlusion Simultaneous Appearance Modeling and Segmentation for Matching People under Occlusion Zhe Lin, Larry S. Davis, David Doermann, and Daniel DeMenthon Institute for Advanced Computer Studies University of

More information

A 3D Pattern for Post Estimation for Object Capture

A 3D Pattern for Post Estimation for Object Capture A 3D Pattern for Post Estimation for Object Capture Lei Wang, Cindy Grimm, and Robert Pless Department of Computer Science and Engineering Washington University One Brookings Drive, St. Louis, MO, 63130

More information

CS443: Digital Imaging and Multimedia Perceptual Grouping Detecting Lines and Simple Curves

CS443: Digital Imaging and Multimedia Perceptual Grouping Detecting Lines and Simple Curves CS443: Digital Imaging and Multimedia Perceptual Grouping Detecting Lines and Simple Curves Spring 2008 Ahmed Elgammal Dept. of Computer Science Rutgers University Outlines Perceptual Grouping and Segmentation

More information

Supplemental Material: Detailed, accurate, human shape estimation from clothed 3D scan sequences

Supplemental Material: Detailed, accurate, human shape estimation from clothed 3D scan sequences Supplemental Material: Detailed, accurate, human shape estimation from clothed 3D scan sequences Chao Zhang 1,2, Sergi Pujades 1, Michael Black 1, and Gerard Pons-Moll 1 1 MPI for Intelligent Systems,

More information

Correcting User Guided Image Segmentation

Correcting User Guided Image Segmentation Correcting User Guided Image Segmentation Garrett Bernstein (gsb29) Karen Ho (ksh33) Advanced Machine Learning: CS 6780 Abstract We tackle the problem of segmenting an image into planes given user input.

More information

Chapter 3 Image Registration. Chapter 3 Image Registration

Chapter 3 Image Registration. Chapter 3 Image Registration Chapter 3 Image Registration Distributed Algorithms for Introduction (1) Definition: Image Registration Input: 2 images of the same scene but taken from different perspectives Goal: Identify transformation

More information

Extraction and Integration of Window in a 3D Building Model from Ground View images

Extraction and Integration of Window in a 3D Building Model from Ground View images Extraction and Integration of Window in a 3D Building Model from Ground View images Sung Chun Lee and Ram Nevatia Institute for Robotics and Intelligent Systems, University of Southern California Los Angeles,

More information

Supplemental Document for Deep Photo Style Transfer

Supplemental Document for Deep Photo Style Transfer Supplemental Document for Deep Photo Style Transfer Fujun Luan Cornell University Sylvain Paris Adobe Eli Shechtman Adobe Kavita Bala Cornell University fujun@cs.cornell.edu sparis@adobe.com elishe@adobe.com

More information

Analysis of ARES Data using ML-EM

Analysis of ARES Data using ML-EM Analysis of ARES Data using ML-EM Nicole Eikmeier Hosting Site: Lawrence Berkeley National Laboratory Mentor(s): Brian Quiter, Mark Bandstra Abstract. Imaging analysis of background data collected from

More information

Space Filling: A new algorithm for procedural creation of game assets

Space Filling: A new algorithm for procedural creation of game assets Space Filling: A new algorithm for procedural creation of game assets Paul Bourke ivec@uwa, The University of Western Australia, 35 Stirling Hwy, Crawley, Perth, West Australia 6009. Email: paul.bourke@uwa.edu.au

More information

IMA Preprint Series # 2016

IMA Preprint Series # 2016 VIDEO INPAINTING OF OCCLUDING AND OCCLUDED OBJECTS By Kedar A. Patwardhan Guillermo Sapiro and Marcelo Bertalmio IMA Preprint Series # 2016 ( January 2005 ) INSTITUTE FOR MATHEMATICS AND ITS APPLICATIONS

More information

CREATING 3D WRL OBJECT BY USING 2D DATA

CREATING 3D WRL OBJECT BY USING 2D DATA ISSN : 0973-7391 Vol. 3, No. 1, January-June 2012, pp. 139-142 CREATING 3D WRL OBJECT BY USING 2D DATA Isha 1 and Gianetan Singh Sekhon 2 1 Department of Computer Engineering Yadavindra College of Engineering,

More information

Automatic Generation of 3D Building Models for Sustainable Development

Automatic Generation of 3D Building Models for Sustainable Development International review for spatial planning and sustainable development, Vol.3 No.2 (2015), 68-78 ISSN: 2187-3666 (online) DOI: http://dx.doi.org/10.14246/irspsd.3.2_68 Copyright@SPSD Press from 2010, SPSD

More information

Robotics Programming Laboratory

Robotics Programming Laboratory Chair of Software Engineering Robotics Programming Laboratory Bertrand Meyer Jiwon Shin Lecture 8: Robot Perception Perception http://pascallin.ecs.soton.ac.uk/challenges/voc/databases.html#caltech car

More information

Specular 3D Object Tracking by View Generative Learning

Specular 3D Object Tracking by View Generative Learning Specular 3D Object Tracking by View Generative Learning Yukiko Shinozuka, Francois de Sorbier and Hideo Saito Keio University 3-14-1 Hiyoshi, Kohoku-ku 223-8522 Yokohama, Japan shinozuka@hvrl.ics.keio.ac.jp

More information

Unwrapping of Urban Surface Models

Unwrapping of Urban Surface Models Unwrapping of Urban Surface Models Generation of virtual city models using laser altimetry and 2D GIS Abstract In this paper we present an approach for the geometric reconstruction of urban areas. It is

More information

Project Report for EE7700

Project Report for EE7700 Project Report for EE7700 Name: Jing Chen, Shaoming Chen Student ID: 89-507-3494, 89-295-9668 Face Tracking 1. Objective of the study Given a video, this semester project aims at implementing algorithms

More information

Segmentation and Tracking of Partial Planar Templates

Segmentation and Tracking of Partial Planar Templates Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract

More information

Query-Sensitive Similarity Measure for Content-Based Image Retrieval

Query-Sensitive Similarity Measure for Content-Based Image Retrieval Query-Sensitive Similarity Measure for Content-Based Image Retrieval Zhi-Hua Zhou Hong-Bin Dai National Laboratory for Novel Software Technology Nanjing University, Nanjing 2193, China {zhouzh, daihb}@lamda.nju.edu.cn

More information

Department of Electrical Engineering, Keio University Hiyoshi Kouhoku-ku Yokohama 223, Japan

Department of Electrical Engineering, Keio University Hiyoshi Kouhoku-ku Yokohama 223, Japan Shape Modeling from Multiple View Images Using GAs Satoshi KIRIHARA and Hideo SAITO Department of Electrical Engineering, Keio University 3-14-1 Hiyoshi Kouhoku-ku Yokohama 223, Japan TEL +81-45-563-1141

More information

Shape from Silhouettes I CV book Szelisky

Shape from Silhouettes I CV book Szelisky Shape from Silhouettes I CV book Szelisky 11.6.2 Guido Gerig CS 6320, Spring 2012 (slides modified from Marc Pollefeys UNC Chapel Hill, some of the figures and slides are adapted from M. Pollefeys, J.S.

More information

CS 534: Computer Vision Segmentation and Perceptual Grouping

CS 534: Computer Vision Segmentation and Perceptual Grouping CS 534: Computer Vision Segmentation and Perceptual Grouping Ahmed Elgammal Dept of Computer Science CS 534 Segmentation - 1 Outlines Mid-level vision What is segmentation Perceptual Grouping Segmentation

More information

Façade Reconstruction An Interactive Image-Based Approach

Façade Reconstruction An Interactive Image-Based Approach Façade Reconstruction An Interactive Image-Based Approach Przemyslaw Musialski Institute of Computer Graphics and Algorithms Vienna University of Technology What is Façade Reconstruction? Part of Urban

More information

B. Tech. Project Second Stage Report on

B. Tech. Project Second Stage Report on B. Tech. Project Second Stage Report on GPU Based Active Contours Submitted by Sumit Shekhar (05007028) Under the guidance of Prof Subhasis Chaudhuri Table of Contents 1. Introduction... 1 1.1 Graphic

More information

S U N G - E U I YO O N, K A I S T R E N D E R I N G F R E E LY A VA I L A B L E O N T H E I N T E R N E T

S U N G - E U I YO O N, K A I S T R E N D E R I N G F R E E LY A VA I L A B L E O N T H E I N T E R N E T S U N G - E U I YO O N, K A I S T R E N D E R I N G F R E E LY A VA I L A B L E O N T H E I N T E R N E T Copyright 2018 Sung-eui Yoon, KAIST freely available on the internet http://sglab.kaist.ac.kr/~sungeui/render

More information

Edge Equalized Treemaps

Edge Equalized Treemaps Edge Equalized Treemaps Aimi Kobayashi Department of Computer Science University of Tsukuba Ibaraki, Japan kobayashi@iplab.cs.tsukuba.ac.jp Kazuo Misue Faculty of Engineering, Information and Systems University

More information

Nonrigid Surface Modelling. and Fast Recovery. Department of Computer Science and Engineering. Committee: Prof. Leo J. Jia and Prof. K. H.

Nonrigid Surface Modelling. and Fast Recovery. Department of Computer Science and Engineering. Committee: Prof. Leo J. Jia and Prof. K. H. Nonrigid Surface Modelling and Fast Recovery Zhu Jianke Supervisor: Prof. Michael R. Lyu Committee: Prof. Leo J. Jia and Prof. K. H. Wong Department of Computer Science and Engineering May 11, 2007 1 2

More information

Subpixel Corner Detection Using Spatial Moment 1)

Subpixel Corner Detection Using Spatial Moment 1) Vol.31, No.5 ACTA AUTOMATICA SINICA September, 25 Subpixel Corner Detection Using Spatial Moment 1) WANG She-Yang SONG Shen-Min QIANG Wen-Yi CHEN Xing-Lin (Department of Control Engineering, Harbin Institute

More information

Final Exam Assigned: 11/21/02 Due: 12/05/02 at 2:30pm

Final Exam Assigned: 11/21/02 Due: 12/05/02 at 2:30pm 6.801/6.866 Machine Vision Final Exam Assigned: 11/21/02 Due: 12/05/02 at 2:30pm Problem 1 Line Fitting through Segmentation (Matlab) a) Write a Matlab function to generate noisy line segment data with

More information

Optimal Segmentation and Understanding of Motion Capture Data

Optimal Segmentation and Understanding of Motion Capture Data Optimal Segmentation and Understanding of Motion Capture Data Xiang Huang, M.A.Sc Candidate Department of Electrical and Computer Engineering McMaster University Supervisor: Dr. Xiaolin Wu 7 Apr, 2005

More information

Estimating Gothic Facade Architecture from Imagery

Estimating Gothic Facade Architecture from Imagery Estimating Gothic Facade Architecture from Imagery Andrew Willis University of North Carlolina at Charlotte arwillis@uncc.edu Katharina Galor Brown University kgalor@brown.edu Yunfeng Sui University of

More information

Chaplin, Modern Times, 1936

Chaplin, Modern Times, 1936 Chaplin, Modern Times, 1936 [A Bucket of Water and a Glass Matte: Special Effects in Modern Times; bonus feature on The Criterion Collection set] Multi-view geometry problems Structure: Given projections

More information

CSE/EE-576, Final Project

CSE/EE-576, Final Project 1 CSE/EE-576, Final Project Torso tracking Ke-Yu Chen Introduction Human 3D modeling and reconstruction from 2D sequences has been researcher s interests for years. Torso is the main part of the human

More information

Multi-view stereo. Many slides adapted from S. Seitz

Multi-view stereo. Many slides adapted from S. Seitz Multi-view stereo Many slides adapted from S. Seitz Beyond two-view stereo The third eye can be used for verification Multiple-baseline stereo Pick a reference image, and slide the corresponding window

More information

On Skeletons Attached to Grey Scale Images. Institute for Studies in Theoretical Physics and Mathematics Tehran, Iran ABSTRACT

On Skeletons Attached to Grey Scale Images. Institute for Studies in Theoretical Physics and Mathematics Tehran, Iran ABSTRACT On Skeletons Attached to Grey Scale Images M. Karimi Behbahani, Arash Rafiey, 2 Mehrdad Shahshahani 3 Institute for Studies in Theoretical Physics and Mathematics Tehran, Iran ABSTRACT In [2], [3] and

More information

(Urban) Forward Procedural and Inverse Procedural Modeling

(Urban) Forward Procedural and Inverse Procedural Modeling (Urban) Forward Procedural and Inverse Procedural Modeling www.cs.purdue.edu/cgvlab/urban Daniel Aliaga, CGVLAB Department of Computer Science Purdue University What is the main challenge? Solving the

More information

Biometrics Technology: Image Processing & Pattern Recognition (by Dr. Dickson Tong)

Biometrics Technology: Image Processing & Pattern Recognition (by Dr. Dickson Tong) Biometrics Technology: Image Processing & Pattern Recognition (by Dr. Dickson Tong) References: [1] http://homepages.inf.ed.ac.uk/rbf/hipr2/index.htm [2] http://www.cs.wisc.edu/~dyer/cs540/notes/vision.html

More information

Bilevel Sparse Coding

Bilevel Sparse Coding Adobe Research 345 Park Ave, San Jose, CA Mar 15, 2013 Outline 1 2 The learning model The learning algorithm 3 4 Sparse Modeling Many types of sensory data, e.g., images and audio, are in high-dimensional

More information

Evaluation of Moving Object Tracking Techniques for Video Surveillance Applications

Evaluation of Moving Object Tracking Techniques for Video Surveillance Applications International Journal of Current Engineering and Technology E-ISSN 2277 4106, P-ISSN 2347 5161 2015INPRESSCO, All Rights Reserved Available at http://inpressco.com/category/ijcet Research Article Evaluation

More information

Bayesian Methods in Vision: MAP Estimation, MRFs, Optimization

Bayesian Methods in Vision: MAP Estimation, MRFs, Optimization Bayesian Methods in Vision: MAP Estimation, MRFs, Optimization CS 650: Computer Vision Bryan S. Morse Optimization Approaches to Vision / Image Processing Recurring theme: Cast vision problem as an optimization

More information

Fast HDR Image-Based Lighting Using Summed-Area Tables

Fast HDR Image-Based Lighting Using Summed-Area Tables Fast HDR Image-Based Lighting Using Summed-Area Tables Justin Hensley 1, Thorsten Scheuermann 2, Montek Singh 1 and Anselmo Lastra 1 1 University of North Carolina, Chapel Hill, NC, USA {hensley, montek,

More information

Stereo imaging ideal geometry

Stereo imaging ideal geometry Stereo imaging ideal geometry (X,Y,Z) Z f (x L,y L ) f (x R,y R ) Optical axes are parallel Optical axes separated by baseline, b. Line connecting lens centers is perpendicular to the optical axis, and

More information

The SIFT (Scale Invariant Feature

The SIFT (Scale Invariant Feature The SIFT (Scale Invariant Feature Transform) Detector and Descriptor developed by David Lowe University of British Columbia Initial paper ICCV 1999 Newer journal paper IJCV 2004 Review: Matt Brown s Canonical

More information

Textureless Layers CMU-RI-TR Qifa Ke, Simon Baker, and Takeo Kanade

Textureless Layers CMU-RI-TR Qifa Ke, Simon Baker, and Takeo Kanade Textureless Layers CMU-RI-TR-04-17 Qifa Ke, Simon Baker, and Takeo Kanade The Robotics Institute Carnegie Mellon University 5000 Forbes Avenue Pittsburgh, PA 15213 Abstract Layers are one of the most well

More information

Keywords:Synthetic Data, IBR, Data Generation Tool. Abstract

Keywords:Synthetic Data, IBR, Data Generation Tool. Abstract Data Generation Toolkit for Image Based Rendering Algorithms V Vamsi Krishna, P J Narayanan Center for Visual Information Technology International Institute of Information Technology, Hyderabad, India

More information

Segmentation with non-linear constraints on appearance, complexity, and geometry

Segmentation with non-linear constraints on appearance, complexity, and geometry IPAM February 2013 Western Univesity Segmentation with non-linear constraints on appearance, complexity, and geometry Yuri Boykov Andrew Delong Lena Gorelick Hossam Isack Anton Osokin Frank Schmidt Olga

More information