Efficient 3D Reconstruction for Urban Scenes
|
|
- Meagan Collins
- 5 years ago
- Views:
Transcription
1 Efficient 3D Reconstruction for Urban Scenes Weichao Fu, Lin Zhang, Hongyu Li, Xinfeng Zhang, and Di Wu School of Software Engineering, Tongji University, Shanghai, China Abstract. Recently, researchers working in the fields of computer graphics and computer vision have shown tremendous interests in reconstructing urban scenes. For this task, the acquisition of the 3D point clouds is the first step, for which scans are usually widely utilized. Nevertheless, on seeing the potential drawbacks of scans, in this paper, we propose a novel urban scene reconstruction system based on the Multi-View Stereo (MVS). Given a set of calibrated photographs, we first generate point clouds using an existing MVS algorithm, and then reconstruct the sub-structures that often regularly repeat in urban buildings. Finally, we recover the entire architectural models through an automatic growing algorithm of the sub-structures in dominant directions. Experimental results on regular urban buildings show the practicality and high efficiency of the proposed reconstruction method. Keywords: Urban Scene, Multi-View Stereo, sub-structures. 1 Introduction Reconstruction of urban scenes is attracting increasing attention these days, motivated by ambitious applications that aim to build digital copies of real cities (e.g., Microsoft Virtual Earth 3D and Google Earth 3D) [1]. It is a very complex problem with a lot of research history, open problems, and possible solutions. Usually, the first step for an urban scene reconstruction system is the acquisition of 3D point clouds and the state-of-the-art methods, such as [2], utilized laser scans for this step. However, as an approach for 3D point clouds acquisition from urban scenes, laser scans has some inherent drawbacks with respect to the cost and the easiness for distribution. In addition, the main difficulty of laser scans of large-scale urban environments is the data quality. Large distances between the scanner and the scanned objects reduce the precision and imply higher level of noise. The obtained point clouds usually exhibit significant missing data. As a result, in this paper, we propose a new urban scene reconstruction scheme based on another 3D point clouds acquisition approach, multi-view stereo (MVS). Compared with laser scans, MVS has the advantage of low cost to get data sources (e.g., Internet photo sharing sites like Flickr and Google Images). And for the urban scene, the quality of point clouds obtained by MVS is usually satisfactory. According to a quantitative evaluation on the Middlebury benchmark, a patch-based Corresponding author. D.-S. Huang et al. (Eds.): ICIC 2013, LNCS 7995, pp , Springer-Verlag Berlin Heidelberg 2013
2 Efficient 3D Reconstruction for Urban Scenes 547 MVS method [3] outperforms all others for most of the datasets. Hence, we make use of this PMVS algorithm to recover point clouds. Urban landscapes exhibit a high degree of self-similarity and redundancy [1]. This characteristic of urban scenes is not a chance occurrence, but is demonstrated universally across countries and cultures. Such large scale repetitions arise from aesthetics, build-ability, manufacturing ease, etc. Also urban buildings are mostly comprised of flat or near-planar faces because of functional requirements and constraints. Consequently, the presence of such characteristics suggests opportunities for simplifying the reconstruction task. (a) (b) (c) Fig. 1. Images of urban scenes with texture-poor but highly structured surfaces are increasingly ubiquitous on the Internet In this paper, we propose a 3D urban scene reconstruction method based on exploration of properties of architectural scenes. Our method is inspired by recent works of [2], which assemble an architectural model over a 3D point cloud by interactive operations. The key idea is to replace the smoothness prior used in traditional methods with priors that are more appropriate for urban buildings. Our approach operates as follows. Given a set of calibrated photographs, we first generate point clouds using an existing MVS algorithm, and then reconstruct the sub-structures that often regularly repeat in urban buildings. At this step, regularity and self-symmetry properties of urban buildings [1] are explored. Finally, we recover the entire architectural models through an automatic growing algorithm of the sub-structures in dominant directions. Our method, exploits properties of urban scenes, offers the following advantages. 1) It is remarkably robust to lack of texture, and able to model flat painted walls shown in Fig. 1(b). 2) It can work with poor quality point clouds generated from photographs due to occlusion, as shown in Fig. 1(c). 3) It produces remarkably clean and simple models as outputs. This paper is arranged as follows. In section 2 we review the PMVS algorithm and preprocess the point clouds generated by PMVS. In section 3 we reconstruct the substructures that often regularly repeat in urban buildings and in section 4 we recover the entire architectural models through an automatic growing algorithm of the substructures. Experimental results are presented in section 5, followed by conclusions in section 6.
3 548 W. Fu et al. 2 MVS Preprocessing PMVS algorithm presented in [3] is implemented as a match, expand, and filter procedure, starting from a sparse set of matched keypoints, and repeatedly expanding these before using visibility constraints to filter away false matches. Given a set of calibrated photographs, the first step of our method is to use the PMVS algorithm [3] to generate a set of oriented 3D points (positions and normal). We retain only highconfidence points in textured areas. Associated with each point P i, several attributes can be extracted, including the 3D location, a surface normal N i, a set of visible images V i, and a photometric consistency score (normalized cross correlation) C(P i ). 2.1 Outliers Removing Although the PMVS algorithm proposed in [3] enforces local photometric consistency and global visibility constraints to remove outliers, the point clouds generated by PMVS still have sparse outliers which will corrupt the estimation of local point cloud attributes such as surface normal or curvature changes. We solve these irregularities by performing a statistical analysis on each point s neighborhood, and trimming those which do not meet a certain criteria. Our sparse outlier removal scheme is based on the distribution of distances from a point to its neighbors in the PMVS point clouds. For each point, we compute the mean distance from it to all its neighbors. By assuming that the resulted distribution is Gaussian with a mean and a standard deviation, all points whose mean distances are outside an interval defined by the global distances mean and standard deviation can be considered as outliers and trimmed from the point clouds. An example of the outliers-removal is shown in Fig. 2. Fig. 2(a) shows the original point cloud obtained by PMVS, while Fig. 2(b) shows the result of outliers-removal. We set the number of nearest neighbors as 50 for mean distance estimation. (a) (b) (c) Fig. 2. (a) The original point cloud obtained by PMVS. (b) The resultant one after outliers removing. (c) The filtered point cloud with normals.
4 Efficient 3D Reconstruction for Urban Scenes Estimating Normals and Curvature The PMVS algorithm in [3] optimizes the normal of each patch by simply minimizing the photometric discrepancy score, and constraining the number of degrees of freedom to two (yaw and pitch). However, the normal optimized in PMVS is not good enough for urban scene, because of the large distance between building surface and the optical center of the camera, and the initialization of PMVS algorithm works poorly for the large scale urban buildings. Though many different normal estimation methods exist, the one we used in this paper is one of the simplest, and is formulated as follows. The problem of determining the normal to a point on the surface is approximated by the problem of estimating the normal of a plane tangent to the surface, which in turn reduces to an analysis of the eigenvectors and eigenvalues of a covariance matrix created from the nearest neighbors of the query point. More specifically, for each point P i, we assemble the covariance matrix C as follows: 1 C = P P P P C V =λ V j k T ( i )( i ), j j j, { 0,1,2} k i = 1 (1) where k is the number of point neighbors considered in the neighborhood of P i, P represents the 3D centroid of the nearest neighbors, λ j is the j-th eigenvalue of the covariance matrix, and V j is the j-th eigenvector. The point cloud with normals is shown in Fig. 2(c). The surface curvature change is estimated from the eigenvalues as: λ0, where λ 0 <λ 1 <λ2 λ +λ +λ (2) 2.3 Extracting Dominant Directions We could require the dominant axes of buildings in an automatic growing algorithm of the sub-structures. Hence, we employ a simple greedy algorithm using the normal estimates N i recovered above (See [7, 8, 9] for similar ideas). We compute a histogram of normal directions over a unit sphere, subdivided into 1000 bins. We then set the first dominant axis d 1 to the average of the normals within the largest bin. Next, we find the largest bin which is in the range 80 to 100 degrees away from d 1 and set the second dominant axis d 2 dominant axis d 3 to the average normal within that bin. Finally, the third is computed in the same way. We allow for some deviation from orthogonality to compensate for possible errors in normal estimation and to handle urban building that itself is not composed of exactly orthogonal planes.
5 550 W. Fu et al. 3 Sub-structure Reconstruction While in recent years many techniques have been developed to detect repeated parts in models [5, 11, 12, 13] and regularity directly on 3D geometry [6], in this paper, we attempt to learn the repetitions pattern directly from high level user guidance for the existence of occlusion or missing data. Due to the generally poor quality of generated data, we argue that some level of user interaction is required to understand the semantics of the building model. Hence, at the beginning of this section, we select the points that imply the sub-structures by our interactive tool. Then, a region growing algorithm is utilized for simplifying the reconstruction of the sub-structures. 3.1 Region Growing Algorithm The purpose of the region growing algorithm is to merge the points that are close enough in terms of the smoothness constraint. Thereby, the output of this algorithm is a set of clusters, and each cluster has a set of points that are considered to be a part of the same smooth surface. We first sort the points by their curvature value. It is necessary because the region begins its growth from the point that has the minimum curvature value. The reason for this is that the point with the minimum curvature is located in the flattest area. Then, as long as there are unlabeled points in the cloud, we pick up the point with minimum curvature value and start the growth of the region. First, the picked point is added to a set called seeds. Then, for every seed point we find its neighboring points, and test the angle between the normal of the current seed point and its neighbors normals. If the angle is less than a threshold θ, the current point is added to the current region. After that every neighbor is tested for the curvature value. If the curvature is less than a threshold δ, then this point is added to the seeds. Finally, current seed is removed from the seeds, and if the seed set becomes empty which means that a certain region has grown, we repeat the process from the beginning. The overall pseudo code for this growing algorithm is given in Table Reconstruction After the region growing algorithm, we employ an efficient RANSAC method for point-cloud shape detection in each region of the sub-structures. The RANSAC algorithm we used is presented in [9]. It decomposed the point cloud into a concise, hybrid structure of inherent shapes and a set of remaining points. The method can detect planes, spheres, cylinders, cones and tori. However, because of functional requirements and constraints, urban buildings are mostly comprised of flat or near-planar faces. Consequently, the RANSAC method with plane model detection is usually sufficient. Also, the region with too little points can be rejected or determined by high level user guidance which RANSAC model can be adopted. Eventually, the substructure is generated by enforcing alignment among the RANSAC models. This step will resolve the inconsistencies we wish to repair, e.g., model intersections, small gaps, and other forms of misalignments.
6 Efficient 3D Reconstruction for Urban Scenes 551 Table 1. The pseudo code for the region growing algorithm Input: Point cloud = { P}, Point normals = { N}, Point curvatures = { C}, Neighbor finding function Ω ( ), Curvature threshold δ, Angle threshold θ. Output: Region list { R }. R φ, Available points list { A} { P1,..., Pn } While { A } is not empty do Current region{ R c } φ, Current seeds { S c } φ Point with minimum curvature in { A} Pmin { Sc} { Sc} Pmin,{ Rc} { Rc} Pmin,{ A} { A} \ Pmin for i = 0 to size ({ Sc} ) do Find nearest neighbors of current seed point { Bc} Ω ( Sc{ i} ) for j = 0 to size ({ Bc }) do Current neighbor point Pj Bc{ j} 1 If { A } contains Pj and cos ( N{ Sc{} i }, N{ Bc{} j }) < θ then { Rc} { Rc} Pj { A} { A} \ Pj If C{ P j } < δ then { S } { S } P end if end if end for end for c c j Add current region to global region list { R} { R} { R c } end while 4 Automatic Continuation Our sub-structure automatic continuation algorithm is inspired by recent works in [2]. However, in [2], some essential interactive operations are utilized for creating complex structures through the propagation of a repeated structure. In our case, the automatic continuation of sub-structure is constrained along the dominant directions computed in Section 2, which can minimize the required user interaction. Also, in [2], an optimization which balances between the data-fitting and contextual forces is adopted. We continue this trend and present our data-fitting term and contextual term below. Given a sub-structure S and the point cloud, the datafitting term, also known as the fitting error is computed by measuring the one-sided Euclidean distance from the sub-structure to points [14]. The contextual sub-structure S required to define the contextual force is a previously positioned sub-structure in a reconstruction sequence. The contextual term is defined as the sum of two terms: the interval term and the alignment term. The interval term I( S, S ) measures how well
7 552 W. Fu et al. the interval length between S and S agrees with the expected interval length due to regularity constraints. The alignment term A( SS, ) measures how well corresponding edges of S and S align. These two terms are normalized independently and weighted equally to form the contextual term. The automatic growing algorithm is achieved by finding an optimal linear transformation T, consisting of translation and scaling of the sub-structure S, to minim- * ize a weighted sum of data-fitting and contextual terms, { ω ( ( ) ) ( ω) ( ( ) )} * T arg min D T S, P 1 C T S, S T = + (3) where ω is a weight which balances between the data-fitting and contextual forces. The choice of the weight ω should correlate with an assessment of data quality and contextual regularity in the urban building model to be reconstructed. In our method, the default setting of ω is 0.5. Nevertheless, the user can manually adjust it according to data quality and contextual regularity perceived by them. 5 Experimental Results The proposed reconstruction method was tested on both a publicly available data set: Hall [4] and our own images of Building F in Tongji University (taken by cell phone Fig. 3. Top left: a target image. Top middle: the point cloud of the sub-structure. Top right: sub-structure reconstruction. Middle: sub-structure growing among the dominant direction. Bottom: the reconstructed model with texture mapping.
8 Efficient 3D Reconstruction for Urban Scenes 553 Fig. 4. Top left: a target image. Top right: the point cloud of the sub-structure. Bottom: the reconstructed model with texture mapping. camera). The Hall dataset consists of 61 images which are used for PMVS algorithm to generate point cloud. This results in a total of points. In our case we use only the high-confidence points, thus, we have points after the MVS preprocessing step. We set θ = and δ = 1.0 for the thresholds used in region growing algorithm. The interval term used in automatic continuation is initialized as the length of the sub-structure s projection among the dominant direction. Our own
9 554 W. Fu et al. Building F dataset consists of 15 images taken by cell phone camera, and some of the images have occlusion as shown in top left of Fig. 4. The reconstruction results are shown in Fig. 3 and Fig. 4, which illustrate the convenience and robustness of our proposed method. 6 Conclusions and the Future Work In this paper, we have demonstrated a Multi-View stereopsis based 3D urban scene reconstruction method that utilizes the large scale repetitions and self-similarities of urban buildings. Our method is remarkably robust to buildings lack of texture, and also able to work with poor quality point clouds by exploiting properties of urban scenes. In future work, we plan to incorporate any parametric primitive that can be detected using standard RANSAC in our sub-structure reconstruction step. This may lead to a better model and exhibit the diversity of modern urban scenes. Acknowledgement. This work is supported by the Fundamental Research Funds for the Central Universities under grant no , the Natural Science Foundation of China under grant no , and the Innovation Program of Shanghai Municipal Education Commission under grant no. 12ZZ029. References 1. Zheng, Q., Sharf, A., Wan, G., Li, Y., Mitra, N.J., Cohen-Or, D., Chen, B.: Non-local Scan Consolidation for 3D Urban Scenes. ACM Transactions on Graphics 29, 94 (2010) 2. Nan, L., Sharf, A., Zhang, H., Cohen-Or, D., Chen, B.: SmartBoxes for Interactive Urban Reconstruction. ACM Transactions on Graphics 29, 93 (2010) 3. Furukawa, Y., Ponce, J.: Accurate, dense, and robust Multi-View stereopsis. In: CVPR, pp. 1 8 (2007) 4. Furukawa, Y., Curless, B., Seitz, S.M., Szeliski, R.: Manhattan-world Stereo. In: CVPR, pp (2009) 5. Liu, J., Musialski, P., Wonka, P., Ye, J.: Tensor Completion for Estimating Missing Values in Visual Data. In: ICCV, pp (2009) 6. Pauly, M., Mitra, N.J., Wallner, J., Pottmann, H., Guibas, L.J.: Discovering Structural Regularity in 3D Geometry. ACM Transactions on Graphics 27, 43 (2008) 7. Coorg, S., Teller, S.: Extracting Textured Vertical Facades from Controlled Close-Range Imagery. In: CVPR, pp (1999) 8. Pollefeys, M., Nister, D., Frahm, J.M., Akbarzadeh, A., Mordohai, P., Clipp, B., Engels, C., Gallup, D., Kim, S.J., Merrell, P.: Detailed Real-Time Urban 3D Reconstruction from Video. In: IJCV 2008, pp (2008) 9. Werner, T., Zisserman, A.: New Techniques for Automated Architectural Reconstruction from Photographs. In: Heyden, A., Sparr, G., Nielsen, M., Johansen, P. (eds.) ECCV 2002, Part II. LNCS, vol. 2351, pp Springer, Heidelberg (2002) 10. Schnabel, R., Wahl, R., Klein, R.: Efficient RANSAC for Point-Cloud Shape Detection. Computer Graphics Forum 26, (2007)
10 Efficient 3D Reconstruction for Urban Scenes Mitra, N.J., Guibas, L.J., Pauly, M.: Partial And Approximate Symmetry Detection for 3D Geometry. ACM Transactions on Graphics 25, (2006) 12. Korah, T., Rasmussen, C.: Spatiotemporal Inpainting for Recovering Texture Maps of Occluded Building Facades. IEEE Transactions on Image Processing 16, (2007) 13. Pauly, M., Mitra, N.J., Wallner, J., Pottmann, H., Guibas, L.J.: Discovering Structural Regularity in 3D Geometry. ACM Transactions on Graphics 27, 43 (2008) 14. Nan, L., Xie, K., Sharf, A.: A Search Classify Approach for Cluttered Indoor Scene Understanding. ACM Transactions on Graphics 31, 137 (2012)
SIMPLE ROOM SHAPE MODELING WITH SPARSE 3D POINT INFORMATION USING PHOTOGRAMMETRY AND APPLICATION SOFTWARE
SIMPLE ROOM SHAPE MODELING WITH SPARSE 3D POINT INFORMATION USING PHOTOGRAMMETRY AND APPLICATION SOFTWARE S. Hirose R&D Center, TOPCON CORPORATION, 75-1, Hasunuma-cho, Itabashi-ku, Tokyo, Japan Commission
More informationReconstruction of Polygonal Faces from Large-Scale Point-Clouds of Engineering Plants
1 Reconstruction of Polygonal Faces from Large-Scale Point-Clouds of Engineering Plants Hiroshi Masuda 1, Takeru Niwa 2, Ichiro Tanaka 3 and Ryo Matsuoka 4 1 The University of Electro-Communications, h.masuda@euc.ac.jp
More informationMulti-view Stereo. Ivo Boyadzhiev CS7670: September 13, 2011
Multi-view Stereo Ivo Boyadzhiev CS7670: September 13, 2011 What is stereo vision? Generic problem formulation: given several images of the same object or scene, compute a representation of its 3D shape
More informationA DATA DRIVEN METHOD FOR FLAT ROOF BUILDING RECONSTRUCTION FROM LiDAR POINT CLOUDS
A DATA DRIVEN METHOD FOR FLAT ROOF BUILDING RECONSTRUCTION FROM LiDAR POINT CLOUDS A. Mahphood, H. Arefi *, School of Surveying and Geospatial Engineering, College of Engineering, University of Tehran,
More informationDense 3D Reconstruction. Christiano Gava
Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Today: dense 3D reconstruction The matching problem
More informationDETECTION AND ROBUST ESTIMATION OF CYLINDER FEATURES IN POINT CLOUDS INTRODUCTION
DETECTION AND ROBUST ESTIMATION OF CYLINDER FEATURES IN POINT CLOUDS Yun-Ting Su James Bethel Geomatics Engineering School of Civil Engineering Purdue University 550 Stadium Mall Drive, West Lafayette,
More informationDense 3D Reconstruction. Christiano Gava
Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Wide baseline matching (SIFT) Today: dense 3D reconstruction
More informationBUILDING POINT GROUPING USING VIEW-GEOMETRY RELATIONS INTRODUCTION
BUILDING POINT GROUPING USING VIEW-GEOMETRY RELATIONS I-Chieh Lee 1, Shaojun He 1, Po-Lun Lai 2, Alper Yilmaz 2 1 Mapping and GIS Laboratory 2 Photogrammetric Computer Vision Laboratory Dept. of Civil
More informationMulti-view stereo. Many slides adapted from S. Seitz
Multi-view stereo Many slides adapted from S. Seitz Beyond two-view stereo The third eye can be used for verification Multiple-baseline stereo Pick a reference image, and slide the corresponding window
More informationStructured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov
Structured Light II Johannes Köhler Johannes.koehler@dfki.de Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Introduction Previous lecture: Structured Light I Active Scanning Camera/emitter
More informationLarge Scale 3D Reconstruction by Structure from Motion
Large Scale 3D Reconstruction by Structure from Motion Devin Guillory Ziang Xie CS 331B 7 October 2013 Overview Rome wasn t built in a day Overview of SfM Building Rome in a Day Building Rome on a Cloudless
More information3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.
3D Computer Vision Structured Light II Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de 1 Introduction
More information3D Computer Vision. Dense 3D Reconstruction II. Prof. Didier Stricker. Christiano Gava
3D Computer Vision Dense 3D Reconstruction II Prof. Didier Stricker Christiano Gava Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de
More information2D-3D Fusion for Layer Decomposition of Urban Facades
2D-3D Fusion for Layer Decomposition of Urban Facades Yangyan Li Qian Zheng Andrei Sharf, Daniel Cohen-Or Baoquan Chen Niloy J. Mitra SIAT, China Ben-Gurion Univ. Tel Aviv Univ. UCL Abstract We present
More informationManhattan-world Stereo and Surface Reconstruction
Manhattan-world Stereo and Surface Reconstruction Kuan-Ting Yu CSAIL MIT Cambridge, USA peterkty@csail.mit.edu Abstract Depth estimation from 2D images has been extensively studied in the computer vision
More informationUrban Scene Segmentation, Recognition and Remodeling. Part III. Jinglu Wang 11/24/2016 ACCV 2016 TUTORIAL
Part III Jinglu Wang Urban Scene Segmentation, Recognition and Remodeling 102 Outline Introduction Related work Approaches Conclusion and future work o o - - ) 11/7/16 103 Introduction Motivation Motivation
More informationA System of Image Matching and 3D Reconstruction
A System of Image Matching and 3D Reconstruction CS231A Project Report 1. Introduction Xianfeng Rui Given thousands of unordered images of photos with a variety of scenes in your gallery, you will find
More informationRobotics Programming Laboratory
Chair of Software Engineering Robotics Programming Laboratory Bertrand Meyer Jiwon Shin Lecture 8: Robot Perception Perception http://pascallin.ecs.soton.ac.uk/challenges/voc/databases.html#caltech car
More informationImage Based Reconstruction II
Image Based Reconstruction II Qixing Huang Feb. 2 th 2017 Slide Credit: Yasutaka Furukawa Image-Based Geometry Reconstruction Pipeline Last Lecture: Multi-View SFM Multi-View SFM This Lecture: Multi-View
More informationStep-by-Step Model Buidling
Step-by-Step Model Buidling Review Feature selection Feature selection Feature correspondence Camera Calibration Euclidean Reconstruction Landing Augmented Reality Vision Based Control Sparse Structure
More informationMiniature faking. In close-up photo, the depth of field is limited.
Miniature faking In close-up photo, the depth of field is limited. http://en.wikipedia.org/wiki/file:jodhpur_tilt_shift.jpg Miniature faking Miniature faking http://en.wikipedia.org/wiki/file:oregon_state_beavers_tilt-shift_miniature_greg_keene.jpg
More informationarxiv: v1 [cs.cv] 28 Sep 2018
Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,
More informationA Systems View of Large- Scale 3D Reconstruction
Lecture 23: A Systems View of Large- Scale 3D Reconstruction Visual Computing Systems Goals and motivation Construct a detailed 3D model of the world from unstructured photographs (e.g., Flickr, Facebook)
More informationFeature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies
Feature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies M. Lourakis, S. Tzurbakis, A. Argyros, S. Orphanoudakis Computer Vision and Robotics Lab (CVRL) Institute of
More informationSupplementary Material : Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision
Supplementary Material : Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision Due to space limitation in the main paper, we present additional experimental results in this supplementary
More informationFAST REGISTRATION OF TERRESTRIAL LIDAR POINT CLOUD AND SEQUENCE IMAGES
FAST REGISTRATION OF TERRESTRIAL LIDAR POINT CLOUD AND SEQUENCE IMAGES Jie Shao a, Wuming Zhang a, Yaqiao Zhu b, Aojie Shen a a State Key Laboratory of Remote Sensing Science, Institute of Remote Sensing
More informationLocal features: detection and description. Local invariant features
Local features: detection and description Local invariant features Detection of interest points Harris corner detection Scale invariant blob detection: LoG Description of local patches SIFT : Histograms
More informationEECS 442 Computer vision. Stereo systems. Stereo vision Rectification Correspondence problem Active stereo vision systems
EECS 442 Computer vision Stereo systems Stereo vision Rectification Correspondence problem Active stereo vision systems Reading: [HZ] Chapter: 11 [FP] Chapter: 11 Stereo vision P p p O 1 O 2 Goal: estimate
More informationPhoto Tourism: Exploring Photo Collections in 3D
Photo Tourism: Exploring Photo Collections in 3D SIGGRAPH 2006 Noah Snavely Steven M. Seitz University of Washington Richard Szeliski Microsoft Research 2006 2006 Noah Snavely Noah Snavely Reproduced with
More informationLearning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009
Learning and Inferring Depth from Monocular Images Jiyan Pan April 1, 2009 Traditional ways of inferring depth Binocular disparity Structure from motion Defocus Given a single monocular image, how to infer
More informationModel-based segmentation and recognition from range data
Model-based segmentation and recognition from range data Jan Boehm Institute for Photogrammetry Universität Stuttgart Germany Keywords: range image, segmentation, object recognition, CAD ABSTRACT This
More informationThe SIFT (Scale Invariant Feature
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor developed by David Lowe University of British Columbia Initial paper ICCV 1999 Newer journal paper IJCV 2004 Review: Matt Brown s Canonical
More informationRobot localization method based on visual features and their geometric relationship
, pp.46-50 http://dx.doi.org/10.14257/astl.2015.85.11 Robot localization method based on visual features and their geometric relationship Sangyun Lee 1, Changkyung Eem 2, and Hyunki Hong 3 1 Department
More informationCS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching
Stereo Matching Fundamental matrix Let p be a point in left image, p in right image l l Epipolar relation p maps to epipolar line l p maps to epipolar line l p p Epipolar mapping described by a 3x3 matrix
More informationWhat is Computer Vision?
Perceptual Grouping in Computer Vision Gérard Medioni University of Southern California What is Computer Vision? Computer Vision Attempt to emulate Human Visual System Perceive visual stimuli with cameras
More informationAlgorithm research of 3D point cloud registration based on iterative closest point 1
Acta Technica 62, No. 3B/2017, 189 196 c 2017 Institute of Thermomechanics CAS, v.v.i. Algorithm research of 3D point cloud registration based on iterative closest point 1 Qian Gao 2, Yujian Wang 2,3,
More information3D object recognition used by team robotto
3D object recognition used by team robotto Workshop Juliane Hoebel February 1, 2016 Faculty of Computer Science, Otto-von-Guericke University Magdeburg Content 1. Introduction 2. Depth sensor 3. 3D object
More informationAutomatic Pipeline Generation by the Sequential Segmentation and Skelton Construction of Point Cloud
, pp.43-47 http://dx.doi.org/10.14257/astl.2014.67.11 Automatic Pipeline Generation by the Sequential Segmentation and Skelton Construction of Point Cloud Ashok Kumar Patil, Seong Sill Park, Pavitra Holi,
More informationLecture 10: Multi view geometry
Lecture 10: Multi view geometry Professor Fei Fei Li Stanford Vision Lab 1 What we will learn today? Stereo vision Correspondence problem (Problem Set 2 (Q3)) Active stereo vision systems Structure from
More information3D model search and pose estimation from single images using VIP features
3D model search and pose estimation from single images using VIP features Changchang Wu 2, Friedrich Fraundorfer 1, 1 Department of Computer Science ETH Zurich, Switzerland {fraundorfer, marc.pollefeys}@inf.ethz.ch
More informationTranslation Symmetry Detection: A Repetitive Pattern Analysis Approach
2013 IEEE Conference on Computer Vision and Pattern Recognition Workshops Translation Symmetry Detection: A Repetitive Pattern Analysis Approach Yunliang Cai and George Baciu GAMA Lab, Department of Computing
More informationLocal invariant features
Local invariant features Tuesday, Oct 28 Kristen Grauman UT-Austin Today Some more Pset 2 results Pset 2 returned, pick up solutions Pset 3 is posted, due 11/11 Local invariant features Detection of interest
More informationMotion Tracking and Event Understanding in Video Sequences
Motion Tracking and Event Understanding in Video Sequences Isaac Cohen Elaine Kang, Jinman Kang Institute for Robotics and Intelligent Systems University of Southern California Los Angeles, CA Objectives!
More informationChaplin, Modern Times, 1936
Chaplin, Modern Times, 1936 [A Bucket of Water and a Glass Matte: Special Effects in Modern Times; bonus feature on The Criterion Collection set] Multi-view geometry problems Structure: Given projections
More informationStatic Scene Reconstruction
GPU supported Real-Time Scene Reconstruction with a Single Camera Jan-Michael Frahm, 3D Computer Vision group, University of North Carolina at Chapel Hill Static Scene Reconstruction 1 Capture on campus
More informationCS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching
Stereo Matching Fundamental matrix Let p be a point in left image, p in right image l l Epipolar relation p maps to epipolar line l p maps to epipolar line l p p Epipolar mapping described by a 3x3 matrix
More informationSimultaneous Vanishing Point Detection and Camera Calibration from Single Images
Simultaneous Vanishing Point Detection and Camera Calibration from Single Images Bo Li, Kun Peng, Xianghua Ying, and Hongbin Zha The Key Lab of Machine Perception (Ministry of Education), Peking University,
More informationBIL Computer Vision Apr 16, 2014
BIL 719 - Computer Vision Apr 16, 2014 Binocular Stereo (cont d.), Structure from Motion Aykut Erdem Dept. of Computer Engineering Hacettepe University Slide credit: S. Lazebnik Basic stereo matching algorithm
More informationAUTOMATED 3D MODELING OF URBAN ENVIRONMENTS
AUTOMATED 3D MODELING OF URBAN ENVIRONMENTS Ioannis Stamos Department of Computer Science Hunter College, City University of New York 695 Park Avenue, New York NY 10065 istamos@hunter.cuny.edu http://www.cs.hunter.cuny.edu/
More informationGraph Matching Iris Image Blocks with Local Binary Pattern
Graph Matching Iris Image Blocs with Local Binary Pattern Zhenan Sun, Tieniu Tan, and Xianchao Qiu Center for Biometrics and Security Research, National Laboratory of Pattern Recognition, Institute of
More informationImproving PMVS Algorithm for 3D Scene Reconstruction from Sparse Stereo Pairs
Improving PMVS Algorithm for 3D Scene Reconstruction from Sparse Stereo Pairs Bo Li 1,Y.V.Venkatesh 2, Ashraf Kassim 3,andYijuanLu 1 1 Department of Computer Science, Texas State University, San Marcos,
More informationSmartBoxes for Interactive Urban Reconstruction
SmartBoxes for Interactive Urban Reconstruction Liangliang Nan 1 Andrei Sharf 1 Hao Zhang 2 Daniel Cohen-Or 3 Baoquan Chen 1 1 Shenzhen Institutes of Advanced Technology (SIAT), China 2 Simon Fraser Univ.
More informationFeature Based Registration - Image Alignment
Feature Based Registration - Image Alignment Image Registration Image registration is the process of estimating an optimal transformation between two or more images. Many slides from Alexei Efros http://graphics.cs.cmu.edu/courses/15-463/2007_fall/463.html
More informationMerging the Unmatchable: Stitching Visually Disconnected SfM Models
This is a preprint for the paper accepted for publication in ICCV 2015. c 2015 IEEE Merging the Unmatchable: Stitching Visually Disconnected SfM Models Andrea Cohen, Torsten Sattler, Marc Pollefeys Department
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 14 130307 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Stereo Dense Motion Estimation Translational
More informationViewpoint Invariant Features from Single Images Using 3D Geometry
Viewpoint Invariant Features from Single Images Using 3D Geometry Yanpeng Cao and John McDonald Department of Computer Science National University of Ireland, Maynooth, Ireland {y.cao,johnmcd}@cs.nuim.ie
More informationLecture 10: Multi-view geometry
Lecture 10: Multi-view geometry Professor Stanford Vision Lab 1 What we will learn today? Review for stereo vision Correspondence problem (Problem Set 2 (Q3)) Active stereo vision systems Structure from
More informationSOME stereo image-matching methods require a user-selected
IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, VOL. 3, NO. 2, APRIL 2006 207 Seed Point Selection Method for Triangle Constrained Image Matching Propagation Qing Zhu, Bo Wu, and Zhi-Xiang Xu Abstract In order
More information3D Reconstruction Using an n-layer Heightmap
3D Reconstruction Using an n-layer Heightmap David Gallup 1, Marc Pollefeys 2, and Jan-Michael Frahm 1 1 Department of Computer Science, University of North Carolina {gallup,jmf}@cs.unc.edu 2 Department
More informationLayered Scene Decomposition via the Occlusion-CRF Supplementary material
Layered Scene Decomposition via the Occlusion-CRF Supplementary material Chen Liu 1 Pushmeet Kohli 2 Yasutaka Furukawa 1 1 Washington University in St. Louis 2 Microsoft Research Redmond 1. Additional
More informationImage Parallax based Modeling of Depth-layer Architecture
Image Parallax based Modeling of Depth-layer Architecture Yong Hu, Bei Chu, Yue Qi State Key Laboratory of Virtual Reality Technology and Systems, Beihang University School of New Media Art and Design,
More informationCRF Based Point Cloud Segmentation Jonathan Nation
CRF Based Point Cloud Segmentation Jonathan Nation jsnation@stanford.edu 1. INTRODUCTION The goal of the project is to use the recently proposed fully connected conditional random field (CRF) model to
More informationCamera Registration in a 3D City Model. Min Ding CS294-6 Final Presentation Dec 13, 2006
Camera Registration in a 3D City Model Min Ding CS294-6 Final Presentation Dec 13, 2006 Goal: Reconstruct 3D city model usable for virtual walk- and fly-throughs Virtual reality Urban planning Simulation
More informationMarcel Worring Intelligent Sensory Information Systems
Marcel Worring worring@science.uva.nl Intelligent Sensory Information Systems University of Amsterdam Information and Communication Technology archives of documentaries, film, or training material, video
More informationGeometry based Repetition Detection for Urban Scene
Geometry based Repetition Detection for Urban Scene Changchang Wu University of Washington Jan Michael Frahm UNC Chapel Hill Marc Pollefeys ETH Zürich Related Work Sparse Feature Matching [Loy et al. 06,
More informationFeatures Points. Andrea Torsello DAIS Università Ca Foscari via Torino 155, Mestre (VE)
Features Points Andrea Torsello DAIS Università Ca Foscari via Torino 155, 30172 Mestre (VE) Finding Corners Edge detectors perform poorly at corners. Corners provide repeatable points for matching, so
More informationBSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy
BSB663 Image Processing Pinar Duygulu Slides are adapted from Selim Aksoy Image matching Image matching is a fundamental aspect of many problems in computer vision. Object or scene recognition Solving
More informationSpecular 3D Object Tracking by View Generative Learning
Specular 3D Object Tracking by View Generative Learning Yukiko Shinozuka, Francois de Sorbier and Hideo Saito Keio University 3-14-1 Hiyoshi, Kohoku-ku 223-8522 Yokohama, Japan shinozuka@hvrl.ics.keio.ac.jp
More informationLocal Feature Detectors
Local Feature Detectors Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr Slides adapted from Cordelia Schmid and David Lowe, CVPR 2003 Tutorial, Matthew Brown,
More informationMultiview Reconstruction
Multiview Reconstruction Why More Than 2 Views? Baseline Too short low accuracy Too long matching becomes hard Why More Than 2 Views? Ambiguity with 2 views Camera 1 Camera 2 Camera 3 Trinocular Stereo
More informationFast K-nearest neighbors searching algorithms for point clouds data of 3D scanning system 1
Acta Technica 62 No. 3B/2017, 141 148 c 2017 Institute of Thermomechanics CAS, v.v.i. Fast K-nearest neighbors searching algorithms for point clouds data of 3D scanning system 1 Zhang Fan 2, 3, Tan Yuegang
More informationContinuous Multi-View Tracking using Tensor Voting
Continuous Multi-View Tracking using Tensor Voting Jinman Kang, Isaac Cohen and Gerard Medioni Institute for Robotics and Intelligent Systems University of Southern California {jinmanka, icohen, medioni}@iris.usc.edu
More informationStructured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov
Structured Light II Johannes Köhler Johannes.koehler@dfki.de Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Introduction Previous lecture: Structured Light I Active Scanning Camera/emitter
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 12 130228 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Panoramas, Mosaics, Stitching Two View Geometry
More informationSpecular Reflection Separation using Dark Channel Prior
2013 IEEE Conference on Computer Vision and Pattern Recognition Specular Reflection Separation using Dark Channel Prior Hyeongwoo Kim KAIST hyeongwoo.kim@kaist.ac.kr Hailin Jin Adobe Research hljin@adobe.com
More informationRansac Based Out-of-Core Point-Cloud Shape Detection for City-Modeling
Ransac Based Out-of-Core Point-Cloud Shape Detection for City-Modeling Ruwen Schnabel, Roland Wahl, Reinhard Klein Institut für Informatik II, Universität Bonn Römerstr. 117 53117 Bonn {schnabel,wahl,rk}@cs.uni-bonn.de
More informationManhattan-world Urban Reconstruction from Point Clouds
Manhattan-world Urban Reconstruction from Point Clouds Minglei Li 1,2, Peter Wonka 1, and Liangliang Nan 1 1 Visual Computing Center, KAUST, Saudi Arabia {mingleili87, liangliang.nan, pwonka}@gmail.com
More informationLocal features and image matching. Prof. Xin Yang HUST
Local features and image matching Prof. Xin Yang HUST Last time RANSAC for robust geometric transformation estimation Translation, Affine, Homography Image warping Given a 2D transformation T and a source
More informationA Statistical Consistency Check for the Space Carving Algorithm.
A Statistical Consistency Check for the Space Carving Algorithm. A. Broadhurst and R. Cipolla Dept. of Engineering, Univ. of Cambridge, Cambridge, CB2 1PZ aeb29 cipolla @eng.cam.ac.uk Abstract This paper
More information3D Reconstruction of a Hopkins Landmark
3D Reconstruction of a Hopkins Landmark Ayushi Sinha (461), Hau Sze (461), Diane Duros (361) Abstract - This paper outlines a method for 3D reconstruction from two images. Our procedure is based on known
More information3D Perception. CS 4495 Computer Vision K. Hawkins. CS 4495 Computer Vision. 3D Perception. Kelsey Hawkins Robotics
CS 4495 Computer Vision Kelsey Hawkins Robotics Motivation What do animals, people, and robots want to do with vision? Detect and recognize objects/landmarks Find location of objects with respect to themselves
More informationREFINEMENT OF COLORED MOBILE MAPPING DATA USING INTENSITY IMAGES
REFINEMENT OF COLORED MOBILE MAPPING DATA USING INTENSITY IMAGES T. Yamakawa a, K. Fukano a,r. Onodera a, H. Masuda a, * a Dept. of Mechanical Engineering and Intelligent Systems, The University of Electro-Communications,
More informationInstance-level recognition
Instance-level recognition 1) Local invariant features 2) Matching and recognition with local features 3) Efficient visual search 4) Very large scale indexing Matching of descriptors Matching and 3D reconstruction
More informationHow to Compute the Pose of an Object without a Direct View?
How to Compute the Pose of an Object without a Direct View? Peter Sturm and Thomas Bonfort INRIA Rhône-Alpes, 38330 Montbonnot St Martin, France {Peter.Sturm, Thomas.Bonfort}@inrialpes.fr Abstract. We
More informationStereo and Epipolar geometry
Previously Image Primitives (feature points, lines, contours) Today: Stereo and Epipolar geometry How to match primitives between two (multiple) views) Goals: 3D reconstruction, recognition Jana Kosecka
More informationPreviously. Part-based and local feature models for generic object recognition. Bag-of-words model 4/20/2011
Previously Part-based and local feature models for generic object recognition Wed, April 20 UT-Austin Discriminative classifiers Boosting Nearest neighbors Support vector machines Useful for object recognition
More informationRegistration of Dynamic Range Images
Registration of Dynamic Range Images Tan-Chi Ho 1,2 Jung-Hong Chuang 1 Wen-Wei Lin 2 Song-Sun Lin 2 1 Department of Computer Science National Chiao-Tung University 2 Department of Applied Mathematics National
More informationTextured Mesh Surface Reconstruction of Large Buildings with Multi-View Stereo
The Visual Computer manuscript No. (will be inserted by the editor) Textured Mesh Surface Reconstruction of Large Buildings with Multi-View Stereo Chen Zhu Wee Kheng Leow Received: date / Accepted: date
More informationContinuous Multi-Views Tracking using Tensor Voting
Continuous Multi-Views racking using ensor Voting Jinman Kang, Isaac Cohen and Gerard Medioni Institute for Robotics and Intelligent Systems University of Southern California Los Angeles, CA 90089-073.
More informationHomographies and RANSAC
Homographies and RANSAC Computer vision 6.869 Bill Freeman and Antonio Torralba March 30, 2011 Homographies and RANSAC Homographies RANSAC Building panoramas Phototourism 2 Depth-based ambiguity of position
More informationSimilarity Image Retrieval System Using Hierarchical Classification
Similarity Image Retrieval System Using Hierarchical Classification Experimental System on Mobile Internet with Cellular Phone Masahiro Tada 1, Toshikazu Kato 1, and Isao Shinohara 2 1 Department of Industrial
More informationA Desktop 3D Scanner Exploiting Rotation and Visual Rectification of Laser Profiles
A Desktop 3D Scanner Exploiting Rotation and Visual Rectification of Laser Profiles Carlo Colombo, Dario Comanducci, and Alberto Del Bimbo Dipartimento di Sistemi ed Informatica Via S. Marta 3, I-5139
More informationCS4495/6495 Introduction to Computer Vision
CS4495/6495 Introduction to Computer Vision 9C-L1 3D perception Some slides by Kelsey Hawkins Motivation Why do animals, people & robots need vision? To detect and recognize objects/landmarks Is that a
More informationK-Means Based Matching Algorithm for Multi-Resolution Feature Descriptors
K-Means Based Matching Algorithm for Multi-Resolution Feature Descriptors Shao-Tzu Huang, Chen-Chien Hsu, Wei-Yen Wang International Science Index, Electrical and Computer Engineering waset.org/publication/0007607
More informationLOAM: LiDAR Odometry and Mapping in Real Time
LOAM: LiDAR Odometry and Mapping in Real Time Aayush Dwivedi (14006), Akshay Sharma (14062), Mandeep Singh (14363) Indian Institute of Technology Kanpur 1 Abstract This project deals with online simultaneous
More informationMultiple View Geometry
Multiple View Geometry Martin Quinn with a lot of slides stolen from Steve Seitz and Jianbo Shi 15-463: Computational Photography Alexei Efros, CMU, Fall 2007 Our Goal The Plenoptic Function P(θ,φ,λ,t,V
More informationInstance-level recognition part 2
Visual Recognition and Machine Learning Summer School Paris 2011 Instance-level recognition part 2 Josef Sivic http://www.di.ens.fr/~josef INRIA, WILLOW, ENS/INRIA/CNRS UMR 8548 Laboratoire d Informatique,
More informationCombining Appearance and Topology for Wide
Combining Appearance and Topology for Wide Baseline Matching Dennis Tell and Stefan Carlsson Presented by: Josh Wills Image Point Correspondences Critical foundation for many vision applications 3-D reconstruction,
More information3D Photography: Stereo
3D Photography: Stereo Marc Pollefeys, Torsten Sattler Spring 2016 http://www.cvg.ethz.ch/teaching/3dvision/ 3D Modeling with Depth Sensors Today s class Obtaining depth maps / range images unstructured
More informationMulti-View Stereo for Static and Dynamic Scenes
Multi-View Stereo for Static and Dynamic Scenes Wolfgang Burgard Jan 6, 2010 Main references Yasutaka Furukawa and Jean Ponce, Accurate, Dense and Robust Multi-View Stereopsis, 2007 C.L. Zitnick, S.B.
More information