Efficient 3D Reconstruction for Urban Scenes

Size: px
Start display at page:

Download "Efficient 3D Reconstruction for Urban Scenes"

Transcription

1 Efficient 3D Reconstruction for Urban Scenes Weichao Fu, Lin Zhang, Hongyu Li, Xinfeng Zhang, and Di Wu School of Software Engineering, Tongji University, Shanghai, China Abstract. Recently, researchers working in the fields of computer graphics and computer vision have shown tremendous interests in reconstructing urban scenes. For this task, the acquisition of the 3D point clouds is the first step, for which scans are usually widely utilized. Nevertheless, on seeing the potential drawbacks of scans, in this paper, we propose a novel urban scene reconstruction system based on the Multi-View Stereo (MVS). Given a set of calibrated photographs, we first generate point clouds using an existing MVS algorithm, and then reconstruct the sub-structures that often regularly repeat in urban buildings. Finally, we recover the entire architectural models through an automatic growing algorithm of the sub-structures in dominant directions. Experimental results on regular urban buildings show the practicality and high efficiency of the proposed reconstruction method. Keywords: Urban Scene, Multi-View Stereo, sub-structures. 1 Introduction Reconstruction of urban scenes is attracting increasing attention these days, motivated by ambitious applications that aim to build digital copies of real cities (e.g., Microsoft Virtual Earth 3D and Google Earth 3D) [1]. It is a very complex problem with a lot of research history, open problems, and possible solutions. Usually, the first step for an urban scene reconstruction system is the acquisition of 3D point clouds and the state-of-the-art methods, such as [2], utilized laser scans for this step. However, as an approach for 3D point clouds acquisition from urban scenes, laser scans has some inherent drawbacks with respect to the cost and the easiness for distribution. In addition, the main difficulty of laser scans of large-scale urban environments is the data quality. Large distances between the scanner and the scanned objects reduce the precision and imply higher level of noise. The obtained point clouds usually exhibit significant missing data. As a result, in this paper, we propose a new urban scene reconstruction scheme based on another 3D point clouds acquisition approach, multi-view stereo (MVS). Compared with laser scans, MVS has the advantage of low cost to get data sources (e.g., Internet photo sharing sites like Flickr and Google Images). And for the urban scene, the quality of point clouds obtained by MVS is usually satisfactory. According to a quantitative evaluation on the Middlebury benchmark, a patch-based Corresponding author. D.-S. Huang et al. (Eds.): ICIC 2013, LNCS 7995, pp , Springer-Verlag Berlin Heidelberg 2013

2 Efficient 3D Reconstruction for Urban Scenes 547 MVS method [3] outperforms all others for most of the datasets. Hence, we make use of this PMVS algorithm to recover point clouds. Urban landscapes exhibit a high degree of self-similarity and redundancy [1]. This characteristic of urban scenes is not a chance occurrence, but is demonstrated universally across countries and cultures. Such large scale repetitions arise from aesthetics, build-ability, manufacturing ease, etc. Also urban buildings are mostly comprised of flat or near-planar faces because of functional requirements and constraints. Consequently, the presence of such characteristics suggests opportunities for simplifying the reconstruction task. (a) (b) (c) Fig. 1. Images of urban scenes with texture-poor but highly structured surfaces are increasingly ubiquitous on the Internet In this paper, we propose a 3D urban scene reconstruction method based on exploration of properties of architectural scenes. Our method is inspired by recent works of [2], which assemble an architectural model over a 3D point cloud by interactive operations. The key idea is to replace the smoothness prior used in traditional methods with priors that are more appropriate for urban buildings. Our approach operates as follows. Given a set of calibrated photographs, we first generate point clouds using an existing MVS algorithm, and then reconstruct the sub-structures that often regularly repeat in urban buildings. At this step, regularity and self-symmetry properties of urban buildings [1] are explored. Finally, we recover the entire architectural models through an automatic growing algorithm of the sub-structures in dominant directions. Our method, exploits properties of urban scenes, offers the following advantages. 1) It is remarkably robust to lack of texture, and able to model flat painted walls shown in Fig. 1(b). 2) It can work with poor quality point clouds generated from photographs due to occlusion, as shown in Fig. 1(c). 3) It produces remarkably clean and simple models as outputs. This paper is arranged as follows. In section 2 we review the PMVS algorithm and preprocess the point clouds generated by PMVS. In section 3 we reconstruct the substructures that often regularly repeat in urban buildings and in section 4 we recover the entire architectural models through an automatic growing algorithm of the substructures. Experimental results are presented in section 5, followed by conclusions in section 6.

3 548 W. Fu et al. 2 MVS Preprocessing PMVS algorithm presented in [3] is implemented as a match, expand, and filter procedure, starting from a sparse set of matched keypoints, and repeatedly expanding these before using visibility constraints to filter away false matches. Given a set of calibrated photographs, the first step of our method is to use the PMVS algorithm [3] to generate a set of oriented 3D points (positions and normal). We retain only highconfidence points in textured areas. Associated with each point P i, several attributes can be extracted, including the 3D location, a surface normal N i, a set of visible images V i, and a photometric consistency score (normalized cross correlation) C(P i ). 2.1 Outliers Removing Although the PMVS algorithm proposed in [3] enforces local photometric consistency and global visibility constraints to remove outliers, the point clouds generated by PMVS still have sparse outliers which will corrupt the estimation of local point cloud attributes such as surface normal or curvature changes. We solve these irregularities by performing a statistical analysis on each point s neighborhood, and trimming those which do not meet a certain criteria. Our sparse outlier removal scheme is based on the distribution of distances from a point to its neighbors in the PMVS point clouds. For each point, we compute the mean distance from it to all its neighbors. By assuming that the resulted distribution is Gaussian with a mean and a standard deviation, all points whose mean distances are outside an interval defined by the global distances mean and standard deviation can be considered as outliers and trimmed from the point clouds. An example of the outliers-removal is shown in Fig. 2. Fig. 2(a) shows the original point cloud obtained by PMVS, while Fig. 2(b) shows the result of outliers-removal. We set the number of nearest neighbors as 50 for mean distance estimation. (a) (b) (c) Fig. 2. (a) The original point cloud obtained by PMVS. (b) The resultant one after outliers removing. (c) The filtered point cloud with normals.

4 Efficient 3D Reconstruction for Urban Scenes Estimating Normals and Curvature The PMVS algorithm in [3] optimizes the normal of each patch by simply minimizing the photometric discrepancy score, and constraining the number of degrees of freedom to two (yaw and pitch). However, the normal optimized in PMVS is not good enough for urban scene, because of the large distance between building surface and the optical center of the camera, and the initialization of PMVS algorithm works poorly for the large scale urban buildings. Though many different normal estimation methods exist, the one we used in this paper is one of the simplest, and is formulated as follows. The problem of determining the normal to a point on the surface is approximated by the problem of estimating the normal of a plane tangent to the surface, which in turn reduces to an analysis of the eigenvectors and eigenvalues of a covariance matrix created from the nearest neighbors of the query point. More specifically, for each point P i, we assemble the covariance matrix C as follows: 1 C = P P P P C V =λ V j k T ( i )( i ), j j j, { 0,1,2} k i = 1 (1) where k is the number of point neighbors considered in the neighborhood of P i, P represents the 3D centroid of the nearest neighbors, λ j is the j-th eigenvalue of the covariance matrix, and V j is the j-th eigenvector. The point cloud with normals is shown in Fig. 2(c). The surface curvature change is estimated from the eigenvalues as: λ0, where λ 0 <λ 1 <λ2 λ +λ +λ (2) 2.3 Extracting Dominant Directions We could require the dominant axes of buildings in an automatic growing algorithm of the sub-structures. Hence, we employ a simple greedy algorithm using the normal estimates N i recovered above (See [7, 8, 9] for similar ideas). We compute a histogram of normal directions over a unit sphere, subdivided into 1000 bins. We then set the first dominant axis d 1 to the average of the normals within the largest bin. Next, we find the largest bin which is in the range 80 to 100 degrees away from d 1 and set the second dominant axis d 2 dominant axis d 3 to the average normal within that bin. Finally, the third is computed in the same way. We allow for some deviation from orthogonality to compensate for possible errors in normal estimation and to handle urban building that itself is not composed of exactly orthogonal planes.

5 550 W. Fu et al. 3 Sub-structure Reconstruction While in recent years many techniques have been developed to detect repeated parts in models [5, 11, 12, 13] and regularity directly on 3D geometry [6], in this paper, we attempt to learn the repetitions pattern directly from high level user guidance for the existence of occlusion or missing data. Due to the generally poor quality of generated data, we argue that some level of user interaction is required to understand the semantics of the building model. Hence, at the beginning of this section, we select the points that imply the sub-structures by our interactive tool. Then, a region growing algorithm is utilized for simplifying the reconstruction of the sub-structures. 3.1 Region Growing Algorithm The purpose of the region growing algorithm is to merge the points that are close enough in terms of the smoothness constraint. Thereby, the output of this algorithm is a set of clusters, and each cluster has a set of points that are considered to be a part of the same smooth surface. We first sort the points by their curvature value. It is necessary because the region begins its growth from the point that has the minimum curvature value. The reason for this is that the point with the minimum curvature is located in the flattest area. Then, as long as there are unlabeled points in the cloud, we pick up the point with minimum curvature value and start the growth of the region. First, the picked point is added to a set called seeds. Then, for every seed point we find its neighboring points, and test the angle between the normal of the current seed point and its neighbors normals. If the angle is less than a threshold θ, the current point is added to the current region. After that every neighbor is tested for the curvature value. If the curvature is less than a threshold δ, then this point is added to the seeds. Finally, current seed is removed from the seeds, and if the seed set becomes empty which means that a certain region has grown, we repeat the process from the beginning. The overall pseudo code for this growing algorithm is given in Table Reconstruction After the region growing algorithm, we employ an efficient RANSAC method for point-cloud shape detection in each region of the sub-structures. The RANSAC algorithm we used is presented in [9]. It decomposed the point cloud into a concise, hybrid structure of inherent shapes and a set of remaining points. The method can detect planes, spheres, cylinders, cones and tori. However, because of functional requirements and constraints, urban buildings are mostly comprised of flat or near-planar faces. Consequently, the RANSAC method with plane model detection is usually sufficient. Also, the region with too little points can be rejected or determined by high level user guidance which RANSAC model can be adopted. Eventually, the substructure is generated by enforcing alignment among the RANSAC models. This step will resolve the inconsistencies we wish to repair, e.g., model intersections, small gaps, and other forms of misalignments.

6 Efficient 3D Reconstruction for Urban Scenes 551 Table 1. The pseudo code for the region growing algorithm Input: Point cloud = { P}, Point normals = { N}, Point curvatures = { C}, Neighbor finding function Ω ( ), Curvature threshold δ, Angle threshold θ. Output: Region list { R }. R φ, Available points list { A} { P1,..., Pn } While { A } is not empty do Current region{ R c } φ, Current seeds { S c } φ Point with minimum curvature in { A} Pmin { Sc} { Sc} Pmin,{ Rc} { Rc} Pmin,{ A} { A} \ Pmin for i = 0 to size ({ Sc} ) do Find nearest neighbors of current seed point { Bc} Ω ( Sc{ i} ) for j = 0 to size ({ Bc }) do Current neighbor point Pj Bc{ j} 1 If { A } contains Pj and cos ( N{ Sc{} i }, N{ Bc{} j }) < θ then { Rc} { Rc} Pj { A} { A} \ Pj If C{ P j } < δ then { S } { S } P end if end if end for end for c c j Add current region to global region list { R} { R} { R c } end while 4 Automatic Continuation Our sub-structure automatic continuation algorithm is inspired by recent works in [2]. However, in [2], some essential interactive operations are utilized for creating complex structures through the propagation of a repeated structure. In our case, the automatic continuation of sub-structure is constrained along the dominant directions computed in Section 2, which can minimize the required user interaction. Also, in [2], an optimization which balances between the data-fitting and contextual forces is adopted. We continue this trend and present our data-fitting term and contextual term below. Given a sub-structure S and the point cloud, the datafitting term, also known as the fitting error is computed by measuring the one-sided Euclidean distance from the sub-structure to points [14]. The contextual sub-structure S required to define the contextual force is a previously positioned sub-structure in a reconstruction sequence. The contextual term is defined as the sum of two terms: the interval term and the alignment term. The interval term I( S, S ) measures how well

7 552 W. Fu et al. the interval length between S and S agrees with the expected interval length due to regularity constraints. The alignment term A( SS, ) measures how well corresponding edges of S and S align. These two terms are normalized independently and weighted equally to form the contextual term. The automatic growing algorithm is achieved by finding an optimal linear transformation T, consisting of translation and scaling of the sub-structure S, to minim- * ize a weighted sum of data-fitting and contextual terms, { ω ( ( ) ) ( ω) ( ( ) )} * T arg min D T S, P 1 C T S, S T = + (3) where ω is a weight which balances between the data-fitting and contextual forces. The choice of the weight ω should correlate with an assessment of data quality and contextual regularity in the urban building model to be reconstructed. In our method, the default setting of ω is 0.5. Nevertheless, the user can manually adjust it according to data quality and contextual regularity perceived by them. 5 Experimental Results The proposed reconstruction method was tested on both a publicly available data set: Hall [4] and our own images of Building F in Tongji University (taken by cell phone Fig. 3. Top left: a target image. Top middle: the point cloud of the sub-structure. Top right: sub-structure reconstruction. Middle: sub-structure growing among the dominant direction. Bottom: the reconstructed model with texture mapping.

8 Efficient 3D Reconstruction for Urban Scenes 553 Fig. 4. Top left: a target image. Top right: the point cloud of the sub-structure. Bottom: the reconstructed model with texture mapping. camera). The Hall dataset consists of 61 images which are used for PMVS algorithm to generate point cloud. This results in a total of points. In our case we use only the high-confidence points, thus, we have points after the MVS preprocessing step. We set θ = and δ = 1.0 for the thresholds used in region growing algorithm. The interval term used in automatic continuation is initialized as the length of the sub-structure s projection among the dominant direction. Our own

9 554 W. Fu et al. Building F dataset consists of 15 images taken by cell phone camera, and some of the images have occlusion as shown in top left of Fig. 4. The reconstruction results are shown in Fig. 3 and Fig. 4, which illustrate the convenience and robustness of our proposed method. 6 Conclusions and the Future Work In this paper, we have demonstrated a Multi-View stereopsis based 3D urban scene reconstruction method that utilizes the large scale repetitions and self-similarities of urban buildings. Our method is remarkably robust to buildings lack of texture, and also able to work with poor quality point clouds by exploiting properties of urban scenes. In future work, we plan to incorporate any parametric primitive that can be detected using standard RANSAC in our sub-structure reconstruction step. This may lead to a better model and exhibit the diversity of modern urban scenes. Acknowledgement. This work is supported by the Fundamental Research Funds for the Central Universities under grant no , the Natural Science Foundation of China under grant no , and the Innovation Program of Shanghai Municipal Education Commission under grant no. 12ZZ029. References 1. Zheng, Q., Sharf, A., Wan, G., Li, Y., Mitra, N.J., Cohen-Or, D., Chen, B.: Non-local Scan Consolidation for 3D Urban Scenes. ACM Transactions on Graphics 29, 94 (2010) 2. Nan, L., Sharf, A., Zhang, H., Cohen-Or, D., Chen, B.: SmartBoxes for Interactive Urban Reconstruction. ACM Transactions on Graphics 29, 93 (2010) 3. Furukawa, Y., Ponce, J.: Accurate, dense, and robust Multi-View stereopsis. In: CVPR, pp. 1 8 (2007) 4. Furukawa, Y., Curless, B., Seitz, S.M., Szeliski, R.: Manhattan-world Stereo. In: CVPR, pp (2009) 5. Liu, J., Musialski, P., Wonka, P., Ye, J.: Tensor Completion for Estimating Missing Values in Visual Data. In: ICCV, pp (2009) 6. Pauly, M., Mitra, N.J., Wallner, J., Pottmann, H., Guibas, L.J.: Discovering Structural Regularity in 3D Geometry. ACM Transactions on Graphics 27, 43 (2008) 7. Coorg, S., Teller, S.: Extracting Textured Vertical Facades from Controlled Close-Range Imagery. In: CVPR, pp (1999) 8. Pollefeys, M., Nister, D., Frahm, J.M., Akbarzadeh, A., Mordohai, P., Clipp, B., Engels, C., Gallup, D., Kim, S.J., Merrell, P.: Detailed Real-Time Urban 3D Reconstruction from Video. In: IJCV 2008, pp (2008) 9. Werner, T., Zisserman, A.: New Techniques for Automated Architectural Reconstruction from Photographs. In: Heyden, A., Sparr, G., Nielsen, M., Johansen, P. (eds.) ECCV 2002, Part II. LNCS, vol. 2351, pp Springer, Heidelberg (2002) 10. Schnabel, R., Wahl, R., Klein, R.: Efficient RANSAC for Point-Cloud Shape Detection. Computer Graphics Forum 26, (2007)

10 Efficient 3D Reconstruction for Urban Scenes Mitra, N.J., Guibas, L.J., Pauly, M.: Partial And Approximate Symmetry Detection for 3D Geometry. ACM Transactions on Graphics 25, (2006) 12. Korah, T., Rasmussen, C.: Spatiotemporal Inpainting for Recovering Texture Maps of Occluded Building Facades. IEEE Transactions on Image Processing 16, (2007) 13. Pauly, M., Mitra, N.J., Wallner, J., Pottmann, H., Guibas, L.J.: Discovering Structural Regularity in 3D Geometry. ACM Transactions on Graphics 27, 43 (2008) 14. Nan, L., Xie, K., Sharf, A.: A Search Classify Approach for Cluttered Indoor Scene Understanding. ACM Transactions on Graphics 31, 137 (2012)

SIMPLE ROOM SHAPE MODELING WITH SPARSE 3D POINT INFORMATION USING PHOTOGRAMMETRY AND APPLICATION SOFTWARE

SIMPLE ROOM SHAPE MODELING WITH SPARSE 3D POINT INFORMATION USING PHOTOGRAMMETRY AND APPLICATION SOFTWARE SIMPLE ROOM SHAPE MODELING WITH SPARSE 3D POINT INFORMATION USING PHOTOGRAMMETRY AND APPLICATION SOFTWARE S. Hirose R&D Center, TOPCON CORPORATION, 75-1, Hasunuma-cho, Itabashi-ku, Tokyo, Japan Commission

More information

Reconstruction of Polygonal Faces from Large-Scale Point-Clouds of Engineering Plants

Reconstruction of Polygonal Faces from Large-Scale Point-Clouds of Engineering Plants 1 Reconstruction of Polygonal Faces from Large-Scale Point-Clouds of Engineering Plants Hiroshi Masuda 1, Takeru Niwa 2, Ichiro Tanaka 3 and Ryo Matsuoka 4 1 The University of Electro-Communications, h.masuda@euc.ac.jp

More information

Multi-view Stereo. Ivo Boyadzhiev CS7670: September 13, 2011

Multi-view Stereo. Ivo Boyadzhiev CS7670: September 13, 2011 Multi-view Stereo Ivo Boyadzhiev CS7670: September 13, 2011 What is stereo vision? Generic problem formulation: given several images of the same object or scene, compute a representation of its 3D shape

More information

A DATA DRIVEN METHOD FOR FLAT ROOF BUILDING RECONSTRUCTION FROM LiDAR POINT CLOUDS

A DATA DRIVEN METHOD FOR FLAT ROOF BUILDING RECONSTRUCTION FROM LiDAR POINT CLOUDS A DATA DRIVEN METHOD FOR FLAT ROOF BUILDING RECONSTRUCTION FROM LiDAR POINT CLOUDS A. Mahphood, H. Arefi *, School of Surveying and Geospatial Engineering, College of Engineering, University of Tehran,

More information

Dense 3D Reconstruction. Christiano Gava

Dense 3D Reconstruction. Christiano Gava Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Today: dense 3D reconstruction The matching problem

More information

DETECTION AND ROBUST ESTIMATION OF CYLINDER FEATURES IN POINT CLOUDS INTRODUCTION

DETECTION AND ROBUST ESTIMATION OF CYLINDER FEATURES IN POINT CLOUDS INTRODUCTION DETECTION AND ROBUST ESTIMATION OF CYLINDER FEATURES IN POINT CLOUDS Yun-Ting Su James Bethel Geomatics Engineering School of Civil Engineering Purdue University 550 Stadium Mall Drive, West Lafayette,

More information

Dense 3D Reconstruction. Christiano Gava

Dense 3D Reconstruction. Christiano Gava Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Wide baseline matching (SIFT) Today: dense 3D reconstruction

More information

BUILDING POINT GROUPING USING VIEW-GEOMETRY RELATIONS INTRODUCTION

BUILDING POINT GROUPING USING VIEW-GEOMETRY RELATIONS INTRODUCTION BUILDING POINT GROUPING USING VIEW-GEOMETRY RELATIONS I-Chieh Lee 1, Shaojun He 1, Po-Lun Lai 2, Alper Yilmaz 2 1 Mapping and GIS Laboratory 2 Photogrammetric Computer Vision Laboratory Dept. of Civil

More information

Multi-view stereo. Many slides adapted from S. Seitz

Multi-view stereo. Many slides adapted from S. Seitz Multi-view stereo Many slides adapted from S. Seitz Beyond two-view stereo The third eye can be used for verification Multiple-baseline stereo Pick a reference image, and slide the corresponding window

More information

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Structured Light II Johannes Köhler Johannes.koehler@dfki.de Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Introduction Previous lecture: Structured Light I Active Scanning Camera/emitter

More information

Large Scale 3D Reconstruction by Structure from Motion

Large Scale 3D Reconstruction by Structure from Motion Large Scale 3D Reconstruction by Structure from Motion Devin Guillory Ziang Xie CS 331B 7 October 2013 Overview Rome wasn t built in a day Overview of SfM Building Rome in a Day Building Rome on a Cloudless

More information

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University. 3D Computer Vision Structured Light II Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de 1 Introduction

More information

3D Computer Vision. Dense 3D Reconstruction II. Prof. Didier Stricker. Christiano Gava

3D Computer Vision. Dense 3D Reconstruction II. Prof. Didier Stricker. Christiano Gava 3D Computer Vision Dense 3D Reconstruction II Prof. Didier Stricker Christiano Gava Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de

More information

2D-3D Fusion for Layer Decomposition of Urban Facades

2D-3D Fusion for Layer Decomposition of Urban Facades 2D-3D Fusion for Layer Decomposition of Urban Facades Yangyan Li Qian Zheng Andrei Sharf, Daniel Cohen-Or Baoquan Chen Niloy J. Mitra SIAT, China Ben-Gurion Univ. Tel Aviv Univ. UCL Abstract We present

More information

Manhattan-world Stereo and Surface Reconstruction

Manhattan-world Stereo and Surface Reconstruction Manhattan-world Stereo and Surface Reconstruction Kuan-Ting Yu CSAIL MIT Cambridge, USA peterkty@csail.mit.edu Abstract Depth estimation from 2D images has been extensively studied in the computer vision

More information

Urban Scene Segmentation, Recognition and Remodeling. Part III. Jinglu Wang 11/24/2016 ACCV 2016 TUTORIAL

Urban Scene Segmentation, Recognition and Remodeling. Part III. Jinglu Wang 11/24/2016 ACCV 2016 TUTORIAL Part III Jinglu Wang Urban Scene Segmentation, Recognition and Remodeling 102 Outline Introduction Related work Approaches Conclusion and future work o o - - ) 11/7/16 103 Introduction Motivation Motivation

More information

A System of Image Matching and 3D Reconstruction

A System of Image Matching and 3D Reconstruction A System of Image Matching and 3D Reconstruction CS231A Project Report 1. Introduction Xianfeng Rui Given thousands of unordered images of photos with a variety of scenes in your gallery, you will find

More information

Robotics Programming Laboratory

Robotics Programming Laboratory Chair of Software Engineering Robotics Programming Laboratory Bertrand Meyer Jiwon Shin Lecture 8: Robot Perception Perception http://pascallin.ecs.soton.ac.uk/challenges/voc/databases.html#caltech car

More information

Image Based Reconstruction II

Image Based Reconstruction II Image Based Reconstruction II Qixing Huang Feb. 2 th 2017 Slide Credit: Yasutaka Furukawa Image-Based Geometry Reconstruction Pipeline Last Lecture: Multi-View SFM Multi-View SFM This Lecture: Multi-View

More information

Step-by-Step Model Buidling

Step-by-Step Model Buidling Step-by-Step Model Buidling Review Feature selection Feature selection Feature correspondence Camera Calibration Euclidean Reconstruction Landing Augmented Reality Vision Based Control Sparse Structure

More information

Miniature faking. In close-up photo, the depth of field is limited.

Miniature faking. In close-up photo, the depth of field is limited. Miniature faking In close-up photo, the depth of field is limited. http://en.wikipedia.org/wiki/file:jodhpur_tilt_shift.jpg Miniature faking Miniature faking http://en.wikipedia.org/wiki/file:oregon_state_beavers_tilt-shift_miniature_greg_keene.jpg

More information

arxiv: v1 [cs.cv] 28 Sep 2018

arxiv: v1 [cs.cv] 28 Sep 2018 Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,

More information

A Systems View of Large- Scale 3D Reconstruction

A Systems View of Large- Scale 3D Reconstruction Lecture 23: A Systems View of Large- Scale 3D Reconstruction Visual Computing Systems Goals and motivation Construct a detailed 3D model of the world from unstructured photographs (e.g., Flickr, Facebook)

More information

Feature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies

Feature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies Feature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies M. Lourakis, S. Tzurbakis, A. Argyros, S. Orphanoudakis Computer Vision and Robotics Lab (CVRL) Institute of

More information

Supplementary Material : Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision

Supplementary Material : Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision Supplementary Material : Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision Due to space limitation in the main paper, we present additional experimental results in this supplementary

More information

FAST REGISTRATION OF TERRESTRIAL LIDAR POINT CLOUD AND SEQUENCE IMAGES

FAST REGISTRATION OF TERRESTRIAL LIDAR POINT CLOUD AND SEQUENCE IMAGES FAST REGISTRATION OF TERRESTRIAL LIDAR POINT CLOUD AND SEQUENCE IMAGES Jie Shao a, Wuming Zhang a, Yaqiao Zhu b, Aojie Shen a a State Key Laboratory of Remote Sensing Science, Institute of Remote Sensing

More information

Local features: detection and description. Local invariant features

Local features: detection and description. Local invariant features Local features: detection and description Local invariant features Detection of interest points Harris corner detection Scale invariant blob detection: LoG Description of local patches SIFT : Histograms

More information

EECS 442 Computer vision. Stereo systems. Stereo vision Rectification Correspondence problem Active stereo vision systems

EECS 442 Computer vision. Stereo systems. Stereo vision Rectification Correspondence problem Active stereo vision systems EECS 442 Computer vision Stereo systems Stereo vision Rectification Correspondence problem Active stereo vision systems Reading: [HZ] Chapter: 11 [FP] Chapter: 11 Stereo vision P p p O 1 O 2 Goal: estimate

More information

Photo Tourism: Exploring Photo Collections in 3D

Photo Tourism: Exploring Photo Collections in 3D Photo Tourism: Exploring Photo Collections in 3D SIGGRAPH 2006 Noah Snavely Steven M. Seitz University of Washington Richard Szeliski Microsoft Research 2006 2006 Noah Snavely Noah Snavely Reproduced with

More information

Learning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009

Learning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009 Learning and Inferring Depth from Monocular Images Jiyan Pan April 1, 2009 Traditional ways of inferring depth Binocular disparity Structure from motion Defocus Given a single monocular image, how to infer

More information

Model-based segmentation and recognition from range data

Model-based segmentation and recognition from range data Model-based segmentation and recognition from range data Jan Boehm Institute for Photogrammetry Universität Stuttgart Germany Keywords: range image, segmentation, object recognition, CAD ABSTRACT This

More information

The SIFT (Scale Invariant Feature

The SIFT (Scale Invariant Feature The SIFT (Scale Invariant Feature Transform) Detector and Descriptor developed by David Lowe University of British Columbia Initial paper ICCV 1999 Newer journal paper IJCV 2004 Review: Matt Brown s Canonical

More information

Robot localization method based on visual features and their geometric relationship

Robot localization method based on visual features and their geometric relationship , pp.46-50 http://dx.doi.org/10.14257/astl.2015.85.11 Robot localization method based on visual features and their geometric relationship Sangyun Lee 1, Changkyung Eem 2, and Hyunki Hong 3 1 Department

More information

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching Stereo Matching Fundamental matrix Let p be a point in left image, p in right image l l Epipolar relation p maps to epipolar line l p maps to epipolar line l p p Epipolar mapping described by a 3x3 matrix

More information

What is Computer Vision?

What is Computer Vision? Perceptual Grouping in Computer Vision Gérard Medioni University of Southern California What is Computer Vision? Computer Vision Attempt to emulate Human Visual System Perceive visual stimuli with cameras

More information

Algorithm research of 3D point cloud registration based on iterative closest point 1

Algorithm research of 3D point cloud registration based on iterative closest point 1 Acta Technica 62, No. 3B/2017, 189 196 c 2017 Institute of Thermomechanics CAS, v.v.i. Algorithm research of 3D point cloud registration based on iterative closest point 1 Qian Gao 2, Yujian Wang 2,3,

More information

3D object recognition used by team robotto

3D object recognition used by team robotto 3D object recognition used by team robotto Workshop Juliane Hoebel February 1, 2016 Faculty of Computer Science, Otto-von-Guericke University Magdeburg Content 1. Introduction 2. Depth sensor 3. 3D object

More information

Automatic Pipeline Generation by the Sequential Segmentation and Skelton Construction of Point Cloud

Automatic Pipeline Generation by the Sequential Segmentation and Skelton Construction of Point Cloud , pp.43-47 http://dx.doi.org/10.14257/astl.2014.67.11 Automatic Pipeline Generation by the Sequential Segmentation and Skelton Construction of Point Cloud Ashok Kumar Patil, Seong Sill Park, Pavitra Holi,

More information

Lecture 10: Multi view geometry

Lecture 10: Multi view geometry Lecture 10: Multi view geometry Professor Fei Fei Li Stanford Vision Lab 1 What we will learn today? Stereo vision Correspondence problem (Problem Set 2 (Q3)) Active stereo vision systems Structure from

More information

3D model search and pose estimation from single images using VIP features

3D model search and pose estimation from single images using VIP features 3D model search and pose estimation from single images using VIP features Changchang Wu 2, Friedrich Fraundorfer 1, 1 Department of Computer Science ETH Zurich, Switzerland {fraundorfer, marc.pollefeys}@inf.ethz.ch

More information

Translation Symmetry Detection: A Repetitive Pattern Analysis Approach

Translation Symmetry Detection: A Repetitive Pattern Analysis Approach 2013 IEEE Conference on Computer Vision and Pattern Recognition Workshops Translation Symmetry Detection: A Repetitive Pattern Analysis Approach Yunliang Cai and George Baciu GAMA Lab, Department of Computing

More information

Local invariant features

Local invariant features Local invariant features Tuesday, Oct 28 Kristen Grauman UT-Austin Today Some more Pset 2 results Pset 2 returned, pick up solutions Pset 3 is posted, due 11/11 Local invariant features Detection of interest

More information

Motion Tracking and Event Understanding in Video Sequences

Motion Tracking and Event Understanding in Video Sequences Motion Tracking and Event Understanding in Video Sequences Isaac Cohen Elaine Kang, Jinman Kang Institute for Robotics and Intelligent Systems University of Southern California Los Angeles, CA Objectives!

More information

Chaplin, Modern Times, 1936

Chaplin, Modern Times, 1936 Chaplin, Modern Times, 1936 [A Bucket of Water and a Glass Matte: Special Effects in Modern Times; bonus feature on The Criterion Collection set] Multi-view geometry problems Structure: Given projections

More information

Static Scene Reconstruction

Static Scene Reconstruction GPU supported Real-Time Scene Reconstruction with a Single Camera Jan-Michael Frahm, 3D Computer Vision group, University of North Carolina at Chapel Hill Static Scene Reconstruction 1 Capture on campus

More information

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching Stereo Matching Fundamental matrix Let p be a point in left image, p in right image l l Epipolar relation p maps to epipolar line l p maps to epipolar line l p p Epipolar mapping described by a 3x3 matrix

More information

Simultaneous Vanishing Point Detection and Camera Calibration from Single Images

Simultaneous Vanishing Point Detection and Camera Calibration from Single Images Simultaneous Vanishing Point Detection and Camera Calibration from Single Images Bo Li, Kun Peng, Xianghua Ying, and Hongbin Zha The Key Lab of Machine Perception (Ministry of Education), Peking University,

More information

BIL Computer Vision Apr 16, 2014

BIL Computer Vision Apr 16, 2014 BIL 719 - Computer Vision Apr 16, 2014 Binocular Stereo (cont d.), Structure from Motion Aykut Erdem Dept. of Computer Engineering Hacettepe University Slide credit: S. Lazebnik Basic stereo matching algorithm

More information

AUTOMATED 3D MODELING OF URBAN ENVIRONMENTS

AUTOMATED 3D MODELING OF URBAN ENVIRONMENTS AUTOMATED 3D MODELING OF URBAN ENVIRONMENTS Ioannis Stamos Department of Computer Science Hunter College, City University of New York 695 Park Avenue, New York NY 10065 istamos@hunter.cuny.edu http://www.cs.hunter.cuny.edu/

More information

Graph Matching Iris Image Blocks with Local Binary Pattern

Graph Matching Iris Image Blocks with Local Binary Pattern Graph Matching Iris Image Blocs with Local Binary Pattern Zhenan Sun, Tieniu Tan, and Xianchao Qiu Center for Biometrics and Security Research, National Laboratory of Pattern Recognition, Institute of

More information

Improving PMVS Algorithm for 3D Scene Reconstruction from Sparse Stereo Pairs

Improving PMVS Algorithm for 3D Scene Reconstruction from Sparse Stereo Pairs Improving PMVS Algorithm for 3D Scene Reconstruction from Sparse Stereo Pairs Bo Li 1,Y.V.Venkatesh 2, Ashraf Kassim 3,andYijuanLu 1 1 Department of Computer Science, Texas State University, San Marcos,

More information

SmartBoxes for Interactive Urban Reconstruction

SmartBoxes for Interactive Urban Reconstruction SmartBoxes for Interactive Urban Reconstruction Liangliang Nan 1 Andrei Sharf 1 Hao Zhang 2 Daniel Cohen-Or 3 Baoquan Chen 1 1 Shenzhen Institutes of Advanced Technology (SIAT), China 2 Simon Fraser Univ.

More information

Feature Based Registration - Image Alignment

Feature Based Registration - Image Alignment Feature Based Registration - Image Alignment Image Registration Image registration is the process of estimating an optimal transformation between two or more images. Many slides from Alexei Efros http://graphics.cs.cmu.edu/courses/15-463/2007_fall/463.html

More information

Merging the Unmatchable: Stitching Visually Disconnected SfM Models

Merging the Unmatchable: Stitching Visually Disconnected SfM Models This is a preprint for the paper accepted for publication in ICCV 2015. c 2015 IEEE Merging the Unmatchable: Stitching Visually Disconnected SfM Models Andrea Cohen, Torsten Sattler, Marc Pollefeys Department

More information

EE795: Computer Vision and Intelligent Systems

EE795: Computer Vision and Intelligent Systems EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 14 130307 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Stereo Dense Motion Estimation Translational

More information

Viewpoint Invariant Features from Single Images Using 3D Geometry

Viewpoint Invariant Features from Single Images Using 3D Geometry Viewpoint Invariant Features from Single Images Using 3D Geometry Yanpeng Cao and John McDonald Department of Computer Science National University of Ireland, Maynooth, Ireland {y.cao,johnmcd}@cs.nuim.ie

More information

Lecture 10: Multi-view geometry

Lecture 10: Multi-view geometry Lecture 10: Multi-view geometry Professor Stanford Vision Lab 1 What we will learn today? Review for stereo vision Correspondence problem (Problem Set 2 (Q3)) Active stereo vision systems Structure from

More information

SOME stereo image-matching methods require a user-selected

SOME stereo image-matching methods require a user-selected IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, VOL. 3, NO. 2, APRIL 2006 207 Seed Point Selection Method for Triangle Constrained Image Matching Propagation Qing Zhu, Bo Wu, and Zhi-Xiang Xu Abstract In order

More information

3D Reconstruction Using an n-layer Heightmap

3D Reconstruction Using an n-layer Heightmap 3D Reconstruction Using an n-layer Heightmap David Gallup 1, Marc Pollefeys 2, and Jan-Michael Frahm 1 1 Department of Computer Science, University of North Carolina {gallup,jmf}@cs.unc.edu 2 Department

More information

Layered Scene Decomposition via the Occlusion-CRF Supplementary material

Layered Scene Decomposition via the Occlusion-CRF Supplementary material Layered Scene Decomposition via the Occlusion-CRF Supplementary material Chen Liu 1 Pushmeet Kohli 2 Yasutaka Furukawa 1 1 Washington University in St. Louis 2 Microsoft Research Redmond 1. Additional

More information

Image Parallax based Modeling of Depth-layer Architecture

Image Parallax based Modeling of Depth-layer Architecture Image Parallax based Modeling of Depth-layer Architecture Yong Hu, Bei Chu, Yue Qi State Key Laboratory of Virtual Reality Technology and Systems, Beihang University School of New Media Art and Design,

More information

CRF Based Point Cloud Segmentation Jonathan Nation

CRF Based Point Cloud Segmentation Jonathan Nation CRF Based Point Cloud Segmentation Jonathan Nation jsnation@stanford.edu 1. INTRODUCTION The goal of the project is to use the recently proposed fully connected conditional random field (CRF) model to

More information

Camera Registration in a 3D City Model. Min Ding CS294-6 Final Presentation Dec 13, 2006

Camera Registration in a 3D City Model. Min Ding CS294-6 Final Presentation Dec 13, 2006 Camera Registration in a 3D City Model Min Ding CS294-6 Final Presentation Dec 13, 2006 Goal: Reconstruct 3D city model usable for virtual walk- and fly-throughs Virtual reality Urban planning Simulation

More information

Marcel Worring Intelligent Sensory Information Systems

Marcel Worring Intelligent Sensory Information Systems Marcel Worring worring@science.uva.nl Intelligent Sensory Information Systems University of Amsterdam Information and Communication Technology archives of documentaries, film, or training material, video

More information

Geometry based Repetition Detection for Urban Scene

Geometry based Repetition Detection for Urban Scene Geometry based Repetition Detection for Urban Scene Changchang Wu University of Washington Jan Michael Frahm UNC Chapel Hill Marc Pollefeys ETH Zürich Related Work Sparse Feature Matching [Loy et al. 06,

More information

Features Points. Andrea Torsello DAIS Università Ca Foscari via Torino 155, Mestre (VE)

Features Points. Andrea Torsello DAIS Università Ca Foscari via Torino 155, Mestre (VE) Features Points Andrea Torsello DAIS Università Ca Foscari via Torino 155, 30172 Mestre (VE) Finding Corners Edge detectors perform poorly at corners. Corners provide repeatable points for matching, so

More information

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy BSB663 Image Processing Pinar Duygulu Slides are adapted from Selim Aksoy Image matching Image matching is a fundamental aspect of many problems in computer vision. Object or scene recognition Solving

More information

Specular 3D Object Tracking by View Generative Learning

Specular 3D Object Tracking by View Generative Learning Specular 3D Object Tracking by View Generative Learning Yukiko Shinozuka, Francois de Sorbier and Hideo Saito Keio University 3-14-1 Hiyoshi, Kohoku-ku 223-8522 Yokohama, Japan shinozuka@hvrl.ics.keio.ac.jp

More information

Local Feature Detectors

Local Feature Detectors Local Feature Detectors Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr Slides adapted from Cordelia Schmid and David Lowe, CVPR 2003 Tutorial, Matthew Brown,

More information

Multiview Reconstruction

Multiview Reconstruction Multiview Reconstruction Why More Than 2 Views? Baseline Too short low accuracy Too long matching becomes hard Why More Than 2 Views? Ambiguity with 2 views Camera 1 Camera 2 Camera 3 Trinocular Stereo

More information

Fast K-nearest neighbors searching algorithms for point clouds data of 3D scanning system 1

Fast K-nearest neighbors searching algorithms for point clouds data of 3D scanning system 1 Acta Technica 62 No. 3B/2017, 141 148 c 2017 Institute of Thermomechanics CAS, v.v.i. Fast K-nearest neighbors searching algorithms for point clouds data of 3D scanning system 1 Zhang Fan 2, 3, Tan Yuegang

More information

Continuous Multi-View Tracking using Tensor Voting

Continuous Multi-View Tracking using Tensor Voting Continuous Multi-View Tracking using Tensor Voting Jinman Kang, Isaac Cohen and Gerard Medioni Institute for Robotics and Intelligent Systems University of Southern California {jinmanka, icohen, medioni}@iris.usc.edu

More information

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Structured Light II Johannes Köhler Johannes.koehler@dfki.de Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Introduction Previous lecture: Structured Light I Active Scanning Camera/emitter

More information

EE795: Computer Vision and Intelligent Systems

EE795: Computer Vision and Intelligent Systems EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 12 130228 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Panoramas, Mosaics, Stitching Two View Geometry

More information

Specular Reflection Separation using Dark Channel Prior

Specular Reflection Separation using Dark Channel Prior 2013 IEEE Conference on Computer Vision and Pattern Recognition Specular Reflection Separation using Dark Channel Prior Hyeongwoo Kim KAIST hyeongwoo.kim@kaist.ac.kr Hailin Jin Adobe Research hljin@adobe.com

More information

Ransac Based Out-of-Core Point-Cloud Shape Detection for City-Modeling

Ransac Based Out-of-Core Point-Cloud Shape Detection for City-Modeling Ransac Based Out-of-Core Point-Cloud Shape Detection for City-Modeling Ruwen Schnabel, Roland Wahl, Reinhard Klein Institut für Informatik II, Universität Bonn Römerstr. 117 53117 Bonn {schnabel,wahl,rk}@cs.uni-bonn.de

More information

Manhattan-world Urban Reconstruction from Point Clouds

Manhattan-world Urban Reconstruction from Point Clouds Manhattan-world Urban Reconstruction from Point Clouds Minglei Li 1,2, Peter Wonka 1, and Liangliang Nan 1 1 Visual Computing Center, KAUST, Saudi Arabia {mingleili87, liangliang.nan, pwonka}@gmail.com

More information

Local features and image matching. Prof. Xin Yang HUST

Local features and image matching. Prof. Xin Yang HUST Local features and image matching Prof. Xin Yang HUST Last time RANSAC for robust geometric transformation estimation Translation, Affine, Homography Image warping Given a 2D transformation T and a source

More information

A Statistical Consistency Check for the Space Carving Algorithm.

A Statistical Consistency Check for the Space Carving Algorithm. A Statistical Consistency Check for the Space Carving Algorithm. A. Broadhurst and R. Cipolla Dept. of Engineering, Univ. of Cambridge, Cambridge, CB2 1PZ aeb29 cipolla @eng.cam.ac.uk Abstract This paper

More information

3D Reconstruction of a Hopkins Landmark

3D Reconstruction of a Hopkins Landmark 3D Reconstruction of a Hopkins Landmark Ayushi Sinha (461), Hau Sze (461), Diane Duros (361) Abstract - This paper outlines a method for 3D reconstruction from two images. Our procedure is based on known

More information

3D Perception. CS 4495 Computer Vision K. Hawkins. CS 4495 Computer Vision. 3D Perception. Kelsey Hawkins Robotics

3D Perception. CS 4495 Computer Vision K. Hawkins. CS 4495 Computer Vision. 3D Perception. Kelsey Hawkins Robotics CS 4495 Computer Vision Kelsey Hawkins Robotics Motivation What do animals, people, and robots want to do with vision? Detect and recognize objects/landmarks Find location of objects with respect to themselves

More information

REFINEMENT OF COLORED MOBILE MAPPING DATA USING INTENSITY IMAGES

REFINEMENT OF COLORED MOBILE MAPPING DATA USING INTENSITY IMAGES REFINEMENT OF COLORED MOBILE MAPPING DATA USING INTENSITY IMAGES T. Yamakawa a, K. Fukano a,r. Onodera a, H. Masuda a, * a Dept. of Mechanical Engineering and Intelligent Systems, The University of Electro-Communications,

More information

Instance-level recognition

Instance-level recognition Instance-level recognition 1) Local invariant features 2) Matching and recognition with local features 3) Efficient visual search 4) Very large scale indexing Matching of descriptors Matching and 3D reconstruction

More information

How to Compute the Pose of an Object without a Direct View?

How to Compute the Pose of an Object without a Direct View? How to Compute the Pose of an Object without a Direct View? Peter Sturm and Thomas Bonfort INRIA Rhône-Alpes, 38330 Montbonnot St Martin, France {Peter.Sturm, Thomas.Bonfort}@inrialpes.fr Abstract. We

More information

Stereo and Epipolar geometry

Stereo and Epipolar geometry Previously Image Primitives (feature points, lines, contours) Today: Stereo and Epipolar geometry How to match primitives between two (multiple) views) Goals: 3D reconstruction, recognition Jana Kosecka

More information

Previously. Part-based and local feature models for generic object recognition. Bag-of-words model 4/20/2011

Previously. Part-based and local feature models for generic object recognition. Bag-of-words model 4/20/2011 Previously Part-based and local feature models for generic object recognition Wed, April 20 UT-Austin Discriminative classifiers Boosting Nearest neighbors Support vector machines Useful for object recognition

More information

Registration of Dynamic Range Images

Registration of Dynamic Range Images Registration of Dynamic Range Images Tan-Chi Ho 1,2 Jung-Hong Chuang 1 Wen-Wei Lin 2 Song-Sun Lin 2 1 Department of Computer Science National Chiao-Tung University 2 Department of Applied Mathematics National

More information

Textured Mesh Surface Reconstruction of Large Buildings with Multi-View Stereo

Textured Mesh Surface Reconstruction of Large Buildings with Multi-View Stereo The Visual Computer manuscript No. (will be inserted by the editor) Textured Mesh Surface Reconstruction of Large Buildings with Multi-View Stereo Chen Zhu Wee Kheng Leow Received: date / Accepted: date

More information

Continuous Multi-Views Tracking using Tensor Voting

Continuous Multi-Views Tracking using Tensor Voting Continuous Multi-Views racking using ensor Voting Jinman Kang, Isaac Cohen and Gerard Medioni Institute for Robotics and Intelligent Systems University of Southern California Los Angeles, CA 90089-073.

More information

Homographies and RANSAC

Homographies and RANSAC Homographies and RANSAC Computer vision 6.869 Bill Freeman and Antonio Torralba March 30, 2011 Homographies and RANSAC Homographies RANSAC Building panoramas Phototourism 2 Depth-based ambiguity of position

More information

Similarity Image Retrieval System Using Hierarchical Classification

Similarity Image Retrieval System Using Hierarchical Classification Similarity Image Retrieval System Using Hierarchical Classification Experimental System on Mobile Internet with Cellular Phone Masahiro Tada 1, Toshikazu Kato 1, and Isao Shinohara 2 1 Department of Industrial

More information

A Desktop 3D Scanner Exploiting Rotation and Visual Rectification of Laser Profiles

A Desktop 3D Scanner Exploiting Rotation and Visual Rectification of Laser Profiles A Desktop 3D Scanner Exploiting Rotation and Visual Rectification of Laser Profiles Carlo Colombo, Dario Comanducci, and Alberto Del Bimbo Dipartimento di Sistemi ed Informatica Via S. Marta 3, I-5139

More information

CS4495/6495 Introduction to Computer Vision

CS4495/6495 Introduction to Computer Vision CS4495/6495 Introduction to Computer Vision 9C-L1 3D perception Some slides by Kelsey Hawkins Motivation Why do animals, people & robots need vision? To detect and recognize objects/landmarks Is that a

More information

K-Means Based Matching Algorithm for Multi-Resolution Feature Descriptors

K-Means Based Matching Algorithm for Multi-Resolution Feature Descriptors K-Means Based Matching Algorithm for Multi-Resolution Feature Descriptors Shao-Tzu Huang, Chen-Chien Hsu, Wei-Yen Wang International Science Index, Electrical and Computer Engineering waset.org/publication/0007607

More information

LOAM: LiDAR Odometry and Mapping in Real Time

LOAM: LiDAR Odometry and Mapping in Real Time LOAM: LiDAR Odometry and Mapping in Real Time Aayush Dwivedi (14006), Akshay Sharma (14062), Mandeep Singh (14363) Indian Institute of Technology Kanpur 1 Abstract This project deals with online simultaneous

More information

Multiple View Geometry

Multiple View Geometry Multiple View Geometry Martin Quinn with a lot of slides stolen from Steve Seitz and Jianbo Shi 15-463: Computational Photography Alexei Efros, CMU, Fall 2007 Our Goal The Plenoptic Function P(θ,φ,λ,t,V

More information

Instance-level recognition part 2

Instance-level recognition part 2 Visual Recognition and Machine Learning Summer School Paris 2011 Instance-level recognition part 2 Josef Sivic http://www.di.ens.fr/~josef INRIA, WILLOW, ENS/INRIA/CNRS UMR 8548 Laboratoire d Informatique,

More information

Combining Appearance and Topology for Wide

Combining Appearance and Topology for Wide Combining Appearance and Topology for Wide Baseline Matching Dennis Tell and Stefan Carlsson Presented by: Josh Wills Image Point Correspondences Critical foundation for many vision applications 3-D reconstruction,

More information

3D Photography: Stereo

3D Photography: Stereo 3D Photography: Stereo Marc Pollefeys, Torsten Sattler Spring 2016 http://www.cvg.ethz.ch/teaching/3dvision/ 3D Modeling with Depth Sensors Today s class Obtaining depth maps / range images unstructured

More information

Multi-View Stereo for Static and Dynamic Scenes

Multi-View Stereo for Static and Dynamic Scenes Multi-View Stereo for Static and Dynamic Scenes Wolfgang Burgard Jan 6, 2010 Main references Yasutaka Furukawa and Jean Ponce, Accurate, Dense and Robust Multi-View Stereopsis, 2007 C.L. Zitnick, S.B.

More information