URBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES

Similar documents
arxiv: v1 [cs.cv] 28 Sep 2018

Planar Structures from Line Correspondences in a Manhattan World

3D Spatial Layout Propagation in a Video Sequence

arxiv: v1 [cs.cv] 28 Sep 2018

Viewpoint Invariant Features from Single Images Using 3D Geometry

Flexible Calibration of a Portable Structured Light System through Surface Plane

3D Reconstruction from Two Views

Augmenting Reality, Naturally:

Stereo and Epipolar geometry

Dense 3D Reconstruction. Christiano Gava

3D Reconstruction of a Hopkins Landmark

Octree-Based Obstacle Representation and Registration for Real-Time

Visualization 2D-to-3D Photo Rendering for 3D Displays

Automatic Feature Extraction of Pose-measuring System Based on Geometric Invariants

Unit 3 Multiple View Geometry

Camera Calibration Using Line Correspondences

An Algorithm for Seamless Image Stitching and Its Application

LOCAL AND GLOBAL DESCRIPTORS FOR PLACE RECOGNITION IN ROBOTICS

3D Reconstruction from Scene Knowledge

Feature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies

Today. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography

Dense 3D Reconstruction. Christiano Gava

Two-view geometry Computer Vision Spring 2018, Lecture 10

Joint Vanishing Point Extraction and Tracking. 9. June 2015 CVPR 2015 Till Kroeger, Dengxin Dai, Luc Van Gool, Computer Vision ETH Zürich

An Image Based 3D Reconstruction System for Large Indoor Scenes

1 Projective Geometry

Robot localization method based on visual features and their geometric relationship

Homographies and RANSAC

Comments on Consistent Depth Maps Recovery from a Video Sequence

Segmentation and Tracking of Partial Planar Templates

calibrated coordinates Linear transformation pixel coordinates

Multiple View Geometry in Computer Vision Second Edition

International Journal for Research in Applied Science & Engineering Technology (IJRASET) A Review: 3D Image Reconstruction From Multiple Images

Monocular Vision Based Autonomous Navigation for Arbitrarily Shaped Urban Roads

CS 231A: Computer Vision (Winter 2018) Problem Set 2

But First: Multi-View Projective Geometry

Indoor Calibration using Segment Chains

A New Approach For 3D Image Reconstruction From Multiple Images

3D Line Segment Based Model Generation by RGB-D Camera for Camera Pose Estimation

Guang-Hui Wang et al.: Single View Based Measurement on Space Planes 375 parallelism of lines and planes. In [1, ], a keypoint-based approach to calcu

Image correspondences and structure from motion

arxiv: v1 [cs.cv] 18 Sep 2017

Multiple View Geometry

Detecting vanishing points by segment clustering on the projective plane for single-view photogrammetry

A Summary of Projective Geometry

Stereo Image Rectification for Simple Panoramic Image Generation

Camera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration

Estimation of Camera Pose with Respect to Terrestrial LiDAR Data

Camera Calibration from the Quasi-affine Invariance of Two Parallel Circles

Scene Modeling for a Single View

DEPTH AND GEOMETRY FROM A SINGLE 2D IMAGE USING TRIANGULATION

Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision

CSCI 5980/8980: Assignment #4. Fundamental Matrix

Scene Modeling for a Single View

EECS 442: Final Project

Accurate Motion Estimation and High-Precision 3D Reconstruction by Sensor Fusion

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

Image Based Rendering. D.A. Forsyth, with slides from John Hart

Absolute Scale Structure from Motion Using a Refractive Plate

Text Block Detection and Segmentation for Mobile Robot Vision System Applications

Epipolar Geometry Prof. D. Stricker. With slides from A. Zisserman, S. Lazebnik, Seitz

Pose Estimation from Circle or Parallel Lines in a Single Image

CS 231A Computer Vision (Winter 2014) Problem Set 3

Non-rigid body Object Tracking using Fuzzy Neural System based on Multiple ROIs and Adaptive Motion Frame Method

Epipolar Geometry and Stereo Vision

Line-Based Relative Pose Estimation

Static Scene Reconstruction

IMPACT OF SUBPIXEL PARADIGM ON DETERMINATION OF 3D POSITION FROM 2D IMAGE PAIR Lukas Sroba, Rudolf Ravas

Coplanar circles, quasi-affine invariance and calibration

E27 Computer Vision - Final Project: Creating Panoramas David Nahmias, Dan Spagnolo, Vincent Stigliani Professor Zucker Due 5/10/13

BUILDING POINT GROUPING USING VIEW-GEOMETRY RELATIONS INTRODUCTION

Plane rectification in real time on an Android device Final report

Stereo Epipolar Geometry for General Cameras. Sanja Fidler CSC420: Intro to Image Understanding 1 / 33

Srikumar Ramalingam. Review. 3D Reconstruction. Pose Estimation Revisited. School of Computing University of Utah

Geometry based Repetition Detection for Urban Scene

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION

Structure from Motion. Introduction to Computer Vision CSE 152 Lecture 10

Incremental Line-based 3D Reconstruction using Geometric Constraints

Utilizing Semantic Interpretation of Junctions for 3D-2D Pose Estimation

Srikumar Ramalingam. Review. 3D Reconstruction. Pose Estimation Revisited. School of Computing University of Utah

Simultaneous Pose and Correspondence Determination using Line Features

Homography-Based Change Detection for Space- Based Satellite Inspection

Video Mosaics for Virtual Environments, R. Szeliski. Review by: Christopher Rasmussen

Specular 3D Object Tracking by View Generative Learning

3D Visualization through Planar Pattern Based Augmented Reality

Visual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry

Thin Plate Spline Feature Point Matching for Organ Surfaces in Minimally Invasive Surgery Imaging

Camera Registration in a 3D City Model. Min Ding CS294-6 Final Presentation Dec 13, 2006

Computational Optical Imaging - Optique Numerique. -- Multiple View Geometry and Stereo --

Research on an Adaptive Terrain Reconstruction of Sequence Images in Deep Space Exploration

Overview. Augmented reality and applications Marker-based augmented reality. Camera model. Binary markers Textured planar markers

3D reconstruction how accurate can it be?

METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS

Stereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman

Visual Odometry for Non-Overlapping Views Using Second-Order Cone Programming

Parameter estimation. Christiano Gava Gabriele Bleser

High Level Landmark-Based Visual Navigation Using Unsupervised Geometric Constraints in Local Bundle Adjustment

Robot Vision: Projective Geometry

Building a Panorama. Matching features. Matching with Features. How do we build a panorama? Computational Photography, 6.882

Transcription:

URBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES An Undergraduate Research Scholars Thesis by RUI LIU Submitted to Honors and Undergraduate Research Texas A&M University in partial fulfillment of the requirements for the designation as an UNDERGRADUATE RESEARCH SCHOLAR Approved by Research Advisor: Dr. Dezhen Song May 2015 Major: Computer Engineering

TABLE OF CONTENTS ABSTRACT... 1 ACKNOWLEDGEMENTS... 2 SECTION I INTRODUCTION... 3 II RELATED WORK... 4 III PROBLEM STATEMENT... 5 Notations and Assumptions... 5 Problem Definition... 5 Page IV SYSTEM DESIGN... 6 V METHODS... 8 Feature Detection and Matching... 8 Feature Point Detection... 8 Line Segment Detection... 9 Line Segment Matching... 9 Homogeneous Region Detection... 10 Vanishing Point Detection... 11 Parallel Line Detection... 12 Region-based feature grouping... 12 Homography Detection... 12 3D Structure Estimation... 13 Line Triangulation... 13 Optimization with geometric constraints... 14 Triangulation Experiments... 14 VI CONCLUSION AND FUTURE WORK... 16

REFERENCES... 17

ABSTRACT Urban Structure Estimation Using Parallel and Orthogonal Lines. (May 2015) Rui Liu Electrical and Computer Engineering Department Texas A&M University Research Advisor: Dr. Dezhen Song Computer and Electrical Engineering Department In urban areas, buildings will block the line-of-sight of GPS signals while mobile robots are navigating through an environment which will result in localization errors of the robot. Robot localization using camera sensor is another way for exploring the surrounding area. Depth uncertainty in camera images could affect the accuracy of the estimated structure of the surrounding environment. Man-made structures contain many parallel and orthogonal lines. In this study, I use parallel and orthogonal lines in the scene to improve the estimation of the environment. The proposed method improves structure estimation compared with the standard reprojection error method. 1

ACKNOWLEDGEMENTS I would like to thank Joseph Lee and Dr. Dezhen Song for their tremendous help for this research project. Their teaching and support motivate me to make continuous development to this thesis and my personal improvement. 2

SECTION I INTRODUCTION In vision-based robot navigation, the robot keeps updating its location and map by measuring the distance from itself to surrounding objects. In man-made environment, building reconstruction based on multiple images can be used to help robots to locate its position and construct its 3D map for its surrounding environment. Man-made objects usually have many line segments. If properly used, line segments can be used to help the building reconstruction process. Urban buildings generally have two geometric properties: line segments on a plane can be parallel or orthogonal. The following figures were taken in Texas A&M University campus; there are many orthogonal and parallel line segments on different building planes. This study introduces the method used to improve 3D structure reconstruction by using orthogonality and parallelism of coplanar line segments. Fig. 1 Man-made environments have many parallel and orthogonal lines 3

SECTION II RELATED WORK Scene recovery based on line segments is a popular topic in computer vision. Previous work has been done to study reconstruction problems by using lines. In [1], a method that uses line features from image correspondences of multiple views to do line-based scene reconstruction by linearizing the Plücker constraint is proposed. Roberto Manduchi in [2] develops an algorithm to reconstruct the planar structures and relative camera pose in a Manhattan world by using matched lines from two images taken from distinct viewpoints. Both reprojection error method and line segments based method are used in [3] to explore a new algorithm to do planar building facade segmentation. Ali Elqursh and Ahmed Elgammal in [4] proposed a new algorithm to calibrated camera relative pose estimation by using lines, especially orthogonal and parallel line segments in physical indoor and urban environments. Having all present features in distinct views of a same scene been accurately detected and exactly matched can be very difficult, which finally make reconstruction be a tough task. 4

SECTION III PROBLEM STATEMENT 1. Notations and Assumptions Notations: Let I and I represent a pair of images from different viewpoints. Prime symbol ( ) represents the entities of the second image. I denote a pair of lines by l = (a, b, c) and l = (a, b, c ). Let P and P represent the camera matrix. Let y represent the intersection point of two lines. Let x i represent the i-th measured image coordinate in image I. Let x i represent an estimated point for x i, and let x i represent the correspondence point in image I. 2. Problem Definition Given two image views I and I, detect matched parallel and orthogonal line segments and estimate the 3D structure of the scene. 5

SECTION IV SYSTEM DESIGN Fig. 2 shows a high level overview of our system, which consist of three main blocks: (1) feature detection and matching, (2) Parallel Line Detection, and (3) 3D structure estimation. I, I 1. Feature Detection and Matching 1.1 Feature point detection 2. Parallel Line Detection 2.2 Region-based feature grouping 1.2 Line segment detection 2.3 Homography detection 1.3 Line segment matching 3. 3D Structure Estimation 1.4 Vanishing point detection 1.5 Homogeneous region detection 3.1 Line triangulation 3.2 Optimization with constraints Optimized Triangulation Fig.2 System diagram 6

Given two input images I and I, Fig. 2 Block 1 performs SIFT and line segment detection and matching. After matching line segments from Block 1, Block 2 performs initial coplanar parallel line detection with image segmentation. The homography computed from point correspondence is used to identify coplanar line segments. From the grouped coplanar parallel line information, Block 3 performs line triangulation and applies the geometric constraints. 7

SECTION V METHODS 1. Feature Detection and Matching We choose scale-invariant feature transform (SIFT) points and line segments as the features here and we use existing software to accomplish these tasks. 1. Feature Point Detection SIFT, designed by David Lowe in [5], is the algorithm used to detect and describe local features in images. It can effectively extract many featured points from images. Here is an example by using SIFT for two image views with matched feature points shown. Fig. 3 182 matched SIFT points are shown 8

2. Line Segment Detection Based on the SIFT, line segment as another local feature can be detected by using [6]. Each line segment is expressed by recording two end points. Fig. 4 shows an example with detected line segments plotted. Fig.4 1038 line segments are detected and plotted 3. Line Segment Matching After detecting line segments for two image views, they are matched by using the method proposed in [7]. This method uses feature point matches in Section 1.1 to match line segments. Fig. 5 shows an example with matched line segments plotted with same index number and color. 9

Fig. 5 175 line segments are matched for two image views for same scene 4. Homogeneous Region Detection Planes usually have homogeneous appearance. The optimization we proposed will be applied to coplanar line segments on the same physical planes. Thus it is important to detect the physical planes in a given scene. We use the image segmentation method proposed in [8]. Fig. 6 is a result for the two images view for same scene. Fig. 6 Image segmentation using method [8] 10

5. Vanishing Point Detection We can identify if two line segments as parallel if they pass the same point at infinity. For each line l and l (each line is expressed with 3 element column vector), the intersection point can be computed by y = l l = [ i j k a b c a b c ] (1) If y is near the vanishing point we determine l and l parallel. Vanishing point detection is the first step we do to explore coplanar parallel line segments. Vanishing point can be detected by using [9]. Fig. 7 is an example with vanishing point plotted. Parallel line segments are plotted with same color. Fig. 7 4 Vanishing points are found and plotted with method [9] 11

2. Parallel Line Detection We have two steps for the coplanar parallel line segments detection as shown in the system diagram. 6. Region-based feature grouping We first find the boundary for each single color area by finding their contours. Fig. 8 shows the homogeneous region contours. Fig. 8 Contours for each homogeneous region is shown We set a criterion for grouping coplanar line segments on each homogeneous region. We use two criterions: (a) those line segments surrounding a same contour are in the group represented by the contour and (b) a line segment can belong to multiple contours. 7. Homography Detection A single color covered area may not represent a single physical plane. If four points are on the same plane, and any three of them are not collinear, then we can compute a homograph matrix H, which represents the transformation from point x i on image I to point x i on image I. By using this theory, we can verify coplanar line segments. For a 12

sequence of four matched point pairs of two image views, Random sample consensus algorithm (RANSAC) is used to compute the homography. Then, we remove the detected homography plane and continue to use RANSAC for the rest. 3. 3D Structure Estimation To estimate the 3D structure, we follow two steps: (1) initialization using line triangulation and (2) using orthogonality, parallelism, and the reprojection error constraint to refine the structure. 1. Line Triangulation In two views, a line is the intersection of two planes, where each plane is generated with the camera center and two points on the image plane of the camera center. Given the camera matrix P and the two image lines l and l, we use the following span representation to represent the line segment L generated from the two intersected planes according to the formula provided from [10]. Fig. 9 from [10] illustrates this theory. L = [ lt P l T P ] (2) 13

Fig. 9 Line triangulation Planes are defined as π = P T l and π = P T l, P and P represent the camera matrixes for the two cameras. 2. Optimization with geometric constraints Fig. 10 shows how we apply the parallelism constraint to two line segments. Two parallel lines in 3D space intersect at ideal point on a plane located infinitely from the camera. Back-projection from that intersection on infinite plane to image plane gives us the vanishing point. We minimize the distance from vanishing point to the measured one. Fig. 10 The parallelism constraint 3. Triangulation Experiments We applied most of the previous theories and methods to triangulate the corridor scene shown in figure 3. The short line segments make a lot noise for the whole reconstruction result, thus we removed the 124 short line segments and mismatched line segments. Figure 11 is shows the colored line segment we want to triangulate. After figure 11, the 14

following pictures show how our program plots the result line by line, and how the noise comes to the triangulation. Fig. 11 The plotted 51 colored line segments will be triangulated Fig. 12 Process of drawing each triangulated line and noise appearance As we can see, the result shows good orthogonality and parallelism at the beginning, however, noise comes in when we take in more line segments. The reason for abnormal result can be due to the bad selection of the scene (we get mismatches for line segments and points) or the failure of visualization software (MATLAB). 15

SECTION VI CONCLUSION AND FUTURE WORK A lot previous work has been done in line segments related topics, but there are still a lot unsolved issues as well as potential improvements like exact line segment matching. This project is aimed at improving line triangulation by enforcing orthogonality and parallelism, however, only the triangulation part has been done and the result tells us that triangulation we have done here should be improved to make the basic scene be seen. Half-year long research for this topic is pressing, but we do obtain a lot related knowledge in computer vision research area. This knowledge could be very useful for future research. For the next several months, continuous work will be done on applying orthogonality and parallelism constraints to improve the performance of line triangulation. Meanwhile, we will collect more ground truth data sets to test the performance of the optimized triangulation. We will consider using new visualization software (C++ and OpenGL) and programming language (C++ and OpenCV) to do the visualization job. 16

REFERENCES [1]. Bartoli, A.; Sturm, P., "Multiple-view structure and motion from line correspondences," Computer Vision, 2003. Proceedings. Ninth IEEE International Conference on, vol., no., pp.207,212 vol.1, 13-16 Oct. 2003 [2]. Kim, Chellhwon; & Manduchi, Roberto, Planar Structures from Line Correspondences in a Manhattan World. Computer Vision, UC Santa Cruz. 2014. [3]. Joseph Lee, Yan Lu, and Dezhen Song, Planar Building Facade Segmentation and Mapping Using Appearance and Geometric Constraints, IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Chicago, US, Sept., 2014 [4]. Elqursh, A.; Elgammal, A., "Line-based relative pose estimation," Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on, vol., no., pp.3049,3056, 20-25 June 2011. [5]. David G. Lowe, "Distinctive image features from scale-invariant keypoints," International Journal of Computer Vision, 60, pp. 91-110, February 2004. [6]. R. G. von Gioi, J. Jakubowicz, J.-M. Morel, and G. Randall, LSD: a line segment detector, Image Processing On Line, 2012. [7]. B. Fan, F. Wu, and Z. Hu, Robust line matching through line-point invariants Pattern Recognition, vol. 45, no. 2, pp. 794-805, Feb. 2012. [8]. Grundmann, M.; Kwatra, V.; Mei Han; Essa, I., "Efficient hierarchical graph-based video segmentation," Computer Vision and Pattern Recognition (CVPR), 2010 IEEE Conference on, vol., no., pp.2141,2148, 13-18 June 2010. [9]. M. Nieto and L. Salgado, Real-time robust estimation of vanishing points through nonlinear optimization, in IS&T/SPIE Int. Conf. on Real-Time Image and Video Processing, SPIE vol.7724, 772402, 2010. (DOI 10.1117/12.854592) [10]. R. I. Hartley and A. Zisserman, Multiple View Geometry in Computer Vision, 2 nd ed. Cambridge University Press, 2004. 17