RGBD Point Cloud Alignment Using Lucas-Kanade Data Association and Automatic Error Metric Selection

Size: px
Start display at page:

Download "RGBD Point Cloud Alignment Using Lucas-Kanade Data Association and Automatic Error Metric Selection"

Transcription

1 IEEE TRANSACTIONS ON ROBOTICS, VOL. 31, NO. 6, DECEMBER 15 1 RGBD Point Cloud Alignment Using Lucas-Kanade Data Association and Automatic Error Metric Selection Brian Peasley and Stan Birchfield, Senior Member, IEEE Abstract We propose to overcome a significant limitation of the ICP algorithm used by KinectFusion, namely, its sole reliance upon geometric information. Our approach uses both geometric and color information in a direct manner that uses all the data in order to accurately estimate camera pose. Data association is performed by Lucas-Kanade to compute an affine warp between the color images associated with two RGBD point clouds. A subsequent step then estimates the Euclidean transformation between the point clouds using either a point-to-point or point-to-plane error metric, with a novel method based on a normal covariance test for automatically selecting between them. Together, Lucas- Kanade data association (LKDA) with covariance testing enables robust camera tracking through areas of low geometric features, without sacrificing accuracy in environments in which the existing ICP technique succeeds. Experimental results on several publicly available datasets demonstrate the improved performance both qualitatively and quantitatively. Index Terms Kinect, KinectFusion, ICP, Lucas-Kanade, Camera Tracking, Mapping I. INTRODUCTION THREE dimensional (3D) reconstruction of an environment is an important problem in robotics that has received much attention in recent years. The advent of inexpensive RGBD sensing (e.g., the Kinect sensor) has greatly improved the fidelity and speed with which such reconstructions can be made. These capabilities have far-reaching implications for the exploration and navigation of unknown environments, as well as for the manipulation of objects in those environments. Of the many 3D modeling techniques that have been developed, the landmark KinectFusion method [19], [13] has established itself as perhaps the most accurate, real-time, dense modeling system using an inexpensive sensor. Recent work by others [9] has extended this algorithm to operate over large-scale environments. However, both the standard and the extended version use geometric information alone to align the camera with the model, thus requiring the environment to contain sufficient geometric features in order for the iterative closest point (ICP) algorithm to accurately estimate the camera pose. This deficiency was noted in [9], where preliminary experiments were conducted to explore the advantage of replacing the ICP algorithm with the output of a featurebased visual odometry system called FOVIS [11] to provide a more stable estimate of camera pose. The -to- approach of FOVIS, however, loses one of the key advantages of KinectFusion, namely the reduced drift that results from matching the current image to the model rather than to the previous image. In this paper we present an approach to RGBD camera tracking in environments with low geometric information. The proposed approach uses all the data (both color and depth), preserves the advantages of KinectFusion s image-to-model matching, and obviates the need for the feature point extraction and feature correspondence steps inherent in feature-based visual odometry systems. The key to the approach is to replace the projective data association (PDA) point matching algorithm [3] used by KinectFusion with a data association technique driven by Lucas-Kanade [18], [6], []. We introduced this matching algorithm, which we call Lucas-Kanade data association (LKDA), in [1]. In this paper we extend that work with an automatic method for selecting the appropriate error metric based on the geometry of the scene, a GPU implementation of the matching, and more thorough experimental results to demonstrate the improvement that results from LKDA as well as to compare it with alternative techniques. These results on standard datasets demonstrate that LKDA combined with the automatic error metric enable camera tracking to succeed in areas of low geometry, without sacrificing either computational efficiency or accurate camera tracking in highly geometric environments. 1 II. PREVIOUS WORK Before the advent of recent inexpensive RGBD sensors, several researchers addressed the problem of improving ICP by incorporating color information. Both geometric and color information are incorporated into the distance metric used by ICP in a number of systems [5], [1], [14]. Similarly, Druon et al. [6] segment colored point clouds based on hue, then perform ICP while requiring matching points to belong to the same color class. More recently, a popular way to use visual information to improve alignment is to match feature points, then apply RANSAC to the features with depth values. This is the approach employed by Henry et al. [1], in which the visual feature associations from RANSAC are then combined with dense point associations in an ICP work. This same procedure of feature matching followed by RANSAC is adopted in both the FOVIS visual odometry system [11] and the algorithm of 1 Although our LKDA approach will fail when there is no visual texture (e.g., a white wall), such scenes have low geometric variation as well, since geometry usually generates visual texture via shading and shadows.

2 IEEE TRANSACTIONS ON ROBOTICS, VOL. 31, NO. 6, DECEMBER 15 Engelhard et al. [8], the latter of which also employs ICP and is evaluated in [7]. As expected, the primary drawback of these techniques is the failure of feature detection and matching when the scene contains little photometric texture. An approach that is closely related to our own is to register RGBD images via direct image warping. Tykkälä et al. [7] formulate the 3D registration of colored point clouds as a direct minimization of both depth and color information. To reduce the computational load, in followup work the same researchers [1] minimize only the image intensities, without using depth. In related work, Steinbrücker et al. [4] align two colored point clouds by employing Lie algebra to solve for the twist coordinates representing the 3D transformation that maximizes the photoconsistency of the projections. This work has been extended by Kerl et al. [16] to achieve realtime performance (3 Hz) on a single CPU, as well as to combine both photoconsistency and depth consistency in the minimization process [15]. Another direct approach is DTAM [], which uses dense methods for whole image aligment without a depth sensor but stores the model as a set of keys rather than as a dense volume. Other researchers have focused on extending KinectFusion. Henry et al. [9] propose a system consisting of local patch volumes, which are similar to KinectFusion volumes, connected together in a graph optimization work to ensure global consistency. In another approach that generates large maps, Whelan et al. [8] combine KinectFusion s ICP algorithm with a GPU implementation of the RGBD alignment approach of [4], and they provide a switching mechanism between the resulting estimate and that of the feature-based FOVIS. As expected, they show that both FOVIS and RGBD alignment fail when the scene contains little photometric texture, KinectFusion s ICP fails when there is little geometric texture, and a combination of these methods yields improved tracking robustness. Like the work of Whelan et al. [8], we focus upon improving the camera alignment step of KinectFusion by considering both visual and depth information. However, instead of combining off-the-shelf algorithms with a switching mechanism between them, we derive a novel algorithm that first aligns the intensity images then uses the resulting correspondences to align the depth images with an automatic selection of either point-to-point or point-to-plane error metrics based on the scene geometry. In Section V we show that our relatively straightforward, direct approach produces more consistent results than those obtained by more complicated methods, without sacrificing computational speed. III. LUCAS-KANADE DATA ASSOCIATION Let P = {p i } N i=1 and Q = {q i} N i=1, where p i,q i R 3, be two point clouds in 3D Euclidean space. The goal of data association is to compute a function f that maps indices in one point cloud to those of the other, so that f(i) = j indicates a corresponding pair of points p i q j. Let I P and I Q be the current and reference RGB images, respectively, associated with the point clouds, and d P and d Q the current and reference depth maps. In our implementation I Q is the previous image, Fig. 1. LEFT: PDA establishes correspondence between two point clouds, P and Q, by projecting each point p i P of one cloud onto the other s depth camera, D Q, to find the nearest projection of q j Q. RIGHT: LKDA establishes correspondence by projecting each point p i onto the RGB camera C P, then warping the projected point onto the other RGB camera C Q using the estimated warp from Lucas-Kanade to find the nearest projection of q j. (The normals n j are used in the point-to-plane error metric.) while d Q arises from projecting the truncated signed distance function (TSDF) according to the previous camera pose. Projective data association (PDA) [17] establishes correspondence between P and Q by projecting the points from both point clouds onto the same image plane, then considering points that project onto the same pixel to be corresponding points. See Figure 1. In a similar manner, Lucas-Kanade data association (LKDA) also finds the closest point on the image plane, but only after first transforming the projected coordinates according to the warp function found by Lucas- Kanade: f(i) = argmin j ϕ(c Q q j ) ϕ( W(C P p i ;ζ)), (1) where ζ contains the warp parameters, C P and C Q are 4 4 homogeneous matrices describing the 3D position and orientation of the RGB cameras, the tilde indicates homogeneous coordinates, and ϕ dehomogenizes the coordinates. The PDA approach relies solely upon geometric information, that is, the spatial coordinates of the points in the two clouds. As a result, when the environment does not provide sufficient geometric texture (e.g., the camera is looking at a planar wall), or when most of the depth readings are invalid due to the limited range of the sensor or occur outside the TSDF maintained by KinectFusion (e.g., the camera looks down a hallway), PDA will fail to provide enough correspondences to track the camera. The LKDA approach improves the alignment of point clouds in such environments, since the photometric information is not affected by the scene geometry, the range of the sensor, or the size of the TSDF. The Lucas-Kanade algorithm is a differential method for computing the optical flow of an image by finding the parameters ζ that minimize the sum of squared distance error ǫ LK = x ( IQ (W 1 (x;ζ)) I P (x) ), () where the summation is over all the pixels x = [ x y ] T in the image, and W(x;ζ) is a parametric warp function that brings the two images into alignment. This error is minimized by linearizing about the current estimate and repeatedly solving

3 IEEE TRANSACTIONS ON ROBOTICS, VOL. 31, NO. 6, DECEMBER 15 3 the following linear system ( ) gg T ζ = g ( I Q (W 1 (x;ζ)) I P (x) ) (3) x x ( ) T W for ζ, where g = ζ IP, and I Q contains the gradient of I P. For efficiency, we use the inverse compositional algorithm []. Common choices for the warp function are translation, affine, or projective. Since we are interested in a global, featureless mapping technique that warps an entire image into another, the translation model is not expressive enough; either an affine or projective warp is thus required. We have found little difference in accuracy between these two alternatives (see the experimental results), but the projective warp incurs a substantial increase in computation due to the fact that the Hessian is dependent upon the current state of the system and therefore must be recomputed each iteration. As a result we use an affine warp, which is given by [ x(rxx +1)+yr W(x,y;ζ) = xy +a x xr yx +y(r yy +1)+a y where ζ = [ r xx r xy r yx r yy a x a y ] T. IV. ESTIMATING THE CAMERA POSE ], (4) Each iteration of ICP can be broken into six steps, and different choices within these steps lead to different variations of the algorithm []. Figure illustrates these steps within the traditional ICP approach used by KinectFusion, compared with our LKDA approach. Instead of recomputing data association after each alignment, we first perform data association, then perform alignment. This decoupling, which is enabled by our separate use of RGB and depth images, leads to significant speedup without sacrificing robustness. At the same time, the correspondences used in each iteration of the algorithm are not necessarily the same, because the rejection step retains only a subset of them. In this section we detail the steps that are different between our approach and KinectFusion. A. Point Matching As already mentioned, projective data association (PDA) establishes correspondence by finding the closest point from the other cloud as projected onto the image plane. In a similar manner, Lucas-Kanade data association (LKDA) also finds the closest point on the image plane, but only after first transforming the projected coordinates according to the warp function found by Lucas-Kanade. Unlike traditional ICP, however, the LKDA point matching step is performed only once per pair of image s, and subsets of the correspondences found in this step are then used in the iterations of the alignment process. B. Outlier rejection As in [19], our approach rejects point correspondences that exceed a distance or angular threshold. In addition, corresponding points whose color difference exceeds a predefined (a) ICP implementation in KinectFusion (b) Our LKDA approach Fig.. TOP: Given the current and reference depth images, KinectFusion iterates through the six ICP steps. Steps 1-3 perform data association (PDA), while steps 4-6 perform alignment. BOTTOM: In contrast, our approach iteratively finds the correspondences using the RGB images (LKDA, steps 1-3), then iteratively estimates the alignment parameters using the depth images (steps 4-6). The steps marked with an asterisk (*) are different in the two approaches. color threshold ρ are rejected: I Q (x,y) I P (W(x,y;ζ)) > ρ. (5) Note that the rejected points in each iteration of the alignment process are not necessarily the same, so that a different subset of points contribute to the error metric in each iteration. In our implementation the distance, angular, and color thresholds are.1 m,, and ρ = 5, respectively. C. Automatic Selection of the Error Metric Two common choices of the metric used for alignment are point-to-point and point-to-plane [4]. If there is sufficient geometric texture in the environment, then the latter converges faster than the former [], which explains its adoption by KinectFusion. On the other hand, if all point correspondences are on a flat wall then the point-to-plane error metric will not result in correct alignment, because there will be no mechanism to induce a lateral shift between the point clouds in the direction perpendicular to the normals. See Figure 3. To overcome this limitation we use a linear combination of the 3D Euclidean transformations computed using the pointto-point (T pp ) and the point-to-plane (T pπ ) error metric: T = (1 λ)t pp +λt pπ, (6) where λ 1 is a weighting factor. The value of λ is determined automatically at run time by solving for the condition number κ of the covariance matrix of the geometric normals seen by the current. In areas of low geometry the value of γ = κ 1 is low, while in areas of high geometry γ is high. We model λ as a sigmoid ( λ = 1+e α(γ ˆγ)) 1, (7)

4 IEEE TRANSACTIONS ON ROBOTICS, VOL. 31, NO. 6, DECEMBER 15 4 D. Error minimization Given the correspondences between the two image s that survived the rejection step, along with the value λ, the two point clouds can be aligned. The transformation T pp is computed in closed-form using orthogonal Procrustes analysis, while T pπ is computed iteratively as described in [17]. The camera pose is then given by (6). Fig. 3. When the scene is planar, the point-to-plane error metric fails to correctly align point clouds, even with perfect correspondence. error (meters) inverse condition number point to point point to plane Fig. 4. Scatter plot of γ, the inverse of the condition number of the covariance matrix of the normals, versus the error of the estimated translation (averaged over all s of several sequences). The red dots indicate the results using the point-to-plane error metric (T pπ) to estimate odometry, while the blue dots show the results from the point-to-point error metric (T pp). Robust exponential curves were fit to both data sets to determine their intersection. where α is a large constant (determined experimentally to be 3), and ˆγ indicates the value of γ for which the two metrics are weighted equally, i.e., λ =.5. To determine ˆγ, for each over several sequences we computed T pπ, T pp, and κ 1, comparing the former two with the ground truth. The resulting scatter plot can been seen in Figure 4, along with the intersection of robust exponential curves fit to both data sets, yielding ˆγ =.1. Figure 5 shows the results of both metrics on two scenes, illustrating that T pp yields less error on scenes with low geometry, while T pπ tends to perform slightly better on scenes with high geometry. cumulative error point to plane point to point 1 3 cumulative error point to plane point to point 1 3 Fig. 5. Cumulative camera translation error for both metrics when reconstructing a relatively flat environment (left) versus a geometrically rich environment (right). V. EXPERIMENTAL RESULTS To implement the proposed approach, we modified the opensource version of KinectFusion known as KinFu in the PCL library [3] with our own C++ and GPGPU code. In our first set of experiments we compared the unmodified PDA version with six different versions of LKDA using either grayscale or color images, and either a translation, affine, or projective warp. To facilitate quantitative comparison, we used the data from the RGB-D SLAM Dataset and Benchmark [5], which provides synchronized depth and color data from a Kinect or Xtion sensor as well as the ground truth sensor trajectory from a high-accuracy motion capture system. For these experiments we used 5 different datasets exhibiting a variety of small environments, which can roughly be categorized as follows: high-geometry. The environment contains a large amount of geometric variation throughout, enabling geometrybased sensor tracking. (fr1/xyz, fr1/floor, fr1/desk) planar. The environment is largely planar (that is, without much geometric variation), making geometry-based sensor tracking difficult. (fr3/ntnw 3 ) out-of-bounds. For at least some s of the sequence, the sensor is located so that a majority of the depth values are outside the TSDF volume. (fr1/rpy) We used a m TSDF volume divided into 51 3 voxels (each voxel is 5.86 mm per side). The results on these 5 datasets are shown in Table I, where the RMSE translational drift (in meters per second) and RMSE rotational drift (in radians per second) are displayed. The drift is also known as the Relative Pose Error (RPE) [5]. For the high-geometry sequences, all the methods succeed, and there is little difference in accuracy in either position or orientation. For the planar (fr3/ntnw) or out-of-bounds (fr1/rpy) sequences, however, the standard KinectFusion algorithm fails to maintain sensor tracking. There is a noticeable improvement using affine over translation, but almost none by using projective over affine. To save space we show only the grayscale results, since the color results are nearly identical. We directly compared our method with the approach of Whelan et al. [8] and the Dense Visual Odometry (DVO) algorithm of Kerl et al. [15], [16]. Results of these comparisons are displayed in Tables II and III, respectively. 4 While the algorithm of [8] performs better on a -to- basis, our approach yields a better overall estimated trajectory, leading 3 A shortened form of the name fr3/nostructure texture near withloop. 4 To reduce space, Table II only includes results from the FOVIS/ICP+RGB- D version of the algorithm in [8]; other versions are similar. For fair comparison, the final column of Table III shows the RGB+D+KF version of [15], which includes keys but not graph optimization.

5 IEEE TRANSACTIONS ON ROBOTICS, VOL. 31, NO. 6, DECEMBER 15 5 fr1/xyz fr1/floor fr1/desk fr3/ntnw fr1/rpy PDA F F (.31) (.58) (.7) (.461) F (1.87) F LKDA F.61 Translation (.31) (.63) (.7) (.533) F (.57) LKDA Affine (.3) (.56) (.71) (.31) (.57) LKDA Projective (.3) (.57) (.71) (.31) (.56) TABLE I RPE RMSE IN TRANSLATION (M/S) (TOP OF EACH CELL) AND ROTATION (RAD/S) (BOTTOM, IN PARENTHESES). THE BEST RESULT IS IN BOLD, AND FAILED TRACKING RESULTS ARE INDICATED BY F. LKDA Whelan ATE RPE ATE RPE med max RMSE med max RMSE fr1/desk fr/desk fr1/room fr/lnl TABLE II COMPARISON WITH WHELAN ET AL. [8], SHOWING MEDIAN AND MAXIMUM VALUES OF THE ABSOLUTE TRANSLATION ERROR (ATE) (M), AND RMSE TRANSLATIONAL DRIFT (M/S). KinectFusion with PDA to lower absolute trajectory errors. Compared with [15], our approach is competitive, yielding the lowest error on several sequences. Plots of the estimated camera position versus number for the planar (fr3/ntnw) and out-of-bounds (fr1/rpy) environments for the different algorithms are shown in Figure 6. Notice that KinectFusion with PDA is unable to maintain tracking in such environments, while our LKDA algorithm (grayscale affine) and the approaches of [8] and [15] are able to maintain camera tracking. Other variations of LKDA also perform well, except that the translation motion model is insufficient for the planar scene. Reconstructions of the outof-bounds environment by rendering the point cloud from two approaches are shown in Figure 7 for comparison. Notice that LKDA achieves an accurate reconstruction of the scene, while the reconstruction from PDA is difficult to interpret. The importance of the automatic error selection is shown in Table IV. Seven sequences were run using either a pointto-point or point-to-plane error metric, or with automatic selection between the two. As expected, the point-to-point error LKDA DVO ATE RPE ATE RPE RMSE RMSE RMSE RMSE fr1/desk fr1/desk fr1/room fr1/ fr1/teddy fr1/floor fr1/xyz fr1/rpy fr1/plant TABLE III COMPARISON WITH DVO [15], SHOWING THE RMSE ABSOLUTE TRAJECTORY ERROR (M) AND THE RMSE TRANSLATIONAL DRIFT (M/S). KinectFusion with LKDA Fig. 7. Reconstruction of the out-of-bounds environment (fr1/rpy) by standard KinectFusion with PDA (top) and KinectFusion with LKDA (bottom). The inability of the former to handle depth values outside the TSDF volume causes a severe loss in camera tracking, thus resulting in a poor reconstruction; while our approach yields accurate reconstruction. High Low point-to-point point-to-plane automatic fr1/rpy fr1/floor fr1/teddy fr1/xyz fr1/plant fr3/ntnw fr3/ntf TABLE IV COMPARISON OF THE RMSE ATE OF THE POINT-TO-POINT ERROR METRIC (LEFT) AGAINST THE POINT-TO-PLANE ERROR METRIC (MIDDLE). THE FORMER PERFORMED BETTER IN SCENES OF LOW GEOMETRY, WHILE THE LATTER PERFORMED BETTER IN SCENES WITH HIGH GEOMETRY. OUR AUTOMATIC SELECTION METHOD HANDLES BOTH TYPES OF SCENES. metric performed better in environments with little geometrical structure, while the point-to-plane error metric always yielded less error on sequences with high geometrical structure. The automatic selection mechanism enabled our approach to yield the minimum error on all sequences. Timing comparisons of the different variations of our algorithm were conducted on several machines with varying computational power and GPUs. For standard KinectFusion

6 IEEE TRANSACTIONS ON ROBOTICS, VOL. 31, NO. 6, DECEMBER 15 6 X (meters) X (meters) PDA + KF Whelan et al. DVO LKDA + KF Ground Truth Y (meters) Y (meters) Z (meters) 1.5 Z (meters) Fig. 6. Camera position estimation in a planar environment (fr3/ntnw, top) and out-of-bounds environment (fr1/rpy, bottom). Shown is a comparison of the proposed grayscale affine LKDA, KinectFusion with PDA, the approach of Whelan et al. [8], and DVO [15]. The solid black line is ground truth. with PDA, nearly all the computation (approximately 85%) is devoted to the alignment, with the remaining computation required for PDA. Our approach retains the alignment computation but replaces PDA with a more expensive LKDA step. On average, the ratio of the time taken by LKDA to PDA is 3:1 (grayscale translation), 6:1 (grayscale affine), and 45:1 (grayscale projective). (Color versions require nearly 3 times as much computation.) Therefore, our grayscale affine approach (the recommended version) is about 1.8 times as expensive as KinectFusion (since (6) = 1.8). On a machine capable of running standard KinectFusion at 3 Hz, our algorithm runs at approximately 17 Hz; on a machine that runs KinectFusion at 1 Hz, our algorithm runs at nearly 6 Hz. VI. CONCLUSION In this paper we have presented a direct technique to robustly estimate camera motion using an RGBD sensor, even when there is little geometric information in the scene. The approach retains the drift-free advantages of the matchto-model approach of KinectFusion, without requiring the estimation of visual features. The approach operates by incrementally computing the affine warp between successive images using Lucas-Kanade, yielding a correspondence between point clouds. This approach to correspondence, which we call LKDA, replaces the PDA technique implemented in the variant of ICP found in KinectFusion. In contrast with PDA, LKDA was shown on a number of publicly available sequences that it is able to maintain camera tracking in environments with limited geometric structure, as well as in environments that extend beyond the TSDF volume or contain depths outside the range of the sensor. Moreover, we have proposed an automatic mechanism for selecting between point-to-point or point-to-plane error metrics that properly handles a variety of environments. Future work will be aimed at extending this approach to large-scale environments. VII. ACKNOWLEDGMENTS The authors thank T. Whelan and C. Kerl for providing their data for comparison, and the anonymous reviewers for their helpful comments. This work was partially supported by NSF grant IIS REFERENCES [1] C. Audras, A. I. Comport, M. Meilland, and P. Rives. Real-time dense RGB-D localisation and mapping. In Australian Conference on Robotics and Automation, Dec. 11. [] S. Baker and I. Matthews. Lucas-Kanade years on: A unifying work. International Journal of Computer Vision, 56(3):1 55, 4. [3] G. Blais and M. Levine. Registering multiview range data to create 3D computer objects. IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI), 17(8):8 84, Aug [4] Y. Chen and G. Medioni. Object modeling by registration of multiple range images. In Proceedings of the IEEE International Conference on Robotics and Automation, volume 3, pages 74 79, Apr [5] L. Douadi, M.-J. Aldon, and A. Crosnier. Pair-wise registration of 3D/color data sets with ICP. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages , Oct. 6. [6] S. Druon, M. Aldon, and A. Crosnier. Color constrained ICP for registration of large unstructured 3D color data sets. In IEEE International Conference on Information Acquisition, pages 49 55, Aug 6. [7] F. Endres, J. Hess, N. Engelhard, J. Sturm, D. Cremers, and W. Burgard. An evaluation of the RGB-D SLAM system. In Proceedings of the International Conference on Robotics and Automation (ICRA), May 1. [8] N. Engelhard, F. Endres, J. Hess, J. Sturm, and W. Burgard. Realtime 3D visual SLAM with a hand-held RGB-D camera. In Proc. of the RGB-D Workshop on 3D Perception in Robotics at the European Robotics Forum, Apr. 11. [9] P. Henry, D. Fox, A. Bhowmik, and R. Mongia. Patch volumes: Segmentation-based consistent mapping with RGB-D cameras. In Proceedings of the International Conference on 3D Vision (3DV), June 13. [1] P. Henry, M. Krainin, E. Herbst, X. Ren, and D. Fox. RGB-D mapping: Using Kinect-style depth cameras for dense 3D modeling of indoor environments. International Journal of Robotics Research, 31(5): , Apr. 1. [11] A. Huang, A. Bachrach, P. Henry, M. Krainin, D. Maturana, D. Fox, and N. Roy. Visual odometry and mapping for autonomous flight using an RGB-D camera. In International Symposium on Robotics Research (ISRR), Aug. 11.

7 IEEE TRANSACTIONS ON ROBOTICS, VOL. 31, NO. 6, DECEMBER 15 7 [1] B. Huhle, M. Magnusson, W. Strasser, and A. J. Lilienthal. Registration of colored 3D point clouds with a kernel-based extension to the normal distributions transform. In Proceedings of the International Conference on Robotics and Automation (ICRA), pages 45 43, May 8. [13] S. Izadi, D. Kim, O. Hilliges, D. Molyneaux, R. Newcombe, P. Kohli, J. Shotton, S. Hodges, D. Freeman, A. Davison, and A. Fitzgibbon. KinectFusion: Real-time 3D reconstruction and interaction using a moving depth camera. In Proceedings of the 4th ACM Symposium on User Interface Software and Technology, pages , 11. [14] J. H. Joung, K. H. An, J. W. Kang, M. J. Chung, and W. Yu. 3D environment reconstruction using modified color ICP algorithm by fusion of a camera and a 3D laser range finder. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages , Oct. 9. [15] C. Kerl, J. Sturm, and D. Cremers. Dense visual SLAM for RGB-D cameras. In Proceedings of the International Conference on Intelligent Robot Systems (IROS), Nov. 13. [16] C. Kerl, J. Sturm, and D. Cremers. Robust odometry estimation for RGB-D cameras. In Proceedings of the International Conference on Robotics and Automation (ICRA), May 13. [17] K.-L. Low. Linear least-squares optimization for point-to-plane ICP surface registration. Technical Report TR4-4, University of North Carolina, Feb. 4. [18] B. D. Lucas and T. Kanade. An iterative image registration technique with an application to stereo vision. In Proceedings of the 7th International Joint Conference on Artificial Intelligence, pages , Aug [19] R. A. Newcombe, S. Izadi, O. Hilliges, D. Molyneaux, D. Kim, A. J. Davison, P. Kohli, J. Shotton, S. Hodges, and A. Fitzgibbon. Kinect- Fusion: Real-time dense surface mapping and tracking. In Proceedings of the 1th IEEE International Symposium on Mixed and Augmented Reality (ISMAR), pages , 11. [] R. A. Newcombe, S. J. Lovegrove, and A. J. Davison. DTAM: Dense tracking and mapping in real-time. In Proceedings of the International Conference on Computer Vision, pages 3 37, 11. [1] B. Peasley and S. Birchfield. Replacing projective data association with Lucas-Kanade for KinectFusion. In Proceedings of the International Conference on Robotics and Automation (ICRA), May 13. [] S. Rusinkiewicz and M. Levoy. Efficient variants of the ICP algorithm. In Third International Conference on 3D Digital Imaging and Modeling (3DIM), June 1. [3] R. B. Rusu and S. Cousins. 3D is here: Point Cloud Library (PCL). In Proceedings of the International Conference on Robotics and Automation (ICRA), May 11. [4] F. Steinbrücker, J. Sturm, and D. Cremers. Real-time visual odometry from dense RGB-D images. In Workshop on Live Dense Reconstruction with Moving Cameras (at ICCV), Nov. 11. [5] J. Sturm, N. Engelhard, F. Endres, W. Burgard, and D. Cremers. A benchmark for the evaluation of RGB-D SLAM systems. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages , Oct. 1. [6] C. Tomasi and T. Kanade. Detection and tracking of point features. Technical Report CMU-CS-91-13, Carnegie Mellon University, Apr [7] T. Tykkälä, C. Audras, and A. I. Comport. Direct iterative closest point for real-time visual odometry. In Second International Workshop on Computer Vision in Vehicle Technology, Nov. 11. [8] T. Whelan, H. Johannsson, M. Kaess, J. J. Leonard, and J. McDonald. Robust real-time visual odometry for dense RGB-D mapping. In Proceedings of the International Conference on Robotics and Automation (ICRA), May 13. [9] T. Whelan, J. B. McDonald, M. Kaess, M. F. Fallon, H. Johannsson, and J. J. Leonard. Kintinuous: Spatially extended KinectFusion. In RSS Workshop on RGB-D: Advanced Reasoning with Depth Cameras, Sydney, Australia, July 1.

Replacing Projective Data Association with Lucas-Kanade for KinectFusion

Replacing Projective Data Association with Lucas-Kanade for KinectFusion Replacing Projective Data Association with Lucas-Kanade for KinectFusion Brian Peasley and Stan Birchfield Electrical and Computer Engineering Dept. Clemson University, Clemson, SC 29634 {bpeasle,stb}@clemson.edu

More information

A Real-Time RGB-D Registration and Mapping Approach by Heuristically Switching Between Photometric And Geometric Information

A Real-Time RGB-D Registration and Mapping Approach by Heuristically Switching Between Photometric And Geometric Information A Real-Time RGB-D Registration and Mapping Approach by Heuristically Switching Between Photometric And Geometric Information The 17th International Conference on Information Fusion (Fusion 2014) Khalid

More information

CVPR 2014 Visual SLAM Tutorial Kintinuous

CVPR 2014 Visual SLAM Tutorial Kintinuous CVPR 2014 Visual SLAM Tutorial Kintinuous kaess@cmu.edu The Robotics Institute Carnegie Mellon University Recap: KinectFusion [Newcombe et al., ISMAR 2011] RGB-D camera GPU 3D/color model RGB TSDF (volumetric

More information

Tracking an RGB-D Camera Using Points and Planes

Tracking an RGB-D Camera Using Points and Planes MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Tracking an RGB-D Camera Using Points and Planes Ataer-Cansizoglu, E.; Taguchi, Y.; Ramalingam, S.; Garaas, T. TR2013-106 December 2013 Abstract

More information

High-speed Three-dimensional Mapping by Direct Estimation of a Small Motion Using Range Images

High-speed Three-dimensional Mapping by Direct Estimation of a Small Motion Using Range Images MECATRONICS - REM 2016 June 15-17, 2016 High-speed Three-dimensional Mapping by Direct Estimation of a Small Motion Using Range Images Shinta Nozaki and Masashi Kimura School of Science and Engineering

More information

Efficient Compositional Approaches for Real-Time Robust Direct Visual Odometry from RGB-D Data

Efficient Compositional Approaches for Real-Time Robust Direct Visual Odometry from RGB-D Data Efficient Compositional Approaches for Real-Time Robust Direct Visual Odometry from RGB-D Data Sebastian Klose 1, Philipp Heise 1 and Alois Knoll 1 Abstract In this paper we give an evaluation of different

More information

Large-Scale Multi-Resolution Surface Reconstruction from RGB-D Sequences

Large-Scale Multi-Resolution Surface Reconstruction from RGB-D Sequences Large-Scale Multi-Resolution Surface Reconstruction from RGB-D Sequences Frank Steinbru cker, Christian Kerl, Ju rgen Sturm, and Daniel Cremers Technical University of Munich Boltzmannstrasse 3, 85748

More information

Dense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm

Dense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm Computer Vision Group Prof. Daniel Cremers Dense Tracking and Mapping for Autonomous Quadrocopters Jürgen Sturm Joint work with Frank Steinbrücker, Jakob Engel, Christian Kerl, Erik Bylow, and Daniel Cremers

More information

Feature-based RGB-D camera pose optimization for real-time 3D reconstruction

Feature-based RGB-D camera pose optimization for real-time 3D reconstruction Computational Visual Media DOI 10.1007/s41095-016-0072-2 Research Article Feature-based RGB-D camera pose optimization for real-time 3D reconstruction Chao Wang 1, Xiaohu Guo 1 ( ) c The Author(s) 2016.

More information

Robust Online 3D Reconstruction Combining a Depth Sensor and Sparse Feature Points

Robust Online 3D Reconstruction Combining a Depth Sensor and Sparse Feature Points 2016 23rd International Conference on Pattern Recognition (ICPR) Cancún Center, Cancún, México, December 4-8, 2016 Robust Online 3D Reconstruction Combining a Depth Sensor and Sparse Feature Points Erik

More information

Large-Scale Multi-Resolution Surface Reconstruction from RGB-D Sequences

Large-Scale Multi-Resolution Surface Reconstruction from RGB-D Sequences 2013 IEEE International Conference on Computer Vision Large-Scale Multi-Resolution Surface Reconstruction from RGB-D Sequences Frank Steinbrücker, Christian Kerl, Jürgen Sturm, and Daniel Cremers Technical

More information

Colored Point Cloud Registration Revisited Supplementary Material

Colored Point Cloud Registration Revisited Supplementary Material Colored Point Cloud Registration Revisited Supplementary Material Jaesik Park Qian-Yi Zhou Vladlen Koltun Intel Labs A. RGB-D Image Alignment Section introduced a joint photometric and geometric objective

More information

Using RGB Information to Improve NDT Distribution Generation and Registration Convergence

Using RGB Information to Improve NDT Distribution Generation and Registration Convergence Using RGB Information to Improve NDT Distribution Generation and Registration Convergence James Servos and Steven L. Waslander University of Waterloo, Waterloo, ON, Canada, N2L 3G1 Abstract Unmanned vehicles

More information

Sturm, Jürgen; Bylow, Erik; Kerl, Christian; Kahl, Fredrik; Cremers, Daniel

Sturm, Jürgen; Bylow, Erik; Kerl, Christian; Kahl, Fredrik; Cremers, Daniel Dense Tracking and Mapping with a Quadrocopter Sturm, Jürgen; Bylow, Erik; Kerl, Christian; Kahl, Fredrik; Cremers, Daniel 213 Link to publication Citation for published version (APA): Sturm, J., Bylow,

More information

Dense 3D Reconstruction from Autonomous Quadrocopters

Dense 3D Reconstruction from Autonomous Quadrocopters Dense 3D Reconstruction from Autonomous Quadrocopters Computer Science & Mathematics TU Munich Martin Oswald, Jakob Engel, Christian Kerl, Frank Steinbrücker, Jan Stühmer & Jürgen Sturm Autonomous Quadrocopters

More information

DeReEs: Real-Time Registration of RGBD Images Using Image-Based Feature Detection And Robust 3D Correspondence Estimation and Refinement

DeReEs: Real-Time Registration of RGBD Images Using Image-Based Feature Detection And Robust 3D Correspondence Estimation and Refinement DeReEs: Real-Time Registration of RGBD Images Using Image-Based Feature Detection And Robust 3D Correspondence Estimation and Refinement Sahand Seifi Memorial University of Newfoundland sahands[at]mun.ca

More information

DNA-SLAM: Dense Noise Aware SLAM for ToF RGB-D Cameras

DNA-SLAM: Dense Noise Aware SLAM for ToF RGB-D Cameras DNA-SLAM: Dense Noise Aware SLAM for ToF RGB-D Cameras Oliver Wasenmüller, Mohammad Dawud Ansari, Didier Stricker German Research Center for Artificial Intelligence (DFKI) University of Kaiserslautern

More information

3D Environment Reconstruction

3D Environment Reconstruction 3D Environment Reconstruction Using Modified Color ICP Algorithm by Fusion of a Camera and a 3D Laser Range Finder The 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems October 11-15,

More information

A Benchmark for RGB-D Visual Odometry, 3D Reconstruction and SLAM

A Benchmark for RGB-D Visual Odometry, 3D Reconstruction and SLAM 2014 IEEE International Conference on Robotics & Automation (ICRA) Hong Kong Convention and Exhibition Center May 31 - June 7, 2014. Hong Kong, China A Benchmark for RGB-D Visual Odometry, 3D Reconstruction

More information

Efficient SLAM Scheme Based ICP Matching Algorithm Using Image and Laser Scan Information

Efficient SLAM Scheme Based ICP Matching Algorithm Using Image and Laser Scan Information Proceedings of the World Congress on Electrical Engineering and Computer Systems and Science (EECSS 2015) Barcelona, Spain July 13-14, 2015 Paper No. 335 Efficient SLAM Scheme Based ICP Matching Algorithm

More information

Incremental compact 3D maps of planar patches from RGBD points

Incremental compact 3D maps of planar patches from RGBD points Incremental compact 3D maps of planar patches from RGBD points Juan Navarro and José M. Cañas Universidad Rey Juan Carlos, Spain Abstract. The RGBD sensors have opened the door to low cost perception capabilities

More information

Volumetric 3D Mapping in Real-Time on a CPU

Volumetric 3D Mapping in Real-Time on a CPU Volumetric 3D Mapping in Real-Time on a CPU Frank Steinbrücker, Jürgen Sturm, and Daniel Cremers Abstract In this paper we propose a novel volumetric multi-resolution mapping system for RGB-D images that

More information

3D PLANE-BASED MAPS SIMPLIFICATION FOR RGB-D SLAM SYSTEMS

3D PLANE-BASED MAPS SIMPLIFICATION FOR RGB-D SLAM SYSTEMS 3D PLANE-BASED MAPS SIMPLIFICATION FOR RGB-D SLAM SYSTEMS 1,2 Hakim ELCHAOUI ELGHOR, 1 David ROUSSEL, 1 Fakhreddine ABABSA and 2 El-Houssine BOUYAKHF 1 IBISC Lab, Evry Val d'essonne University, Evry, France

More information

Scanning and Printing Objects in 3D Jürgen Sturm

Scanning and Printing Objects in 3D Jürgen Sturm Scanning and Printing Objects in 3D Jürgen Sturm Metaio (formerly Technical University of Munich) My Research Areas Visual navigation for mobile robots RoboCup Kinematic Learning Articulated Objects Quadrocopters

More information

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Structured Light II Johannes Köhler Johannes.koehler@dfki.de Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Introduction Previous lecture: Structured Light I Active Scanning Camera/emitter

More information

Stereo and Kinect fusion for continuous. 3D reconstruction and visual odometry

Stereo and Kinect fusion for continuous. 3D reconstruction and visual odometry 1 Stereo and Kinect fusion for continuous 3D reconstruction and visual odometry Ozgur YILMAZ #*1, Fatih KARAKUS *2 # Department of Computer Engineering, Turgut Özal University Ankara, TURKEY 1 ozyilmaz@turgutozal.edu.tr

More information

Integrating Depth and Color Cues for Dense Multi-Resolution Scene Mapping Using RGB-D Cameras

Integrating Depth and Color Cues for Dense Multi-Resolution Scene Mapping Using RGB-D Cameras In Proc. of the IEEE Int. Conf. on Multisensor Fusion and Information Integration (MFI), Hamburg, Germany, 2012 Integrating Depth and Color Cues for Dense Multi-Resolution Scene Mapping Using RGB-D Cameras

More information

Simultaneous Localization and Calibration: Self-Calibration of Consumer Depth Cameras

Simultaneous Localization and Calibration: Self-Calibration of Consumer Depth Cameras Simultaneous Localization and Calibration: Self-Calibration of Consumer Depth Cameras Qian-Yi Zhou Vladlen Koltun Abstract We describe an approach for simultaneous localization and calibration of a stream

More information

Memory Management Method for 3D Scanner Using GPGPU

Memory Management Method for 3D Scanner Using GPGPU GPGPU 3D 1 2 KinectFusion GPGPU 3D., GPU., GPGPU Octree. GPU,,. Memory Management Method for 3D Scanner Using GPGPU TATSUYA MATSUMOTO 1 SATORU FUJITA 2 This paper proposes an efficient memory management

More information

NICP: Dense Normal Based Point Cloud Registration

NICP: Dense Normal Based Point Cloud Registration NICP: Dense Normal Based Point Cloud Registration Jacopo Serafin 1 and Giorgio Grisetti 1 Abstract In this paper we present a novel on-line method to recursively align point clouds. By considering each

More information

Robust Real-Time Visual Odometry for Dense RGB-D Mapping

Robust Real-Time Visual Odometry for Dense RGB-D Mapping Robust Real-Time Visual Odometry for Dense RGB-D Mapping Thomas Whelan 1, Hordur Johannsson, Michael Kaess, John J. Leonard and John McDonald 1 Abstract This paper describes extensions to the Kintinuous

More information

Mobile Point Fusion. Real-time 3d surface reconstruction out of depth images on a mobile platform

Mobile Point Fusion. Real-time 3d surface reconstruction out of depth images on a mobile platform Mobile Point Fusion Real-time 3d surface reconstruction out of depth images on a mobile platform Aaron Wetzler Presenting: Daniel Ben-Hoda Supervisors: Prof. Ron Kimmel Gal Kamar Yaron Honen Supported

More information

Segmentation and Tracking of Partial Planar Templates

Segmentation and Tracking of Partial Planar Templates Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract

More information

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University. 3D Computer Vision Structured Light II Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de 1 Introduction

More information

Super-Resolution Keyframe Fusion for 3D Modeling with High-Quality Textures

Super-Resolution Keyframe Fusion for 3D Modeling with High-Quality Textures Super-Resolution Keyframe Fusion for 3D Modeling with High-Quality Textures Robert Maier, Jörg Stückler, Daniel Cremers International Conference on 3D Vision (3DV) October 2015, Lyon, France Motivation

More information

arxiv: v1 [cs.ro] 23 Nov 2015

arxiv: v1 [cs.ro] 23 Nov 2015 Multi-Volume High Resolution RGB-D Mapping with Dynamic Volume Placement Michael Salvato 1, Ross Finman 1, and John J. Leonard 1 arxiv:1511.07106v1 [cs.ro] 23 Nov 2015 I. ABSTRACT Abstract We present a

More information

Using Augmented Measurements to Improve the Convergence of ICP. Jacopo Serafin and Giorgio Grisetti

Using Augmented Measurements to Improve the Convergence of ICP. Jacopo Serafin and Giorgio Grisetti Jacopo Serafin and Giorgio Grisetti Point Cloud Registration We want to find the rotation and the translation that maximize the overlap between two point clouds Page 2 Point Cloud Registration We want

More information

Kintinuous: Spatially Extended KinectFusion Thomas Whelan, Michael Kaess, Maurice Fallon, Hordur Johannsson, John Leonard, and John McDonald

Kintinuous: Spatially Extended KinectFusion Thomas Whelan, Michael Kaess, Maurice Fallon, Hordur Johannsson, John Leonard, and John McDonald Computer Science and Artificial Intelligence Laboratory Technical Report MIT-CSAIL-TR-2012-020 July 19, 2012 Kintinuous: Spatially Extended KinectFusion Thomas Whelan, Michael Kaess, Maurice Fallon, Hordur

More information

Efficient Online Surface Correction for Real-time Large-Scale 3D Reconstruction

Efficient Online Surface Correction for Real-time Large-Scale 3D Reconstruction MAIER ET AL.: EFFICIENT LARGE-SCALE ONLINE SURFACE CORRECTION 1 Efficient Online Surface Correction for Real-time Large-Scale 3D Reconstruction Robert Maier maierr@in.tum.de Raphael Schaller schaller@in.tum.de

More information

Direct Methods in Visual Odometry

Direct Methods in Visual Odometry Direct Methods in Visual Odometry July 24, 2017 Direct Methods in Visual Odometry July 24, 2017 1 / 47 Motivation for using Visual Odometry Wheel odometry is affected by wheel slip More accurate compared

More information

Removing Moving Objects from Point Cloud Scenes

Removing Moving Objects from Point Cloud Scenes Removing Moving Objects from Point Cloud Scenes Krystof Litomisky and Bir Bhanu University of California, Riverside krystof@litomisky.com, bhanu@ee.ucr.edu Abstract. Three-dimensional simultaneous localization

More information

Photorealistic 3D Mapping of Indoors by RGB-D Scanning Process

Photorealistic 3D Mapping of Indoors by RGB-D Scanning Process Photorealistic 3D Mapping of Indoors by RGB-D Scanning Process Tommi Tykkälä 1, Andrew I. Comport 2, and Joni-Kristian Kämäräinen 3 Abstract In this work, a RGB-D input stream is utilized for GPU-boosted

More information

MonoRGBD-SLAM: Simultaneous Localization and Mapping Using Both Monocular and RGBD Cameras

MonoRGBD-SLAM: Simultaneous Localization and Mapping Using Both Monocular and RGBD Cameras MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com MonoRGBD-SLAM: Simultaneous Localization and Mapping Using Both Monocular and RGBD Cameras Yousif, K.; Taguchi, Y.; Ramalingam, S. TR2017-068

More information

Geometric Reconstruction Dense reconstruction of scene geometry

Geometric Reconstruction Dense reconstruction of scene geometry Lecture 5. Dense Reconstruction and Tracking with Real-Time Applications Part 2: Geometric Reconstruction Dr Richard Newcombe and Dr Steven Lovegrove Slide content developed from: [Newcombe, Dense Visual

More information

arxiv: v1 [cs.cv] 28 Sep 2018

arxiv: v1 [cs.cv] 28 Sep 2018 Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,

More information

A Flexible Scene Representation for 3D Reconstruction Using an RGB-D Camera

A Flexible Scene Representation for 3D Reconstruction Using an RGB-D Camera 2013 IEEE International Conference on Computer Vision A Flexible Scene Representation for 3D Reconstruction Using an RGB-D Camera Diego Thomas National Institute of Informatics Chiyoda, Tokyo, Japan diego

More information

Motion Capture using Body Mounted Cameras in an Unknown Environment

Motion Capture using Body Mounted Cameras in an Unknown Environment Motion Capture using Body Mounted Cameras in an Unknown Environment Nam Vo Taeyoung Kim Siddharth Choudhary 1. The Problem Motion capture has been recently used to provide much of character motion in several

More information

Hierarchical Sparse Coded Surface Models

Hierarchical Sparse Coded Surface Models Hierarchical Sparse Coded Surface Models Michael Ruhnke Liefeng Bo Dieter Fox Wolfram Burgard Abstract In this paper, we describe a novel approach to construct textured 3D environment models in a hierarchical

More information

arxiv: v1 [cs.cv] 2 May 2016

arxiv: v1 [cs.cv] 2 May 2016 16-811 Math Fundamentals for Robotics Comparison of Optimization Methods in Optical Flow Estimation Final Report, Fall 2015 arxiv:1605.00572v1 [cs.cv] 2 May 2016 Contents Noranart Vesdapunt Master of Computer

More information

Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation

Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation ÖGAI Journal 24/1 11 Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation Michael Bleyer, Margrit Gelautz, Christoph Rhemann Vienna University of Technology

More information

Jakob Engel, Thomas Schöps, Daniel Cremers Technical University Munich. LSD-SLAM: Large-Scale Direct Monocular SLAM

Jakob Engel, Thomas Schöps, Daniel Cremers Technical University Munich. LSD-SLAM: Large-Scale Direct Monocular SLAM Computer Vision Group Technical University of Munich Jakob Engel LSD-SLAM: Large-Scale Direct Monocular SLAM Jakob Engel, Thomas Schöps, Daniel Cremers Technical University Munich Monocular Video Engel,

More information

Multi Channel Generalized-ICP

Multi Channel Generalized-ICP Multi Channel Generalized-ICP James Servos and Steven L. Waslander University of Waterloo, Waterloo, ON, Canada, N2L 3G1 Abstract Current state of the art scan registration algorithms which use only positional

More information

Robust Edge-based Visual Odometry using Machine-Learned Edges

Robust Edge-based Visual Odometry using Machine-Learned Edges Robust Edge-based Visual Odometry using Machine-Learned Edges Fabian Schenk and Friedrich Fraundorfer Abstract In this work, we present a real-time robust edgebased visual odometry framework for RGBD sensors

More information

Real-time Depth Enhanced Monocular Odometry

Real-time Depth Enhanced Monocular Odometry 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2014) September 14-18, 2014, Chicago, IL, USA Real-time Depth Enhanced Monocular Odometry Ji Zhang, Michael Kaess, and Sanjiv

More information

SDF-TAR: Parallel Tracking and Refinement in RGB-D Data using Volumetric Registration

SDF-TAR: Parallel Tracking and Refinement in RGB-D Data using Volumetric Registration M. SLAVCHEVA, S. ILIC: SDF-TAR 1 SDF-TAR: Parallel Tracking and Refinement in RGB-D Data using Volumetric Registration Miroslava Slavcheva mira.slavcheva@tum.de Slobodan Ilic slobodan.ilic@siemens.com

More information

Object Reconstruction

Object Reconstruction B. Scholz Object Reconstruction 1 / 39 MIN-Fakultät Fachbereich Informatik Object Reconstruction Benjamin Scholz Universität Hamburg Fakultät für Mathematik, Informatik und Naturwissenschaften Fachbereich

More information

Color Supported Generalized-ICP 0

Color Supported Generalized-ICP 0 Color Supported Generalized-ICP 0 Michael Korn, Martin Holzkothen and Josef Pauli Lehrstuhl Intelligente Systeme, Universität Duisburg-Essen, 47057 Duisburg, Germany michael.korn@uni-due.de Keywords: Abstract:

More information

Occluded Facial Expression Tracking

Occluded Facial Expression Tracking Occluded Facial Expression Tracking Hugo Mercier 1, Julien Peyras 2, and Patrice Dalle 1 1 Institut de Recherche en Informatique de Toulouse 118, route de Narbonne, F-31062 Toulouse Cedex 9 2 Dipartimento

More information

3D Object Recognition and Scene Understanding from RGB-D Videos. Yu Xiang Postdoctoral Researcher University of Washington

3D Object Recognition and Scene Understanding from RGB-D Videos. Yu Xiang Postdoctoral Researcher University of Washington 3D Object Recognition and Scene Understanding from RGB-D Videos Yu Xiang Postdoctoral Researcher University of Washington 1 2 Act in the 3D World Sensing & Understanding Acting Intelligent System 3D World

More information

Deformation-based Loop Closure for Large Scale Dense RGB-D SLAM

Deformation-based Loop Closure for Large Scale Dense RGB-D SLAM 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) November 3-7, 2013. Tokyo, Japan Deformation-based Loop Closure for Large Scale Dense RGB-D SLAM Thomas Whelan 1, Michael

More information

Real-Time RGB-D Registration and Mapping in Texture-less Environments Using Ranked Order Statistics

Real-Time RGB-D Registration and Mapping in Texture-less Environments Using Ranked Order Statistics Real-Time RGB-D Registration and Mapping in Texture-less Environments Using Ranked Order Statistics Khalid Yousif 1, Alireza Bab-Hadiashar 2, Senior Member, IEEE and Reza Hoseinnezhad 3 Abstract In this

More information

Accurate 3D Face and Body Modeling from a Single Fixed Kinect

Accurate 3D Face and Body Modeling from a Single Fixed Kinect Accurate 3D Face and Body Modeling from a Single Fixed Kinect Ruizhe Wang*, Matthias Hernandez*, Jongmoo Choi, Gérard Medioni Computer Vision Lab, IRIS University of Southern California Abstract In this

More information

Using the Kinect as a Navigation Sensor for Mobile Robotics

Using the Kinect as a Navigation Sensor for Mobile Robotics Using the Kinect as a Navigation Sensor for Mobile Robotics ABSTRACT Ayrton Oliver Dept. of Electrical and Computer Engineering aoli009@aucklanduni.ac.nz Burkhard C. Wünsche Dept. of Computer Science burkhard@cs.auckland.ac.nz

More information

A Catadioptric Extension for RGB-D Cameras

A Catadioptric Extension for RGB-D Cameras A Catadioptric Extension for RGB-D Cameras Felix Endres Christoph Sprunk Rainer Kümmerle Wolfram Burgard Abstract The typically restricted field of view of visual sensors often imposes limitations on the

More information

FLaME: Fast Lightweight Mesh Estimation using Variational Smoothing on Delaunay Graphs

FLaME: Fast Lightweight Mesh Estimation using Variational Smoothing on Delaunay Graphs FLaME: Fast Lightweight Mesh Estimation using Variational Smoothing on Delaunay Graphs W. Nicholas Greene Robust Robotics Group, MIT CSAIL LPM Workshop IROS 2017 September 28, 2017 with Nicholas Roy 1

More information

arxiv: v1 [cs.ro] 24 Nov 2018

arxiv: v1 [cs.ro] 24 Nov 2018 BENCHMARKING AND COMPARING POPULAR VISUAL SLAM ALGORITHMS arxiv:1811.09895v1 [cs.ro] 24 Nov 2018 Amey Kasar Department of Electronics and Telecommunication Engineering Pune Institute of Computer Technology

More information

Perceiving the 3D World from Images and Videos. Yu Xiang Postdoctoral Researcher University of Washington

Perceiving the 3D World from Images and Videos. Yu Xiang Postdoctoral Researcher University of Washington Perceiving the 3D World from Images and Videos Yu Xiang Postdoctoral Researcher University of Washington 1 2 Act in the 3D World Sensing & Understanding Acting Intelligent System 3D World 3 Understand

More information

Inertial-Kinect Fusion for Outdoor 3D Navigation

Inertial-Kinect Fusion for Outdoor 3D Navigation Proceedings of Australasian Conference on Robotics and Automation, 2-4 Dec 213, University of New South Wales, Sydney Australia Inertial-Kinect Fusion for Outdoor 3D Navigation Usman Qayyum and Jonghyuk

More information

Outline. 1 Why we re interested in Real-Time tracking and mapping. 3 Kinect Fusion System Overview. 4 Real-time Surface Mapping

Outline. 1 Why we re interested in Real-Time tracking and mapping. 3 Kinect Fusion System Overview. 4 Real-time Surface Mapping Outline CSE 576 KinectFusion: Real-Time Dense Surface Mapping and Tracking PhD. work from Imperial College, London Microsoft Research, Cambridge May 6, 2013 1 Why we re interested in Real-Time tracking

More information

Robustness to Lighting Variations: An RGB-D Indoor Visual Odometry Using Line Segments

Robustness to Lighting Variations: An RGB-D Indoor Visual Odometry Using Line Segments Robustness to Lighting Variations: An RGB-D Indoor Visual Odometry Using Line Segments Yan Lu and Dezhen Song Abstract Large lighting variation challenges all visual odometry methods, even with RGB-D cameras.

More information

arxiv: v1 [cs.ro] 23 Feb 2018

arxiv: v1 [cs.ro] 23 Feb 2018 IMLS-SLAM: scan-to-model matching based on 3D data Jean-Emmanuel Deschaud 1 1 MINES ParisTech, PSL Research University, Centre for Robotics, 60 Bd St Michel 75006 Paris, France arxiv:1802.08633v1 [cs.ro]

More information

Visual Odometry Algorithm Using an RGB-D Sensor and IMU in a Highly Dynamic Environment

Visual Odometry Algorithm Using an RGB-D Sensor and IMU in a Highly Dynamic Environment Visual Odometry Algorithm Using an RGB-D Sensor and IMU in a Highly Dynamic Environment Deok-Hwa Kim, Seung-Beom Han, and Jong-Hwan Kim Department of Electrical Engineering, KAIST 29 Daehak-ro, Yuseong-gu,

More information

arxiv: v1 [cs.cv] 1 Aug 2017

arxiv: v1 [cs.cv] 1 Aug 2017 Dense Piecewise Planar RGB-D SLAM for Indoor Environments Phi-Hung Le and Jana Kosecka arxiv:1708.00514v1 [cs.cv] 1 Aug 2017 Abstract The paper exploits weak Manhattan constraints to parse the structure

More information

METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS

METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS M. Lefler, H. Hel-Or Dept. of CS, University of Haifa, Israel Y. Hel-Or School of CS, IDC, Herzliya, Israel ABSTRACT Video analysis often requires

More information

RGB-D Edge Detection and Edge-based Registration

RGB-D Edge Detection and Edge-based Registration 23 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) November 3-7, 23. Tokyo, Japan RGB-D Edge Detection and Edge-based Registration Changhyun Choi, Alexander J. B. Trevor, and

More information

Toward Online 3-D Object Segmentation and Mapping

Toward Online 3-D Object Segmentation and Mapping Toward Online 3-D Object Segmentation and Mapping Evan Herbst Peter Henry Dieter Fox Abstract We build on recent fast and accurate 3-D reconstruction techniques to segment objects during scene reconstruction.

More information

The Lucas & Kanade Algorithm

The Lucas & Kanade Algorithm The Lucas & Kanade Algorithm Instructor - Simon Lucey 16-423 - Designing Computer Vision Apps Today Registration, Registration, Registration. Linearizing Registration. Lucas & Kanade Algorithm. 3 Biggest

More information

Approximate Surface Reconstruction and Registration for RGB-D SLAM

Approximate Surface Reconstruction and Registration for RGB-D SLAM European Conference on Mobile Robotics (ECMR), Lincoln, UK, September 2015 Approximate Surface Reconstruction and Registration for RGB-D SLAM Dirk Holz and Sven Behnke Abstract RGB-D cameras have attracted

More information

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction Ham Rara, Shireen Elhabian, Asem Ali University of Louisville Louisville, KY {hmrara01,syelha01,amali003}@louisville.edu Mike Miller,

More information

CopyMe3D: Scanning and Printing Persons in 3D

CopyMe3D: Scanning and Printing Persons in 3D CopyMe3D: Scanning and Printing Persons in 3D Sturm, Jürgen; Bylow, Erik; Kahl, Fredrik; Cremers, Daniel Published in: Lecture Notes in Computer Science Vol. 8142 (Pattern Recognition, 35th German Conference,

More information

Simultaneous Localization and Mapping (SLAM)

Simultaneous Localization and Mapping (SLAM) Simultaneous Localization and Mapping (SLAM) RSS Lecture 16 April 8, 2013 Prof. Teller Text: Siegwart and Nourbakhsh S. 5.8 SLAM Problem Statement Inputs: No external coordinate reference Time series of

More information

SPURRED by the ready availability of depth sensors and

SPURRED by the ready availability of depth sensors and IEEE ROBOTICS AND AUTOMATION LETTERS. PREPRINT VERSION. ACCEPTED DECEMBER, 2015 1 Hierarchical Hashing for Efficient Integration of Depth Images Olaf Kähler, Victor Prisacariu, Julien Valentin and David

More information

Dual Back-to-Back Kinects for 3-D Reconstruction

Dual Back-to-Back Kinects for 3-D Reconstruction Ho Chuen Kam, Kin Hong Wong and Baiwu Zhang, Dual Back-to-Back Kinects for 3-D Reconstruction, ISVC'16 12th International Symposium on Visual Computing December 12-14, 2016, Las Vegas, Nevada, USA. Dual

More information

3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller

3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller 3D Computer Vision Depth Cameras Prof. Didier Stricker Oliver Wasenmüller Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de

More information

AN INDOOR SLAM METHOD BASED ON KINECT AND MULTI-FEATURE EXTENDED INFORMATION FILTER

AN INDOOR SLAM METHOD BASED ON KINECT AND MULTI-FEATURE EXTENDED INFORMATION FILTER AN INDOOR SLAM METHOD BASED ON KINECT AND MULTI-FEATURE EXTENDED INFORMATION FILTER M. Chang a, b, Z. Kang a, a School of Land Science and Technology, China University of Geosciences, 29 Xueyuan Road,

More information

CuFusion: Accurate real-time camera tracking and volumetric scene reconstruction with a cuboid

CuFusion: Accurate real-time camera tracking and volumetric scene reconstruction with a cuboid CuFusion: Accurate real-time camera tracking and volumetric scene reconstruction with a cuboid Chen Zhang * and Yu Hu College of Computer Science and Technology, Zhejiang University, Hangzhou 310027, China;

More information

Lecture 10 Multi-view Stereo (3D Dense Reconstruction) Davide Scaramuzza

Lecture 10 Multi-view Stereo (3D Dense Reconstruction) Davide Scaramuzza Lecture 10 Multi-view Stereo (3D Dense Reconstruction) Davide Scaramuzza REMODE: Probabilistic, Monocular Dense Reconstruction in Real Time, ICRA 14, by Pizzoli, Forster, Scaramuzza [M. Pizzoli, C. Forster,

More information

Surface Registration. Gianpaolo Palma

Surface Registration. Gianpaolo Palma Surface Registration Gianpaolo Palma The problem 3D scanning generates multiple range images Each contain 3D points for different parts of the model in the local coordinates of the scanner Find a rigid

More information

Motion Estimation. There are three main types (or applications) of motion estimation:

Motion Estimation. There are three main types (or applications) of motion estimation: Members: D91922016 朱威達 R93922010 林聖凱 R93922044 謝俊瑋 Motion Estimation There are three main types (or applications) of motion estimation: Parametric motion (image alignment) The main idea of parametric motion

More information

The Template Update Problem

The Template Update Problem The Template Update Problem Iain Matthews, Takahiro Ishikawa, and Simon Baker The Robotics Institute Carnegie Mellon University Abstract Template tracking dates back to the 1981 Lucas-Kanade algorithm.

More information

CRF Based Point Cloud Segmentation Jonathan Nation

CRF Based Point Cloud Segmentation Jonathan Nation CRF Based Point Cloud Segmentation Jonathan Nation jsnation@stanford.edu 1. INTRODUCTION The goal of the project is to use the recently proposed fully connected conditional random field (CRF) model to

More information

Incremental Structured ICP Algorithm

Incremental Structured ICP Algorithm Incremental Structured ICP Algorithm Haokun Geng, Johnny Chien, Radu Nicolescu, and Reinhard Klette The.enpeda.. Project, Tamaki Campus The University of Auckland, New Zealand Abstract. Variants of the

More information

Mesh from Depth Images Using GR 2 T

Mesh from Depth Images Using GR 2 T Mesh from Depth Images Using GR 2 T Mairead Grogan & Rozenn Dahyot School of Computer Science and Statistics Trinity College Dublin Dublin, Ireland mgrogan@tcd.ie, Rozenn.Dahyot@tcd.ie www.scss.tcd.ie/

More information

ABSTRACT. KinectFusion is a surface reconstruction method to allow a user to rapidly

ABSTRACT. KinectFusion is a surface reconstruction method to allow a user to rapidly ABSTRACT Title of Thesis: A REAL TIME IMPLEMENTATION OF 3D SYMMETRIC OBJECT RECONSTRUCTION Liangchen Xi, Master of Science, 2017 Thesis Directed By: Professor Yiannis Aloimonos Department of Computer Science

More information

Lucas-Kanade Image Registration Using Camera Parameters

Lucas-Kanade Image Registration Using Camera Parameters Lucas-Kanade Image Registration Using Camera Parameters Sunghyun Cho a, Hojin Cho a, Yu-Wing Tai b, Young Su Moon c, Junguk Cho c, Shihwa Lee c, and Seungyong Lee a a POSTECH, Pohang, Korea b KAIST, Daejeon,

More information

Stereo and Epipolar geometry

Stereo and Epipolar geometry Previously Image Primitives (feature points, lines, contours) Today: Stereo and Epipolar geometry How to match primitives between two (multiple) views) Goals: 3D reconstruction, recognition Jana Kosecka

More information

Efficient Onboard RGBD-SLAM for Autonomous MAVs

Efficient Onboard RGBD-SLAM for Autonomous MAVs Efficient Onboard RGBD-SLAM for Autonomous MAVs Sebastian A. Scherer 1 and Andreas Zell 1 Abstract We present a computationally inexpensive RGBD- SLAM solution taylored to the application on autonomous

More information

Intrinsic3D: High-Quality 3D Reconstruction by Joint Appearance and Geometry Optimization with Spatially-Varying Lighting

Intrinsic3D: High-Quality 3D Reconstruction by Joint Appearance and Geometry Optimization with Spatially-Varying Lighting Intrinsic3D: High-Quality 3D Reconstruction by Joint Appearance and Geometry Optimization with Spatially-Varying Lighting R. Maier 1,2, K. Kim 1, D. Cremers 2, J. Kautz 1, M. Nießner 2,3 Fusion Ours 1

More information

A Proof that Fusing Measurements Using Point-to-Hyperplane Registration is Invariant to Relative Scale

A Proof that Fusing Measurements Using Point-to-Hyperplane Registration is Invariant to Relative Scale A Proof that Fusing Measurements Using Point-to-Hyperplane Registration is Invariant to Relative Scale Fernando Ireta Munoz, Andrew Comport To cite this version: Fernando Ireta Munoz, Andrew Comport. A

More information

A Low Power, High Throughput, Fully Event-Based Stereo System: Supplementary Documentation

A Low Power, High Throughput, Fully Event-Based Stereo System: Supplementary Documentation A Low Power, High Throughput, Fully Event-Based Stereo System: Supplementary Documentation Alexander Andreopoulos, Hirak J. Kashyap, Tapan K. Nayak, Arnon Amir, Myron D. Flickner IBM Research March 25,

More information