RGBD Point Cloud Alignment Using Lucas-Kanade Data Association and Automatic Error Metric Selection
|
|
- Leslie Bennett
- 6 years ago
- Views:
Transcription
1 IEEE TRANSACTIONS ON ROBOTICS, VOL. 31, NO. 6, DECEMBER 15 1 RGBD Point Cloud Alignment Using Lucas-Kanade Data Association and Automatic Error Metric Selection Brian Peasley and Stan Birchfield, Senior Member, IEEE Abstract We propose to overcome a significant limitation of the ICP algorithm used by KinectFusion, namely, its sole reliance upon geometric information. Our approach uses both geometric and color information in a direct manner that uses all the data in order to accurately estimate camera pose. Data association is performed by Lucas-Kanade to compute an affine warp between the color images associated with two RGBD point clouds. A subsequent step then estimates the Euclidean transformation between the point clouds using either a point-to-point or point-to-plane error metric, with a novel method based on a normal covariance test for automatically selecting between them. Together, Lucas- Kanade data association (LKDA) with covariance testing enables robust camera tracking through areas of low geometric features, without sacrificing accuracy in environments in which the existing ICP technique succeeds. Experimental results on several publicly available datasets demonstrate the improved performance both qualitatively and quantitatively. Index Terms Kinect, KinectFusion, ICP, Lucas-Kanade, Camera Tracking, Mapping I. INTRODUCTION THREE dimensional (3D) reconstruction of an environment is an important problem in robotics that has received much attention in recent years. The advent of inexpensive RGBD sensing (e.g., the Kinect sensor) has greatly improved the fidelity and speed with which such reconstructions can be made. These capabilities have far-reaching implications for the exploration and navigation of unknown environments, as well as for the manipulation of objects in those environments. Of the many 3D modeling techniques that have been developed, the landmark KinectFusion method [19], [13] has established itself as perhaps the most accurate, real-time, dense modeling system using an inexpensive sensor. Recent work by others [9] has extended this algorithm to operate over large-scale environments. However, both the standard and the extended version use geometric information alone to align the camera with the model, thus requiring the environment to contain sufficient geometric features in order for the iterative closest point (ICP) algorithm to accurately estimate the camera pose. This deficiency was noted in [9], where preliminary experiments were conducted to explore the advantage of replacing the ICP algorithm with the output of a featurebased visual odometry system called FOVIS [11] to provide a more stable estimate of camera pose. The -to- approach of FOVIS, however, loses one of the key advantages of KinectFusion, namely the reduced drift that results from matching the current image to the model rather than to the previous image. In this paper we present an approach to RGBD camera tracking in environments with low geometric information. The proposed approach uses all the data (both color and depth), preserves the advantages of KinectFusion s image-to-model matching, and obviates the need for the feature point extraction and feature correspondence steps inherent in feature-based visual odometry systems. The key to the approach is to replace the projective data association (PDA) point matching algorithm [3] used by KinectFusion with a data association technique driven by Lucas-Kanade [18], [6], []. We introduced this matching algorithm, which we call Lucas-Kanade data association (LKDA), in [1]. In this paper we extend that work with an automatic method for selecting the appropriate error metric based on the geometry of the scene, a GPU implementation of the matching, and more thorough experimental results to demonstrate the improvement that results from LKDA as well as to compare it with alternative techniques. These results on standard datasets demonstrate that LKDA combined with the automatic error metric enable camera tracking to succeed in areas of low geometry, without sacrificing either computational efficiency or accurate camera tracking in highly geometric environments. 1 II. PREVIOUS WORK Before the advent of recent inexpensive RGBD sensors, several researchers addressed the problem of improving ICP by incorporating color information. Both geometric and color information are incorporated into the distance metric used by ICP in a number of systems [5], [1], [14]. Similarly, Druon et al. [6] segment colored point clouds based on hue, then perform ICP while requiring matching points to belong to the same color class. More recently, a popular way to use visual information to improve alignment is to match feature points, then apply RANSAC to the features with depth values. This is the approach employed by Henry et al. [1], in which the visual feature associations from RANSAC are then combined with dense point associations in an ICP work. This same procedure of feature matching followed by RANSAC is adopted in both the FOVIS visual odometry system [11] and the algorithm of 1 Although our LKDA approach will fail when there is no visual texture (e.g., a white wall), such scenes have low geometric variation as well, since geometry usually generates visual texture via shading and shadows.
2 IEEE TRANSACTIONS ON ROBOTICS, VOL. 31, NO. 6, DECEMBER 15 Engelhard et al. [8], the latter of which also employs ICP and is evaluated in [7]. As expected, the primary drawback of these techniques is the failure of feature detection and matching when the scene contains little photometric texture. An approach that is closely related to our own is to register RGBD images via direct image warping. Tykkälä et al. [7] formulate the 3D registration of colored point clouds as a direct minimization of both depth and color information. To reduce the computational load, in followup work the same researchers [1] minimize only the image intensities, without using depth. In related work, Steinbrücker et al. [4] align two colored point clouds by employing Lie algebra to solve for the twist coordinates representing the 3D transformation that maximizes the photoconsistency of the projections. This work has been extended by Kerl et al. [16] to achieve realtime performance (3 Hz) on a single CPU, as well as to combine both photoconsistency and depth consistency in the minimization process [15]. Another direct approach is DTAM [], which uses dense methods for whole image aligment without a depth sensor but stores the model as a set of keys rather than as a dense volume. Other researchers have focused on extending KinectFusion. Henry et al. [9] propose a system consisting of local patch volumes, which are similar to KinectFusion volumes, connected together in a graph optimization work to ensure global consistency. In another approach that generates large maps, Whelan et al. [8] combine KinectFusion s ICP algorithm with a GPU implementation of the RGBD alignment approach of [4], and they provide a switching mechanism between the resulting estimate and that of the feature-based FOVIS. As expected, they show that both FOVIS and RGBD alignment fail when the scene contains little photometric texture, KinectFusion s ICP fails when there is little geometric texture, and a combination of these methods yields improved tracking robustness. Like the work of Whelan et al. [8], we focus upon improving the camera alignment step of KinectFusion by considering both visual and depth information. However, instead of combining off-the-shelf algorithms with a switching mechanism between them, we derive a novel algorithm that first aligns the intensity images then uses the resulting correspondences to align the depth images with an automatic selection of either point-to-point or point-to-plane error metrics based on the scene geometry. In Section V we show that our relatively straightforward, direct approach produces more consistent results than those obtained by more complicated methods, without sacrificing computational speed. III. LUCAS-KANADE DATA ASSOCIATION Let P = {p i } N i=1 and Q = {q i} N i=1, where p i,q i R 3, be two point clouds in 3D Euclidean space. The goal of data association is to compute a function f that maps indices in one point cloud to those of the other, so that f(i) = j indicates a corresponding pair of points p i q j. Let I P and I Q be the current and reference RGB images, respectively, associated with the point clouds, and d P and d Q the current and reference depth maps. In our implementation I Q is the previous image, Fig. 1. LEFT: PDA establishes correspondence between two point clouds, P and Q, by projecting each point p i P of one cloud onto the other s depth camera, D Q, to find the nearest projection of q j Q. RIGHT: LKDA establishes correspondence by projecting each point p i onto the RGB camera C P, then warping the projected point onto the other RGB camera C Q using the estimated warp from Lucas-Kanade to find the nearest projection of q j. (The normals n j are used in the point-to-plane error metric.) while d Q arises from projecting the truncated signed distance function (TSDF) according to the previous camera pose. Projective data association (PDA) [17] establishes correspondence between P and Q by projecting the points from both point clouds onto the same image plane, then considering points that project onto the same pixel to be corresponding points. See Figure 1. In a similar manner, Lucas-Kanade data association (LKDA) also finds the closest point on the image plane, but only after first transforming the projected coordinates according to the warp function found by Lucas- Kanade: f(i) = argmin j ϕ(c Q q j ) ϕ( W(C P p i ;ζ)), (1) where ζ contains the warp parameters, C P and C Q are 4 4 homogeneous matrices describing the 3D position and orientation of the RGB cameras, the tilde indicates homogeneous coordinates, and ϕ dehomogenizes the coordinates. The PDA approach relies solely upon geometric information, that is, the spatial coordinates of the points in the two clouds. As a result, when the environment does not provide sufficient geometric texture (e.g., the camera is looking at a planar wall), or when most of the depth readings are invalid due to the limited range of the sensor or occur outside the TSDF maintained by KinectFusion (e.g., the camera looks down a hallway), PDA will fail to provide enough correspondences to track the camera. The LKDA approach improves the alignment of point clouds in such environments, since the photometric information is not affected by the scene geometry, the range of the sensor, or the size of the TSDF. The Lucas-Kanade algorithm is a differential method for computing the optical flow of an image by finding the parameters ζ that minimize the sum of squared distance error ǫ LK = x ( IQ (W 1 (x;ζ)) I P (x) ), () where the summation is over all the pixels x = [ x y ] T in the image, and W(x;ζ) is a parametric warp function that brings the two images into alignment. This error is minimized by linearizing about the current estimate and repeatedly solving
3 IEEE TRANSACTIONS ON ROBOTICS, VOL. 31, NO. 6, DECEMBER 15 3 the following linear system ( ) gg T ζ = g ( I Q (W 1 (x;ζ)) I P (x) ) (3) x x ( ) T W for ζ, where g = ζ IP, and I Q contains the gradient of I P. For efficiency, we use the inverse compositional algorithm []. Common choices for the warp function are translation, affine, or projective. Since we are interested in a global, featureless mapping technique that warps an entire image into another, the translation model is not expressive enough; either an affine or projective warp is thus required. We have found little difference in accuracy between these two alternatives (see the experimental results), but the projective warp incurs a substantial increase in computation due to the fact that the Hessian is dependent upon the current state of the system and therefore must be recomputed each iteration. As a result we use an affine warp, which is given by [ x(rxx +1)+yr W(x,y;ζ) = xy +a x xr yx +y(r yy +1)+a y where ζ = [ r xx r xy r yx r yy a x a y ] T. IV. ESTIMATING THE CAMERA POSE ], (4) Each iteration of ICP can be broken into six steps, and different choices within these steps lead to different variations of the algorithm []. Figure illustrates these steps within the traditional ICP approach used by KinectFusion, compared with our LKDA approach. Instead of recomputing data association after each alignment, we first perform data association, then perform alignment. This decoupling, which is enabled by our separate use of RGB and depth images, leads to significant speedup without sacrificing robustness. At the same time, the correspondences used in each iteration of the algorithm are not necessarily the same, because the rejection step retains only a subset of them. In this section we detail the steps that are different between our approach and KinectFusion. A. Point Matching As already mentioned, projective data association (PDA) establishes correspondence by finding the closest point from the other cloud as projected onto the image plane. In a similar manner, Lucas-Kanade data association (LKDA) also finds the closest point on the image plane, but only after first transforming the projected coordinates according to the warp function found by Lucas-Kanade. Unlike traditional ICP, however, the LKDA point matching step is performed only once per pair of image s, and subsets of the correspondences found in this step are then used in the iterations of the alignment process. B. Outlier rejection As in [19], our approach rejects point correspondences that exceed a distance or angular threshold. In addition, corresponding points whose color difference exceeds a predefined (a) ICP implementation in KinectFusion (b) Our LKDA approach Fig.. TOP: Given the current and reference depth images, KinectFusion iterates through the six ICP steps. Steps 1-3 perform data association (PDA), while steps 4-6 perform alignment. BOTTOM: In contrast, our approach iteratively finds the correspondences using the RGB images (LKDA, steps 1-3), then iteratively estimates the alignment parameters using the depth images (steps 4-6). The steps marked with an asterisk (*) are different in the two approaches. color threshold ρ are rejected: I Q (x,y) I P (W(x,y;ζ)) > ρ. (5) Note that the rejected points in each iteration of the alignment process are not necessarily the same, so that a different subset of points contribute to the error metric in each iteration. In our implementation the distance, angular, and color thresholds are.1 m,, and ρ = 5, respectively. C. Automatic Selection of the Error Metric Two common choices of the metric used for alignment are point-to-point and point-to-plane [4]. If there is sufficient geometric texture in the environment, then the latter converges faster than the former [], which explains its adoption by KinectFusion. On the other hand, if all point correspondences are on a flat wall then the point-to-plane error metric will not result in correct alignment, because there will be no mechanism to induce a lateral shift between the point clouds in the direction perpendicular to the normals. See Figure 3. To overcome this limitation we use a linear combination of the 3D Euclidean transformations computed using the pointto-point (T pp ) and the point-to-plane (T pπ ) error metric: T = (1 λ)t pp +λt pπ, (6) where λ 1 is a weighting factor. The value of λ is determined automatically at run time by solving for the condition number κ of the covariance matrix of the geometric normals seen by the current. In areas of low geometry the value of γ = κ 1 is low, while in areas of high geometry γ is high. We model λ as a sigmoid ( λ = 1+e α(γ ˆγ)) 1, (7)
4 IEEE TRANSACTIONS ON ROBOTICS, VOL. 31, NO. 6, DECEMBER 15 4 D. Error minimization Given the correspondences between the two image s that survived the rejection step, along with the value λ, the two point clouds can be aligned. The transformation T pp is computed in closed-form using orthogonal Procrustes analysis, while T pπ is computed iteratively as described in [17]. The camera pose is then given by (6). Fig. 3. When the scene is planar, the point-to-plane error metric fails to correctly align point clouds, even with perfect correspondence. error (meters) inverse condition number point to point point to plane Fig. 4. Scatter plot of γ, the inverse of the condition number of the covariance matrix of the normals, versus the error of the estimated translation (averaged over all s of several sequences). The red dots indicate the results using the point-to-plane error metric (T pπ) to estimate odometry, while the blue dots show the results from the point-to-point error metric (T pp). Robust exponential curves were fit to both data sets to determine their intersection. where α is a large constant (determined experimentally to be 3), and ˆγ indicates the value of γ for which the two metrics are weighted equally, i.e., λ =.5. To determine ˆγ, for each over several sequences we computed T pπ, T pp, and κ 1, comparing the former two with the ground truth. The resulting scatter plot can been seen in Figure 4, along with the intersection of robust exponential curves fit to both data sets, yielding ˆγ =.1. Figure 5 shows the results of both metrics on two scenes, illustrating that T pp yields less error on scenes with low geometry, while T pπ tends to perform slightly better on scenes with high geometry. cumulative error point to plane point to point 1 3 cumulative error point to plane point to point 1 3 Fig. 5. Cumulative camera translation error for both metrics when reconstructing a relatively flat environment (left) versus a geometrically rich environment (right). V. EXPERIMENTAL RESULTS To implement the proposed approach, we modified the opensource version of KinectFusion known as KinFu in the PCL library [3] with our own C++ and GPGPU code. In our first set of experiments we compared the unmodified PDA version with six different versions of LKDA using either grayscale or color images, and either a translation, affine, or projective warp. To facilitate quantitative comparison, we used the data from the RGB-D SLAM Dataset and Benchmark [5], which provides synchronized depth and color data from a Kinect or Xtion sensor as well as the ground truth sensor trajectory from a high-accuracy motion capture system. For these experiments we used 5 different datasets exhibiting a variety of small environments, which can roughly be categorized as follows: high-geometry. The environment contains a large amount of geometric variation throughout, enabling geometrybased sensor tracking. (fr1/xyz, fr1/floor, fr1/desk) planar. The environment is largely planar (that is, without much geometric variation), making geometry-based sensor tracking difficult. (fr3/ntnw 3 ) out-of-bounds. For at least some s of the sequence, the sensor is located so that a majority of the depth values are outside the TSDF volume. (fr1/rpy) We used a m TSDF volume divided into 51 3 voxels (each voxel is 5.86 mm per side). The results on these 5 datasets are shown in Table I, where the RMSE translational drift (in meters per second) and RMSE rotational drift (in radians per second) are displayed. The drift is also known as the Relative Pose Error (RPE) [5]. For the high-geometry sequences, all the methods succeed, and there is little difference in accuracy in either position or orientation. For the planar (fr3/ntnw) or out-of-bounds (fr1/rpy) sequences, however, the standard KinectFusion algorithm fails to maintain sensor tracking. There is a noticeable improvement using affine over translation, but almost none by using projective over affine. To save space we show only the grayscale results, since the color results are nearly identical. We directly compared our method with the approach of Whelan et al. [8] and the Dense Visual Odometry (DVO) algorithm of Kerl et al. [15], [16]. Results of these comparisons are displayed in Tables II and III, respectively. 4 While the algorithm of [8] performs better on a -to- basis, our approach yields a better overall estimated trajectory, leading 3 A shortened form of the name fr3/nostructure texture near withloop. 4 To reduce space, Table II only includes results from the FOVIS/ICP+RGB- D version of the algorithm in [8]; other versions are similar. For fair comparison, the final column of Table III shows the RGB+D+KF version of [15], which includes keys but not graph optimization.
5 IEEE TRANSACTIONS ON ROBOTICS, VOL. 31, NO. 6, DECEMBER 15 5 fr1/xyz fr1/floor fr1/desk fr3/ntnw fr1/rpy PDA F F (.31) (.58) (.7) (.461) F (1.87) F LKDA F.61 Translation (.31) (.63) (.7) (.533) F (.57) LKDA Affine (.3) (.56) (.71) (.31) (.57) LKDA Projective (.3) (.57) (.71) (.31) (.56) TABLE I RPE RMSE IN TRANSLATION (M/S) (TOP OF EACH CELL) AND ROTATION (RAD/S) (BOTTOM, IN PARENTHESES). THE BEST RESULT IS IN BOLD, AND FAILED TRACKING RESULTS ARE INDICATED BY F. LKDA Whelan ATE RPE ATE RPE med max RMSE med max RMSE fr1/desk fr/desk fr1/room fr/lnl TABLE II COMPARISON WITH WHELAN ET AL. [8], SHOWING MEDIAN AND MAXIMUM VALUES OF THE ABSOLUTE TRANSLATION ERROR (ATE) (M), AND RMSE TRANSLATIONAL DRIFT (M/S). KinectFusion with PDA to lower absolute trajectory errors. Compared with [15], our approach is competitive, yielding the lowest error on several sequences. Plots of the estimated camera position versus number for the planar (fr3/ntnw) and out-of-bounds (fr1/rpy) environments for the different algorithms are shown in Figure 6. Notice that KinectFusion with PDA is unable to maintain tracking in such environments, while our LKDA algorithm (grayscale affine) and the approaches of [8] and [15] are able to maintain camera tracking. Other variations of LKDA also perform well, except that the translation motion model is insufficient for the planar scene. Reconstructions of the outof-bounds environment by rendering the point cloud from two approaches are shown in Figure 7 for comparison. Notice that LKDA achieves an accurate reconstruction of the scene, while the reconstruction from PDA is difficult to interpret. The importance of the automatic error selection is shown in Table IV. Seven sequences were run using either a pointto-point or point-to-plane error metric, or with automatic selection between the two. As expected, the point-to-point error LKDA DVO ATE RPE ATE RPE RMSE RMSE RMSE RMSE fr1/desk fr1/desk fr1/room fr1/ fr1/teddy fr1/floor fr1/xyz fr1/rpy fr1/plant TABLE III COMPARISON WITH DVO [15], SHOWING THE RMSE ABSOLUTE TRAJECTORY ERROR (M) AND THE RMSE TRANSLATIONAL DRIFT (M/S). KinectFusion with LKDA Fig. 7. Reconstruction of the out-of-bounds environment (fr1/rpy) by standard KinectFusion with PDA (top) and KinectFusion with LKDA (bottom). The inability of the former to handle depth values outside the TSDF volume causes a severe loss in camera tracking, thus resulting in a poor reconstruction; while our approach yields accurate reconstruction. High Low point-to-point point-to-plane automatic fr1/rpy fr1/floor fr1/teddy fr1/xyz fr1/plant fr3/ntnw fr3/ntf TABLE IV COMPARISON OF THE RMSE ATE OF THE POINT-TO-POINT ERROR METRIC (LEFT) AGAINST THE POINT-TO-PLANE ERROR METRIC (MIDDLE). THE FORMER PERFORMED BETTER IN SCENES OF LOW GEOMETRY, WHILE THE LATTER PERFORMED BETTER IN SCENES WITH HIGH GEOMETRY. OUR AUTOMATIC SELECTION METHOD HANDLES BOTH TYPES OF SCENES. metric performed better in environments with little geometrical structure, while the point-to-plane error metric always yielded less error on sequences with high geometrical structure. The automatic selection mechanism enabled our approach to yield the minimum error on all sequences. Timing comparisons of the different variations of our algorithm were conducted on several machines with varying computational power and GPUs. For standard KinectFusion
6 IEEE TRANSACTIONS ON ROBOTICS, VOL. 31, NO. 6, DECEMBER 15 6 X (meters) X (meters) PDA + KF Whelan et al. DVO LKDA + KF Ground Truth Y (meters) Y (meters) Z (meters) 1.5 Z (meters) Fig. 6. Camera position estimation in a planar environment (fr3/ntnw, top) and out-of-bounds environment (fr1/rpy, bottom). Shown is a comparison of the proposed grayscale affine LKDA, KinectFusion with PDA, the approach of Whelan et al. [8], and DVO [15]. The solid black line is ground truth. with PDA, nearly all the computation (approximately 85%) is devoted to the alignment, with the remaining computation required for PDA. Our approach retains the alignment computation but replaces PDA with a more expensive LKDA step. On average, the ratio of the time taken by LKDA to PDA is 3:1 (grayscale translation), 6:1 (grayscale affine), and 45:1 (grayscale projective). (Color versions require nearly 3 times as much computation.) Therefore, our grayscale affine approach (the recommended version) is about 1.8 times as expensive as KinectFusion (since (6) = 1.8). On a machine capable of running standard KinectFusion at 3 Hz, our algorithm runs at approximately 17 Hz; on a machine that runs KinectFusion at 1 Hz, our algorithm runs at nearly 6 Hz. VI. CONCLUSION In this paper we have presented a direct technique to robustly estimate camera motion using an RGBD sensor, even when there is little geometric information in the scene. The approach retains the drift-free advantages of the matchto-model approach of KinectFusion, without requiring the estimation of visual features. The approach operates by incrementally computing the affine warp between successive images using Lucas-Kanade, yielding a correspondence between point clouds. This approach to correspondence, which we call LKDA, replaces the PDA technique implemented in the variant of ICP found in KinectFusion. In contrast with PDA, LKDA was shown on a number of publicly available sequences that it is able to maintain camera tracking in environments with limited geometric structure, as well as in environments that extend beyond the TSDF volume or contain depths outside the range of the sensor. Moreover, we have proposed an automatic mechanism for selecting between point-to-point or point-to-plane error metrics that properly handles a variety of environments. Future work will be aimed at extending this approach to large-scale environments. VII. ACKNOWLEDGMENTS The authors thank T. Whelan and C. Kerl for providing their data for comparison, and the anonymous reviewers for their helpful comments. This work was partially supported by NSF grant IIS REFERENCES [1] C. Audras, A. I. Comport, M. Meilland, and P. Rives. Real-time dense RGB-D localisation and mapping. In Australian Conference on Robotics and Automation, Dec. 11. [] S. Baker and I. Matthews. Lucas-Kanade years on: A unifying work. International Journal of Computer Vision, 56(3):1 55, 4. [3] G. Blais and M. Levine. Registering multiview range data to create 3D computer objects. IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI), 17(8):8 84, Aug [4] Y. Chen and G. Medioni. Object modeling by registration of multiple range images. In Proceedings of the IEEE International Conference on Robotics and Automation, volume 3, pages 74 79, Apr [5] L. Douadi, M.-J. Aldon, and A. Crosnier. Pair-wise registration of 3D/color data sets with ICP. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages , Oct. 6. [6] S. Druon, M. Aldon, and A. Crosnier. Color constrained ICP for registration of large unstructured 3D color data sets. In IEEE International Conference on Information Acquisition, pages 49 55, Aug 6. [7] F. Endres, J. Hess, N. Engelhard, J. Sturm, D. Cremers, and W. Burgard. An evaluation of the RGB-D SLAM system. In Proceedings of the International Conference on Robotics and Automation (ICRA), May 1. [8] N. Engelhard, F. Endres, J. Hess, J. Sturm, and W. Burgard. Realtime 3D visual SLAM with a hand-held RGB-D camera. In Proc. of the RGB-D Workshop on 3D Perception in Robotics at the European Robotics Forum, Apr. 11. [9] P. Henry, D. Fox, A. Bhowmik, and R. Mongia. Patch volumes: Segmentation-based consistent mapping with RGB-D cameras. In Proceedings of the International Conference on 3D Vision (3DV), June 13. [1] P. Henry, M. Krainin, E. Herbst, X. Ren, and D. Fox. RGB-D mapping: Using Kinect-style depth cameras for dense 3D modeling of indoor environments. International Journal of Robotics Research, 31(5): , Apr. 1. [11] A. Huang, A. Bachrach, P. Henry, M. Krainin, D. Maturana, D. Fox, and N. Roy. Visual odometry and mapping for autonomous flight using an RGB-D camera. In International Symposium on Robotics Research (ISRR), Aug. 11.
7 IEEE TRANSACTIONS ON ROBOTICS, VOL. 31, NO. 6, DECEMBER 15 7 [1] B. Huhle, M. Magnusson, W. Strasser, and A. J. Lilienthal. Registration of colored 3D point clouds with a kernel-based extension to the normal distributions transform. In Proceedings of the International Conference on Robotics and Automation (ICRA), pages 45 43, May 8. [13] S. Izadi, D. Kim, O. Hilliges, D. Molyneaux, R. Newcombe, P. Kohli, J. Shotton, S. Hodges, D. Freeman, A. Davison, and A. Fitzgibbon. KinectFusion: Real-time 3D reconstruction and interaction using a moving depth camera. In Proceedings of the 4th ACM Symposium on User Interface Software and Technology, pages , 11. [14] J. H. Joung, K. H. An, J. W. Kang, M. J. Chung, and W. Yu. 3D environment reconstruction using modified color ICP algorithm by fusion of a camera and a 3D laser range finder. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages , Oct. 9. [15] C. Kerl, J. Sturm, and D. Cremers. Dense visual SLAM for RGB-D cameras. In Proceedings of the International Conference on Intelligent Robot Systems (IROS), Nov. 13. [16] C. Kerl, J. Sturm, and D. Cremers. Robust odometry estimation for RGB-D cameras. In Proceedings of the International Conference on Robotics and Automation (ICRA), May 13. [17] K.-L. Low. Linear least-squares optimization for point-to-plane ICP surface registration. Technical Report TR4-4, University of North Carolina, Feb. 4. [18] B. D. Lucas and T. Kanade. An iterative image registration technique with an application to stereo vision. In Proceedings of the 7th International Joint Conference on Artificial Intelligence, pages , Aug [19] R. A. Newcombe, S. Izadi, O. Hilliges, D. Molyneaux, D. Kim, A. J. Davison, P. Kohli, J. Shotton, S. Hodges, and A. Fitzgibbon. Kinect- Fusion: Real-time dense surface mapping and tracking. In Proceedings of the 1th IEEE International Symposium on Mixed and Augmented Reality (ISMAR), pages , 11. [] R. A. Newcombe, S. J. Lovegrove, and A. J. Davison. DTAM: Dense tracking and mapping in real-time. In Proceedings of the International Conference on Computer Vision, pages 3 37, 11. [1] B. Peasley and S. Birchfield. Replacing projective data association with Lucas-Kanade for KinectFusion. In Proceedings of the International Conference on Robotics and Automation (ICRA), May 13. [] S. Rusinkiewicz and M. Levoy. Efficient variants of the ICP algorithm. In Third International Conference on 3D Digital Imaging and Modeling (3DIM), June 1. [3] R. B. Rusu and S. Cousins. 3D is here: Point Cloud Library (PCL). In Proceedings of the International Conference on Robotics and Automation (ICRA), May 11. [4] F. Steinbrücker, J. Sturm, and D. Cremers. Real-time visual odometry from dense RGB-D images. In Workshop on Live Dense Reconstruction with Moving Cameras (at ICCV), Nov. 11. [5] J. Sturm, N. Engelhard, F. Endres, W. Burgard, and D. Cremers. A benchmark for the evaluation of RGB-D SLAM systems. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages , Oct. 1. [6] C. Tomasi and T. Kanade. Detection and tracking of point features. Technical Report CMU-CS-91-13, Carnegie Mellon University, Apr [7] T. Tykkälä, C. Audras, and A. I. Comport. Direct iterative closest point for real-time visual odometry. In Second International Workshop on Computer Vision in Vehicle Technology, Nov. 11. [8] T. Whelan, H. Johannsson, M. Kaess, J. J. Leonard, and J. McDonald. Robust real-time visual odometry for dense RGB-D mapping. In Proceedings of the International Conference on Robotics and Automation (ICRA), May 13. [9] T. Whelan, J. B. McDonald, M. Kaess, M. F. Fallon, H. Johannsson, and J. J. Leonard. Kintinuous: Spatially extended KinectFusion. In RSS Workshop on RGB-D: Advanced Reasoning with Depth Cameras, Sydney, Australia, July 1.
Replacing Projective Data Association with Lucas-Kanade for KinectFusion
Replacing Projective Data Association with Lucas-Kanade for KinectFusion Brian Peasley and Stan Birchfield Electrical and Computer Engineering Dept. Clemson University, Clemson, SC 29634 {bpeasle,stb}@clemson.edu
More informationA Real-Time RGB-D Registration and Mapping Approach by Heuristically Switching Between Photometric And Geometric Information
A Real-Time RGB-D Registration and Mapping Approach by Heuristically Switching Between Photometric And Geometric Information The 17th International Conference on Information Fusion (Fusion 2014) Khalid
More informationCVPR 2014 Visual SLAM Tutorial Kintinuous
CVPR 2014 Visual SLAM Tutorial Kintinuous kaess@cmu.edu The Robotics Institute Carnegie Mellon University Recap: KinectFusion [Newcombe et al., ISMAR 2011] RGB-D camera GPU 3D/color model RGB TSDF (volumetric
More informationTracking an RGB-D Camera Using Points and Planes
MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Tracking an RGB-D Camera Using Points and Planes Ataer-Cansizoglu, E.; Taguchi, Y.; Ramalingam, S.; Garaas, T. TR2013-106 December 2013 Abstract
More informationHigh-speed Three-dimensional Mapping by Direct Estimation of a Small Motion Using Range Images
MECATRONICS - REM 2016 June 15-17, 2016 High-speed Three-dimensional Mapping by Direct Estimation of a Small Motion Using Range Images Shinta Nozaki and Masashi Kimura School of Science and Engineering
More informationEfficient Compositional Approaches for Real-Time Robust Direct Visual Odometry from RGB-D Data
Efficient Compositional Approaches for Real-Time Robust Direct Visual Odometry from RGB-D Data Sebastian Klose 1, Philipp Heise 1 and Alois Knoll 1 Abstract In this paper we give an evaluation of different
More informationLarge-Scale Multi-Resolution Surface Reconstruction from RGB-D Sequences
Large-Scale Multi-Resolution Surface Reconstruction from RGB-D Sequences Frank Steinbru cker, Christian Kerl, Ju rgen Sturm, and Daniel Cremers Technical University of Munich Boltzmannstrasse 3, 85748
More informationDense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm
Computer Vision Group Prof. Daniel Cremers Dense Tracking and Mapping for Autonomous Quadrocopters Jürgen Sturm Joint work with Frank Steinbrücker, Jakob Engel, Christian Kerl, Erik Bylow, and Daniel Cremers
More informationFeature-based RGB-D camera pose optimization for real-time 3D reconstruction
Computational Visual Media DOI 10.1007/s41095-016-0072-2 Research Article Feature-based RGB-D camera pose optimization for real-time 3D reconstruction Chao Wang 1, Xiaohu Guo 1 ( ) c The Author(s) 2016.
More informationRobust Online 3D Reconstruction Combining a Depth Sensor and Sparse Feature Points
2016 23rd International Conference on Pattern Recognition (ICPR) Cancún Center, Cancún, México, December 4-8, 2016 Robust Online 3D Reconstruction Combining a Depth Sensor and Sparse Feature Points Erik
More informationLarge-Scale Multi-Resolution Surface Reconstruction from RGB-D Sequences
2013 IEEE International Conference on Computer Vision Large-Scale Multi-Resolution Surface Reconstruction from RGB-D Sequences Frank Steinbrücker, Christian Kerl, Jürgen Sturm, and Daniel Cremers Technical
More informationColored Point Cloud Registration Revisited Supplementary Material
Colored Point Cloud Registration Revisited Supplementary Material Jaesik Park Qian-Yi Zhou Vladlen Koltun Intel Labs A. RGB-D Image Alignment Section introduced a joint photometric and geometric objective
More informationUsing RGB Information to Improve NDT Distribution Generation and Registration Convergence
Using RGB Information to Improve NDT Distribution Generation and Registration Convergence James Servos and Steven L. Waslander University of Waterloo, Waterloo, ON, Canada, N2L 3G1 Abstract Unmanned vehicles
More informationSturm, Jürgen; Bylow, Erik; Kerl, Christian; Kahl, Fredrik; Cremers, Daniel
Dense Tracking and Mapping with a Quadrocopter Sturm, Jürgen; Bylow, Erik; Kerl, Christian; Kahl, Fredrik; Cremers, Daniel 213 Link to publication Citation for published version (APA): Sturm, J., Bylow,
More informationDense 3D Reconstruction from Autonomous Quadrocopters
Dense 3D Reconstruction from Autonomous Quadrocopters Computer Science & Mathematics TU Munich Martin Oswald, Jakob Engel, Christian Kerl, Frank Steinbrücker, Jan Stühmer & Jürgen Sturm Autonomous Quadrocopters
More informationDeReEs: Real-Time Registration of RGBD Images Using Image-Based Feature Detection And Robust 3D Correspondence Estimation and Refinement
DeReEs: Real-Time Registration of RGBD Images Using Image-Based Feature Detection And Robust 3D Correspondence Estimation and Refinement Sahand Seifi Memorial University of Newfoundland sahands[at]mun.ca
More informationDNA-SLAM: Dense Noise Aware SLAM for ToF RGB-D Cameras
DNA-SLAM: Dense Noise Aware SLAM for ToF RGB-D Cameras Oliver Wasenmüller, Mohammad Dawud Ansari, Didier Stricker German Research Center for Artificial Intelligence (DFKI) University of Kaiserslautern
More information3D Environment Reconstruction
3D Environment Reconstruction Using Modified Color ICP Algorithm by Fusion of a Camera and a 3D Laser Range Finder The 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems October 11-15,
More informationA Benchmark for RGB-D Visual Odometry, 3D Reconstruction and SLAM
2014 IEEE International Conference on Robotics & Automation (ICRA) Hong Kong Convention and Exhibition Center May 31 - June 7, 2014. Hong Kong, China A Benchmark for RGB-D Visual Odometry, 3D Reconstruction
More informationEfficient SLAM Scheme Based ICP Matching Algorithm Using Image and Laser Scan Information
Proceedings of the World Congress on Electrical Engineering and Computer Systems and Science (EECSS 2015) Barcelona, Spain July 13-14, 2015 Paper No. 335 Efficient SLAM Scheme Based ICP Matching Algorithm
More informationIncremental compact 3D maps of planar patches from RGBD points
Incremental compact 3D maps of planar patches from RGBD points Juan Navarro and José M. Cañas Universidad Rey Juan Carlos, Spain Abstract. The RGBD sensors have opened the door to low cost perception capabilities
More informationVolumetric 3D Mapping in Real-Time on a CPU
Volumetric 3D Mapping in Real-Time on a CPU Frank Steinbrücker, Jürgen Sturm, and Daniel Cremers Abstract In this paper we propose a novel volumetric multi-resolution mapping system for RGB-D images that
More information3D PLANE-BASED MAPS SIMPLIFICATION FOR RGB-D SLAM SYSTEMS
3D PLANE-BASED MAPS SIMPLIFICATION FOR RGB-D SLAM SYSTEMS 1,2 Hakim ELCHAOUI ELGHOR, 1 David ROUSSEL, 1 Fakhreddine ABABSA and 2 El-Houssine BOUYAKHF 1 IBISC Lab, Evry Val d'essonne University, Evry, France
More informationScanning and Printing Objects in 3D Jürgen Sturm
Scanning and Printing Objects in 3D Jürgen Sturm Metaio (formerly Technical University of Munich) My Research Areas Visual navigation for mobile robots RoboCup Kinematic Learning Articulated Objects Quadrocopters
More informationStructured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov
Structured Light II Johannes Köhler Johannes.koehler@dfki.de Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Introduction Previous lecture: Structured Light I Active Scanning Camera/emitter
More informationStereo and Kinect fusion for continuous. 3D reconstruction and visual odometry
1 Stereo and Kinect fusion for continuous 3D reconstruction and visual odometry Ozgur YILMAZ #*1, Fatih KARAKUS *2 # Department of Computer Engineering, Turgut Özal University Ankara, TURKEY 1 ozyilmaz@turgutozal.edu.tr
More informationIntegrating Depth and Color Cues for Dense Multi-Resolution Scene Mapping Using RGB-D Cameras
In Proc. of the IEEE Int. Conf. on Multisensor Fusion and Information Integration (MFI), Hamburg, Germany, 2012 Integrating Depth and Color Cues for Dense Multi-Resolution Scene Mapping Using RGB-D Cameras
More informationSimultaneous Localization and Calibration: Self-Calibration of Consumer Depth Cameras
Simultaneous Localization and Calibration: Self-Calibration of Consumer Depth Cameras Qian-Yi Zhou Vladlen Koltun Abstract We describe an approach for simultaneous localization and calibration of a stream
More informationMemory Management Method for 3D Scanner Using GPGPU
GPGPU 3D 1 2 KinectFusion GPGPU 3D., GPU., GPGPU Octree. GPU,,. Memory Management Method for 3D Scanner Using GPGPU TATSUYA MATSUMOTO 1 SATORU FUJITA 2 This paper proposes an efficient memory management
More informationNICP: Dense Normal Based Point Cloud Registration
NICP: Dense Normal Based Point Cloud Registration Jacopo Serafin 1 and Giorgio Grisetti 1 Abstract In this paper we present a novel on-line method to recursively align point clouds. By considering each
More informationRobust Real-Time Visual Odometry for Dense RGB-D Mapping
Robust Real-Time Visual Odometry for Dense RGB-D Mapping Thomas Whelan 1, Hordur Johannsson, Michael Kaess, John J. Leonard and John McDonald 1 Abstract This paper describes extensions to the Kintinuous
More informationMobile Point Fusion. Real-time 3d surface reconstruction out of depth images on a mobile platform
Mobile Point Fusion Real-time 3d surface reconstruction out of depth images on a mobile platform Aaron Wetzler Presenting: Daniel Ben-Hoda Supervisors: Prof. Ron Kimmel Gal Kamar Yaron Honen Supported
More informationSegmentation and Tracking of Partial Planar Templates
Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract
More information3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.
3D Computer Vision Structured Light II Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de 1 Introduction
More informationSuper-Resolution Keyframe Fusion for 3D Modeling with High-Quality Textures
Super-Resolution Keyframe Fusion for 3D Modeling with High-Quality Textures Robert Maier, Jörg Stückler, Daniel Cremers International Conference on 3D Vision (3DV) October 2015, Lyon, France Motivation
More informationarxiv: v1 [cs.ro] 23 Nov 2015
Multi-Volume High Resolution RGB-D Mapping with Dynamic Volume Placement Michael Salvato 1, Ross Finman 1, and John J. Leonard 1 arxiv:1511.07106v1 [cs.ro] 23 Nov 2015 I. ABSTRACT Abstract We present a
More informationUsing Augmented Measurements to Improve the Convergence of ICP. Jacopo Serafin and Giorgio Grisetti
Jacopo Serafin and Giorgio Grisetti Point Cloud Registration We want to find the rotation and the translation that maximize the overlap between two point clouds Page 2 Point Cloud Registration We want
More informationKintinuous: Spatially Extended KinectFusion Thomas Whelan, Michael Kaess, Maurice Fallon, Hordur Johannsson, John Leonard, and John McDonald
Computer Science and Artificial Intelligence Laboratory Technical Report MIT-CSAIL-TR-2012-020 July 19, 2012 Kintinuous: Spatially Extended KinectFusion Thomas Whelan, Michael Kaess, Maurice Fallon, Hordur
More informationEfficient Online Surface Correction for Real-time Large-Scale 3D Reconstruction
MAIER ET AL.: EFFICIENT LARGE-SCALE ONLINE SURFACE CORRECTION 1 Efficient Online Surface Correction for Real-time Large-Scale 3D Reconstruction Robert Maier maierr@in.tum.de Raphael Schaller schaller@in.tum.de
More informationDirect Methods in Visual Odometry
Direct Methods in Visual Odometry July 24, 2017 Direct Methods in Visual Odometry July 24, 2017 1 / 47 Motivation for using Visual Odometry Wheel odometry is affected by wheel slip More accurate compared
More informationRemoving Moving Objects from Point Cloud Scenes
Removing Moving Objects from Point Cloud Scenes Krystof Litomisky and Bir Bhanu University of California, Riverside krystof@litomisky.com, bhanu@ee.ucr.edu Abstract. Three-dimensional simultaneous localization
More informationPhotorealistic 3D Mapping of Indoors by RGB-D Scanning Process
Photorealistic 3D Mapping of Indoors by RGB-D Scanning Process Tommi Tykkälä 1, Andrew I. Comport 2, and Joni-Kristian Kämäräinen 3 Abstract In this work, a RGB-D input stream is utilized for GPU-boosted
More informationMonoRGBD-SLAM: Simultaneous Localization and Mapping Using Both Monocular and RGBD Cameras
MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com MonoRGBD-SLAM: Simultaneous Localization and Mapping Using Both Monocular and RGBD Cameras Yousif, K.; Taguchi, Y.; Ramalingam, S. TR2017-068
More informationGeometric Reconstruction Dense reconstruction of scene geometry
Lecture 5. Dense Reconstruction and Tracking with Real-Time Applications Part 2: Geometric Reconstruction Dr Richard Newcombe and Dr Steven Lovegrove Slide content developed from: [Newcombe, Dense Visual
More informationarxiv: v1 [cs.cv] 28 Sep 2018
Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,
More informationA Flexible Scene Representation for 3D Reconstruction Using an RGB-D Camera
2013 IEEE International Conference on Computer Vision A Flexible Scene Representation for 3D Reconstruction Using an RGB-D Camera Diego Thomas National Institute of Informatics Chiyoda, Tokyo, Japan diego
More informationMotion Capture using Body Mounted Cameras in an Unknown Environment
Motion Capture using Body Mounted Cameras in an Unknown Environment Nam Vo Taeyoung Kim Siddharth Choudhary 1. The Problem Motion capture has been recently used to provide much of character motion in several
More informationHierarchical Sparse Coded Surface Models
Hierarchical Sparse Coded Surface Models Michael Ruhnke Liefeng Bo Dieter Fox Wolfram Burgard Abstract In this paper, we describe a novel approach to construct textured 3D environment models in a hierarchical
More informationarxiv: v1 [cs.cv] 2 May 2016
16-811 Math Fundamentals for Robotics Comparison of Optimization Methods in Optical Flow Estimation Final Report, Fall 2015 arxiv:1605.00572v1 [cs.cv] 2 May 2016 Contents Noranart Vesdapunt Master of Computer
More informationColour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation
ÖGAI Journal 24/1 11 Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation Michael Bleyer, Margrit Gelautz, Christoph Rhemann Vienna University of Technology
More informationJakob Engel, Thomas Schöps, Daniel Cremers Technical University Munich. LSD-SLAM: Large-Scale Direct Monocular SLAM
Computer Vision Group Technical University of Munich Jakob Engel LSD-SLAM: Large-Scale Direct Monocular SLAM Jakob Engel, Thomas Schöps, Daniel Cremers Technical University Munich Monocular Video Engel,
More informationMulti Channel Generalized-ICP
Multi Channel Generalized-ICP James Servos and Steven L. Waslander University of Waterloo, Waterloo, ON, Canada, N2L 3G1 Abstract Current state of the art scan registration algorithms which use only positional
More informationRobust Edge-based Visual Odometry using Machine-Learned Edges
Robust Edge-based Visual Odometry using Machine-Learned Edges Fabian Schenk and Friedrich Fraundorfer Abstract In this work, we present a real-time robust edgebased visual odometry framework for RGBD sensors
More informationReal-time Depth Enhanced Monocular Odometry
2014 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2014) September 14-18, 2014, Chicago, IL, USA Real-time Depth Enhanced Monocular Odometry Ji Zhang, Michael Kaess, and Sanjiv
More informationSDF-TAR: Parallel Tracking and Refinement in RGB-D Data using Volumetric Registration
M. SLAVCHEVA, S. ILIC: SDF-TAR 1 SDF-TAR: Parallel Tracking and Refinement in RGB-D Data using Volumetric Registration Miroslava Slavcheva mira.slavcheva@tum.de Slobodan Ilic slobodan.ilic@siemens.com
More informationObject Reconstruction
B. Scholz Object Reconstruction 1 / 39 MIN-Fakultät Fachbereich Informatik Object Reconstruction Benjamin Scholz Universität Hamburg Fakultät für Mathematik, Informatik und Naturwissenschaften Fachbereich
More informationColor Supported Generalized-ICP 0
Color Supported Generalized-ICP 0 Michael Korn, Martin Holzkothen and Josef Pauli Lehrstuhl Intelligente Systeme, Universität Duisburg-Essen, 47057 Duisburg, Germany michael.korn@uni-due.de Keywords: Abstract:
More informationOccluded Facial Expression Tracking
Occluded Facial Expression Tracking Hugo Mercier 1, Julien Peyras 2, and Patrice Dalle 1 1 Institut de Recherche en Informatique de Toulouse 118, route de Narbonne, F-31062 Toulouse Cedex 9 2 Dipartimento
More information3D Object Recognition and Scene Understanding from RGB-D Videos. Yu Xiang Postdoctoral Researcher University of Washington
3D Object Recognition and Scene Understanding from RGB-D Videos Yu Xiang Postdoctoral Researcher University of Washington 1 2 Act in the 3D World Sensing & Understanding Acting Intelligent System 3D World
More informationDeformation-based Loop Closure for Large Scale Dense RGB-D SLAM
2013 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) November 3-7, 2013. Tokyo, Japan Deformation-based Loop Closure for Large Scale Dense RGB-D SLAM Thomas Whelan 1, Michael
More informationReal-Time RGB-D Registration and Mapping in Texture-less Environments Using Ranked Order Statistics
Real-Time RGB-D Registration and Mapping in Texture-less Environments Using Ranked Order Statistics Khalid Yousif 1, Alireza Bab-Hadiashar 2, Senior Member, IEEE and Reza Hoseinnezhad 3 Abstract In this
More informationAccurate 3D Face and Body Modeling from a Single Fixed Kinect
Accurate 3D Face and Body Modeling from a Single Fixed Kinect Ruizhe Wang*, Matthias Hernandez*, Jongmoo Choi, Gérard Medioni Computer Vision Lab, IRIS University of Southern California Abstract In this
More informationUsing the Kinect as a Navigation Sensor for Mobile Robotics
Using the Kinect as a Navigation Sensor for Mobile Robotics ABSTRACT Ayrton Oliver Dept. of Electrical and Computer Engineering aoli009@aucklanduni.ac.nz Burkhard C. Wünsche Dept. of Computer Science burkhard@cs.auckland.ac.nz
More informationA Catadioptric Extension for RGB-D Cameras
A Catadioptric Extension for RGB-D Cameras Felix Endres Christoph Sprunk Rainer Kümmerle Wolfram Burgard Abstract The typically restricted field of view of visual sensors often imposes limitations on the
More informationFLaME: Fast Lightweight Mesh Estimation using Variational Smoothing on Delaunay Graphs
FLaME: Fast Lightweight Mesh Estimation using Variational Smoothing on Delaunay Graphs W. Nicholas Greene Robust Robotics Group, MIT CSAIL LPM Workshop IROS 2017 September 28, 2017 with Nicholas Roy 1
More informationarxiv: v1 [cs.ro] 24 Nov 2018
BENCHMARKING AND COMPARING POPULAR VISUAL SLAM ALGORITHMS arxiv:1811.09895v1 [cs.ro] 24 Nov 2018 Amey Kasar Department of Electronics and Telecommunication Engineering Pune Institute of Computer Technology
More informationPerceiving the 3D World from Images and Videos. Yu Xiang Postdoctoral Researcher University of Washington
Perceiving the 3D World from Images and Videos Yu Xiang Postdoctoral Researcher University of Washington 1 2 Act in the 3D World Sensing & Understanding Acting Intelligent System 3D World 3 Understand
More informationInertial-Kinect Fusion for Outdoor 3D Navigation
Proceedings of Australasian Conference on Robotics and Automation, 2-4 Dec 213, University of New South Wales, Sydney Australia Inertial-Kinect Fusion for Outdoor 3D Navigation Usman Qayyum and Jonghyuk
More informationOutline. 1 Why we re interested in Real-Time tracking and mapping. 3 Kinect Fusion System Overview. 4 Real-time Surface Mapping
Outline CSE 576 KinectFusion: Real-Time Dense Surface Mapping and Tracking PhD. work from Imperial College, London Microsoft Research, Cambridge May 6, 2013 1 Why we re interested in Real-Time tracking
More informationRobustness to Lighting Variations: An RGB-D Indoor Visual Odometry Using Line Segments
Robustness to Lighting Variations: An RGB-D Indoor Visual Odometry Using Line Segments Yan Lu and Dezhen Song Abstract Large lighting variation challenges all visual odometry methods, even with RGB-D cameras.
More informationarxiv: v1 [cs.ro] 23 Feb 2018
IMLS-SLAM: scan-to-model matching based on 3D data Jean-Emmanuel Deschaud 1 1 MINES ParisTech, PSL Research University, Centre for Robotics, 60 Bd St Michel 75006 Paris, France arxiv:1802.08633v1 [cs.ro]
More informationVisual Odometry Algorithm Using an RGB-D Sensor and IMU in a Highly Dynamic Environment
Visual Odometry Algorithm Using an RGB-D Sensor and IMU in a Highly Dynamic Environment Deok-Hwa Kim, Seung-Beom Han, and Jong-Hwan Kim Department of Electrical Engineering, KAIST 29 Daehak-ro, Yuseong-gu,
More informationarxiv: v1 [cs.cv] 1 Aug 2017
Dense Piecewise Planar RGB-D SLAM for Indoor Environments Phi-Hung Le and Jana Kosecka arxiv:1708.00514v1 [cs.cv] 1 Aug 2017 Abstract The paper exploits weak Manhattan constraints to parse the structure
More informationMETRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS
METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS M. Lefler, H. Hel-Or Dept. of CS, University of Haifa, Israel Y. Hel-Or School of CS, IDC, Herzliya, Israel ABSTRACT Video analysis often requires
More informationRGB-D Edge Detection and Edge-based Registration
23 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) November 3-7, 23. Tokyo, Japan RGB-D Edge Detection and Edge-based Registration Changhyun Choi, Alexander J. B. Trevor, and
More informationToward Online 3-D Object Segmentation and Mapping
Toward Online 3-D Object Segmentation and Mapping Evan Herbst Peter Henry Dieter Fox Abstract We build on recent fast and accurate 3-D reconstruction techniques to segment objects during scene reconstruction.
More informationThe Lucas & Kanade Algorithm
The Lucas & Kanade Algorithm Instructor - Simon Lucey 16-423 - Designing Computer Vision Apps Today Registration, Registration, Registration. Linearizing Registration. Lucas & Kanade Algorithm. 3 Biggest
More informationApproximate Surface Reconstruction and Registration for RGB-D SLAM
European Conference on Mobile Robotics (ECMR), Lincoln, UK, September 2015 Approximate Surface Reconstruction and Registration for RGB-D SLAM Dirk Holz and Sven Behnke Abstract RGB-D cameras have attracted
More informationFace Recognition At-a-Distance Based on Sparse-Stereo Reconstruction
Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction Ham Rara, Shireen Elhabian, Asem Ali University of Louisville Louisville, KY {hmrara01,syelha01,amali003}@louisville.edu Mike Miller,
More informationCopyMe3D: Scanning and Printing Persons in 3D
CopyMe3D: Scanning and Printing Persons in 3D Sturm, Jürgen; Bylow, Erik; Kahl, Fredrik; Cremers, Daniel Published in: Lecture Notes in Computer Science Vol. 8142 (Pattern Recognition, 35th German Conference,
More informationSimultaneous Localization and Mapping (SLAM)
Simultaneous Localization and Mapping (SLAM) RSS Lecture 16 April 8, 2013 Prof. Teller Text: Siegwart and Nourbakhsh S. 5.8 SLAM Problem Statement Inputs: No external coordinate reference Time series of
More informationSPURRED by the ready availability of depth sensors and
IEEE ROBOTICS AND AUTOMATION LETTERS. PREPRINT VERSION. ACCEPTED DECEMBER, 2015 1 Hierarchical Hashing for Efficient Integration of Depth Images Olaf Kähler, Victor Prisacariu, Julien Valentin and David
More informationDual Back-to-Back Kinects for 3-D Reconstruction
Ho Chuen Kam, Kin Hong Wong and Baiwu Zhang, Dual Back-to-Back Kinects for 3-D Reconstruction, ISVC'16 12th International Symposium on Visual Computing December 12-14, 2016, Las Vegas, Nevada, USA. Dual
More information3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller
3D Computer Vision Depth Cameras Prof. Didier Stricker Oliver Wasenmüller Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de
More informationAN INDOOR SLAM METHOD BASED ON KINECT AND MULTI-FEATURE EXTENDED INFORMATION FILTER
AN INDOOR SLAM METHOD BASED ON KINECT AND MULTI-FEATURE EXTENDED INFORMATION FILTER M. Chang a, b, Z. Kang a, a School of Land Science and Technology, China University of Geosciences, 29 Xueyuan Road,
More informationCuFusion: Accurate real-time camera tracking and volumetric scene reconstruction with a cuboid
CuFusion: Accurate real-time camera tracking and volumetric scene reconstruction with a cuboid Chen Zhang * and Yu Hu College of Computer Science and Technology, Zhejiang University, Hangzhou 310027, China;
More informationLecture 10 Multi-view Stereo (3D Dense Reconstruction) Davide Scaramuzza
Lecture 10 Multi-view Stereo (3D Dense Reconstruction) Davide Scaramuzza REMODE: Probabilistic, Monocular Dense Reconstruction in Real Time, ICRA 14, by Pizzoli, Forster, Scaramuzza [M. Pizzoli, C. Forster,
More informationSurface Registration. Gianpaolo Palma
Surface Registration Gianpaolo Palma The problem 3D scanning generates multiple range images Each contain 3D points for different parts of the model in the local coordinates of the scanner Find a rigid
More informationMotion Estimation. There are three main types (or applications) of motion estimation:
Members: D91922016 朱威達 R93922010 林聖凱 R93922044 謝俊瑋 Motion Estimation There are three main types (or applications) of motion estimation: Parametric motion (image alignment) The main idea of parametric motion
More informationThe Template Update Problem
The Template Update Problem Iain Matthews, Takahiro Ishikawa, and Simon Baker The Robotics Institute Carnegie Mellon University Abstract Template tracking dates back to the 1981 Lucas-Kanade algorithm.
More informationCRF Based Point Cloud Segmentation Jonathan Nation
CRF Based Point Cloud Segmentation Jonathan Nation jsnation@stanford.edu 1. INTRODUCTION The goal of the project is to use the recently proposed fully connected conditional random field (CRF) model to
More informationIncremental Structured ICP Algorithm
Incremental Structured ICP Algorithm Haokun Geng, Johnny Chien, Radu Nicolescu, and Reinhard Klette The.enpeda.. Project, Tamaki Campus The University of Auckland, New Zealand Abstract. Variants of the
More informationMesh from Depth Images Using GR 2 T
Mesh from Depth Images Using GR 2 T Mairead Grogan & Rozenn Dahyot School of Computer Science and Statistics Trinity College Dublin Dublin, Ireland mgrogan@tcd.ie, Rozenn.Dahyot@tcd.ie www.scss.tcd.ie/
More informationABSTRACT. KinectFusion is a surface reconstruction method to allow a user to rapidly
ABSTRACT Title of Thesis: A REAL TIME IMPLEMENTATION OF 3D SYMMETRIC OBJECT RECONSTRUCTION Liangchen Xi, Master of Science, 2017 Thesis Directed By: Professor Yiannis Aloimonos Department of Computer Science
More informationLucas-Kanade Image Registration Using Camera Parameters
Lucas-Kanade Image Registration Using Camera Parameters Sunghyun Cho a, Hojin Cho a, Yu-Wing Tai b, Young Su Moon c, Junguk Cho c, Shihwa Lee c, and Seungyong Lee a a POSTECH, Pohang, Korea b KAIST, Daejeon,
More informationStereo and Epipolar geometry
Previously Image Primitives (feature points, lines, contours) Today: Stereo and Epipolar geometry How to match primitives between two (multiple) views) Goals: 3D reconstruction, recognition Jana Kosecka
More informationEfficient Onboard RGBD-SLAM for Autonomous MAVs
Efficient Onboard RGBD-SLAM for Autonomous MAVs Sebastian A. Scherer 1 and Andreas Zell 1 Abstract We present a computationally inexpensive RGBD- SLAM solution taylored to the application on autonomous
More informationIntrinsic3D: High-Quality 3D Reconstruction by Joint Appearance and Geometry Optimization with Spatially-Varying Lighting
Intrinsic3D: High-Quality 3D Reconstruction by Joint Appearance and Geometry Optimization with Spatially-Varying Lighting R. Maier 1,2, K. Kim 1, D. Cremers 2, J. Kautz 1, M. Nießner 2,3 Fusion Ours 1
More informationA Proof that Fusing Measurements Using Point-to-Hyperplane Registration is Invariant to Relative Scale
A Proof that Fusing Measurements Using Point-to-Hyperplane Registration is Invariant to Relative Scale Fernando Ireta Munoz, Andrew Comport To cite this version: Fernando Ireta Munoz, Andrew Comport. A
More informationA Low Power, High Throughput, Fully Event-Based Stereo System: Supplementary Documentation
A Low Power, High Throughput, Fully Event-Based Stereo System: Supplementary Documentation Alexander Andreopoulos, Hirak J. Kashyap, Tapan K. Nayak, Arnon Amir, Myron D. Flickner IBM Research March 25,
More information