Real-Time RGB-D Registration and Mapping in Texture-less Environments Using Ranked Order Statistics

Size: px
Start display at page:

Download "Real-Time RGB-D Registration and Mapping in Texture-less Environments Using Ranked Order Statistics"

Transcription

1

2 Real-Time RGB-D Registration and Mapping in Texture-less Environments Using Ranked Order Statistics Khalid Yousif 1, Alireza Bab-Hadiashar 2, Senior Member, IEEE and Reza Hoseinnezhad 3 Abstract In this paper, we present a real time 3D SLAM system for texture-less scenes using only the depth information provided by a low cost RGB-D sensor. The proposed registration method is based on a novel informative sampling scheme that is able to extract the points carrying the most useful information from two consecutive frames. Those points are assigned 3D feature descriptors and matched with their correspondences from the other frame. A robust estimator is then used to refine the matches and estimate a rigid transformation between the frames. A global pose for the camera and associated 3D position of the points are computed by concatenating the relative transformations and a map is constructed. As a result, we are able to achieve very accurate registration in real-time using the sparse keypoints. Experimental evaluations show that our proposed method outperforms the state of the art registration algorithms in terms of accuracy and efficiency. I. INTRODUCTION Simultaneous Localization and Mapping (SLAM) [1] is the problem of simultaneous estimation of the pose of a mobile robot or a moving camera and mapping an unknown environment. The SLAM problem has been intensely studied over the past couple of decades and various solutions have been developed for specific applications [2]. The invention of cheap RGB-D sensors (e.g. Microsoft Kinect) in recent times has caused an intense interest in obtaining a dense 3D model of the environment instead of the sparse 2D map [3][4][5]. The main motivation is that dense 3D maps are rich in information and can be effectively used to improve object recognition and manipulation, navigation, collision avoidance and path planning algorithms. A 3D representation of the environment is also very useful in augmented reality applications that facilitate interactions between the real and virtual environments [3]. Despite the above advantages, a number of challenges arise when using RGB-D cameras for 3D registration and mapping applications. For instance, the registration of frames in a texture-less scene is very difficult because the existing visual information would not be sufficient to align the corresponding 3D points between different frames. In addition, the huge amount of data provided by RGB-D sensors pose significant challenges for acquiring, processing and visualizing data in real-time. Many types of environments in which mobile robots are used in (e.g. offices, warehouses and homes) predominantly inlclude texture-less objects (plain walls, floors, ceilings, Authors are with school of Aerospace, Mechanical and Manufacturing Engineering, RMIT University, Melbourne, VIC 3001, Australia 1 s @student.rmit.edu.au 2 abh@rmit.edu.au 3 rezah@rmit.edu.au Fig. 1: Constructed 3D map of a RMIT university postgraduate office using the proposed 3D registration and mapping algorithm. desks etc..). Fig. 2 shows a typical corridor in an office building. Registration in these environments using visual feature extraction methods would often fail as there is not enough constraining visual features. In this paper, we outline the problem of 3D registration and mapping in an environment that may contain textureless scenes and propose a fast and accurate method that does not rely on extracting any visual information from the RGB image. Instead, we extract geometric 3D features from sequential frames and assign a descriptor to each feature, match them using their descriptors and finally, refine the matches and align them using a robust estimator [6]. Due to the hard real time constraints and high computational cost of feature descriptor matching, this is a challenging task. Our main contribution here is the development of an informative sampling based 3D feature extraction technique. The method is able to exploit the geometric information of the points and their neighbors to identify points that carry the most useful information. We call the points resulting from our informative sampling scheme: Ranked Order Statistics (ROS) keypoints. We will show that the proposed feature extraction method is able obtain a subset of points of the original point cloud that can deliver a very accurate registration compared to using a point cloud containing 10 times more points. The main advantage of using this sampling technique is that it would reduce computational time significantly. In fact, we will show that our method outperforms the state of the art registration methods both in accuracy and computational efficiency. The rest of the paper is organized as follows. In section II

3 (b) (a) Fig. 2: (a) A typical corridor in an office building with limited texture information (b) Alignment using the proposed method (c) Alignment using our previous method [7] that relies on extracting visual features. we will discuss the related work in this area. We will present our informative sampling based feature extraction method in section III and our 3D registration method in section IV. Results are presented in section V followed by a conclusion in section VI. II. RELATED WORK A. RGB-D Registration and Mapping RGB-D based mapping was first introduced by Henry et al. [4] who employed a Microsoft Kinect to capture RGB and depth information in sequential frames in order to estimate the camera pose and obtain a 3D reconstructed environment. They extracted and matched FAST features [8] from sequential RGB images, and used RANSAC to refine those matches and estimate an initial transformation. This transformation was refined using a generalized-icp [9] algorithm. Global consistency was achieved by applying a Sparse Bundle Adjustment method and loop closure was detected by matching the current frame to previously collected keyframes. Endres et al. [5] proposed a similar method called RGB-D-SLAM in which SIFT, SURF [10] or ORB [11] features were used in place of FAST features and posegraph optimization was used instead of Bundle Adjustment for global optimization. Du et al. [12] extended Henry et al. s method by incorporating on-line user interaction and feedback during the map building step. User input enabled the system to recover from registration failures which may be caused by a fast moving camera. Newcombe et al. [3] proposed a GPU based RGB-D mapping algorithm called KinectFusion. The method was designed to avoid using the RGB images and instead, they proposed an ICP variant for registration in which the current measurement is matched with the full growing surface model as opposed to matching sequential frames. They also implemented a segmentation step which divided the scene into (c) foreground and background and was able to detect moving humans that would be present in the scene. This allowed user interaction during the map building procedure without deteriorating the accuracy of the estimated transformations and the map. The major downside to their approach is that it is computationally expensive because of the requirement to store and process the full dense volumetric representation of the constructed model in GPU memory and as such, is not scalable for large environments. Whelan et al. [13] proposed an extension to KinectFusion which outlined those problems and allowed the region of the environment that is mapped by the KinectFusion algorithm to vary dynamically, allowing their approach to map larger scenes. Our approach is similar as we also do not rely on the availability of RGB features for registration. However, since ICP only works when there is a good initial alignment, we advocate the use of geometric 3D features for registration. Hu et al. [14] proposed a heuristic switching algorithm that chooses between an RGB-D mapping approach and a 8- point RANSAC monocular SLAM, based on the availability of depth information. A 3D Iterative Sparse Local Submap Joining Filter (I-SLSJF) was then used to merge the two generated maps. In our previous work [7], we outlined the problem of 3D registration in dark environments and proposed a method that switches between RGB and IR images for visual feature extraction and matching based on the brightness of the RGB image. A highly robust estimator [6] was then employed to estimate the transformation between frames. The estimate was then refined using an ICP algorithm. This method fails in cases where both the RGB and the IR images contain insufficient visual information, since ICP alone fails to provide an accurate estimate without being provided with a good initial guess. The method proposed in this paper is designed to overcome solely based on using depth information. B. 3D Features The well known Harris corner detector [15] was implemented in 3D and is available in the point cloud library (PCL) [16]. In the 3D implementation, the surface normals were used in place of the image gradients which were used in the original algorithm [17]. SIFT [18] is a blob detector which extracts image patterns that differ from their immediate neighborhood in terms of intensity, color and texture. SIFT was also implemented in 3D and is available in PCL. The main difference between the original method and the 3D implementation is that instead of comparing a pixel with its neighboring pixels in the image plane, a 3D point in a point cloud is compared to its euclidean distance based nearest neighbors found using a kd-tree search. Gelfand et al. [19] proposed a sampling strategy that groups the point cloud data into stable points (constraining points) and non-stable points (that have no effect on the transformation estimation). This is achieved by analyzing the contribution of the force (translational) and torque (rotational) components by each pair of points using a point-to-plane error metric. The main shortcoming of this method is that it is optimized

4 for a specific metric and it becomes suboptimal when other metrics (such as the squared Euclidean distance error metric employed here) are used. An example of this method is illustrated in Fig. 3 (c) which shows some points with significant local curvature being omitted from the point cloud (such as the top part of a chair). Zhong [20] proposed a 3D feature detector and descriptor called Intrinsic Shape Signatures. This method relies on exploiting the information provided by the covariance matrix that is constructed using surface normals estimates. The method extracts keypoints based on two measures: points that correspond to the smallest eigenvalues of the covariance matrix representing those with large variations, and the ratio between successive eigenvalues in order to discard redundant points that have a similar spread. Despite the availability of 3D feature extraction methods, the above mentioned implementations seem to be focused on general 3D modeling applications and none of those seem to be particularly optimized for SLAM purposes. For instance, our experiments showed that Zhong s method [20] finds many feature points on commonly encountered planar surfaces such as office walls and doors (see Fig. 3 (b)). Those points are ill-conditioned for registration purposes and can deteriorate the estimation outcome. In this work, we address the problem of aligning RGB- D images in an environment that may not contain sufficient visual information for 3D registration using approaches that rely on visual feature extraction and matching tasks. Our proposed method extracts unique 3D features from sequential frames using a novel ranked order statistics based informative sampling segmentation method. The extracted 3D features are then matched using SHOT descriptors [21] and finally, the modified selective statistical estimator (MSSE) [6] is employed to segment the good matches and estimate a 6 degree of freedom (6DOF) transformation between the two frames. The estimated transformations are concatenated up to the current time in order to obtain a global transformation of the camera and a global map. These steps are outlined in the following sections. III. EXTRACTING GEOMETRIC FEATURES USING RANKED ORDER STATISTICS The aim of our proposed feature extraction method is to informatively down-sample a point cloud into a subset of points that geometrically differ from their immediate neighborhood. This is somewhat similar to the idea behind geometrically stable sampling for ICP [19]. However, we present an efficient method that finds the sample with the most information directly using the statistical analysis of their flatness. The main idea is to segment the points into two main groups (a 3rd group is defined later): points that are locally flat and points that have significant local curvature. The details of the implementation of this method is presented in Algorithm 1. The input to this method is a point cloud with a normal vector calculated at every point. We then calculate the average of the squared angles between each point s normal and the normals of its N nearest neighbors (a) (c) Fig. 3: A comparison between different sampling techniques. The point clouds after: (a) uniform sampling (b) ISS (c) covariance keypoints (d) ROS. which are found using a kd-tree search. The angles are calculated using the following equations: θ i = (b) (d) φ = arccos(p i.q f ) (1) N j=1 φ 2 j N ; i = 1... n (2) where φ is the angle between two normal vectors p i (a query point) and q f (a neighboring point), n is the number of points in the point cloud and θ i is the average angle between normals (of the query points and all its neighbors). The average squared angle θ i which is calculated for every point in the point cloud, is considered as an error measure that specifies the similarity of a point s curvature to its neighbors. Therefore, for points associated with a θ i value close to zero, the orientations of normals of the point and its neighbors are similar (such as points lying on a flat plane) regardless of the normal vector s orientation at those points (i.e. regardless of which plane they lay on). On the other hand, the points associated with higher residual values differ from their neighborhood and carry more useful information for registration purposes. To further clarify this point, let us assume that there are two planes perpendicular to each other. Those points with the smallest residuals correspond to points lying on either one of the two planes. However, we are mainly looking for the points that lay close to the intersection between the two planes. In this scheme, those correspond to points with higher residual values. In the next step, we store all the θ i angles in the residual vector r θ and sort those in ascending order. The final step is to find the transition point that separates the two aforementioned groups. A ranked order statistics technique (MSSE) [6] is employed for this segmentation. The segmentation is conducted iteratively by first calculating the standard deviation of sorted residuals using the first k-th

5 Residuals Group 1 (Points with similar normal vector orientations to neighbors) Group 3 (Points with lowest resdiuals) T σ i Sorted resdiuals (r θ(i) ) Group 2 (points with significant local curvature) Transiction Point (k ) Sorted point index (i) Fig. 4: Classification of points based on the calculated residuals using the MSSE constraint. points (initial k-th order that corresponds to the assumed minimum percentage of points included in the first segment): σ 2 k = k i=1 r 2 θ i n p where r θi is ith residual in the sorted residual vector θ and p is the dimension of the model. The transition point k is found when the following condition is met: (3) r θk > T σ k (4) where T is a constant factor which is set to 2.5 and includes around 99% population of a normal distribution. We finally construct a third group, which are chosen to be the first 2% points of the sorted residuals (i.e. the points associated with the 2% smallest residuals in the first group). The points that are included in groups 2 and 3 are the extracted 3D feature points. We found that including points of group 3 improves the accuracy of the registration since adding a few of those points which share similar normal orientations and carry some useful information, helps constrain the transformation estimation. We chose the first 2% for a repeatability purpose, since those points with the lowest residuals, are more likely to be the same points in sequential overlapping frames and as such, be appropriately matched with their correspondences. Fig. 4 shows the classification of different points based on the calculated residuals using the MSSE constraint (4). To meet the stringent time constraint of the registration procedure, the selected keypoints are downsampled to around 350 keypoints. Our extensive experiments showed that having around 350 keypoints provides the right balance between registration accuracy and speed. Fig. 3 shows a comparison between our proposed method and other feature extraction techniques. IV. 3D REGISTRATION AND MAPPING In the following sections, we will discuss the steps of our proposed 3D registration and mapping method which Algorithm 1 Step-by-Step Algorithm of Proposed Feature Extraction Method 1: Input: Point cloud with Normals P n 2: Output: Sampled Point cloud P d 3: Initialize and clear vector α; 4: for i = 0 to length(p n ) do 5: Find N nearest neighbors to query point i; 6: for j = 1 to N do 7: Calculate angle between normals (φ j ) of query point and its neighbor j; 8: Add φ 2 j to the vector α; 9: end for 10: Calculate the average of the angles in vector α (θ); 11: Add θ to residuals vector (r θ ); 12: end for 13: Sort r θ in an ascending order; 14: Apply the MSSE constraint (4) to find transition point (k ) that separates the 2 groups; 15: Store indices associated with points that have sorted indices > k (group 2) in indices vector K Indices ; 16: Store indices of the points associated with the first 2% residuals (group 3) in K Indices ; 17: Down-sample the selected keypoint indices until number of extracted keypoints is less than : Store points from P n associated with K Indices into P d ; 19: return P d ; are outlined in Algorithm 2. A. Preprocessing steps To improve the efficiency and accuracy of the registration, we exclude points that are more than 5 meters away from the sensor (due to the decrease in depth precision the further the points are from the camera). We then uniformly sample the point cloud (containing around 300, 000 points) in a way that all points lying inside a predefined radius are reduced to a single point (their centroid). As a result, the minimum distances between points in the point cloud are constrained and the total number of points are reduced. The point cloud spatial resolution decreases the further the scene is from the RGB-D sensor. To insure that the sampled points are distributed fairly uniformly in depth, we assign a variable search radius to every frame based on the average distance between each point in the point cloud and its nearest neighbors. In our experiments, the above procedure reduces the number of points to around 3500 points. B. Normal vector estimation We estimate the normal to a surface at a given point by fitting a plane to the point and its neighbors in a search area that is calculated in a similar way to the method explained in section IV-A. The search area is intentionally large to include points from adjacent surfaces, therefore affecting the surface normals near edges and corners. By doing so, we differentiate between the angles of normals lying on either plane and those near corners and edges. This concept is illustrated in Fig. 5 in

6 Algorithm 2 Step-by-Step Algorithm of Proposed 3D Registration and Mapping Method 1: Input: Point cloud (previous frame) P t, point cloud (current frame) P s ) 2: Output: Transformation between frames T, Global pose of camera G n, Map M 3: Filter point clouds; 4: Uniformly down-sample P t and P s to approximately 3, 500; 5: Calculate normal vectors at each point in P t and P s ; 6: Extract ROS features from P t and P s as described in Algorithm 1; 7: Assign SHOT descriptors to each feature point in P t and P s ; 8: Initially match the extracted features between P t and P s using their descriptors; 9: Obtain the inliers (good matches) using MSSE and estimate the 6DOF transformation T between the two consecutive frames; 10: Concatenate the estimated transformations to obtain a global pose of the camera G n. 11: Transform and map (M) the points with respect to a global reference frame; 12: return T, G n and M; Fig. 5: Example of estimated surface normals for a subset of points using a small search radius (left) large search radius (right) (Courtesy of and permission granted by Radu B. Rasu [22]). which the effect of using a large search area is demonstrated in the right image. C. 3D Feature Extraction and Matching The process of constructing a descriptor vector, even for around 3500 points, and matching those with their corresponding points from the previous frame is still computationally expensive. To improve the computation efficiency, we need to find a way to reduce the number of points without losing the information required for registration. To achieve this, we developed an informative sampling scheme based on using ranked order statistics (described in section III). After applying this method, the resulting point cloud would contain around 300 to 400 points. A feature descriptor is then computed for each extracted keypoint. We found that SHOT descriptors [21] provided the most accurate matches while it is only slightly more computationally expensive than other descriptors. Matching was performed, using a mutual consistency check, by finding the nearest neighbors in the descriptor vector space from the source point cloud (at time t) to the target point cloud (at time t 1) and vice versa. Only pairs of corresponding points that are mutually matched to each other are considered as the initial correspondences. D. Inlier Detection and Initial Transformation Estimation Matching 3D features between consecutive images using their descriptors usually results in many false matches (outliers). To remove the effect of these false matches, we refine the matches using a high breakdown point robust estimator (e.g. MSSE [6]). The MSSE is an extension of the robust least K-th order statistical estimator and can tolerate much higher ratio of outliers compared to RANSAC (we previously showed that MSSE outperforms RANSAC in terms of registration accuracy [7]). MSSE is comprised of two main steps, first the MSSE fits a given model to a set of data that contains outliers. In the second step, the scale of noise from this data is estimated. In contrast to RANSAC, MSSE is flexible and does not rely on the strong fixed error threshold assumption (which forms the bases of the RANSAC algorithm) [7]. Instead, MSSE assumes that the minimum number of inliers that are required to define a structure is known a priori. After setting the minimum number of inliers (ḱ), the implementation of MSSE is very straightforward and is as follows. Thousand sets of 3-tuples of corresponding initial matches are randomly chosen (although the minimum number of points required to constrain the transformation estimation is 2, we select 3 for obtaining a more accurate estimation [23]) and for each set, the transformation is estimated using the following equation: ( A d ) T = argmin w i T(p i s) p i t 2 T i=1 where T is the estimated transformation, p s and p t are the 3D coordinates of the matched feature points of the source and target frames respectively, A d is the number of correspondences (here A d = 3) and w i is a weighing factor that takes the distance between the point from the source point cloud to the camera into account (due to the decrease in depth precision and resolution with increasing the distance from the sensor). The estimated transformation is applied to all the features in the first frame and the distances between the transformed feature points and the corresponding points in the target frame are calculated (called residuals - r) and their squared values are sorted in ascending order. Having set the ḱ, then the transformation associated with the least ḱ-th order residual (from the 1000 hypotheses) is chosen as the best initial transformation. Using the residuals of that transformation, the inlier group members are chosen by applying (4) iteratively starting at i = ḱ and incrementing this value at each iteration until the condition is met. Finally, (5)

7 TABLE I: Comparison of the accuracy of 3D registration using different keypoint extraction methods. Translational errors are in meters, rotational errors are in degrees and time is in seconds. Method Trans. errorrot. errorextr. timereg. time# of points ROS SIFT3D ISS keypoints Covariance samp Uniform samp the transformation is re-calculated using all the obtained inliers. E. Global Pose Estimation and Mapping The transformation that was calculated in the previous section describes the motion between two RGB-D frames. In order to obtain a global pose of the camera with respect to a fixed reference frame (the initial frame), we concatenate all the transformations up to the current time using: G n = G n 1 T n,n 1 (6) where G n 1 is the previous global transformation with respect to the pose of the initial reference frame G 0 at n = 1 and T n,n 1 is the 6DOF transformation between the current frame and the previous frame. The map is then constructed by aligning the 3D points of the current frame with the previously constructed map (initially the map consists of the first frame) using the above global pose estimate. Fig. 1 shows a reconstructed 3D map of an office environment using the proposed method. V. EXPERIMENTAL RESULTS We evaluate the performance of our proposed 3D registration and mapping method by comparing it with the state of the art techniques. All methods were implemented using a Dell Latitude E6320, powered by an Intel i processor, 8 GB of RAM and running on Ubuntu The Robot Operating System (ROS Hydro) [24] and PCL 1.7 [16] were used for perception and 3D geometry processing. We used a Microsoft Kinect for capturing RGB-D data which operates at a frame-rate of 30 fps in VGA resolution mode. All of the experiments were conducted at typical university offices, which contain many texure-less objects (e.g. plain walls). In all of our experiments, the Microsoft Kinect was rotated counter clockwise until it was back to to its original position (360 ). During this process, the point clouds were acquired, registered, mapped and a global pose of the camera was estimated online. We performed 6 trails for each method and averaged the calculated results and errors. We evaluated the accuracy of all methods by calculating the average translational error (in the x,y and z directions) in meters, average rotational error (roll, pitch and yaw angles) in degrees and the average time of the registration process in seconds. A. Comparison between keypoint extraction methods In this experiment, we compared the accuracy of registration using our proposed keypoint extraction and some of the best available methods. The compared methods are: SIFT3D [16], ISS keypoints [20], geometrically stable points [19] (covariance sampling) and uniform sampling (we set a large search radius to obtain around 350 points). For this evaluation, we used the registration procedure explained in section IV. In addition to the previously mentioned measures, we also compared the average keypoint extraction time (in seconds) and average number of extracted keypoints. The results of these experiments are outlined in table I. The results show that our proposed method significantly outperforms the other methods in terms of both the translational and rotational accuracy. In terms of computational efficiency, our method was slightly slower than ISS keypoints, covariance sampling and uniform sampling. However, this does not significantly affect the registration efficiency, as they all required similar times for registration (covariance sampling was slightly slower as it requires a larger number of constraining feature points). This is due to the fact that the keypoint extraction method accounts for only around 10% of the total processing time. B. Sparse keypoints vs a denser point cloud In this experiment, we demonstrate the effectiveness of our informative sampling keypoint extraction method by comparing its registration accuracy and computational efficiency with another method that uses a much denser point cloud for registration. The point cloud was processed as it was described in section IV-A. Table III shows the results of this experiment. The results show that our method was able to achieve a very similar accuracy, despite using only 10% of the processed point cloud points and 0.11% of the original point cloud. The computational time for registration was also significantly improved (by a factor of 12). C. Comparison between registration methods In the final experiment, we evaluate the performance of our registration method by comparing it to the state of the art registration techniques. In the first of those methods, SIFT3D keypoints were extracted and matched using SHOT descriptors. Then SAC-IA [25] (a sampling consensus method that is similar to RANSAC but instead of selecting random samples between the source and target clouds, samples are selected based on points with the most similar feature histograms) was applied for rejecting false matches and initial transformation estimation, followed by ICP for refining the initial transformation. For the second method, the same procedure is repeated with replacing the ICP with a 3D normal distribution transform (NDT) [26] refinement step. The last method uses the above keypoint extraction and matching, followed by a pre-rejection RANSAC algorithm [27]. This method adds an additional verification step to the standard RANSAC algorithm which eliminates some false matches by analyzing their geometry. Table II shows the result of this comparison.

8 TABLE II: Comparison of the accuracy of different 3D registration methods. Method Trans. error(m)rot. error( )Extr. time(s) Proposed registration SIFT3D+SAC-IA+ICP SIFT3D+SAC-IA+NDT SIFT3D+pre-rejective RANSAC Method TABLE III: Sparse vs dense. Trans. error(m)rot. error( )Reg. time(s)# of points Proposed registration Uniform sampling Our proposed method outperforms the other methods both in terms of accuracy and computational efficiency. VI. CONCLUSIONS In this paper, we presented a novel real-time 3D registration and mapping system that uses only the depth information provided by RGB-D sensors, without relying on any visual information. As such, our method is well suited to perform localization and mapping in environments with insufficient visual texture information, as long as there is sufficient constraining geometric information in the scene. Since registration using the dense point cloud is a very computationally expensive operation, we propose a novel sampling scheme that informatively selects the points carrying the most useful information. We showed that our proposed registration method outperforms other well known registration methods in terms of accuracy and computational efficiency. Having said that, depth only registration methods struggle when registering scenes with very little structure. As such, we plan on aiding the depth information with visual information when available. Another aspect that our method struggled with was the registration in the presence of multiple motions (such as people walking in the scene). Since we use a high-breakdown estimator in our method, we think it should be possible to segment different motions explicitly and register those separately. This is part of our future work. ACKNOWLEDGMENT The first author would like to acknowledge the financial support provided by the Australian Postgraduate Award (APA) scholarship. REFERENCES [1] H. Durrant-Whyte, D. Rye, and E. Nebot, Localization of autonomous guided vehicles, ROBOTICS RESEARCH-INTERNATIONAL SYMPOSIUM-, vol. 7, pp , [2] S. Thrun, Robotic mapping: A survey, Exploring artificial intelligence in the new millennium, pp. 1 35, [3] R. Newcombe, S. Izadi, O. Hilliges, D. Molyneaux, D. Kim, A. Davison, P. Kohli, J. Shotton, S. Hodges, and A. Fitzgibbon, Kinectfusion: Real-time dense surface mapping and tracking, in Mixed and Augmented Reality (ISMAR), th IEEE International Symposium on. IEEE, 2011, pp [4] P. Henry, M. Krainin, E. Herbst, X. Ren, and D. Fox, Rgb-d mapping: Using kinect-style depth cameras for dense 3d modeling of indoor environments, The International Journal of Robotics Research, vol. 31, no. 5, pp , [5] F. Endres, J. Hess, N. Engelhard, J. Sturm, D. Cremers, and W. Burgard, An evaluation of the rgb-d slam system, in Robotics and Automation (ICRA), 2012 IEEE International Conference on. IEEE, 2012, pp [6] A. Bab-Hadiashar and D. Suter, Robust segmentation of visual data using ranked unbiased scale estimate, Robotica, vol. 17, no. 6, pp , [7] K. Yousif, A. Bab-Hadiashar, and R. Hoseinnezhad, 3d registration in dark environments using rgb-d cameras, in Digital Image Computing: Techniques and Applications (DICTA), 2013 International Conference on, 2013, pp [8] E. Rosten and T. Drummond, Machine learning for high-speed corner detection, Computer Vision ECCV 2006, pp , [9] A. Segal, D. Haehnel, and S. Thrun, Generalized-icp. in Robotics: Science and Systems, vol. 2, 2009, p. 4. [10] H. Bay, T. Tuytelaars, and L. Van Gool, Surf: Speeded up robust features, Computer Vision ECCV 2006, pp , [11] E. Rublee, V. Rabaud, K. Konolige, and G. Bradski, Orb: an efficient alternative to sift or surf, in Computer Vision (ICCV), 2011 IEEE International Conference on. IEEE, 2011, pp [12] H. Du, P. Henry, X. Ren, M. Cheng, D. Goldman, S. Seitz, and D. Fox, Interactive 3d modeling of indoor environments with a consumer depth camera, in Proceedings of the 13th international conference on Ubiquitous computing. ACM, 2011, pp [13] T. Whelan, M. Kaess, M. Fallon, H. Johannsson, J. Leonard, and J. McDonald, Kintinuous: Spatially extended kinectfusion, [14] G. Hu, S. Huang, L. Zhao, A. Alempijevic, and G. Dissanayake, A robust rgb-d slam algorithm, in Intelligent Robots and Systems (IROS), 2012 IEEE/RSJ International Conference on. IEEE, 2012, pp [15] C. Harris and M. Stephens, A combined corner and edge detector, in Alvey vision conference, vol. 15. Manchester, UK, 1988, p. 50. [16] R. B. Rusu and S. Cousins, 3d is here: Point cloud library (pcl), in Robotics and Automation (ICRA), 2011 IEEE International Conference on. IEEE, 2011, pp [17] S. Filipe and L. A. Alexandre, A comparative evaluation of 3d keypoint detectors. [18] D. Lowe, Distinctive image features from scale-invariant keypoints, International journal of computer vision, vol. 60, no. 2, pp , [19] N. Gelfand, L. Ikemoto, S. Rusinkiewicz, and M. Levoy, Geometrically stable sampling for the icp algorithm, in 3-D Digital Imaging and Modeling, DIM Proceedings. Fourth International Conference on. IEEE, 2003, pp [20] Y. Zhong, Intrinsic shape signatures: A shape descriptor for 3d object recognition, in Computer Vision Workshops (ICCV Workshops), 2009 IEEE 12th International Conference on. IEEE, 2009, pp [21] F. Tombari, S. Salti, and L. Di Stefano, Unique signatures of histograms for local surface description, in Computer Vision ECCV Springer, 2010, pp [22] R. B. Rusu, Semantic 3d object maps for everyday manipulation in human living environments, KI-Künstliche Intelligenz, vol. 24, no. 4, pp , [23] F. Fraundorfer and D. Scaramuzza, Visual odometry: Part ii: Matching, robustness, optimization, and applications, Robotics & Automation Magazine, IEEE, vol. 19, no. 2, pp , [24] M. Quigley, K. Conley, B. P. Gerkey, J. Faust, T. Foote, J. Leibs, R. Wheeler, and A. Y. Ng, Ros: an open-source robot operating system, in ICRA Workshop on Open Source Software, [25] R. B. Rusu, N. Blodow, and M. Beetz, Fast point feature histograms (fpfh) for 3d registration, in Robotics and Automation, ICRA 09. IEEE International Conference on. IEEE, 2009, pp [26] M. Magnusson, A. Lilienthal, and T. Duckett, Scan registration for autonomous mining vehicles using 3d-ndt, Journal of Field Robotics, vol. 24, no. 10, pp , [27] A. G. Buch, D. Kraft, J.-K. Kämäräinen, H. G. Petersen, and N. Krüger, Pose estimation using local structure-specific shape and appearance context, ICRA (accepted, 2013), 2013.

A Real-Time RGB-D Registration and Mapping Approach by Heuristically Switching Between Photometric And Geometric Information

A Real-Time RGB-D Registration and Mapping Approach by Heuristically Switching Between Photometric And Geometric Information A Real-Time RGB-D Registration and Mapping Approach by Heuristically Switching Between Photometric And Geometric Information The 17th International Conference on Information Fusion (Fusion 2014) Khalid

More information

3D SLAM in texture-less environments using rank order statistics Khalid Yousif, Alireza Bab-Hadiashar and Reza Hoseinnezhad

3D SLAM in texture-less environments using rank order statistics Khalid Yousif, Alireza Bab-Hadiashar and Reza Hoseinnezhad 3D SLAM in texture-less environments using rank order statistics Khalid Yousif, Alireza Bab-Hadiashar and Reza Hoseinnezhad School of Aerospace, Mechanical and Manufacturing Engineering, RMIT University,

More information

Removing Moving Objects from Point Cloud Scenes

Removing Moving Objects from Point Cloud Scenes Removing Moving Objects from Point Cloud Scenes Krystof Litomisky and Bir Bhanu University of California, Riverside krystof@litomisky.com, bhanu@ee.ucr.edu Abstract. Three-dimensional simultaneous localization

More information

3D Environment Reconstruction

3D Environment Reconstruction 3D Environment Reconstruction Using Modified Color ICP Algorithm by Fusion of a Camera and a 3D Laser Range Finder The 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems October 11-15,

More information

MonoRGBD-SLAM: Simultaneous Localization and Mapping Using Both Monocular and RGBD Cameras

MonoRGBD-SLAM: Simultaneous Localization and Mapping Using Both Monocular and RGBD Cameras MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com MonoRGBD-SLAM: Simultaneous Localization and Mapping Using Both Monocular and RGBD Cameras Yousif, K.; Taguchi, Y.; Ramalingam, S. TR2017-068

More information

Filtering and mapping systems for underwater 3D imaging sonar

Filtering and mapping systems for underwater 3D imaging sonar Filtering and mapping systems for underwater 3D imaging sonar Tomohiro Koshikawa 1, a, Shin Kato 1,b, and Hitoshi Arisumi 1,c 1 Field Robotics Research Group, National Institute of Advanced Industrial

More information

High-speed Three-dimensional Mapping by Direct Estimation of a Small Motion Using Range Images

High-speed Three-dimensional Mapping by Direct Estimation of a Small Motion Using Range Images MECATRONICS - REM 2016 June 15-17, 2016 High-speed Three-dimensional Mapping by Direct Estimation of a Small Motion Using Range Images Shinta Nozaki and Masashi Kimura School of Science and Engineering

More information

Mobile Point Fusion. Real-time 3d surface reconstruction out of depth images on a mobile platform

Mobile Point Fusion. Real-time 3d surface reconstruction out of depth images on a mobile platform Mobile Point Fusion Real-time 3d surface reconstruction out of depth images on a mobile platform Aaron Wetzler Presenting: Daniel Ben-Hoda Supervisors: Prof. Ron Kimmel Gal Kamar Yaron Honen Supported

More information

ROS2D: Image Feature Detector Using Rank Order Statistics

ROS2D: Image Feature Detector Using Rank Order Statistics MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com ROSD: Image Feature Detector Using Rank Order Statistics Yousif, K.; Taguchi, Y.; Ramalingam, S.; Bab-Hadiashar, A. TR7-66 May 7 Abstract We

More information

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University. 3D Computer Vision Structured Light II Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de 1 Introduction

More information

Ensemble of Bayesian Filters for Loop Closure Detection

Ensemble of Bayesian Filters for Loop Closure Detection Ensemble of Bayesian Filters for Loop Closure Detection Mohammad Omar Salameh, Azizi Abdullah, Shahnorbanun Sahran Pattern Recognition Research Group Center for Artificial Intelligence Faculty of Information

More information

Object Reconstruction

Object Reconstruction B. Scholz Object Reconstruction 1 / 39 MIN-Fakultät Fachbereich Informatik Object Reconstruction Benjamin Scholz Universität Hamburg Fakultät für Mathematik, Informatik und Naturwissenschaften Fachbereich

More information

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Structured Light II Johannes Köhler Johannes.koehler@dfki.de Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Introduction Previous lecture: Structured Light I Active Scanning Camera/emitter

More information

CVPR 2014 Visual SLAM Tutorial Kintinuous

CVPR 2014 Visual SLAM Tutorial Kintinuous CVPR 2014 Visual SLAM Tutorial Kintinuous kaess@cmu.edu The Robotics Institute Carnegie Mellon University Recap: KinectFusion [Newcombe et al., ISMAR 2011] RGB-D camera GPU 3D/color model RGB TSDF (volumetric

More information

Chapter 3 Image Registration. Chapter 3 Image Registration

Chapter 3 Image Registration. Chapter 3 Image Registration Chapter 3 Image Registration Distributed Algorithms for Introduction (1) Definition: Image Registration Input: 2 images of the same scene but taken from different perspectives Goal: Identify transformation

More information

DeReEs: Real-Time Registration of RGBD Images Using Image-Based Feature Detection And Robust 3D Correspondence Estimation and Refinement

DeReEs: Real-Time Registration of RGBD Images Using Image-Based Feature Detection And Robust 3D Correspondence Estimation and Refinement DeReEs: Real-Time Registration of RGBD Images Using Image-Based Feature Detection And Robust 3D Correspondence Estimation and Refinement Sahand Seifi Memorial University of Newfoundland sahands[at]mun.ca

More information

Robust Online 3D Reconstruction Combining a Depth Sensor and Sparse Feature Points

Robust Online 3D Reconstruction Combining a Depth Sensor and Sparse Feature Points 2016 23rd International Conference on Pattern Recognition (ICPR) Cancún Center, Cancún, México, December 4-8, 2016 Robust Online 3D Reconstruction Combining a Depth Sensor and Sparse Feature Points Erik

More information

5.2 Surface Registration

5.2 Surface Registration Spring 2018 CSCI 621: Digital Geometry Processing 5.2 Surface Registration Hao Li http://cs621.hao-li.com 1 Acknowledgement Images and Slides are courtesy of Prof. Szymon Rusinkiewicz, Princeton University

More information

AN INDOOR SLAM METHOD BASED ON KINECT AND MULTI-FEATURE EXTENDED INFORMATION FILTER

AN INDOOR SLAM METHOD BASED ON KINECT AND MULTI-FEATURE EXTENDED INFORMATION FILTER AN INDOOR SLAM METHOD BASED ON KINECT AND MULTI-FEATURE EXTENDED INFORMATION FILTER M. Chang a, b, Z. Kang a, a School of Land Science and Technology, China University of Geosciences, 29 Xueyuan Road,

More information

Local features and image matching. Prof. Xin Yang HUST

Local features and image matching. Prof. Xin Yang HUST Local features and image matching Prof. Xin Yang HUST Last time RANSAC for robust geometric transformation estimation Translation, Affine, Homography Image warping Given a 2D transformation T and a source

More information

Direct Methods in Visual Odometry

Direct Methods in Visual Odometry Direct Methods in Visual Odometry July 24, 2017 Direct Methods in Visual Odometry July 24, 2017 1 / 47 Motivation for using Visual Odometry Wheel odometry is affected by wheel slip More accurate compared

More information

CRF Based Point Cloud Segmentation Jonathan Nation

CRF Based Point Cloud Segmentation Jonathan Nation CRF Based Point Cloud Segmentation Jonathan Nation jsnation@stanford.edu 1. INTRODUCTION The goal of the project is to use the recently proposed fully connected conditional random field (CRF) model to

More information

3D Line Segment Based Model Generation by RGB-D Camera for Camera Pose Estimation

3D Line Segment Based Model Generation by RGB-D Camera for Camera Pose Estimation 3D Line Segment Based Model Generation by RGB-D Camera for Camera Pose Estimation Yusuke Nakayama, Hideo Saito, Masayoshi Shimizu, and Nobuyasu Yamaguchi Graduate School of Science and Technology, Keio

More information

Feature Based Registration - Image Alignment

Feature Based Registration - Image Alignment Feature Based Registration - Image Alignment Image Registration Image registration is the process of estimating an optimal transformation between two or more images. Many slides from Alexei Efros http://graphics.cs.cmu.edu/courses/15-463/2007_fall/463.html

More information

Uncertainties: Representation and Propagation & Line Extraction from Range data

Uncertainties: Representation and Propagation & Line Extraction from Range data 41 Uncertainties: Representation and Propagation & Line Extraction from Range data 42 Uncertainty Representation Section 4.1.3 of the book Sensing in the real world is always uncertain How can uncertainty

More information

PRECEDING VEHICLE TRACKING IN STEREO IMAGES VIA 3D FEATURE MATCHING

PRECEDING VEHICLE TRACKING IN STEREO IMAGES VIA 3D FEATURE MATCHING PRECEDING VEHICLE TRACKING IN STEREO IMAGES VIA 3D FEATURE MATCHING Daniel Weingerl, Wilfried Kubinger, Corinna Engelhardt-Nowitzki UAS Technikum Wien: Department for Advanced Engineering Technologies,

More information

Lecture 10 Detectors and descriptors

Lecture 10 Detectors and descriptors Lecture 10 Detectors and descriptors Properties of detectors Edge detectors Harris DoG Properties of detectors SIFT Shape context Silvio Savarese Lecture 10-26-Feb-14 From the 3D to 2D & vice versa P =

More information

Object Recognition with Invariant Features

Object Recognition with Invariant Features Object Recognition with Invariant Features Definition: Identify objects or scenes and determine their pose and model parameters Applications Industrial automation and inspection Mobile robots, toys, user

More information

Using Augmented Measurements to Improve the Convergence of ICP. Jacopo Serafin and Giorgio Grisetti

Using Augmented Measurements to Improve the Convergence of ICP. Jacopo Serafin and Giorgio Grisetti Jacopo Serafin and Giorgio Grisetti Point Cloud Registration We want to find the rotation and the translation that maximize the overlap between two point clouds Page 2 Point Cloud Registration We want

More information

Efficient SLAM Scheme Based ICP Matching Algorithm Using Image and Laser Scan Information

Efficient SLAM Scheme Based ICP Matching Algorithm Using Image and Laser Scan Information Proceedings of the World Congress on Electrical Engineering and Computer Systems and Science (EECSS 2015) Barcelona, Spain July 13-14, 2015 Paper No. 335 Efficient SLAM Scheme Based ICP Matching Algorithm

More information

LOCAL AND GLOBAL DESCRIPTORS FOR PLACE RECOGNITION IN ROBOTICS

LOCAL AND GLOBAL DESCRIPTORS FOR PLACE RECOGNITION IN ROBOTICS 8th International DAAAM Baltic Conference "INDUSTRIAL ENGINEERING - 19-21 April 2012, Tallinn, Estonia LOCAL AND GLOBAL DESCRIPTORS FOR PLACE RECOGNITION IN ROBOTICS Shvarts, D. & Tamre, M. Abstract: The

More information

NICP: Dense Normal Based Point Cloud Registration

NICP: Dense Normal Based Point Cloud Registration NICP: Dense Normal Based Point Cloud Registration Jacopo Serafin 1 and Giorgio Grisetti 1 Abstract In this paper we present a novel on-line method to recursively align point clouds. By considering each

More information

Replacing Projective Data Association with Lucas-Kanade for KinectFusion

Replacing Projective Data Association with Lucas-Kanade for KinectFusion Replacing Projective Data Association with Lucas-Kanade for KinectFusion Brian Peasley and Stan Birchfield Electrical and Computer Engineering Dept. Clemson University, Clemson, SC 29634 {bpeasle,stb}@clemson.edu

More information

A NEW FEATURE BASED IMAGE REGISTRATION ALGORITHM INTRODUCTION

A NEW FEATURE BASED IMAGE REGISTRATION ALGORITHM INTRODUCTION A NEW FEATURE BASED IMAGE REGISTRATION ALGORITHM Karthik Krish Stuart Heinrich Wesley E. Snyder Halil Cakir Siamak Khorram North Carolina State University Raleigh, 27695 kkrish@ncsu.edu sbheinri@ncsu.edu

More information

3D PLANE-BASED MAPS SIMPLIFICATION FOR RGB-D SLAM SYSTEMS

3D PLANE-BASED MAPS SIMPLIFICATION FOR RGB-D SLAM SYSTEMS 3D PLANE-BASED MAPS SIMPLIFICATION FOR RGB-D SLAM SYSTEMS 1,2 Hakim ELCHAOUI ELGHOR, 1 David ROUSSEL, 1 Fakhreddine ABABSA and 2 El-Houssine BOUYAKHF 1 IBISC Lab, Evry Val d'essonne University, Evry, France

More information

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Structured Light II Johannes Köhler Johannes.koehler@dfki.de Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Introduction Previous lecture: Structured Light I Active Scanning Camera/emitter

More information

Tracking an RGB-D Camera Using Points and Planes

Tracking an RGB-D Camera Using Points and Planes MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Tracking an RGB-D Camera Using Points and Planes Ataer-Cansizoglu, E.; Taguchi, Y.; Ramalingam, S.; Garaas, T. TR2013-106 December 2013 Abstract

More information

Using RGB Information to Improve NDT Distribution Generation and Registration Convergence

Using RGB Information to Improve NDT Distribution Generation and Registration Convergence Using RGB Information to Improve NDT Distribution Generation and Registration Convergence James Servos and Steven L. Waslander University of Waterloo, Waterloo, ON, Canada, N2L 3G1 Abstract Unmanned vehicles

More information

Segmentation and Tracking of Partial Planar Templates

Segmentation and Tracking of Partial Planar Templates Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract

More information

Inertial-Kinect Fusion for Outdoor 3D Navigation

Inertial-Kinect Fusion for Outdoor 3D Navigation Proceedings of Australasian Conference on Robotics and Automation, 2-4 Dec 213, University of New South Wales, Sydney Australia Inertial-Kinect Fusion for Outdoor 3D Navigation Usman Qayyum and Jonghyuk

More information

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014 SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014 SIFT SIFT: Scale Invariant Feature Transform; transform image

More information

arxiv: v1 [cs.cv] 28 Sep 2018

arxiv: v1 [cs.cv] 28 Sep 2018 Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,

More information

Feature Detectors and Descriptors: Corners, Lines, etc.

Feature Detectors and Descriptors: Corners, Lines, etc. Feature Detectors and Descriptors: Corners, Lines, etc. Edges vs. Corners Edges = maxima in intensity gradient Edges vs. Corners Corners = lots of variation in direction of gradient in a small neighborhood

More information

Surface Registration. Gianpaolo Palma

Surface Registration. Gianpaolo Palma Surface Registration Gianpaolo Palma The problem 3D scanning generates multiple range images Each contain 3D points for different parts of the model in the local coordinates of the scanner Find a rigid

More information

Monocular SLAM for a Small-Size Humanoid Robot

Monocular SLAM for a Small-Size Humanoid Robot Tamkang Journal of Science and Engineering, Vol. 14, No. 2, pp. 123 129 (2011) 123 Monocular SLAM for a Small-Size Humanoid Robot Yin-Tien Wang*, Duen-Yan Hung and Sheng-Hsien Cheng Department of Mechanical

More information

arxiv: v1 [cs.ro] 14 Mar 2016

arxiv: v1 [cs.ro] 14 Mar 2016 : A Novel Rotation, Illumination, Scale Invariant Appearance and Shape Feature Xiaoyang Li, Kanzhi Wu 2, Yong Liu, Ravindra Ranasinghe 2, Gamini Dissanayake 2 and Rong Xiong arxiv:6.44v [cs.ro] 4 Mar 26

More information

RGBD Point Cloud Alignment Using Lucas-Kanade Data Association and Automatic Error Metric Selection

RGBD Point Cloud Alignment Using Lucas-Kanade Data Association and Automatic Error Metric Selection IEEE TRANSACTIONS ON ROBOTICS, VOL. 31, NO. 6, DECEMBER 15 1 RGBD Point Cloud Alignment Using Lucas-Kanade Data Association and Automatic Error Metric Selection Brian Peasley and Stan Birchfield, Senior

More information

Specular 3D Object Tracking by View Generative Learning

Specular 3D Object Tracking by View Generative Learning Specular 3D Object Tracking by View Generative Learning Yukiko Shinozuka, Francois de Sorbier and Hideo Saito Keio University 3-14-1 Hiyoshi, Kohoku-ku 223-8522 Yokohama, Japan shinozuka@hvrl.ics.keio.ac.jp

More information

Computer Vision for HCI. Topics of This Lecture

Computer Vision for HCI. Topics of This Lecture Computer Vision for HCI Interest Points Topics of This Lecture Local Invariant Features Motivation Requirements, Invariances Keypoint Localization Features from Accelerated Segment Test (FAST) Harris Shi-Tomasi

More information

Dual Back-to-Back Kinects for 3-D Reconstruction

Dual Back-to-Back Kinects for 3-D Reconstruction Ho Chuen Kam, Kin Hong Wong and Baiwu Zhang, Dual Back-to-Back Kinects for 3-D Reconstruction, ISVC'16 12th International Symposium on Visual Computing December 12-14, 2016, Las Vegas, Nevada, USA. Dual

More information

3D object recognition used by team robotto

3D object recognition used by team robotto 3D object recognition used by team robotto Workshop Juliane Hoebel February 1, 2016 Faculty of Computer Science, Otto-von-Guericke University Magdeburg Content 1. Introduction 2. Depth sensor 3. 3D object

More information

Topological Mapping. Discrete Bayes Filter

Topological Mapping. Discrete Bayes Filter Topological Mapping Discrete Bayes Filter Vision Based Localization Given a image(s) acquired by moving camera determine the robot s location and pose? Towards localization without odometry What can be

More information

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS Cognitive Robotics Original: David G. Lowe, 004 Summary: Coen van Leeuwen, s1460919 Abstract: This article presents a method to extract

More information

Incremental Structured ICP Algorithm

Incremental Structured ICP Algorithm Incremental Structured ICP Algorithm Haokun Geng, Johnny Chien, Radu Nicolescu, and Reinhard Klette The.enpeda.. Project, Tamaki Campus The University of Auckland, New Zealand Abstract. Variants of the

More information

Reconstruction of 3D Models Consisting of Line Segments

Reconstruction of 3D Models Consisting of Line Segments Reconstruction of 3D Models Consisting of Line Segments Naoto Ienaga (B) and Hideo Saito Department of Information and Computer Science, Keio University, Yokohama, Japan {ienaga,saito}@hvrl.ics.keio.ac.jp

More information

3D Simultaneous Localization and Mapping in Texture-less and Structure-less Environments Using Rank Order Statistics

3D Simultaneous Localization and Mapping in Texture-less and Structure-less Environments Using Rank Order Statistics 3D Simultaneous Localization and Mapping in Texture-less and Structure-less Environments Using Rank Order Statistics Khalid Yousif B.Eng (Robotics and Mechatronics - 1 st Class Honours) Supervisors: Prof.

More information

Robot Localization based on Geo-referenced Images and G raphic Methods

Robot Localization based on Geo-referenced Images and G raphic Methods Robot Localization based on Geo-referenced Images and G raphic Methods Sid Ahmed Berrabah Mechanical Department, Royal Military School, Belgium, sidahmed.berrabah@rma.ac.be Janusz Bedkowski, Łukasz Lubasiński,

More information

Perception. Autonomous Mobile Robots. Sensors Vision Uncertainties, Line extraction from laser scans. Autonomous Systems Lab. Zürich.

Perception. Autonomous Mobile Robots. Sensors Vision Uncertainties, Line extraction from laser scans. Autonomous Systems Lab. Zürich. Autonomous Mobile Robots Localization "Position" Global Map Cognition Environment Model Local Map Path Perception Real World Environment Motion Control Perception Sensors Vision Uncertainties, Line extraction

More information

Robot localization method based on visual features and their geometric relationship

Robot localization method based on visual features and their geometric relationship , pp.46-50 http://dx.doi.org/10.14257/astl.2015.85.11 Robot localization method based on visual features and their geometric relationship Sangyun Lee 1, Changkyung Eem 2, and Hyunki Hong 3 1 Department

More information

Scanning and Printing Objects in 3D Jürgen Sturm

Scanning and Printing Objects in 3D Jürgen Sturm Scanning and Printing Objects in 3D Jürgen Sturm Metaio (formerly Technical University of Munich) My Research Areas Visual navigation for mobile robots RoboCup Kinematic Learning Articulated Objects Quadrocopters

More information

Integrating Depth and Color Cues for Dense Multi-Resolution Scene Mapping Using RGB-D Cameras

Integrating Depth and Color Cues for Dense Multi-Resolution Scene Mapping Using RGB-D Cameras In Proc. of the IEEE Int. Conf. on Multisensor Fusion and Information Integration (MFI), Hamburg, Germany, 2012 Integrating Depth and Color Cues for Dense Multi-Resolution Scene Mapping Using RGB-D Cameras

More information

A Stable and Accurate Marker-less Augmented Reality Registration Method

A Stable and Accurate Marker-less Augmented Reality Registration Method A Stable and Accurate Marker-less Augmented Reality Registration Method Qing Hong Gao College of Electronic and Information Xian Polytechnic University, Xian, China Long Chen Faculty of Science and Technology

More information

Perception IV: Place Recognition, Line Extraction

Perception IV: Place Recognition, Line Extraction Perception IV: Place Recognition, Line Extraction Davide Scaramuzza University of Zurich Margarita Chli, Paul Furgale, Marco Hutter, Roland Siegwart 1 Outline of Today s lecture Place recognition using

More information

Dense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm

Dense Tracking and Mapping for Autonomous Quadrocopters. Jürgen Sturm Computer Vision Group Prof. Daniel Cremers Dense Tracking and Mapping for Autonomous Quadrocopters Jürgen Sturm Joint work with Frank Steinbrücker, Jakob Engel, Christian Kerl, Erik Bylow, and Daniel Cremers

More information

Collecting outdoor datasets for benchmarking vision based robot localization

Collecting outdoor datasets for benchmarking vision based robot localization Collecting outdoor datasets for benchmarking vision based robot localization Emanuele Frontoni*, Andrea Ascani, Adriano Mancini, Primo Zingaretti Department of Ingegneria Infromatica, Gestionale e dell

More information

Personal Navigation and Indoor Mapping: Performance Characterization of Kinect Sensor-based Trajectory Recovery

Personal Navigation and Indoor Mapping: Performance Characterization of Kinect Sensor-based Trajectory Recovery Personal Navigation and Indoor Mapping: Performance Characterization of Kinect Sensor-based Trajectory Recovery 1 Charles TOTH, 1 Dorota BRZEZINSKA, USA 2 Allison KEALY, Australia, 3 Guenther RETSCHER,

More information

Visual Bearing-Only Simultaneous Localization and Mapping with Improved Feature Matching

Visual Bearing-Only Simultaneous Localization and Mapping with Improved Feature Matching Visual Bearing-Only Simultaneous Localization and Mapping with Improved Feature Matching Hauke Strasdat, Cyrill Stachniss, Maren Bennewitz, and Wolfram Burgard Computer Science Institute, University of

More information

Automatic Image Alignment (feature-based)

Automatic Image Alignment (feature-based) Automatic Image Alignment (feature-based) Mike Nese with a lot of slides stolen from Steve Seitz and Rick Szeliski 15-463: Computational Photography Alexei Efros, CMU, Fall 2006 Today s lecture Feature

More information

Instance-level recognition

Instance-level recognition Instance-level recognition 1) Local invariant features 2) Matching and recognition with local features 3) Efficient visual search 4) Very large scale indexing Matching of descriptors Matching and 3D reconstruction

More information

arxiv: v2 [cs.ro] 19 Sep 2016

arxiv: v2 [cs.ro] 19 Sep 2016 : A Novel Rotation, Illumination, Scale Invariant Appearance and Shape Feature Kanzhi Wu, Xiaoyang Li 2, Ravindra Ranasinghe, Gamini Dissanayake and Yong Liu 2 arxiv:6.44v2 [cs.ro] 9 Sep 26 Abstract This

More information

Motion Tracking and Event Understanding in Video Sequences

Motion Tracking and Event Understanding in Video Sequences Motion Tracking and Event Understanding in Video Sequences Isaac Cohen Elaine Kang, Jinman Kang Institute for Robotics and Intelligent Systems University of Southern California Los Angeles, CA Objectives!

More information

Feature Detection and Matching

Feature Detection and Matching and Matching CS4243 Computer Vision and Pattern Recognition Leow Wee Kheng Department of Computer Science School of Computing National University of Singapore Leow Wee Kheng (CS4243) Camera Models 1 /

More information

BRAND: A Robust Appearance and Depth Descriptor for RGB-D Images

BRAND: A Robust Appearance and Depth Descriptor for RGB-D Images BRAND: A Robust Appearance and Depth Descriptor for RGB-D Images Erickson R. Nascimento Gabriel L. Oliveira Mario F. M. Campos Antônio W. Vieira William Robson Schwartz Abstract This work introduces a

More information

A Systems View of Large- Scale 3D Reconstruction

A Systems View of Large- Scale 3D Reconstruction Lecture 23: A Systems View of Large- Scale 3D Reconstruction Visual Computing Systems Goals and motivation Construct a detailed 3D model of the world from unstructured photographs (e.g., Flickr, Facebook)

More information

Determinant of homography-matrix-based multiple-object recognition

Determinant of homography-matrix-based multiple-object recognition Determinant of homography-matrix-based multiple-object recognition 1 Nagachetan Bangalore, Madhu Kiran, Anil Suryaprakash Visio Ingenii Limited F2-F3 Maxet House Liverpool Road Luton, LU1 1RS United Kingdom

More information

Robotics Programming Laboratory

Robotics Programming Laboratory Chair of Software Engineering Robotics Programming Laboratory Bertrand Meyer Jiwon Shin Lecture 8: Robot Perception Perception http://pascallin.ecs.soton.ac.uk/challenges/voc/databases.html#caltech car

More information

3D Perception. CS 4495 Computer Vision K. Hawkins. CS 4495 Computer Vision. 3D Perception. Kelsey Hawkins Robotics

3D Perception. CS 4495 Computer Vision K. Hawkins. CS 4495 Computer Vision. 3D Perception. Kelsey Hawkins Robotics CS 4495 Computer Vision Kelsey Hawkins Robotics Motivation What do animals, people, and robots want to do with vision? Detect and recognize objects/landmarks Find location of objects with respect to themselves

More information

Outline 7/2/201011/6/

Outline 7/2/201011/6/ Outline Pattern recognition in computer vision Background on the development of SIFT SIFT algorithm and some of its variations Computational considerations (SURF) Potential improvement Summary 01 2 Pattern

More information

A New Approach For 3D Image Reconstruction From Multiple Images

A New Approach For 3D Image Reconstruction From Multiple Images International Journal of Electronics Engineering Research. ISSN 0975-6450 Volume 9, Number 4 (2017) pp. 569-574 Research India Publications http://www.ripublication.com A New Approach For 3D Image Reconstruction

More information

Instance-level recognition

Instance-level recognition Instance-level recognition 1) Local invariant features 2) Matching and recognition with local features 3) Efficient visual search 4) Very large scale indexing Matching of descriptors Matching and 3D reconstruction

More information

A System of Image Matching and 3D Reconstruction

A System of Image Matching and 3D Reconstruction A System of Image Matching and 3D Reconstruction CS231A Project Report 1. Introduction Xianfeng Rui Given thousands of unordered images of photos with a variety of scenes in your gallery, you will find

More information

K-Means Based Matching Algorithm for Multi-Resolution Feature Descriptors

K-Means Based Matching Algorithm for Multi-Resolution Feature Descriptors K-Means Based Matching Algorithm for Multi-Resolution Feature Descriptors Shao-Tzu Huang, Chen-Chien Hsu, Wei-Yen Wang International Science Index, Electrical and Computer Engineering waset.org/publication/0007607

More information

IRON: A Fast Interest Point Descriptor for Robust NDT-Map Matching and its Application to Robot Localization*

IRON: A Fast Interest Point Descriptor for Robust NDT-Map Matching and its Application to Robot Localization* IRON: A Fast Interest Point Descriptor for Robust NDT-Map Matching and its Application to Robot Localization* Thomas Schmiedel, Erik Einhorn and Horst-Michael Gross 1 Abstract This work introduces the

More information

Incremental compact 3D maps of planar patches from RGBD points

Incremental compact 3D maps of planar patches from RGBD points Incremental compact 3D maps of planar patches from RGBD points Juan Navarro and José M. Cañas Universidad Rey Juan Carlos, Spain Abstract. The RGBD sensors have opened the door to low cost perception capabilities

More information

Department of Electrical and Electronic Engineering, University of Peradeniya, KY 20400, Sri Lanka

Department of Electrical and Electronic Engineering, University of Peradeniya, KY 20400, Sri Lanka WIT: Window Intensity Test Detector and Descriptor T.W.U.Madhushani, D.H.S.Maithripala, J.V.Wijayakulasooriya Postgraduate and Research Unit, Sri Lanka Technological Campus, CO 10500, Sri Lanka. Department

More information

Data-driven Depth Inference from a Single Still Image

Data-driven Depth Inference from a Single Still Image Data-driven Depth Inference from a Single Still Image Kyunghee Kim Computer Science Department Stanford University kyunghee.kim@stanford.edu Abstract Given an indoor image, how to recover its depth information

More information

Instance-level recognition part 2

Instance-level recognition part 2 Visual Recognition and Machine Learning Summer School Paris 2011 Instance-level recognition part 2 Josef Sivic http://www.di.ens.fr/~josef INRIA, WILLOW, ENS/INRIA/CNRS UMR 8548 Laboratoire d Informatique,

More information

Fast Image Matching Using Multi-level Texture Descriptor

Fast Image Matching Using Multi-level Texture Descriptor Fast Image Matching Using Multi-level Texture Descriptor Hui-Fuang Ng *, Chih-Yang Lin #, and Tatenda Muindisi * Department of Computer Science, Universiti Tunku Abdul Rahman, Malaysia. E-mail: nghf@utar.edu.my

More information

BendIT An Interactive Game with two Robots

BendIT An Interactive Game with two Robots BendIT An Interactive Game with two Robots Tim Niemueller, Stefan Schiffer, Albert Helligrath, Safoura Rezapour Lakani, and Gerhard Lakemeyer Knowledge-based Systems Group RWTH Aachen University, Aachen,

More information

CSC494: Individual Project in Computer Science

CSC494: Individual Project in Computer Science CSC494: Individual Project in Computer Science Seyed Kamyar Seyed Ghasemipour Department of Computer Science University of Toronto kamyar@cs.toronto.edu Abstract One of the most important tasks in computer

More information

RGB-D Edge Detection and Edge-based Registration

RGB-D Edge Detection and Edge-based Registration 23 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) November 3-7, 23. Tokyo, Japan RGB-D Edge Detection and Edge-based Registration Changhyun Choi, Alexander J. B. Trevor, and

More information

Large Scale 3D Reconstruction by Structure from Motion

Large Scale 3D Reconstruction by Structure from Motion Large Scale 3D Reconstruction by Structure from Motion Devin Guillory Ziang Xie CS 331B 7 October 2013 Overview Rome wasn t built in a day Overview of SfM Building Rome in a Day Building Rome on a Cloudless

More information

Visual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

Visual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech Visual Odometry Features, Tracking, Essential Matrix, and RANSAC Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline The

More information

Memory Management Method for 3D Scanner Using GPGPU

Memory Management Method for 3D Scanner Using GPGPU GPGPU 3D 1 2 KinectFusion GPGPU 3D., GPU., GPGPU Octree. GPU,,. Memory Management Method for 3D Scanner Using GPGPU TATSUYA MATSUMOTO 1 SATORU FUJITA 2 This paper proposes an efficient memory management

More information

Visual Tracking (1) Tracking of Feature Points and Planar Rigid Objects

Visual Tracking (1) Tracking of Feature Points and Planar Rigid Objects Intelligent Control Systems Visual Tracking (1) Tracking of Feature Points and Planar Rigid Objects Shingo Kagami Graduate School of Information Sciences, Tohoku University swk(at)ic.is.tohoku.ac.jp http://www.ic.is.tohoku.ac.jp/ja/swk/

More information

Motion illusion, rotating snakes

Motion illusion, rotating snakes Motion illusion, rotating snakes Local features: main components 1) Detection: Find a set of distinctive key points. 2) Description: Extract feature descriptor around each interest point as vector. x 1

More information

Building a Panorama. Matching features. Matching with Features. How do we build a panorama? Computational Photography, 6.882

Building a Panorama. Matching features. Matching with Features. How do we build a panorama? Computational Photography, 6.882 Matching features Building a Panorama Computational Photography, 6.88 Prof. Bill Freeman April 11, 006 Image and shape descriptors: Harris corner detectors and SIFT features. Suggested readings: Mikolajczyk

More information

Using Geometric Blur for Point Correspondence

Using Geometric Blur for Point Correspondence 1 Using Geometric Blur for Point Correspondence Nisarg Vyas Electrical and Computer Engineering Department, Carnegie Mellon University, Pittsburgh, PA Abstract In computer vision applications, point correspondence

More information

Appearance-Based Place Recognition Using Whole-Image BRISK for Collaborative MultiRobot Localization

Appearance-Based Place Recognition Using Whole-Image BRISK for Collaborative MultiRobot Localization Appearance-Based Place Recognition Using Whole-Image BRISK for Collaborative MultiRobot Localization Jung H. Oh, Gyuho Eoh, and Beom H. Lee Electrical and Computer Engineering, Seoul National University,

More information

Application questions. Theoretical questions

Application questions. Theoretical questions The oral exam will last 30 minutes and will consist of one application question followed by two theoretical questions. Please find below a non exhaustive list of possible application questions. The list

More information