Lazy Data Association For Image Sequences Matching Under Substantial Appearance Changes

Size: px
Start display at page:

Download "Lazy Data Association For Image Sequences Matching Under Substantial Appearance Changes"

Transcription

1 IEEE ROBOTICS AND AUTOMATION LETTERS. PREPRINT VERSION. DECEMBER, 25 Lazy Data Association For Image Sequences Matching Under Substantial Appearance Changes Olga Vysotska and Cyrill Stachniss Abstract Localization is an essential capability for mobile robots and the ability to localize in changing environments is key to robust outdoor navigation. Robots operating over extended periods of time should be able to handle substantial appearance changes such as those occurring over seasons or under different weather conditions. In this paper, we investigate the problem of efficiently coping with seasonal appearance changes in online localization. We propose a lazy data association approach for matching streams of incoming images to a reference image sequence in an online fashion. We present a search heuristic to quickly find matches between the current image sequence and a database using a data association graph. Our experiments conducted under substantial seasonal changes suggest that our approach can efficiently match image sequences while requiring a comparably small number of image to image comparisons. Index Terms Localization, Place Recognition, Visual-Based Navigation I. I NTRODUCTION L OCALIZATION is essential for goal-directed navigation. The ability to identify that a robot is at a previously visited place is an important element of localization. Handling large appearance changes such as those depicted in Figure is a challenging problem and cannot be neglected in the context of persistent autonomous navigation. Localization through image matching or through appearance-based approaches for handling seasonal changes has been addressed by different researchers in the past, for example [6], [7], [], [22], [27]. Several visual place recognition systems exploit features such as SURF [2] or SIFT [6]. These features encode local gradients computed at keypoints. Matching approaches relying on such features can deal with viewpoint changes and they show a great performance if the appearance of the environment does not change dramatically. They, however, perform rather poor under extreme perceptual changes. Recently, a series of robust visual localization approaches has been proposed including FAB-MAP2 [7], SeqSLAM [9], SP-ACP [23], as well as [22], [3]. Some of these methods have been shown to robustly recognize previously seen locations even under a wide spectrum of visual changes including dynamic objects, different illumination, and varying weather conditions. Manuscript received: August, 28, 25; Revised October, 26, 25; Accepted December, 7, 25. This paper was recommended for publication by Editor Jingshan Li upon evaluation of the Associate Editor and Reviewers comments. This work has partly been supported by the European Commission under the grant numbers FP7-663-EUROPA2. The authors are with Institute for Geodesy and Geoinformation, University of Bonn, Germany olga.vysotska@uni-bonn.de Digital Object Identifier xxxxxxxxxxxxxxxxxxx Fig. : Examples of typical image pairs taken at the same places within multiple datasets. The image pairs are successively found by our approach. First row: Freiburg dataset; second row: Alderley dataset; third row: Nordland dataset. Fourth row: day/night scene from the VPRiCE 5 Challenge dataset. It is essential for robot navigation that the location information is available in a timely manner, i.e., that the algorithm provides an online solution to the localization problem. Several existing techniques operate either in an online fashion but have issues to deal with strong seasonal changes or they can handle such changes but may not work online. In our work, we address the problem of online image sequence matching tailored to situations with large appearance changes. The contribution of this paper is an online approach to image sequence matching that uses a data association heuristic while searching for matching image sequences in a data association graph. The heuristic estimates the expected cost of matching images considering the best matches found so far. Our directed acyclic data association graph is built incremen-

2 2 IEEE ROBOTICS AND AUTOMATION LETTERS. PREPRINT VERSION. DECEMBER, 25 Fig. 2: Illustration of the fact that SIFT features do not perform well under strong seasonal variations. As a result of the seasonal changes, most SIFT matches illustrated by the lines between the images are outliers as the lines do not connect corresponding points. tally whenever new sensor data arrives and its leaves model the data association hypotheses currently under consideration. For matching images, we rely on learned features from deep convolutional neural networks and the image similarity defines the cost in the data association graph. We furthermore show how additional location information can be exploited in the process and evaluate our method on several real world datasets. II. RELATED WORK Pose estimation is a frequently studied problem in robotics and different approaches have been proposed for visual localization [], [3], [7], [8], [9]. The ability to localize is an essential prerequisite for most autonomous navigation systems. Dealing with substantial variations in the visual input has been recognized as an obstacle for persistent autonomous navigation and this problem has been investigated by different researchers [7], [], [5]. The majority of visual place recognition systems exploit features such as SURF [2] or SIFT [6] and several approaches apply bag-of-words techniques, i.e. they perform matching based on an appearance statistics of such features. To improve the robustness of appearance-based place recognition, Stumm et al. [27] consider the constellations of visual words and keeping track of their covisibility. Often, approaches using SIFT or SURF show a great performance if the appearance of the environment does not change radically. As also reported in our previous work [22], the matching performance of SIFT or SURF degrades under strong perceptual changes. An example, which illustrates this fact, is shown in Figure 2. The two images are taken from the same place and similar view points but during different seasons. As can be seen from the matches illustrated through the yellow lines, most of the correspondences are outliers. This examples illustrates our experience, that SIFT and SURF features are not well-suited for image matching under strong seasonal variations. Across season matching using SIFT and SURF has been investigated by Valgren and Lilienthal [29] by combining features and geometric constraints, which can improve the matching. In previous approaches, we proposed the use of tessellated HOG features [22], [3]. In contrast to that, in this paper we apply deeply learned features proposed by Sermanet et al. [25] and suggested for place recognition by Chen et al. [5]. We use them as an alternative to tessellated HOG features as they provide a better matching performance in our settings. Another recent work [28] suggests a technique for place recognition, where features stem from convolutional neural networks. The authors extract features from the landmark proposals, construct the similarity matrix by comparing the landmark features using the cosine distance and also take into account the size of the bounding boxes for the matched landmarks. The recognition task is then performed by selecting individual matches based on the highest similarity score. In terms of aligning image sequences, several approaches have been proposed. For example, Matsumoto et al. [7] use the image sequences and directional relations between images to perform visual navigation in a corridor environment. SeqSLAM [9], which also aims at matching image sequences under strong seasonal changes, computes an image-by-image matching matrix that stores dissimilarity scores between all images in a query and database sequence. SeqSLAM computes a straight-line path through the matching matrix and select the path with the smallest sum of dissimilarity scores across image pairs to determine the matching route. Milford et al. [8] also present a comprehensive study about the SeqSLAM performance on low resolution images. Related to that, Naseer et al. [22] focus on offline sequence matching using a network flow approach. If odometry is available, this approach can also be combined with a least squares SLAM system to build metric maps [2]. The approach by Neubert et al. [23] aims at predicting the change in appearance on top of a vocabulary. For the vocabulary, the method predicts the change of the visual word over different seasons but the learning phase requires an accurate image alignment over seasons. A recent approach by Johns and Young [4] builds a statistic on the co-occurrence of features under different conditions. It relies on the ability to detect stable and discriminative features over different seasons. Finding such discriminative and stable features under strong changes is however a challenge. To avoid addressing the problems of finding features that are robust under extreme perceptual differences, Churchill and Newman [6] store different appearances for each place. These so-called experiences enable a robot to localize in previously learned experiences and associate a new data to places. A recent extension of that paradigm targets vast-scale localization by exploiting a prioritized recollection of relevant experiences so that the number of matches can be reduced [5]. Biber and Duckett [4] address the problem of dealing with changes in the environments by representing maps at different time scales. Each map is maintained and updated using the sensor data modeling short-term and long-term memories. This enables handling variations. In contrast to that, Stachniss and Burgard [26] model different instances of typical world states using clustering. There are furthermore approaches combining laser and visual information for large-scale localization at city scale. The approach of Pascoe et al. [24] exploits laser data and vision information during the mapping phase with a survey vehicle but can then localize a car only using a camera. To achieve a visual localization in a long term autonomy setup, Furgale and Barfoot [] propose a teach and repeat system that is based on a stereo setup. The approach exploits

3 VYSOTSKA AND STACHNISS: LAZY DATA ASSOCIATION FOR IMAGE SEQUENCES MATCHING UNDER SUBSTANTIAL APPEARANCE CHANGES 3 query images database images : : 2: : source Fig. 3: Schematic illustration of the graph structure for the search. To perform an online localization our algorithm compares only image pairs that correspond to the green nodes and expands the green area on the fly. Red nodes correspond to matches of similar images along the path through the data association graph and blue indicates matches along the path with a low similarity. local submaps and enables a robot to navigate on long trajectories but this method does not address large perceptual changes with respect to the taught path. In this paper, we introduce a lazy data association scheme inspired by the work of Hähnel et al. [3] to come up with an online solution that can be executed on a robot while navigating. A key goal is to reduce the number of imageto-image comparisons with respect to existing methods such as [9], [22], [3]. In contrast to the work by Hähnel et al., we use a heuristic that considers the cost of the path taken so far in order to speed-up the search. Our work is an extension of a short paper [3] presented at the ICRA 25 Workshop on Visual Place Recognition in Changing Environments. III. LAZY MATCHING FOR ONLINE OPERATION The main goal of this paper is to propose an online algorithm that exploits the sequence information to perform a global localization under strong appearance changes. We perform localization in the sense that we match a sequence Q of the images that we receive from sensor with a reference or database sequence of images called D. For every incoming image, we want to know if there is a corresponding match in the database and if so, to which image of the database it corresponds to. The database itself is organized as regular files. In memory, we only keep an index to the images and feature descriptions and load individual images on demand from disk. A. Data Association Graph We use a directed acyclic graph G = (X, E) as our main data structure for modeling the data association problem. We model the sequential image matching task as finding a shortest path in the data association graph G. We build up the data association graph on the fly and therefore only need to compare images if our search algorithm expands the corresponding node in G. The key idea of this data association graph can be explained as follows. A node in the graph represents a potential match between two images. We aim at finding the best combination of matching images by searching a path through this graph, see Figure 3 for an illustration, where the cost of visiting a node depends on the similarity of both images. In more detail, we propose the following graph structure. a) Nodes: We have two types of nodes in X: the root or start node x s and matching nodes. A matching node x ij models a match of the image i Q with the image j D. The more similar two images are, the more likely is the fact that they represent the same place. The similarity of an image pair is defined as z ij [, ], where z ij = means that both images appear identical. The similarity z ij is computed by comparing the images i Q and j D only through their global image descriptor using the cosine distance. As we are building the graph online, new nodes x ij need to be created as soon as a new image i is recorded. Adding a node x ij to the graph, however, comes at a computational cost as we need to compare images to compute z ij. Thus, for building up the graph, we seek to avoid instantiating unnecessary nodes x ij, i.e. nodes, which are not part of the matching sequence. b) Edges: Similar to the nodes, we use two types of edges E = {E s, E X } according to the types of nodes the edges connect. The set of edges E s connects the source node x s with the matching nodes x j corresponding to matching the first query image with any database image j D, i.e., E s = {(x s, x j )} j D. () The second set of edges E X connects the matching nodes. In our approach, we define the set E X of edges in a slightly different way to [3], [22] as E X = {(x ij, x (i+)k )} k=j K,...,j+K, (2) where K is a fanout parameter that influences the nodes that are connected between the query images i and i+. The fanout basically models that the robot can move at different speeds through the environment or that the cameras can operate at different framerates. The larger K is, the larger the branching factor of the graph and, thus, the value of K impacts the speed of the search described in the following sections. In our current implementation, we use a constant fanout parameter of K = 5. In the remainder of this paper, the nodes x (i+)k are referred to as the children ch(x ij ) of the node x ij. c) Weights/Costs: Each edge e E has a weight or cost associated to it. This weight w(e) is related to the similarity score z ij. The weight of an edge e = (x ij, x i j ) EX is inverse proportional to the similarity of the node to which this edges leads to, i.e. w(e) =, (3) z i j where z i j is a similarity score computed when comparing image i and j using the cosine distance. B. Computing Image Similarity based on Features from Deep Convolutional Neural Networks The computation of the similarity of two images has to be done often and thus we are in general interested in a fast computation. Nevertheless, the quality of the similarity

4 4 IEEE ROBOTICS AND AUTOMATION LETTERS. PREPRINT VERSION. DECEMBER, 25 match unlikely pop heuristic possible match expand level reached yes match found no Fig. 5: Illustration of searching for a match for an input image. Fig. 4: Similarity matrix computed using tessellated HOGs as in [22] (left) and OverFeat features (right). As can be seen in the first row, the OverFeat features yield more distinct similarity values. This leads to a smaller number of nodes that are instantiated in the data association graph (green), as depicted in the second row. function is of high importance. The more distinct the value of z ij are for images taken from the same places vs. from different places, the easier the data association problem. As a result, the more distinct such values are, the better the performance of our graph search algorithm as less nodes will need to be expanded. In our previous works [22], [3], we computed the tessellated HOG descriptor for every image and then compared them using the cosine distance. The obtained cost difference between the best match and the worse match was sufficient to find good solution with an exhaustive search. In the context of our lazy data association approach with a search heuristic, we experience problems to find matching sequences reliably without expanding the majority of the nodes in the graph. Therefore, we changed the image descriptors in this work to the pre-trained deep convolutional neural network OverFeat as proposed by Sermanet et al. [25] due to its better matching performance. OverFeat is built using a network trained on the ImageNet dataset consisting of.2 million images. We used the th layer as a global image feature as suggested by Chen et al. [5]. Using OverFeat features instead of HOG directly improves the performance of our algorithm and supports the lazy approach. To give an intuition about the matching similarity, Figure 4 depicts the similarity of comparing all possible combinations of images from database D and query Q computed with the tessellated HOG descriptor (left) and OverFeat (right). Brighter values indicate a higher similarity. As can be seen from the images, the OverFeat features lead to more distinct values (higher contrast). C. Image Sequence Matching through Graph Search The sequence of matching images between Q and D can be computed by a path search from the start node x s to any node x l, with referring to any index in D and l being the most recent image in Q. Every node that is a part of the shortest path corresponds to a selected data association, i.e. a match. The computationally most demanding process for building and searching in such a data association graph is instantiating all nodes as a large number of possible matches has to be computed. For online localization, we are interested in keeping the computational efforts small and in avoiding the creation of nodes that do not represent matches. To address this issue, we propose an approach that limits the number of image comparisons and results in an efficient algorithm. Our work is motivated by the ideas of lazy data associations in the context of SLAM proposed by Hähnel et al. [3] for constructing pose graphs. Hähnel et al. build up a data association tree and expand in each round the node with the highest log likelihood of representing a match between laser range scans. This is similar to a greedy search in a data association tree. In our case, we go a step further and seek to performing an informed search through the graph, while the graph is built on the fly. One popular way to perform an informed search is the A* algorithm using a heuristic, which allows for estimating the cost from the currently expanded node to the goal node. For our matching problem, that means we need to predict how well the images that we will receive in the future will match our database images this is in general a difficult task. Furthermore, A* requires that the heuristic is a predefined function and does not change during the search. In contrast to that, we take a different approach and try to predict the matching cost based on the images that we have received so far. This means, our heuristic is updated during the search, which prevents the application of A*. Our search procedure takes into account the estimated matching cost and works as follows. Similar to A*, we use an open-list F of nodes that are still under consideration. This open-list is realized through a priority queue. In contrast to A*, the key of our priority queue for a node x ij is the cost g(x ij ) of reaching x ij from the source x s. Our search and simultaneous graph construction starts with creating the source node x s and connecting it to the matching nodes according to Eq. (). This step requires to instantiate D nodes if no further information about the first possible match is provided. For a new incoming image referred to as l, we use the following procedure to update the graph as well as the matching sequence (see Figure 5 for a brief illustration): Whenever a new image is obtained, we pop a node from F. We use

5 VYSOTSKA AND STACHNISS: LAZY DATA ASSOCIATION FOR IMAGE SEQUENCES MATCHING UNDER SUBSTANTIAL APPEARANCE CHANGES 5 database query level level Fig. 7: Keeping connectivity though additional edges when using location priors. Green: nodes that are expended and added to the graph; gray: potential neighboring nodes according to the prior, but not encountered in the graph search and thus not instantiated. Fig. 6: Illustration for the graph expanding procedure. Orange nodes are nodes in the F. The red square indicates that the element x ij will be the next one in F. The dashed gray line represent nodes and edges not computed yet. our heuristic, which will be described in the remainder of this section, to estimate if the popped node x ij is worth expanding or if it is unlikely to be part of the matching sequences. If the node is unlikely to be part of the matching sequences, we continue with the next node in F. Otherwise, we expand the node x ij by computing the matching costs for its children ch(x ij ) and connecting the node x ij with ch(x ij ) using the edges define in E X, see Eq. (2). If a node in ch(x ij ) lies on the l th level of the graph, it represents the so far best match for the most recent image and the search terminates for this input image. Otherwise, we proceed expanding nodes from F. The above described method relies on a heuristic to estimate the sum of matching costs for reaching the l th level (the most recent query image). The key problem here is that defining an effective and admissible heuristic is hard due to the small amount of background information that can be exploited to predict future image similarities. Therefore, we take an alternative approach to come up with a heuristic that provides a good estimate of the cost but is not guaranteed to be admissible in the sense of A search. We take a statistical approach and approximate an expected lower bound for the average cost of the unexpanded and thus unknown nodes. We do so by using the average cost of the best path found so far as a prediction of the cost of individual matches along a new path. Furthermore, we exploit the fact that we know the number of images obtained so far and we know that the shortest path will have l + nodes (start node plus one matching node for each image). This allows us to formulate the expected cost f(x ij ) for a node x ij to a node on the l th level, i.e. x l, as the computed cost from x s to x ij expressed through g(x ij ) plus the estimate cost as: f(x ij ) = g(x ij ) + α(l i)µ cost (ˆx), (4) }{{} heuristic where α [, ] is a factor to trade off the quality of the solution and the number of nodes that needs to be expanded. For α =, we obtain a greedy search behavior and for α =, we may not expand enough nodes to find a good solution. The term (l i) is the number of images that have to be matched to end the sequence and µ cost (ˆx) is the average cost of the best path found so far, see also Figure 6. In sum, the data association graph constructed in the proposed way using OverFeat features allows us to design a useful, but not guaranteed to be admissible heuristic for the search for data associations. This in turn means that we are not guaranteed to find the optimal solution but enables a fast search for image matching across seasons that can be executed online. D. Exploiting Location Priors for Online Matching In case a rough location prior, for example from a noisy GPS, is available, we can further improve the matching procedure and can also better deal with loops in the database as well as query sequences, i.e. place revisits. The graph construction described in Section III-A can naturally be extended to account for location prior information. The overall procedure of constructing the graph stays the same but an additional location prior allows us to identify for every query image i the set of possible neighboring images N(i) as N(i) = {j j D dist(i, j) < d max }, (5) where dist(i, j) is the distance between the location at which the images with indices i and j have been taken according the location prior. All elements not in N(i) will be discarded based on the prior information and thus several matching hypotheses do not need to be computed. As reported in [3], incorporating such location prior information may lead to disconnected graph components. For an incremental search, it is however easy to connect different components by extending the definition of the children ch(x ij ) of a node by ch(x ij ) ch(x ij ) C... C n, (6) see Figure 7 for an intuitive definition of the components C,..., C n. Thus, we ensure that if the path stays in the same component, the procedure of building the graph is not changed. If, however, the path may jump to another component, we account for this possibility given the prior. IV. EXPERIMENTAL EVALUATION Our evaluation is designed to illustrate the performance of our approach and to support the three main claims made in this paper. These three claims are: (i) our approach has the ability to run in an incremental fashion so that only few nodes are expanded so that online localization is possible, (ii) our

6 6 IEEE ROBOTICS AND AUTOMATION LETTERS. PREPRINT VERSION. DECEMBER, 25 view point (suburbs) day/night (bike) seasons (Nordland) 8.8 Precision Node expanded, % Online Offline OpenSeqSLAM FLANN Recall.5 Fig. 8: Left: visualization of the graph structure for the dataset with dramatic seasonal changes (Nordland sequence from VPRiCE). The algorithm compares the images only for the nodes marked with green. Other nodes are computed for visualization only. Right: Plot of the dependency between the expansion rate α and the number of matching cost computations, expressed in percentage from total number of nodes. Fig. : Performance evaluation on the Alderley dataset. Euclidean nearest neighbor search to find the most similar image according to the OverFeat features using the FLANN library. This approach is called FLANN in the remainder of this work and is, as we will see later on, not well-suited to solve the across season matching problem. A. Matching Performance Fig. 9: Full matching matrix (left) and the nodes expanded by our algorithm (green nodes in the right image). The similarity matrix is computed for visualization only. The squares highlights an area in which most images are hard to distinguish, which leads to a larger node expansion..8.8 Precision Precision heuristic is well-suited to find a competitive solution in most real world situations and (iii) our algorithm is able to exploit additional location prior information and can in this case also handle loops in robot s trajectories. Throughout our evaluation, we rely on multiple publicly available datasets, see Figure. We use the summer-winter dataset from [22], [3], referred to as Freiburg and the Nordland dataset, which is a four season train ride through Norway. We also use the Alderley dataset [9] recorded during a sunny day and a rainy night. Finally, we used the datasets that have been selected for the VPRiCE Challenge 25. The challenge consists of 422 query and 3756 database images organized as a single sequence although it resembles multiple different datasets stitched together. Besides setting the performance of our online method in relation to full offline matching, we compare it to OpenSeqSLAM as well as to a baseline approach that uses approximate.6.4 Online Offline OpenSeqSLAM FLANN.2.5 Recall.6.4 Online Offline OpenSeqSLAM FLANN.2.5 Recall Fig. : Precision-recall plots for the datasets Nordland (left) and Freiburg (right). The first experiment is designed to illustrate the capabilities of our approach. Figure 8 depicts the full matching matrix from a subset of the VPRiCE dataset with strong seasonal changes. Our algorithm compares 29, 37 image pairs out of 5, 693, 35 possible matching that approaches such as [22] would expand. This yields a reduction of the computation cost of 99.5%, while obtaining a comparable matching performance. We obtain reductions of more than 95% for most datasets. In general, the larger the dataset the bigger the savings. Also the distinctiveness of the similarity score plays a role for our algorithm. As it can be seen in Figure 9, the block of the similarity matrix in the upper left corner shows no distinct matching pattern. As a result, our approach expands a comparably large number of nodes, indicated by the green elements in the right image. Note that our algorithm does not compute the full similarity matrix as it is shown here, we depict it for visualization purposes only. The second set of experiments is designed to show how the proposed heuristic influences the matching performance based on the Freiburg, Nordland, and Alderley datasets. We compare the matches of our online method with those of our previous offline approach [3] using the full matching matrix but replacing the previously used HOG features by OverFeat. The results in Figure and Figure illustrate that our heuristic leads to comparable matching results while the number of image comparisons that need to be performed drops dramatically with increasing expansion parameter α, see Figure 2. As all precision-recall plots illustrate, we outperform OpenSeqSLAM as well as the FLANN baseline. The performance of our algorithm has also been evaluated within the place recognition challenge VPRiCE 25 conducted at ICRA 25 and CVPR 25 workshops. The evaluation has been performed by the challenge organizers. Our online algorithm achieved 3rd place with precision.68 and recall.755 in the test settings. The approach that scored first [2] is an offline method and the second place [2] focuses on the design of new features for image comparisons and thus could even be combined with our method as they are rather orthogonal.

7 VYSOTSKA AND STACHNISS: LAZY DATA ASSOCIATION FOR IMAGE SEQUENCES MATCHING UNDER SUBSTANTIAL APPEARANCE CHANGES 7.8 Node expanded, % F score α =.5 α =.6 α =.8 seasons (Freiburg) seasons (Nordland) seasons (Freiburg) seasons (Nordland) Fig. 2: In overall selecting the bigger expansion parameter α leads to a decrease in node expansion, while preserving the accuracy of the solution. The middle plot also shows that constraining α close to may prevent finding the correct path. This leads to degradation in accuracy (bottom) and may lead to increase in node expansion, depending on the underlying data (middle). B. Node Expansion The third experiment is designed to evaluate the expansion of nodes in the data association graph in more detail. The evaluation illustrates that we can achieve online performance as only a comparably small number of nodes in the data association graph get expanded in every step. The two major factors that influence how the graph expands are the distinctiveness of matching costs and the expansion parameter α. We varied the expansion parameter of our heuristic in Eq. (4) between and. Zero basically leads to a greedy search, while α = approximates the expected cost by the average cost of the best path. Figure 2 (middle) shows the dependency between the graph size and the applied expansion rate for the Freiburg and Nordland dataset. Roughly speaking, the closer the expansion parameter α is to the smaller the resulting graph will be and vice versa. On the other hand, constraining α to the values close to, is likely to prevent the algorithm from finding the optimal path, see Figure 2 (top, right) for an example. In this figure, the matching nodes, which are computed using our approach, are colored in green. All other are depicted for illustration purposes only and do not need to be computed in practice. In these cases, the reductions in the number of expanded nodes are {65%, 95.7%, 95%} (from left to right). The figure also shows the F score illustrating that too large values for α can lead to a decay in matching performance. In all our experiments, we select α =.6 as a good trade off between matching performance and computational savings. Node expand Precision Node expanded, %.6.4 Online (m) Offline (m).2 Offline (no gps) OpenSeqSLAM FLANN Recall GPS prior 5m 8 GPS prior m GPS prior 5m No GPS Fig. 3: Exploiting location priors enables handling the loops in image sequences. Right: Example of the similarity matrix constrained with m GPS prior and overlaid graph search results. Top left: comparison using precision-recall plots. Bottom left: node reductions relative to the uncertainty of the location prior. C. Exploitation of Additional Location Priors The last experiment is designed to show that in presence of additional but potentially noisy location priors, our algorithm is able to handle loops in the query and database trajectories as well as deviations from the database, i.e., visiting unknown areas. Figure 3 (right) depicts a similarity matrix between a query and database sequence for the scenario in which the location of the robot is known up to m, for example obtained from a GPS receiver operating under suboptimal conditions. The green area corresponds to the nodes that are instantiated in the graph construction and search, while black areas correspond to the nodes excluded due to the location prior. Also in this settings, our algorithm is able to avoid instantiating unnecessary nodes while correctly finding the path. Exploiting such location priors furthermore allows us to handle loops, see for example Figure 3 (top). There is almost no decrease in performance in comparison to the offline method also using OverFeat. Additionally, we outperform SeqSLAM and FLANN. Figure 3 (bottom) shows that the gain in node reduction is smaller the better the pose is known from the prior, which is an expected result. D. OverFeat vs. HOG Features We also analyzed the performance of the matching approaches using HOG features, as used in [22], [3] and the pre-trained OverFeat features by Sermanet et al. [25]. We found that the OverFeat features outperform the HOG features for place recognition under strong appearance changes as they provide more distinctive matching scores, see also Figure 4. For the HOG features, the ratio between the best match and the worse match using the cosine distance is.46 compared to 4.28 for OverFeat. Thus, using HOGs would be less effective for the search method presented in this paper as a larger number of nodes of the data association graph would be expanded (see green area in the last row of Figure 4). In

8 8 IEEE ROBOTICS AND AUTOMATION LETTERS. PREPRINT VERSION. DECEMBER, 25 this sense, we confirm the results by Chen et al. [5] that the th layer is well-suited for place recognition tasks. E. Timing Our approach can run online with around fps on a standard notebook. Breaking down the timings of the individual components shows that computing the OverFeat descriptor takes the largest amount of time with approx. 5 ms. Expanding a single node, i.e., comparing two descriptors, takes 8 ms. The incremental update of the shortest path takes around 4 ms on average. V. CONCLUSION We proposed an incremental approach to visual image sequence matching under substantial appearance changes for online operation. The key idea is to apply a lazy data association approach and to define a heuristic for the search in the data association graph that estimates the similarity of images. This enables us to achieve online performance for image sequence matching under substantial appearance changes. We furthermore illustrated that noisy location priors can be exploited during online search. We implemented and tested our approach using real world image sequences acquired in summer and in winter as well as under different weather conditions. Our comparisons to other methods as well as the results from the VPRiCE 25 place recognition challenge suggest that our approach provides competitive results and avoids expanding large portions of the data association graph or building a large matching matrix. ACKNOWLEDGMENT The authors would like to gratefully thank Luciano Spinello for fruitful discussions and feedback. REFERENCES [] M. Agrawal and K. Konolige. Frameslam: From bundle adjustment to real-time visual mapping. IEEE Transactions on Robotics, 24(5), 28. [2] H. Bay, A. Ess, T. Tuytelaars, and L. Van Gool. Speeded-up robust features (SURF). Comput. Vis. Image Underst., (3): , 28. [3] M. Bennewitz, C. Stachniss, W. Burgard, and S. Behnke. Metric localization with scale-invariant visual features using a single perspective camera. In European Robotics Symposium, pages 43 57, 26. [4] P. Biber and T. Duckett. Dynamic maps for long-term operation of mobile service robots. In Proc. of Robotics: Science and Systems, pages 7 24, 25. [5] Z. Chen, O. Lam, A. Jacobson, and M.Milford. Convolutional neural network-based place recognition. Australasian Conference on Robotics and Automation 24. [6] W. Churchill and P. Newman. Experience-based Navigation for Longterm Localisation. Int. Journal of Robotics Research, 23. [7] M. Cummins and P. Newman. Highly scalable appearance-only SLAM - FAB-MAP 2.. In Proc. of Robotics: Science and Systems, 29. [8] A.J. Davison, I.D. Reid, N.D. Molton, and O. Stasse. Monoslam: Realtime single camera slam. IEEE Transactions Pattern Analysis and Machine Intelligence, 29, 27. [9] J. Fuentes-Pacheco, J. Ruiz-Ascencio, and J.M. Rendón-Mancha. Visual simultaneous localization and mapping: a survey. Artificial Intelligence Review, pages 27, 22. [] P.T. Furgale and T.D. Barfoot. Visual teach and repeat for long-range rover autonomy. Int. J. Field Robotics, 27:534 56, 2. [] A.J. Glover, W.P. Maddern, M. Milford, and G.F. Wyeth. FAB-MAP + RatSLAM: Appearance-based slam for multiple times of day. In Proc. of the IEEE Int. Conf. on Robotics and Automation (ICRA), pages , 2. [2] R. Gomez-Ojeda, M. Lopez-Antequera, N. Petkov, and J. Gonzalez- Jimenez. Training a convolutional neural network for appearanceinvariant place recognition. arxiv preprint arxiv: , 25. [3] D. Hähnel, W. Burgard, B. Wegbreit, and S. Thrun. Towards lazy data association in slam. In Proc. of the Int. Symposium of Robotics Research (ISRR), pages 42 43, Siena, Italy, 23. [4] E. Johns and G.-Z. Yang. Feature co-occurrence maps: Appearancebased localisation throughout the day. In Proc. of the IEEE Int. Conf. on Robotics and Automation (ICRA), 23. [5] C. Linegar, W. Churchill, and P. Newman. Work smart, not hard: Recalling relevant experiences for vast-scale but time-constrained localisation. In Proc. of the IEEE Int. Conf. on Robotics & Automation (ICRA), 25. [6] D.G. Lowe. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vision, 6(2):9, 24. [7] Y. Matsumoto, M. Inaba, and H. Inoue. Visual navigation using viewsequenced route representation. In Proc. of the IEEE Int. Conf. on Robotics & Automation (ICRA), 996. [8] M. Milford. Vision-based place recognition: how low can you go? Int. Journal of Robotics Research, 32(7): , 23. [9] M. Milford and G.F. Wyeth. Seqslam: Visual route-based navigation for sunny summer days and stormy winter nights. In Proc. of the IEEE Int. Conf. on Robotics and Automation (ICRA), 22. [2] D. Mishkin, M. Perdoch, and J. Matas. Place recognition with wxbs retrieval. In Proc. of the CVPR 25 Workshop on Visual Place Recognition in Changing Environments, 25. [2] T. Naseer, M. Ruhnke, C. Stachniss, L. Spinello, and W. Burgard. Robust visual slam across seasons. In Proc. of the IEEE/RSJ Int. Conf. on Intelligent Robots and Systems (IROS), 25. [22] T. Naseer, L. Spinello, W. Burgard, and C. Stachniss. Robust visual robot localization across seasons using network flows. In Proc. of the AAAI Conference on Artificial Intelligence, 24. [23] P. Neubert, N. Sunderhauf, and P. Protzel. Appearance change prediction for long-term navigation across seasons. In Proc. of the European Conference on Mobile Robotics (ECMR), 23. [24] G. Pascoe, W. Maddern, A.D. Stewart, and P Newman. Farlap: Fast robust localisation using appearance priors. In Proc. of the IEEE Int. Conf. on Robotics & Automation (ICRA), 25. [25] P. Sermanet, D. Eigen, Z. Zhang, M. Mathieu, R. Fergus, and Y. LeCun. Overfeat: Integrated recognition, localization and detection using convolutional networks. In Intl. Conf. on Learning Representations (ICLR), 24. [26] C. Stachniss and W. Burgard. Mobile robot mapping and localization in non-static environments. In Proc. of the AAAI Conference on Artificial Intelligence, pages , 25. [27] E. Stumm, C. Mei, S. Lacroix, and M. Chli. Location graphs for visual place recognition. In Proc. of the IEEE Int. Conf. on Robotics & Automation (ICRA), 25. [28] N. Sünderhauf, S. Shirazi, A. Jacobson, F. Dayoub, E. Pepperell, B. Upcroft, and M. Milford. Place recognition with convnet landmarks: Viewpoint-robust, condition-robust, training-free. In Proc. of Robotics: Science and Systems, 25. [29] C. Valgren and A.J. Lilienthal. SIFT, SURF & Seasons: Appearancebased long-term localization in outdoor environments. Robotics and Autonomous Systems, 85(2):49 56, 2. [3] O. Vysotska, T. Naseer, L. Spinello, W. Burgard, and C. Stachniss. Efficient and effective matching of image sequences under substantial appearance changes exploiting gps priors. In Proc. of the IEEE Int. Conf. on Robotics & Automation (ICRA), 25. [3] O. Vysotska and C. Stachniss. Lazy sequences matching under substantial appearance changes. In Workshop on Visual Place Recognition in Changing Environments at the IEEE Proceedings of the IEEE Int. Conf. on Robotics & Automation (ICRA), 25. Short paper.

Efficient and Effective Matching of Image Sequences Under Substantial Appearance Changes Exploiting GPS Priors

Efficient and Effective Matching of Image Sequences Under Substantial Appearance Changes Exploiting GPS Priors Efficient and Effective Matching of Image Sequences Under Substantial Appearance Changes Exploiting GPS Priors Olga Vysotska Tayyab Naseer Luciano Spinello Wolfram Burgard Cyrill Stachniss Abstract The

More information

Robust Visual Robot Localization Across Seasons using Network Flows

Robust Visual Robot Localization Across Seasons using Network Flows Robust Visual Robot Localization Across Seasons using Network Flows Tayyab Naseer University of Freiburg naseer@cs.uni-freiburg.de Luciano Spinello University of Freiburg spinello@cs.uni-freiburg.de Wolfram

More information

Robust Visual Robot Localization Across Seasons Using Network Flows

Robust Visual Robot Localization Across Seasons Using Network Flows Proceedings of the Twenty-Eighth AAAI Conference on Artificial Intelligence Robust Visual Robot Localization Across Seasons Using Network Flows Tayyab Naseer University of Freiburg naseer@cs.uni-freiburg.de

More information

Vision-Based Markov Localization Across Large Perceptual Changes

Vision-Based Markov Localization Across Large Perceptual Changes Vision-Based Markov Localization Across Large Perceptual Changes Tayyab Naseer Benjamin Suger Michael Ruhnke Wolfram Burgard Autonomous Intelligent Systems Group, University of Freiburg, Germany {naseer,

More information

Robust Visual SLAM Across Seasons

Robust Visual SLAM Across Seasons Robust Visual SLAM Across Seasons Tayyab Naseer Michael Ruhnke Cyrill Stachniss Luciano Spinello Wolfram Burgard Abstract In this paper, we present an appearance-based visual SLAM approach that focuses

More information

Ensemble of Bayesian Filters for Loop Closure Detection

Ensemble of Bayesian Filters for Loop Closure Detection Ensemble of Bayesian Filters for Loop Closure Detection Mohammad Omar Salameh, Azizi Abdullah, Shahnorbanun Sahran Pattern Recognition Research Group Center for Artificial Intelligence Faculty of Information

More information

Effective Visual Place Recognition Using Multi-Sequence Maps

Effective Visual Place Recognition Using Multi-Sequence Maps IEEE ROBOTICS AND AUTOMATION LETTERS. PREPRINT VERSION. ACCEPTED JANUARY, 2019 1 Effective Visual Place Recognition Using Multi-Sequence Maps Olga Vysotska and Cyrill Stachniss Abstract Visual place recognition

More information

Towards Life-Long Visual Localization using an Efficient Matching of Binary Sequences from Images

Towards Life-Long Visual Localization using an Efficient Matching of Binary Sequences from Images Towards Life-Long Visual Localization using an Efficient Matching of Binary Sequences from Images Roberto Arroyo 1, Pablo F. Alcantarilla 2, Luis M. Bergasa 1 and Eduardo Romera 1 Abstract Life-long visual

More information

Proc. 14th Int. Conf. on Intelligent Autonomous Systems (IAS-14), 2016

Proc. 14th Int. Conf. on Intelligent Autonomous Systems (IAS-14), 2016 Proc. 14th Int. Conf. on Intelligent Autonomous Systems (IAS-14), 2016 Outdoor Robot Navigation Based on View-based Global Localization and Local Navigation Yohei Inoue, Jun Miura, and Shuji Oishi Department

More information

Simultaneous Localization and Mapping

Simultaneous Localization and Mapping Sebastian Lembcke SLAM 1 / 29 MIN Faculty Department of Informatics Simultaneous Localization and Mapping Visual Loop-Closure Detection University of Hamburg Faculty of Mathematics, Informatics and Natural

More information

Visual Bearing-Only Simultaneous Localization and Mapping with Improved Feature Matching

Visual Bearing-Only Simultaneous Localization and Mapping with Improved Feature Matching Visual Bearing-Only Simultaneous Localization and Mapping with Improved Feature Matching Hauke Strasdat, Cyrill Stachniss, Maren Bennewitz, and Wolfram Burgard Computer Science Institute, University of

More information

Collecting outdoor datasets for benchmarking vision based robot localization

Collecting outdoor datasets for benchmarking vision based robot localization Collecting outdoor datasets for benchmarking vision based robot localization Emanuele Frontoni*, Andrea Ascani, Adriano Mancini, Primo Zingaretti Department of Ingegneria Infromatica, Gestionale e dell

More information

arxiv: v1 [cs.ro] 30 Oct 2018

arxiv: v1 [cs.ro] 30 Oct 2018 Pre-print of article that will appear in Proceedings of the Australasian Conference on Robotics and Automation 28. Please cite this paper as: Stephen Hausler, Adam Jacobson, and Michael Milford. Feature

More information

arxiv: v1 [cs.ro] 18 Apr 2017

arxiv: v1 [cs.ro] 18 Apr 2017 Multisensory Omni-directional Long-term Place Recognition: Benchmark Dataset and Analysis arxiv:1704.05215v1 [cs.ro] 18 Apr 2017 Ashwin Mathur, Fei Han, and Hao Zhang Abstract Recognizing a previously

More information

arxiv: v1 [cs.cv] 27 May 2015

arxiv: v1 [cs.cv] 27 May 2015 Training a Convolutional Neural Network for Appearance-Invariant Place Recognition Ruben Gomez-Ojeda 1 Manuel Lopez-Antequera 1,2 Nicolai Petkov 2 Javier Gonzalez-Jimenez 1 arxiv:.7428v1 [cs.cv] 27 May

More information

Improving Vision-based Topological Localization by Combining Local and Global Image Features

Improving Vision-based Topological Localization by Combining Local and Global Image Features Improving Vision-based Topological Localization by Combining Local and Global Image Features Shuai Yang and Han Wang Nanyang Technological University, Singapore Introduction Self-localization is crucial

More information

Visual Place Recognition using HMM Sequence Matching

Visual Place Recognition using HMM Sequence Matching Visual Place Recognition using HMM Sequence Matching Peter Hansen and Brett Browning Abstract Visual place recognition and loop closure is critical for the global accuracy of visual Simultaneous Localization

More information

Single-View Place Recognition under Seasonal Changes

Single-View Place Recognition under Seasonal Changes Single-View Place Recognition under Seasonal Changes Daniel Olid, José M. Fácil and Javier Civera Abstract Single-view place recognition, that we can define as finding an image that corresponds to the

More information

LOCAL AND GLOBAL DESCRIPTORS FOR PLACE RECOGNITION IN ROBOTICS

LOCAL AND GLOBAL DESCRIPTORS FOR PLACE RECOGNITION IN ROBOTICS 8th International DAAAM Baltic Conference "INDUSTRIAL ENGINEERING - 19-21 April 2012, Tallinn, Estonia LOCAL AND GLOBAL DESCRIPTORS FOR PLACE RECOGNITION IN ROBOTICS Shvarts, D. & Tamre, M. Abstract: The

More information

Appearance-Based Place Recognition Using Whole-Image BRISK for Collaborative MultiRobot Localization

Appearance-Based Place Recognition Using Whole-Image BRISK for Collaborative MultiRobot Localization Appearance-Based Place Recognition Using Whole-Image BRISK for Collaborative MultiRobot Localization Jung H. Oh, Gyuho Eoh, and Beom H. Lee Electrical and Computer Engineering, Seoul National University,

More information

Semantics-aware Visual Localization under Challenging Perceptual Conditions

Semantics-aware Visual Localization under Challenging Perceptual Conditions Semantics-aware Visual Localization under Challenging Perceptual Conditions Tayyab Naseer Gabriel L. Oliveira Thomas Brox Wolfram Burgard Abstract Visual place recognition under difficult perceptual conditions

More information

Advanced Techniques for Mobile Robotics Bag-of-Words Models & Appearance-Based Mapping

Advanced Techniques for Mobile Robotics Bag-of-Words Models & Appearance-Based Mapping Advanced Techniques for Mobile Robotics Bag-of-Words Models & Appearance-Based Mapping Wolfram Burgard, Cyrill Stachniss, Kai Arras, Maren Bennewitz Motivation: Analogy to Documents O f a l l t h e s e

More information

Made to Measure: Bespoke Landmarks for 24-Hour, All-Weather Localisation with a Camera

Made to Measure: Bespoke Landmarks for 24-Hour, All-Weather Localisation with a Camera Made to Measure: Bespoke Landmarks for 24-Hour, All-Weather Localisation with a Camera Chris Linegar, Winston Churchill and Paul Newman Abstract This paper is about camera-only localisation in challenging

More information

Fast-SeqSLAM: Place Recognition and Multi-robot Map Merging. Sayem Mohammad Siam

Fast-SeqSLAM: Place Recognition and Multi-robot Map Merging. Sayem Mohammad Siam Fast-SeqSLAM: Place Recognition and Multi-robot Map Merging by Sayem Mohammad Siam A thesis submitted in partial fulfillment of the requirements for the degree of Master of Science Department of Computing

More information

Monocular Camera Localization in 3D LiDAR Maps

Monocular Camera Localization in 3D LiDAR Maps Monocular Camera Localization in 3D LiDAR Maps Tim Caselitz Bastian Steder Michael Ruhnke Wolfram Burgard Abstract Localizing a camera in a given map is essential for vision-based navigation. In contrast

More information

Robotics. Lecture 7: Simultaneous Localisation and Mapping (SLAM)

Robotics. Lecture 7: Simultaneous Localisation and Mapping (SLAM) Robotics Lecture 7: Simultaneous Localisation and Mapping (SLAM) See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information. Andrew Davison Department of Computing Imperial College

More information

Analysis of the CMU Localization Algorithm Under Varied Conditions

Analysis of the CMU Localization Algorithm Under Varied Conditions Analysis of the CMU Localization Algorithm Under Varied Conditions Aayush Bansal, Hernán Badino, Daniel Huber CMU-RI-TR-00-00 July 2014 Robotics Institute Carnegie Mellon University Pittsburgh, Pennsylvania

More information

A Hybrid Feature Extractor using Fast Hessian Detector and SIFT

A Hybrid Feature Extractor using Fast Hessian Detector and SIFT Technologies 2015, 3, 103-110; doi:10.3390/technologies3020103 OPEN ACCESS technologies ISSN 2227-7080 www.mdpi.com/journal/technologies Article A Hybrid Feature Extractor using Fast Hessian Detector and

More information

Work Smart, Not Hard: Recalling Relevant Experiences for Vast-Scale but Time-Constrained Localisation

Work Smart, Not Hard: Recalling Relevant Experiences for Vast-Scale but Time-Constrained Localisation Work Smart, Not Hard: Recalling Relevant Experiences for Vast-Scale but Time-Constrained Localisation Chris Linegar, Winston Churchill and Paul Newman Abstract This paper is about life-long vast-scale

More information

Proceedings of Australasian Conference on Robotics and Automation, 7-9 Dec 2011, Monash University, Melbourne Australia.

Proceedings of Australasian Conference on Robotics and Automation, 7-9 Dec 2011, Monash University, Melbourne Australia. Towards Condition-Invariant Sequence-Based Route Recognition Michael Milford School of Engineering Systems Queensland University of Technology michael.milford@qut.edu.au Abstract In this paper we present

More information

Dynamic Environments Localization via Dimensions Reduction of Deep Learning Features

Dynamic Environments Localization via Dimensions Reduction of Deep Learning Features Dynamic Environments Localization via Dimensions Reduction of Deep Learning Features Hui Zhang 1(B), Xiangwei Wang 1, Xiaoguo Du 1, Ming Liu 2, and Qijun Chen 1 1 RAI-LAB, Tongji University, Shanghai,

More information

K-Means Based Matching Algorithm for Multi-Resolution Feature Descriptors

K-Means Based Matching Algorithm for Multi-Resolution Feature Descriptors K-Means Based Matching Algorithm for Multi-Resolution Feature Descriptors Shao-Tzu Huang, Chen-Chien Hsu, Wei-Yen Wang International Science Index, Electrical and Computer Engineering waset.org/publication/0007607

More information

Visual Place Recognition in Changing Environments with Time-Invariant Image Patch Descriptors

Visual Place Recognition in Changing Environments with Time-Invariant Image Patch Descriptors Visual Place Recognition in Changing Environments with Time-Invariant Image Patch Descriptors Boris Ivanovic Stanford University borisi@cs.stanford.edu Abstract Feature descriptors for images are a mature

More information

Mobile Robot Mapping and Localization in Non-Static Environments

Mobile Robot Mapping and Localization in Non-Static Environments Mobile Robot Mapping and Localization in Non-Static Environments Cyrill Stachniss Wolfram Burgard University of Freiburg, Department of Computer Science, D-790 Freiburg, Germany {stachnis burgard @informatik.uni-freiburg.de}

More information

Robot localization method based on visual features and their geometric relationship

Robot localization method based on visual features and their geometric relationship , pp.46-50 http://dx.doi.org/10.14257/astl.2015.85.11 Robot localization method based on visual features and their geometric relationship Sangyun Lee 1, Changkyung Eem 2, and Hyunki Hong 3 1 Department

More information

Robust Place Recognition for 3D Range Data based on Point Features

Robust Place Recognition for 3D Range Data based on Point Features 21 IEEE International Conference on Robotics and Automation Anchorage Convention District May 3-8, 21, Anchorage, Alaska, USA Robust Place Recognition for 3D Range Data based on Point Features Bastian

More information

Perception IV: Place Recognition, Line Extraction

Perception IV: Place Recognition, Line Extraction Perception IV: Place Recognition, Line Extraction Davide Scaramuzza University of Zurich Margarita Chli, Paul Furgale, Marco Hutter, Roland Siegwart 1 Outline of Today s lecture Place recognition using

More information

Improving Vision-based Topological Localization by Combing Local and Global Image Features

Improving Vision-based Topological Localization by Combing Local and Global Image Features Improving Vision-based Topological Localization by Combing Local and Global Image Features Shuai Yang and Han Wang School of Electrical and Electronic Engineering Nanyang Technological University Singapore

More information

Lecture 12 Recognition

Lecture 12 Recognition Institute of Informatics Institute of Neuroinformatics Lecture 12 Recognition Davide Scaramuzza 1 Lab exercise today replaced by Deep Learning Tutorial Room ETH HG E 1.1 from 13:15 to 15:00 Optional lab

More information

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014 SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014 SIFT SIFT: Scale Invariant Feature Transform; transform image

More information

Proceedings of Australasian Conference on Robotics and Automation, 2-4 Dec 2013, University of New South Wales, Sydney Australia

Proceedings of Australasian Conference on Robotics and Automation, 2-4 Dec 2013, University of New South Wales, Sydney Australia Training-Free Probability Models for Whole-Image Based Place Recognition Stephanie Lowry, Gordon Wyeth and Michael Milford School of Electrical Engineering and Computer Science Queensland University of

More information

Robotics. Lecture 8: Simultaneous Localisation and Mapping (SLAM)

Robotics. Lecture 8: Simultaneous Localisation and Mapping (SLAM) Robotics Lecture 8: Simultaneous Localisation and Mapping (SLAM) See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information. Andrew Davison Department of Computing Imperial College

More information

High-Level Visual Features for Underwater Place Recognition

High-Level Visual Features for Underwater Place Recognition High-Level Visual Features for Underwater Place Recognition Jie Li, Ryan M. Eustice, and Matthew Johnson-Roberson Abstract This paper reports on a method to perform robust visual relocalization between

More information

Experiments in Place Recognition using Gist Panoramas

Experiments in Place Recognition using Gist Panoramas Experiments in Place Recognition using Gist Panoramas A. C. Murillo DIIS-I3A, University of Zaragoza, Spain. acm@unizar.es J. Kosecka Dept. of Computer Science, GMU, Fairfax, USA. kosecka@cs.gmu.edu Abstract

More information

Message Propagation based Place Recognition with Novelty Detection

Message Propagation based Place Recognition with Novelty Detection Message Propagation based Place Recognition with Novelty Detection Lionel Ott and Fabio Ramos 1 Introduction In long-term deployments the robot needs to cope with new and unexpected observations that cannot

More information

A Comparison of SIFT, PCA-SIFT and SURF

A Comparison of SIFT, PCA-SIFT and SURF A Comparison of SIFT, PCA-SIFT and SURF Luo Juan Computer Graphics Lab, Chonbuk National University, Jeonju 561-756, South Korea qiuhehappy@hotmail.com Oubong Gwun Computer Graphics Lab, Chonbuk National

More information

Local Region Detector + CNN based Landmarks for Practical Place Recognition in Changing Environments

Local Region Detector + CNN based Landmarks for Practical Place Recognition in Changing Environments Local Region Detector + CNN based Landmarks for Practical Place Recognition in Changing Environments Peer Neubert and Peter Protzel Technische Universita t Chemnitz 926 Chemnitz, Germany Contact: peer.neubert@etit.tu-chemnitz.de

More information

A Multi-Domain Feature Learning Method for Visual Place Recognition

A Multi-Domain Feature Learning Method for Visual Place Recognition A Multi-Domain Learning Method for Visual Place Recognition Peng Yin 1,, Lingyun Xu 1, Xueqian Li 3, Chen Yin 4, Yingli Li 1, Rangaprasad Arun Srivatsan 3, Lu Li 3, Jianmin Ji 2,, Yuqing He 1 Abstract

More information

Viewpoint Invariant Features from Single Images Using 3D Geometry

Viewpoint Invariant Features from Single Images Using 3D Geometry Viewpoint Invariant Features from Single Images Using 3D Geometry Yanpeng Cao and John McDonald Department of Computer Science National University of Ireland, Maynooth, Ireland {y.cao,johnmcd}@cs.nuim.ie

More information

Visual Recognition and Search April 18, 2008 Joo Hyun Kim

Visual Recognition and Search April 18, 2008 Joo Hyun Kim Visual Recognition and Search April 18, 2008 Joo Hyun Kim Introduction Suppose a stranger in downtown with a tour guide book?? Austin, TX 2 Introduction Look at guide What s this? Found Name of place Where

More information

Appearance-Based Landmark Selection for Efficient Long-Term Visual Localization

Appearance-Based Landmark Selection for Efficient Long-Term Visual Localization Appearance-Based Landmar Selection for Efficient Long-Term Visual Localization Mathias Büri, Igor Gilitschensi, Elena Stumm, Roland Siegwart, and Juan Nieto Autonomous Systems Lab, ETH Zürich {firstname.lastname}@mavt.ethz.ch

More information

Omnidirectional vision for robot localization in urban environments

Omnidirectional vision for robot localization in urban environments Omnidirectional vision for robot localization in urban environments Emanuele Frontoni, Andrea Ascani, Adriano Mancini, Primo Zingaretti Università Politecnica delle Marche Dipartimento di Ingegneria Informatica

More information

Mobile Human Detection Systems based on Sliding Windows Approach-A Review

Mobile Human Detection Systems based on Sliding Windows Approach-A Review Mobile Human Detection Systems based on Sliding Windows Approach-A Review Seminar: Mobile Human detection systems Njieutcheu Tassi cedrique Rovile Department of Computer Engineering University of Heidelberg

More information

Lecture 12 Recognition. Davide Scaramuzza

Lecture 12 Recognition. Davide Scaramuzza Lecture 12 Recognition Davide Scaramuzza Oral exam dates UZH January 19-20 ETH 30.01 to 9.02 2017 (schedule handled by ETH) Exam location Davide Scaramuzza s office: Andreasstrasse 15, 2.10, 8050 Zurich

More information

ICRA 2016 Tutorial on SLAM. Graph-Based SLAM and Sparsity. Cyrill Stachniss

ICRA 2016 Tutorial on SLAM. Graph-Based SLAM and Sparsity. Cyrill Stachniss ICRA 2016 Tutorial on SLAM Graph-Based SLAM and Sparsity Cyrill Stachniss 1 Graph-Based SLAM?? 2 Graph-Based SLAM?? SLAM = simultaneous localization and mapping 3 Graph-Based SLAM?? SLAM = simultaneous

More information

A Novel Extreme Point Selection Algorithm in SIFT

A Novel Extreme Point Selection Algorithm in SIFT A Novel Extreme Point Selection Algorithm in SIFT Ding Zuchun School of Electronic and Communication, South China University of Technolog Guangzhou, China zucding@gmail.com Abstract. This paper proposes

More information

Pictures at an Exhibition

Pictures at an Exhibition Pictures at an Exhibition Han-I Su Department of Electrical Engineering Stanford University, CA, 94305 Abstract We employ an image identification algorithm for interactive museum guide with pictures taken

More information

Robust Visual Localization in Changing Lighting Conditions

Robust Visual Localization in Changing Lighting Conditions 2017 IEEE International Conference on Robotics and Automation (ICRA) Singapore, May 29 - June 3, 2017 Robust Visual Localization in Changing Lighting Conditions Pyojin Kim 1, Brian Coltin 2, Oleg Alexandrov

More information

Large-Scale Robotic SLAM through Visual Mapping

Large-Scale Robotic SLAM through Visual Mapping Large-Scale Robotic SLAM through Visual Mapping Christof Hoppe, Kathrin Pirker, Matthias Ru ther and Horst Bischof Institute for Computer Graphics and Vision Graz University of Technology, Austria {hoppe,

More information

3D object recognition used by team robotto

3D object recognition used by team robotto 3D object recognition used by team robotto Workshop Juliane Hoebel February 1, 2016 Faculty of Computer Science, Otto-von-Guericke University Magdeburg Content 1. Introduction 2. Depth sensor 3. 3D object

More information

Fusion and Binarization of CNN Features for Robust Topological Localization across Seasons

Fusion and Binarization of CNN Features for Robust Topological Localization across Seasons Fusion and Binarization of CNN Features for Robust Topological Localization across Seasons Roberto Arroyo 1, Pablo F. Alcantarilla 2, Luis M. Bergasa 1 and Eduardo Romera 1 Abstract The extreme variability

More information

Department of Electrical and Electronic Engineering, University of Peradeniya, KY 20400, Sri Lanka

Department of Electrical and Electronic Engineering, University of Peradeniya, KY 20400, Sri Lanka WIT: Window Intensity Test Detector and Descriptor T.W.U.Madhushani, D.H.S.Maithripala, J.V.Wijayakulasooriya Postgraduate and Research Unit, Sri Lanka Technological Campus, CO 10500, Sri Lanka. Department

More information

Beyond a Shadow of a Doubt: Place Recognition with Colour-Constant Images

Beyond a Shadow of a Doubt: Place Recognition with Colour-Constant Images Beyond a Shadow of a Doubt: Place Recognition with Colour-Constant Images Kirk MacTavish, Michael Paton, and Timothy D. Barfoot Abstract Colour-constant images have been shown to improve visual navigation

More information

Lightweight Unsupervised Deep Loop Closure

Lightweight Unsupervised Deep Loop Closure Robotics: Science and Systems 2018 Pittsburgh, PA, USA, June 26-30, 2018 Lightweight Unsupervised Deep Loop Closure Nate Merrill and Guoquan Huang Abstract Robust efficient loop closure detection is essential

More information

Real-time Object Detection CS 229 Course Project

Real-time Object Detection CS 229 Course Project Real-time Object Detection CS 229 Course Project Zibo Gong 1, Tianchang He 1, and Ziyi Yang 1 1 Department of Electrical Engineering, Stanford University December 17, 2016 Abstract Objection detection

More information

Memory Management for Real-Time Appearance-Based Loop Closure Detection

Memory Management for Real-Time Appearance-Based Loop Closure Detection Memory Management for Real-Time Appearance-Based Loop Closure Detection Mathieu Labbé and François Michaud Department of Electrical and Computer Engineering Université de Sherbrooke, Sherbrooke, QC, CA

More information

Fast and Effective Visual Place Recognition using Binary Codes and Disparity Information

Fast and Effective Visual Place Recognition using Binary Codes and Disparity Information Fast and Effective Visual Place Recognition using Binary Codes and Disparity Information Roberto Arroyo, Pablo F. Alcantarilla 2, Luis M. Bergasa, J. Javier Yebes and Sebastián Bronte Abstract We present

More information

Grid-Based Models for Dynamic Environments

Grid-Based Models for Dynamic Environments Grid-Based Models for Dynamic Environments Daniel Meyer-Delius Maximilian Beinhofer Wolfram Burgard Abstract The majority of existing approaches to mobile robot mapping assume that the world is, an assumption

More information

A Systems View of Large- Scale 3D Reconstruction

A Systems View of Large- Scale 3D Reconstruction Lecture 23: A Systems View of Large- Scale 3D Reconstruction Visual Computing Systems Goals and motivation Construct a detailed 3D model of the world from unstructured photographs (e.g., Flickr, Facebook)

More information

Visual Localization across Seasons Using Sequence Matching Based on Multi-Feature Combination

Visual Localization across Seasons Using Sequence Matching Based on Multi-Feature Combination sensors Article Visual Localization across Seasons Using Sequence Matching Based on Multi-Feature Combination Yongliang Qiao ID Le2i FRE2005, CNRS, Arts et Métiers, UBFC, Université de technologie de Belfort-Montbéliard,

More information

Improved Coding for Image Feature Location Information

Improved Coding for Image Feature Location Information Improved Coding for Image Feature Location Information Sam S. Tsai, David Chen, Gabriel Takacs, Vijay Chandrasekhar Mina Makar, Radek Grzeszczuk, and Bernd Girod Department of Electrical Engineering, Stanford

More information

Using Geometric Blur for Point Correspondence

Using Geometric Blur for Point Correspondence 1 Using Geometric Blur for Point Correspondence Nisarg Vyas Electrical and Computer Engineering Department, Carnegie Mellon University, Pittsburgh, PA Abstract In computer vision applications, point correspondence

More information

Are you ABLE to perform a life-long visual topological localization?

Are you ABLE to perform a life-long visual topological localization? Noname manuscript No. (will be inserted by the editor) Are you ABLE to perform a life-long visual topological localization? Roberto Arroyo 1 Pablo F. Alcantarilla 2 Luis M. Bergasa 1 Eduardo Romera 1 Received:

More information

Online Tracking Parameter Adaptation based on Evaluation

Online Tracking Parameter Adaptation based on Evaluation 2013 10th IEEE International Conference on Advanced Video and Signal Based Surveillance Online Tracking Parameter Adaptation based on Evaluation Duc Phu Chau Julien Badie François Brémond Monique Thonnat

More information

A High Dynamic Range Vision Approach to Outdoor Localization

A High Dynamic Range Vision Approach to Outdoor Localization A High Dynamic Range Vision Approach to Outdoor Localization Kiyoshi Irie, Tomoaki Yoshida, and Masahiro Tomono Abstract We propose a novel localization method for outdoor mobile robots using High Dynamic

More information

III. VERVIEW OF THE METHODS

III. VERVIEW OF THE METHODS An Analytical Study of SIFT and SURF in Image Registration Vivek Kumar Gupta, Kanchan Cecil Department of Electronics & Telecommunication, Jabalpur engineering college, Jabalpur, India comparing the distance

More information

Feature Detection. Raul Queiroz Feitosa. 3/30/2017 Feature Detection 1

Feature Detection. Raul Queiroz Feitosa. 3/30/2017 Feature Detection 1 Feature Detection Raul Queiroz Feitosa 3/30/2017 Feature Detection 1 Objetive This chapter discusses the correspondence problem and presents approaches to solve it. 3/30/2017 Feature Detection 2 Outline

More information

Place Recognition using Straight Lines for Vision-based SLAM

Place Recognition using Straight Lines for Vision-based SLAM Place Recognition using Straight Lines for Vision-based SLAM Jin Han Lee, Guoxuan Zhang, Jongwoo Lim, and Il Hong Suh Abstract Most visual simultaneous localization and mapping systems use point features

More information

Towards illumination invariance for visual localization

Towards illumination invariance for visual localization Towards illumination invariance for visual localization Ananth Ranganathan Honda Research Institute USA, Inc Mountain View, CA 94043 www.ananth.in Shohei Matsumoto Honda R&D Co. Ltd. Wako-shi, Saitama,

More information

Local features and image matching. Prof. Xin Yang HUST

Local features and image matching. Prof. Xin Yang HUST Local features and image matching Prof. Xin Yang HUST Last time RANSAC for robust geometric transformation estimation Translation, Affine, Homography Image warping Given a 2D transformation T and a source

More information

arxiv: v1 [cs.cv] 1 Jan 2019

arxiv: v1 [cs.cv] 1 Jan 2019 Mapping Areas using Computer Vision Algorithms and Drones Bashar Alhafni Saulo Fernando Guedes Lays Cavalcante Ribeiro Juhyun Park Jeongkyu Lee University of Bridgeport. Bridgeport, CT, 06606. United States

More information

Image Features: Local Descriptors. Sanja Fidler CSC420: Intro to Image Understanding 1/ 58

Image Features: Local Descriptors. Sanja Fidler CSC420: Intro to Image Understanding 1/ 58 Image Features: Local Descriptors Sanja Fidler CSC420: Intro to Image Understanding 1/ 58 [Source: K. Grauman] Sanja Fidler CSC420: Intro to Image Understanding 2/ 58 Local Features Detection: Identify

More information

This chapter explains two techniques which are frequently used throughout

This chapter explains two techniques which are frequently used throughout Chapter 2 Basic Techniques This chapter explains two techniques which are frequently used throughout this thesis. First, we will introduce the concept of particle filters. A particle filter is a recursive

More information

Implementation and Comparison of Feature Detection Methods in Image Mosaicing

Implementation and Comparison of Feature Detection Methods in Image Mosaicing IOSR Journal of Electronics and Communication Engineering (IOSR-JECE) e-issn: 2278-2834,p-ISSN: 2278-8735 PP 07-11 www.iosrjournals.org Implementation and Comparison of Feature Detection Methods in Image

More information

SIFT - scale-invariant feature transform Konrad Schindler

SIFT - scale-invariant feature transform Konrad Schindler SIFT - scale-invariant feature transform Konrad Schindler Institute of Geodesy and Photogrammetry Invariant interest points Goal match points between images with very different scale, orientation, projective

More information

Monocular SLAM for a Small-Size Humanoid Robot

Monocular SLAM for a Small-Size Humanoid Robot Tamkang Journal of Science and Engineering, Vol. 14, No. 2, pp. 123 129 (2011) 123 Monocular SLAM for a Small-Size Humanoid Robot Yin-Tien Wang*, Duen-Yan Hung and Sheng-Hsien Cheng Department of Mechanical

More information

Robot Localization based on Geo-referenced Images and G raphic Methods

Robot Localization based on Geo-referenced Images and G raphic Methods Robot Localization based on Geo-referenced Images and G raphic Methods Sid Ahmed Berrabah Mechanical Department, Royal Military School, Belgium, sidahmed.berrabah@rma.ac.be Janusz Bedkowski, Łukasz Lubasiński,

More information

Introduction to SLAM Part II. Paul Robertson

Introduction to SLAM Part II. Paul Robertson Introduction to SLAM Part II Paul Robertson Localization Review Tracking, Global Localization, Kidnapping Problem. Kalman Filter Quadratic Linear (unless EKF) SLAM Loop closing Scaling: Partition space

More information

Hand Posture Recognition Using Adaboost with SIFT for Human Robot Interaction

Hand Posture Recognition Using Adaboost with SIFT for Human Robot Interaction Hand Posture Recognition Using Adaboost with SIFT for Human Robot Interaction Chieh-Chih Wang and Ko-Chih Wang Department of Computer Science and Information Engineering Graduate Institute of Networking

More information

Salient Visual Features to Help Close the Loop in 6D SLAM

Salient Visual Features to Help Close the Loop in 6D SLAM Visual Features to Help Close the Loop in 6D SLAM Lars Kunze, Kai Lingemann, Andreas Nüchter, and Joachim Hertzberg University of Osnabrück, Institute of Computer Science Knowledge Based Systems Research

More information

A Framework for Multiple Radar and Multiple 2D/3D Camera Fusion

A Framework for Multiple Radar and Multiple 2D/3D Camera Fusion A Framework for Multiple Radar and Multiple 2D/3D Camera Fusion Marek Schikora 1 and Benedikt Romba 2 1 FGAN-FKIE, Germany 2 Bonn University, Germany schikora@fgan.de, romba@uni-bonn.de Abstract: In this

More information

Tensor Decomposition of Dense SIFT Descriptors in Object Recognition

Tensor Decomposition of Dense SIFT Descriptors in Object Recognition Tensor Decomposition of Dense SIFT Descriptors in Object Recognition Tan Vo 1 and Dat Tran 1 and Wanli Ma 1 1- Faculty of Education, Science, Technology and Mathematics University of Canberra, Australia

More information

A NEW FEATURE BASED IMAGE REGISTRATION ALGORITHM INTRODUCTION

A NEW FEATURE BASED IMAGE REGISTRATION ALGORITHM INTRODUCTION A NEW FEATURE BASED IMAGE REGISTRATION ALGORITHM Karthik Krish Stuart Heinrich Wesley E. Snyder Halil Cakir Siamak Khorram North Carolina State University Raleigh, 27695 kkrish@ncsu.edu sbheinri@ncsu.edu

More information

Incremental vision-based topological SLAM

Incremental vision-based topological SLAM Incremental vision-based topological SLAM Adrien Angeli, Stéphane Doncieux, Jean-Arcady Meyer Université Pierre et Marie Curie - Paris 6 FRE 2507, ISIR, 4 place Jussieu, F-75005 Paris, France. firstname.lastname@isir.fr

More information

Toward Part-based Document Image Decoding

Toward Part-based Document Image Decoding 2012 10th IAPR International Workshop on Document Analysis Systems Toward Part-based Document Image Decoding Wang Song, Seiichi Uchida Kyushu University, Fukuoka, Japan wangsong@human.ait.kyushu-u.ac.jp,

More information

IBuILD: Incremental Bag of Binary Words for Appearance Based Loop Closure Detection

IBuILD: Incremental Bag of Binary Words for Appearance Based Loop Closure Detection IBuILD: Incremental Bag of Binary Words for Appearance Based Loop Closure Detection Sheraz Khan and Dirk Wollherr Chair of Automatic Control Engineering Technische Universität München (TUM), 80333 München,

More information

Long-term motion estimation from images

Long-term motion estimation from images Long-term motion estimation from images Dennis Strelow 1 and Sanjiv Singh 2 1 Google, Mountain View, CA, strelow@google.com 2 Carnegie Mellon University, Pittsburgh, PA, ssingh@cmu.edu Summary. Cameras

More information

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy BSB663 Image Processing Pinar Duygulu Slides are adapted from Selim Aksoy Image matching Image matching is a fundamental aspect of many problems in computer vision. Object or scene recognition Solving

More information

Analyzing the Quality of Matched 3D Point Clouds of Objects

Analyzing the Quality of Matched 3D Point Clouds of Objects Analyzing the Quality of Matched 3D Point Clouds of Objects Igor Bogoslavskyi Cyrill Stachniss Abstract 3D laser scanners are frequently used sensors for mobile robots or autonomous cars and they are often

More information

SIFT: Scale Invariant Feature Transform

SIFT: Scale Invariant Feature Transform 1 / 25 SIFT: Scale Invariant Feature Transform Ahmed Othman Systems Design Department University of Waterloo, Canada October, 23, 2012 2 / 25 1 SIFT Introduction Scale-space extrema detection Keypoint

More information