Performance Evaluation of Feature Point Descriptors in the Infrared Domain

Similar documents
Cross-Spectral Image Registration And Fusion: An Evaluation Study

A Novel SIFT-Like-Based Approach for FIR-VS Images Registration

A Hybrid Feature Extractor using Fast Hessian Detector and SIFT

Local features and image matching. Prof. Xin Yang HUST

A Visible-Thermal Fusion based Monocular Visual Odometry

Appearance-Based Place Recognition Using Whole-Image BRISK for Collaborative MultiRobot Localization

An Evaluation of Local Feature Detectors and Descriptors for Infrared Images

sensors Cross-Spectral Local Descriptors via Quadruplet Network Article

A Fuzzy Brute Force Matching Method for Binary Image Features

Instance Detection by Keypoint Matching Beyond the Nearest Neighbor

IMPROVING DISTINCTIVENESS OF BRISK FEATURES USING DEPTH MAPS. Maxim Karpushin, Giuseppe Valenzise, Frédéric Dufaux

Infrared Image Colorization based on a Triplet DCGAN Architecture

Factorization with Missing and Noisy Data

Department of Electrical and Electronic Engineering, University of Peradeniya, KY 20400, Sri Lanka

Edge Point Linking by Means of Global and Local Schemes

Specular 3D Object Tracking by View Generative Learning

Online Learning of Binary Feature Indexing for Real-time SLAM Relocalization

Lecture 10 Detectors and descriptors

An Algorithm for Medical Image Registration using Local Feature Modal Mapping

Lossless Compression of Binary Image Descriptors for Visual Sensor Networks

Learning cross-spectral similarity measures with deep convolutional neural networks

Learning to Colorize Infrared Images

Image Matching Using SIFT, SURF, BRIEF and ORB: Performance Comparison for Distorted Images

A Method to Eliminate Wrongly Matched Points for Image Matching

A NEW ILLUMINATION INVARIANT FEATURE BASED ON FREAK DESCRIPTOR IN RGB COLOR SPACE

LEARNING WEIGHTED HAMMING DISTANCE FOR BINARY DESCRIPTORS

How Many Bits Do I Need For Matching Local Binary Descriptors?

A hardware design of optimized ORB algorithm with reduced hardware cost

The Brightness Clustering Transform and Locally Contrasting Keypoints

Ensemble of Bayesian Filters for Loop Closure Detection

Implementation and Comparison of Feature Detection Methods in Image Mosaicing

Using Geometric Blur for Point Correspondence

Collecting outdoor datasets for benchmarking vision based robot localization

Robust Binary Feature using the Intensity Order

State-of-the-Art: Transformation Invariant Descriptors. Asha S, Sreeraj M

Video Processing for Judicial Applications

Lecture 4.1 Feature descriptors. Trym Vegard Haavardsholm

IJSER. 1. Introduction

A NEW TOMOGRAPHIC BASED KEYPOINT DESCRIPTOR USING HEURISTIC GENETIC ALGORITHM

Fast Image Matching Using Multi-level Texture Descriptor

K-Means Based Matching Algorithm for Multi-Resolution Feature Descriptors

Fast and Effective Visual Place Recognition using Binary Codes and Disparity Information

The Gixel Array Descriptor (GAD) for Multi-Modal Image Matching

Wavelet-Based Visible and Infrared Image Fusion: A Comparative Study

A Study on Low-Cost Representations for Image Feature Extraction on Mobile Devices

A Novel Extreme Point Selection Algorithm in SIFT

LOCAL AND GLOBAL DESCRIPTORS FOR PLACE RECOGNITION IN ROBOTICS

International Journal of Advanced Research in Computer Science and Software Engineering

A System of Image Matching and 3D Reconstruction

Repositorio Institucional de la Universidad Autónoma de Madrid.

PERFORMANCE MEASUREMENTS OF FEATURE TRACKING AND HISTOGRAM BASED TRAFFIC CONGESTION ALGORITHMS

Local Pixel Class Pattern Based on Fuzzy Reasoning for Feature Description

Evaluation and comparison of interest points/regions

Colorizing Infrared Images through a Triplet Conditional DCGAN Architecture

Robot localization method based on visual features and their geometric relationship

An Edge-Based Approach to Motion Detection*

Yudistira Pictures; Universitas Brawijaya

A Keypoint Descriptor Inspired by Retinal Computation

A NEW FEATURE BASED IMAGE REGISTRATION ALGORITHM INTRODUCTION

IMAGE RETRIEVAL USING VLAD WITH MULTIPLE FEATURES

Fast Natural Feature Tracking for Mobile Augmented Reality Applications

Modern to Historic Image Matching: ORB/SURF an Effective Matching Technique

Analysis of Feature Detector and Descriptor Combinations with a Localization Experiment for Various Performance Metrics

ARTVision Tracker 2D

Global localization from a single feature correspondence

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

3D reconstruction how accurate can it be?

Invariant Feature Extraction using 3D Silhouette Modeling

Abstract. 1 Introduction. 2 Motivation. Information and Communication Engineering October 29th 2010

Daniel A. Lavigne Defence Research and Development Canada Valcartier. Mélanie Breton Aerex Avionics Inc. July 27, 2010

FREAK: Fast Retina Keypoint

Accelerating Pattern Matching or HowMuchCanYouSlide?

Chapter 3 Image Registration. Chapter 3 Image Registration

LATCH: Learned Arrangements of Three Patch Codes

Thermal-Visible Registration of Human Silhouettes: a Similarity Measure Performance Evaluation

Thermal Face Recognition using Local Interest Points and Descriptors for HRI Applications *

Motion Estimation and Optical Flow Tracking

A Comparison of SIFT, PCA-SIFT and SURF

Applying FAST & FREAK Algorithms in Selected Object Tracking

Feature Detection. Raul Queiroz Feitosa. 3/30/2017 Feature Detection 1

Quantitative Comparison of Feature Matchers Implemented in OpenCV3

Automatic Logo Detection and Removal

A Robust Feature Descriptor: Signed LBP

III. VERVIEW OF THE METHODS

The raycloud A Vision Beyond the Point Cloud

Occluded Facial Expression Tracking

Benchmarking Keypoint Filtering Approaches for Document Image Matching

Compress-then-analyze vs. analyze-then-compress: Two paradigms for image analysis in visual sensor networks

WATERMARKING FOR LIGHT FIELD RENDERING 1

Keypoint Recognition with Two-Stage Randomized Trees

A Novel Real-Time Feature Matching Scheme

An Evaluation of Volumetric Interest Points

Prof. Feng Liu. Spring /26/2017

From Structure-from-Motion Point Clouds to Fast Location Recognition

Flood-survivors detection using IR imagery on an autonomous drone

FIXED FOCAL LENGTH LENSES

arxiv: v1 [cs.cv] 28 Sep 2018

Depth Range Accuracy for Plenoptic Cameras

A General Framework For Image Feature Matching Without Geometric Constraints

AN INDOOR SLAM METHOD BASED ON KINECT AND MULTI-FEATURE EXTENDED INFORMATION FILTER

Transcription:

Performance Evaluation of Feature Point Descriptors in the Infrared Domain Pablo Ricaurte 1, Carmen Chilán 1, Cristhian A. Aguilera-Carrasco 2 Boris X. Vintimilla 1 and Angel D. Sappa 1,2 1 CIDIS-FIEC, Escuela Superior Politécnica del Litoral (ESPOL), Campus Gustavo Galindo, Km 3 vía Perimetral, P.O. Box 9-1-5863, Guayaquil, Ecuador 2 Computer Vision Center, Universitat Autònoma de Barcelona, 8193 Bellaterra, Barcelona, Spain {paricaur, cchilan, boris.vintimilla}@espol.edu.ec, {caguilera, asappa}@cvc.uab.es Keywords: Abstract: Infrared imaging, Feature point descriptors. This paper presents a comparative evaluation of classical feature point descriptors when they are used in the long-wave infrared spectral band. Robustness to changes in rotation, scaling, blur, and additive noise are evaluated using a state of the art framework. Statistical results using an outdoor image data set are presented together with a discussion about the differences with respect to the results obtained when images from the visible spectrum are considered. 1 INTRODUCTION In general, computer vision applications are based on the use of cameras that work in the visible spectrum. Recent advances in infrared imaging, as well as the reduction on the prices of these cameras, have opened new opportunities to develop novel solutions working in infrared spectral band or in the cross-spectral domain between infrared and visible images (e.g., (Barrera et al., 213), (Aguilera et al., 212), (Barrera et al., 212) and (Felicísimo and Cuartero, 26)). The spectral band of infrared imaging goes from.75µm to 15 µm, which is split up into the following categories: Near-Infrared (NIR:.751.4 µm), Short- Wave Infrared (SWIR: 1.43 µm), Mid-Wave Infrared (MWIR: 38 µm) or Long-Wave Infrared (LWIR: 815 µm). Images from each one of these categories have a particular advantage for a given application; for instance NIR images are generally used in gaze detection and eye tracking applications (Coyle et al., 24); SWIR spectral band has shown its usage in heavy fog environments (Hansen and Malchow, 28); MWIR is generally used to detect temperatures somehow above body temperature in military applications; finally, LWIR images have been used in video surveillance and driver assistance (Krotosky and Trivedi, 27). The current work is focussed on the LWIR domain, which corresponds to the farthest infrared spectral band from the visible spectrum. Following the evolution of visible spectrum based computer vision, in the infrared imaging domain topics such as image registration, pattern recognition or stereo vision, are being addressed. As a first attempt, classical tools from the visible spectrum are just used or little adapted to the new domain. One of these tools is the feature point description, which has been a very active research topic during the last decade in the computer vision community. Due to the large amount of contributions on this topic there were several works on the literature evaluating and comparing their performance in the visible spectrum case (e.g., (Miksik and Mikolajczyk, 212), (Mikolajczyk and Schmid, 25), (Bauer et al., 27) and (Schmid et al., 2)). Similarly to in the visible spectrum case, the current work proposes to study the performance of feature point descriptors when they are considered in the infrared domain. Since there is a large amount of algorithms in the literature, we decided to select the most representative and recent ones. Our study includes: (Lowe, 1999), (Bay et al., 26), (Rublee et al., 211), (Leutenegger et al., 211), (Calonder et al., 212) and (Alahi et al., 212). The study is motivated by the fact that although the appearance of LWIR im-

(a) (b) (c) (d) (e) Figure 1: A LWIR image from the evaluation dataset together with some illustrations of the applied transformations: (a) original image; (b) rotation; (c) scale; (d) blur; (e) noise. ages is similar to the ones from the visible spectrum their nature is different, hence we consider that conclusions from visible spectrum cannot be directly extended to the infrared domain. The manuscript is organized as follow. The evaluation methodology used for the comparison is presented in Section 2. Then, the data set and experimental results are detailed in Section 3. Finally, conclusions and discussions are given in Section 4. 2 EVALUATION FRAMEWORK This section summarizes the framework used to evaluate the performance of the different approaches. It aims at finding the best descriptor for feature point correspondence when common image transformations are considered: rotation in the image plane, changes in the image size, blur and presence of noise in the images. In order to take into account all these possible changes, the given images are modified, then different descriptors are applied. The used framework has been proposed by Khvedchenia 1 for evaluating 1 http://computer-vision-talks.com/211/8/featuredescriptor-comparison-report/ the performance of feature descriptors in the visible spectrum case. The algorithms are evaluated considering as a ground truth those points in the given image. A brute force strategy is used for finding the matchings, together with a L2 norm or Hamming distance, as detailed in Table 1. The percentage of correct matches between the ground truth image and the modified one is used as a criterion for the evaluation. The transformations applied to the given images are detailed below. Figure 1 shows an illustration of a given LWIR image together with some of the images resulting after applying the different transformations. Rotation: the study consists in evaluating the sensibility to rotations of the image. The rotations are in the image plane spanning the 36 degrees, a new image is obtained every 1 degrees. Scale: the size of the given image is changed and the repeatability of a given descriptor is evaluated. The original image is scaled in between to 2 times its size with a step of.1 per test. Pixels of scaled images are obtained through a linear interpolation. Blur: the robustness with respect to blur is evaluated. It consists of a Gaussian filter iteratively applied over the given image. At each iteration

Figure 2: Illustration of infrared images considered in the evaluation framework. the size of the kernel filter (K K) used to blur the image is update as follows: K = 2t + 1, where t = {1, 2,..., 9}. Noise: this final study consists in adding noise to the original image. This process is implemented by adding to the original image a personalized image. The value of the pixels of the personalized image are randomly obtained following a uniform distribution with µ = and σ = t, where t = {, 1, 2,..., 1}. In the original framework proposed by Khvedchenia, lighting changes were also considered, since that study was intended for images in the visible spectrum. In the current work, since it aims at studying the LWIR spectrum, changes in the image intensity values won t follow the same behavior all through the image (like lighting changes in the visible spectrum). Intensity values in LWIR images are related with the material of the objects in the scene. In summary, a study similar to the lighting changes is not considered in the current work. 3 EXPERIMENTAL RESULTS A set of 2 LWIR images has been considered with the evaluation framework presented above. For each one of the LWIR images the corresponding im- age from the visible spectrum is also provided. These visible spectrum images are used to compare the results obtained in the LWIR domain. Figure 2 presents some of the LWIR images contained in our dataset; it is publicly available for further research through our Web site2. For each algorithm and transformation the number of correct matches out of the total number of feature points described in the original image (the number of correspondences used as ground truth) is considered for the evaluation, similar to (Mikolajczyk and Schmid, 25): = #correct matches. #correspondences (1) The algorithms evaluated in the current work are presented in Table 1. In the cases of and the algorithm is used as a detector. In,, and the Hamming distance is used, instead of L2 norm, for speeding up the matching. For each transformation (Section 2) a set of images is obtained; for instance, in the rotation case 36 images are evaluated. Figure 3 depicts results obtained when the given image is rotated 36 degrees. It can be observed that the most robust algorithm is, which is the same conclusion obtained when images from the visible spectrum are considered. On the other hand, the 2 http://www.cvc.uab.es/adas/projects/simeve/

1.9.8.7 Rotation.6.1 5 1 15 2 25 3 35 4 Angle (in degrees) Figure 3: Rotation case study: average results from the evaluation data set. 1.9.8.7 Scale.6.1.6.8 1 1.2 1.4 1.6 1.8 2 Scale Factor Figure 4: Scale case study: average results from the evaluation data set.

Table 1: Algorithms evaluated in the study. Descriptor Alg. ( detector) ( detector) Matcher norm type L2 Norm L2 Norm algorithm (using as a detector) is the most sensible to rotations; actually, its performance drop to zero just after applying ±25 degrees to the given image. This behavior was also observed in the visible spectrum. Regarding the algorithms in between, a slightly better performance was appreciated in the LWIR case. In summary, the ranking from the best to the worst is as follow:,,,,, (the same ranking was observed in both spectrums). In the scale study, although similar results were obtained in both spectrums, the algorithms performance was better in the infrared case, in particular in the case of, which being the worst algorithm in both cases it has a better performance in the LWIR case. The ranking of algorithms performance is as follow (from the best to the worst):,,,, and. Figure 4 shows these results. Figure 5 presents the study of robustness of the different algorithms when the given images are degraded using a Gaussian filter of increasing size. Similarly to in the previous case all the algorithms have a better performance in the LWIR case than in the visible spectrum. In this case the algorithm is the most robust one, the other algorithms are sorted as follow:,,, and ; being the algorithm less robust to noise. Finally, Fig 6 shows the curves obtained when additive noise is considered. In this case, differently than in the previous studies, the algorithm has a better performance in the visible spectrum. Additionally, the performance of in the LWIR spectrum is not as bad as in the visible spectrum. The ranking of algorithms performance is as follow (from the best to the worst):,,,, and. 4 CONCLUSIONS This work presents an empirical evaluation of the performance of the state of the art descriptors when they are used in the LWIR domain. The main objective was to study whether conclusions obtained in the visible spectrum are also valid for the LWIR spectrum. Although results are similar to those obtained in the visible spectrum it can be appreciated that the performance of the algorithm (using as a detector) is better in LWIR spectrum when compared with its performance in the visible spectrum. The relative ranking in between algorithms keep the same in both spectral bands. ACKNOWLEDGEMENTS This work has been partially supported by the Spanish Government under Research Project TIN211-2566 and PROMETEO Project of the Secretaría Nacional de Educación Superior, Ciencia, Tecnología e Innovación de la República del Ecuador. Cristhian A. Aguilera-Carrasco was supported by a grant from Universitat Autònoma de Barcelona. The authors would like to thanks to Mr. Khvedchenia Ievgen for providing them with the evaluation framework. REFERENCES Aguilera, C., Barrera, F., Lumbreras, F., Sappa, A., and Toledo, R. (212). Multispectral image feature points. Sensors, 12(9):12661 12672. Alahi, A., Ortiz, R., and Vandergheynst, P. (212). : Fast retina keypoint. In IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, June 16-21, pages 51 517. Barrera, F., Lumbreras, F., and Sappa, A. (212). Multimodal stereo vision system: 3d data extraction and algorithm evaluation. IEEE Journal of Selected Topics in Signal Processing, 6(5):437 446. Barrera, F., Lumbreras, F., and Sappa, A. (213). Multispectral piecewise planar stereo using manhattanworld assumption. Pattern Recognition Letters, 34(1):52 61. Bauer, J., Snderhauf, N., and Protzel, P. (27). Comparing several implementations of two recently published feature detectors. In Proceedings of the International Conference on Intelligent and Autonomous Systems, Toulouse, France. Bay, H., Tuytelaars, T., and Gool, L. J. V. (26). : Speeded Up Robust Features. In Proceedings of the 9th European Conference on Computer Vision, Graz, Austria, May 7-13, pages 44 417. Calonder, M., Lepetit, V., Özuysal, M., Trzcinski, T., Strecha, C., and Fua, P. (212). : Computing a local binary descriptor very fast. IEEE Trans. Pattern Anal. Mach. Intell., 34(7):1281 1298. Coyle, S., Ward, T., Markham, C., and McDarby, G. (24). On the suitability of near-infrared (NIR) systems for

.8.7.6 Blur.1 3 5 7 9 11 13 15 17 19 Gaussian Blur Kernel Size Figure 5: Blur case study: average results from the evaluation data set. 1.9.8.7 Aditive Noise.6.1 1 2 3 4 5 6 7 8 9 1 Standard Deviation Figure 6: Noise case study: average results from the evaluation data set. next-generation braincomputer interfaces. Physiological Measurement, 25(4). Felicísimo, A. and Cuartero, A. (26). Methodological proposal for multispectral stereo matching. IEEE Trans. on Geoscience and Remote Sensing, 44(9):2534 2538. Hansen, M. P. and Malchow, D. S. (28). Overview of swir detectors, cameras, and applications. In Proceedings of the SPIE 6939, Thermosense, Orlando, FL, USA, March 16. Krotosky, S. and Trivedi, M. (27). On color-, infrared-, and multimodal-stereo approaches to pedestrian detection. IEEE Transactions on Intelligent Transportation Systems, 8:619 629. Leutenegger, S., Chli, M., and Siegwart, R. (211). : Binary Robust Invariant Scalable Keypoints. pages 2548 2555. Lowe, D. G. (1999). Object recognition from local scaleinvariant features. In Proceedings of the IEEE International Conference on Computer Vision, Kerkyra, Greece, September 2-27, pages 115 1157. Mikolajczyk, K. and Schmid, C. (25). A performance evaluation of local descriptors. IEEE Trans. Pattern Anal. Mach. Intell., 27(1):1615 163. Miksik, O. and Mikolajczyk, K. (212). Evaluation of local detectors and descriptors for fast feature matching. In Proceedings of the 21st International Conference on Pattern Recognition, ICPR 212, Tsukuba, Japan, November 11-15, pages 2681 2684. Rublee, E., Rabaud, V., Konolige, K., and Bradski, G. R. (211). : An efficient alternative to or. In IEEE International Conference on Computer Vision, Barcelona, Spain, November 6-13, pages 2564 2571. Schmid, C., Mohr, R., and Bauckhage, C. (2). Evaluation of interest point detectors. International Journal of Computer Vision, 37(2):151 172.