Fuzzy based Multiple Dictionary Bag of Words for Image Classification

Similar documents
Patch Descriptors. CSE 455 Linda Shapiro

CS6670: Computer Vision

Bundling Features for Large Scale Partial-Duplicate Web Image Search

Large Scale Image Retrieval

Patch Descriptors. EE/CSE 576 Linda Shapiro

Large-scale visual recognition The bag-of-words representation

Instance-level recognition part 2

IMAGE RETRIEVAL USING VLAD WITH MULTIPLE FEATURES

Bag of Words Models. CS4670 / 5670: Computer Vision Noah Snavely. Bag-of-words models 11/26/2013

Instance-level recognition II.

Lecture 12 Recognition

Specular 3D Object Tracking by View Generative Learning

CS229: Action Recognition in Tennis

Evaluation and comparison of interest points/regions

SEARCH BY MOBILE IMAGE BASED ON VISUAL AND SPATIAL CONSISTENCY. Xianglong Liu, Yihua Lou, Adams Wei Yu, Bo Lang

FACULTY OF ENGINEERING AND INFORMATION TECHNOLOGY DEPARTMENT OF COMPUTER SCIENCE. Project Plan

Beyond bags of features: Adding spatial information. Many slides adapted from Fei-Fei Li, Rob Fergus, and Antonio Torralba

Visual Object Recognition

Large scale object/scene recognition

Lecture 12 Recognition. Davide Scaramuzza

Binary SIFT: Towards Efficient Feature Matching Verification for Image Search

By Suren Manvelyan,

A Rapid Automatic Image Registration Method Based on Improved SIFT

Computer Vision. Exercise Session 10 Image Categorization

Previously. Part-based and local feature models for generic object recognition. Bag-of-words model 4/20/2011

Local Image Features

Ensemble of Bayesian Filters for Loop Closure Detection

Detecting Printed and Handwritten Partial Copies of Line Drawings Embedded in Complex Backgrounds

Three things everyone should know to improve object retrieval. Relja Arandjelović and Andrew Zisserman (CVPR 2012)

Scale Invariant Feature Transform

Lecture 14: Indexing with local features. Thursday, Nov 1 Prof. Kristen Grauman. Outline

Video Google faces. Josef Sivic, Mark Everingham, Andrew Zisserman. Visual Geometry Group University of Oxford

Comparisson of Fuzzy C Means, K-Means and K-Medoids for Clustering in the Bag Of Visual Words Algorithm

String distance for automatic image classification

Preliminary Local Feature Selection by Support Vector Machine for Bag of Features

Local invariant features

Part based models for recognition. Kristen Grauman

Video Google: A Text Retrieval Approach to Object Matching in Videos

Scale Invariant Feature Transform

3D model search and pose estimation from single images using VIP features

CPPP/UFMS at ImageCLEF 2014: Robot Vision Task

Local Features and Bag of Words Models

Lecture 10 Detectors and descriptors

SURF. Lecture6: SURF and HOG. Integral Image. Feature Evaluation with Integral Image

OBJECT CATEGORIZATION

Shape recognition with edge-based features

Tensor Decomposition of Dense SIFT Descriptors in Object Recognition

Learning a Fine Vocabulary

Descriptors for CV. Introduc)on:

VK Multimedia Information Systems

Image matching on a mobile device

Keypoint-based Recognition and Object Search

Part-based and local feature models for generic object recognition

Similar Fragment Retrieval of Animations by a Bag-of-features Approach

A NEW FEATURE BASED IMAGE REGISTRATION ALGORITHM INTRODUCTION

Visual Recognition and Search April 18, 2008 Joo Hyun Kim

Stereoscopic Images Generation By Monocular Camera

Improving feature based object recognition in service robotics by disparity map based segmentation.

Matching Local Invariant Features with Contextual Information: An Experimental Evaluation.

Hamming embedding and weak geometric consistency for large scale image search

Local Image Features

Selection of Scale-Invariant Parts for Object Class Recognition

Viewpoint Invariant Features from Single Images Using 3D Geometry

Image classification Computer Vision Spring 2018, Lecture 18

A Comparison of SIFT, PCA-SIFT and SURF

Feature Detection. Raul Queiroz Feitosa. 3/30/2017 Feature Detection 1

Lecture 12 Recognition

The most cited papers in Computer Vision

Image Processing. Image Features

Image Feature Evaluation for Contents-based Image Retrieval

Discovering Visual Hierarchy through Unsupervised Learning Haider Razvi

Partial Copy Detection for Line Drawings by Local Feature Matching

Today. Main questions 10/30/2008. Bag of words models. Last time: Local invariant features. Harris corner detector: rotation invariant detection

Sparse coding for image classification

Out-of-Plane Rotated Object Detection using Patch Feature based Classifier

Object Recognition. Computer Vision. Slides from Lana Lazebnik, Fei-Fei Li, Rob Fergus, Antonio Torralba, and Jean Ponce

Human Detection and Action Recognition. in Video Sequences

WISE: Large Scale Content Based Web Image Search. Michael Isard Joint with: Qifa Ke, Jian Sun, Zhong Wu Microsoft Research Silicon Valley

IMPROVING SPATIO-TEMPORAL FEATURE EXTRACTION TECHNIQUES AND THEIR APPLICATIONS IN ACTION CLASSIFICATION. Maral Mesmakhosroshahi, Joohee Kim

Supervised learning. y = f(x) function

Image Retrieval with a Visual Thesaurus

Distributed Kd-Trees for Retrieval from Very Large Image Collections

Image Features: Detection, Description, and Matching and their Applications

Feature Detection and Matching

Local features and image matching. Prof. Xin Yang HUST

Classifying Images with Visual/Textual Cues. By Steven Kappes and Yan Cao

Computer Vision for HCI. Topics of This Lecture

Light-Weight Spatial Distribution Embedding of Adjacent Features for Image Search

Bag-of-features. Cordelia Schmid

Object Classification Problem

TEXTURE CLASSIFICATION METHODS: A REVIEW

A Novel Extreme Point Selection Algorithm in SIFT

Determinant of homography-matrix-based multiple-object recognition

Motion Estimation and Optical Flow Tracking

Fast Image Matching Using Multi-level Texture Descriptor

Announcements. Recognition. Recognition. Recognition. Recognition. Homework 3 is due May 18, 11:59 PM Reading: Computer Vision I CSE 152 Lecture 14

TRECVid 2013 Experiments at Dublin City University

Visual words. Map high-dimensional descriptors to tokens/words by quantizing the feature space.

Patch-based Object Recognition. Basic Idea

Local features: detection and description. Local invariant features

Transcription:

Available online at www.sciencedirect.com Procedia Engineering 38 (2012 ) 2196 2206 International Conference on Modeling Optimisation and Computing Fuzzy based Multiple Dictionary Bag of Words for Image Classification Abstract K.S.Sujatha a P. Keerthana b S. Suga Priya b E.Kaavya b B.Vinod c a Assistant Professor, Department of Electronics and Communication Engineering PSG College of Technology, Coimbatore, India, 641004. E-mail: bsoorya@rediffmail.com b UG students,department of Electronics and Communication Engineering PSG College of Technology, Coimbatore, India, 641004. c Head, Department of Robotics and Automation, PSG College of Technology, Coimbatore, India, 641004. E-mail: bvinod@rediffmail.com Object recognition in a large scale collection of images has become an important application in machine vision. The recent advances in the object or image recognition for classification of objects shows that Bag-of-visual words approach is a better method for image classification problems. An object recognition method based on the Bag-of-Words (BoW) model is implemented were descriptors are quantized to form a visual word dictionary called codebook with the help of soft clustering algorithm. To increase the recognition rate and accuracy of detection, the concept of Multiple Dictionary Bag of Words model (MDBoW) is implemented in which the dictionaries built using soft clustering algorithm from different subsets of the features are combined. The performances of existing BoW model with fuzzy codebook and the proposed MDBoW are evaluated in terms of macro precision, micro precision, accuracy and F1 measure. The proposed algorithm gives an increased recognition rate and accuracy of detection. 2011 Published by Elsevier Ltd. Selection and/or peer-review under responsibility of [name organizer] Open access under CC BY-NC-ND license. Key words: MDBoW; BoW; fuzzy;codebook 1. Introduction Object categorization through Bag of Words model is one of the most popular representation methods for object categorization. Bag of Words (BoW) approach has shown acceptable performance because of its fast run time and low storage requirements [14, 15, 16, 17, 18]. The key idea is to quantize each extracted key point into one of visual word, and then represent each image by a histogram of the visual words. For this purpose, a clustering algorithm like K-means is generally used for generating the visual words. Appropriate datasets are required at all stages of object recognition research, including learning visual 1877-7058 2012 Published by Elsevier Ltd. doi:10.1016/j.proeng.2012.06.264 Open access under CC BY-NC-ND license.

K.S. Sujatha et al. / Procedia Engineering 38 ( 2012 ) 2196 2206 2197 models of object and scene categories, detecting and localizing instances of these models in images, and evaluating the performance of recognition algorithms. Image databases are an essential element of object recognition research. They are required for learning visual object models and for testing the performance of classification, detection, and localization algorithms. The process of object recognition using bag of words has the following stages: Firstly, it extracts local features from images by detectors or dense sampling and then calculates their descriptors. For local feature detection, classic detectors include Harris detector [1] and its extension [2], maximally stable extremal region detector [3], affine invariant salient region detector [4]. For local feature description, we usually use local descriptors such as Haar descriptor [5], scale-invariant feature transform (SIFT) descriptor [6], gradient location and orientation histogram (GLOH) descriptor [7], rotation-invariant feature transform (RIFT) descriptor [8], shape context [9], histogram of gradients (HOG) descriptor [10] and speeded up robust feature descriptor (SURF) [11]. In this paper Bag of Words model has been implemented for visual categorization of images using Harris corner detector for extracting features and Scale Invariant Feature descriptor (SIFT) for representing the extracted features. After obtaining local features called descriptors, a codebook is generated to represent them. The codebook is a group of codes usually obtained by clustering over all descriptors. Clustering is the process of assigning a set of objects into groups so that the objects of similar type will be in one cluster. Clustering can be classified as hard clustering and soft clustering. The performance of BoW depends on the dictionary generation method, dictionary size, histogram weighting, normalization, and distance function. In this paper the method of generation of the dictionary of visual words is being focused. A novel method, Multiple Dictionaries for BoW (MDBoW) [18] using soft clustering algorithm Fuzzy C-means, that uses more visual words is implemented. This method significantly increases the performance of the algorithm when compared to the baseline method for large scale collection of images. Unlike baseline method, more words are used from different independent dictionaries instead of adding more words to the same dictionary. The resulting distribution of descriptors is quantified by using vector quantization against the pre-specified codebook to convert it to a histogram of votes for codebook centers. K nearest neighbor algorithm (KNN) is used to classify images through the resulting global descriptor vector. 2. Base line method In baseline method of Bag of Words model implemented in this paper, features are extracted using Harris corner detector and SIFT descriptor is used for representing the extracted features. The extracted features of the image should be distinctive. Features should be easily detected under changes in pose and lighting. There should be many features per object. Image content is transformed into local feature coordinates that are invariant to translation, rotation, scale, and other imaging parameters. The advantages of SIFT features are locality, distinctiveness, efficiency and extensibility. After feature extraction, clustering of the features is done by FCM clustering. Fuzzy C Mean (FCM) [12] is a data clustering technique in which a data set is grouped into clusters depending on the membership value. Fuzzy C-means is suited to identify clusters of the same geometry or the same order that is the clusters should have homogeneous order. After clustering a codebook with predefined number of visual words will be obtained. In training phase, the input vectors from the feature fool are assigned to one or more classes and any decision rule divides input space into decision regions separated by decision boundaries and histogram is built up. In testing phase, for the test data point, the k closest points from training data is found and classification is done using KNN classifier. It works well for large number of data and the distance metric used is good. The distance function used is Euclidean distance. Fig.1 shows the schematic of baseline method.

2198 K.S. Sujatha et al. / Procedia Engineering 38 ( 2012 ) 2196 2206 INPUT IMAGE FEATURE EXTRACTION USING HARRIS DETECTOR AND SIFT DESCRIPTOR FEATURE POOL FUZZY CLUSTERING TRAINING CLASSIFICATION USING KNN EVALUATION Fig.1 Schematic for Base line method 2.1. Fuzzy C-means Algorithm Given x N }, choose the number of clusters 1 < c < N, the weighting -inducing matrix A. The fuzzy C-means clustering algorithm is based on the minimization of an objective function called C-means functional given by Equation (2.1). (2.1) (2.2) where v i is the cluster prototype or the cluster centre, Dik corresponds to the distance of the kth sample point from the ith cluster centre. The parameter ik shall be interpreted as, the value of the membership function of the ith fuzzy subset for the kth datum. The value of m varies from 1 to which is a real number which indicates the amount of fuzziness. 2.2. Steps for Fuzzy C-means Algorithm The following are the steps to be followed for implementation of the algorithm. Initialize the partition matrix randomly, such that.

K.S. Sujatha et al. / Procedia Engineering 38 ( 2012 ) 2196 2206 2199 1. Compute the cluster prototypes (means) (2.3) f i is the cluster center calculated using the membership function. 2. Compute the distances:, (2.4) where A = I for Euclidean Norm and is the distance matrix containing the square distances between data points and cluster centers. 3. Update the partition matrix: Until (2.5) tolerance and m is the fuzziness weighting exponent. Use of FCM algorithm requires determination of several parameters like c, m, the inner product norm and the matrix norm. In addition, the set of initial cluster centers must be defined. 3. Multiple Dictionary Bag of Words model Searching large scale collections of images has become an important application of machine vision. Multiple Dictionaries for BoW (MDBoW), that uses more visual words has significantly increased the performance for large scale classification of images. Multiple dictionaries can be implemented in two ways Unified and. In single dictionary generation which is the baseline method a single dictionary of visual words is generated from the pool of features, which is used to generate the histogram for the image. In multiple dictionary generation each dictionary D N is generated with a different subset of the image features. In dictionary implementation the image gets a histogram h N from every dictionary D N which is concatenated to form a single histogram h. Every feature gets N entries in the histogram h, one from every dictionary. In Unified dictionary implementation a single unified dictionary is built from the concatenation of visual words from the dictionaries 1,.,N and the image get a single histogram h. Every feature gets only one entry in the histogram h. In this approach, more words are taken from different independent dictionaries where as in base line method more words will be taken from same dictionary. Thus multiple dictionary method has less storage than baseline approach. In this paper dictionary implementation of Multiple Dictionaries for BoW (MDBoW) is implemented. Fig.2 shows the schematic of dictionary implementation. 3.1 Steps for dictionary generation 1. Generate N random possibly overlapping subsets of the image features {S n } 1 N 2. Compute a dictionary D n independently for each subset S n. Each dictionary has a set of K n visual words.

2200 K.S. Sujatha et al. / Procedia Engineering 38 ( 2012 ) 2196 2206 3. Compute the histogram. Every image feature gets its visual word from every dictionary D n. Accumulate these visual words as individual words into individual histograms h n for each dictionary. The final histogram is the concatenation of the individual histograms. IMAGE FEATURES DICTIONARY N HISTOGRAM OF HISTOGRAM OF DICTIONARY N CONCATENATION OF HISTOGRAMS Fig.2. Schematic for dictionary implementation. 3.2 Multiple Dictionary Bag of Words model with FCM Clustering In this paper, dictionary concept has been implemented with Fuzzy C-means algorithm. Fuzzy clustering is the process of assigning membership levels and then using these member ship levels data elements are assigned to one or more clusters. The advantage of soft clustering is that it is insensitive to noise. In many real situations, fuzzy clustering is more natural than hard clustering, as objects on the boundaries between several classes are not forced to fully belong to one of the classes, but rather are assigned membership degrees between 0 and 1 indicating their partial memberships. The schematic for dictionary generation using fuzzy clustering is shown in Fig.3. Features are extracted from the images using Harris corner detector and represented using SIFT descriptor. From the feature pool N subsets of features are taken randomly and N dictionaries are generated using Fuzzy C-means algorithm. For each of the dictionary generated histograms are generated for each image in the dataset and the final histogram is the concatenation of the individual histograms. This is done during the training phase of the algorithm. During the testing phase features are extracted from each image and histogram for the image is generated by the same process as stated above. The KNN classifier then finds the k closest index and gives the classification result. 4. Experimental Result Bag of words model for visual categorization of large scale images has been implemented using Harris corner detector for extracting features and 128 dimensional scale invariant feature descriptor (SIFT) for representing the extracted features. The features extracted are clustered using Fuzzy C-means algorithm and a code book is generated with each vector in it being a visual word which serves as the basis for indexing the images. Images are then represented as histogram counts of these visual words. K nearest neighbour algorithm (KNN) is used to classify images. The performance of Bag of Words depends on dictionary generation method, dictionary size, histogram weighting, normalization, and distance function. In the proposed method the performance of Multiple

K.S. Sujatha et al. / Procedia Engineering 38 ( 2012 ) 2196 2206 2201 Dictionary Bag of Words model using dictionary by varying the word per dictionary and also the number of dictionaries generated is analysed. Fuzzy C means soft clustering algorithm is used to generate dictionary. This paper work is based on the hypothesis that fuzziness in the codebook creation step as well as in the histogram creation process leads to more robust behaviour of the bag of visual words approach in terms of codebook size. The performance of the Multiple Dictionary Bag of Words model using dictionary is compared with base line method by varying the word per dictionary and also by varying the number of individual dictionary generated by taking features randomly. INPUT IMAGE FEATURE EXTRACTION USING HARRIS DETECTOR AND SIFT DESCRIPTOR FEATURE POOL SUBSET 1....... SUBSET N... DICTIONARY N HISTOGRAM OF HISTOGRAM OF CONCATENATION OF HISTOGRAMS CLASSIFICATION USING KNN EVALUATION Fig.3. Schematic for dictionary implementation with FCM Clustering. The parameters used for the evaluation of the different algorithms are: I. Macro precision (4.1) II. Micro precision (4.2)

2202 K.S. Sujatha et al. / Procedia Engineering 38 ( 2012 ) 2196 2206 III. Accuracy (4.3) IV. Macro F1 where (4.4) (4.5) V. Micro F1 In these equations TP indicates true positive, FP false positive, FN false negative and TN true negative of the classification result. Precision and recall are the most common measures for evaluating an information retrieval system. The notable difference between these two calculations is that micro-averaging gives equal weight to every document that is it is called a document-pivoted measure while macro-averaging gives equal weight to every category that is it is categoryaccuracy. It considers both the precision p and recall r of the test to compute the score. For the Fuzzy C means the parameter m = 1.7 and stop condition = 0.001. The test data set includes eight different topics each containing 50 images. 200 images per concept were used during the training phase to build the codebooks. The classifier is trained for another 200 images from each topic. The number of dictionaries formed randomly is varied from 1 to 5 and the word per dictionary is varied from 80 to 200. The distance measure used is Euclidean distance. Since dataset is taken for real time application for visual recognition of objects for a humanoid used in restaurant, it is created from Google images. The images in the dataset used can be categorised as tiny images. The sample images from dataset are as shown in Fig. 4. (4.6) Fig. 4 Sample images from dataset Fig 5 to 9 shows the variation of accuracy rate with words per dictionary by varying the number of dictionary generated randomly from the feature pool from 1 to 5 which is named as dictionary1, dictionary2, dictionary3, dictionary4 and dictionary5. The results obtained are compared with the baseline method implemented in the paper. In both baseline method and Multiple Dictionary Bag of Words model the clustering of words are done using Fuzzy C means soft clustering algorithm. The algorithm was also implemented for Dataset taken from Caltech database which includes four different topics each topic

K.S. Sujatha et al. / Procedia Engineering 38 ( 2012 ) 2196 2206 2203 containing 200 images. It was found that the Multiple Dictionary Bag of Words model works for large scale image search where the number of topics and the number of images per topics are more. ACCURACY RATE 0.92 0.91 0.9 0.89 0.88 0.87 80 120 160 200 WORDS PER DICTIONARY baseline method MD separate ACCURACY RATE 0.92 0.91 0.9 0.89 0.88 0.87 DICTIONARY 2 80 120 160 200 WORDS PER DICTIONARY baseline method MD separate Fig. 5 Accuracy vs. words per dictionary for Dictionary1 ACCURACY RATE 0.92 0.91 0.9 0.89 0.88 0.87 DICTIONARY 3 80 120 160 200 WORDS PER DICTIONARY Fig. 6 Accuracy vs. words per dictionary for Dictionary2 baseline method MD separate Fig. 7 Accuracy vs. words per dictionary for Dictionary3 Table 1. Accuracy rate for word per dictionary 160 for various numbers of dictionaries No: of Dictionary 1 2 3 4 5 Accuracy Rate 0.9137 0.91 0.9144 0.9075 0.92

2204 K.S. Sujatha et al. / Procedia Engineering 38 ( 2012 ) 2196 2206 DICTIONARY 4 DICTIONARY 5 ACCURACY RATE 0.91 0.9 0.89 0.88 0.87 80 120 160 200 WORDS PER DICTIONARY baseline method MD separate ACCURACY RATE 0.92 0.91 0.9 0.89 0.88 0.87 80 120 160 200 WORDS PER DICTIONARY baseline method MD separate Fig. 8 Accuracy vs. words per dictionary for Dictionary4 Fig. 9 Accuracy vs. words per dictionary for Dictionary5 Table 2. Macro Precision for different words per dictionary for Base line method and Dictionary (MDBoW) No: Of Words Per Dictionary Base Line Method Dic 1 Dic 2 Dic3 Dic 4 Dic 5 80 0.5714 0.6122 0.6213 0.6103 0.6472 0.6051 120 0.6332 0.6958 0.6613 0.6542 0.6236 0.6467 160 0.6225 0.6381 0.6478 0.6702 0.6433 0.6842 200 0.5739 0.6666 0.6193 0.6127 0.6144 0.6082 Table 3. Micro Precision for different words per dictionary for Base line method and Dictionary (MDBoW) No: of Words Per Dictionary Base Line Method Dic 1 Dic 2 Dic3 Dic 4 Dic 5 80 0.555 0.5975 0.6075 0.565 0.615 0.5725 120 0.6 0.6475 0.635 0.6325 0.5965 0.61 160 0.6075 0.62 0.64 0.6566 0.63 0.68 200 0.555 0.645 0.605 0.595 0.6075 0.595

K.S. Sujatha et al. / Procedia Engineering 38 ( 2012 ) 2196 2206 2205 The results projected in Tables 2 to 5 shows that Multiple Dictionary Bag of Words model using dictionary shows better performance than baseline method. It can be seen from the results that on an average the method gives maximum accuracy rate for word per dictionary of 160 and the accuracy rate increases as the number of dictionary increases from 1 to 5.The tabulation of this result is given in Table1. The parameters Macro Precision, Micro Precision, Micro F1 and Macro F1 have better values for Multiple Dictionary Bag of Words than baseline method. For word per dictionary of 160 all these parameters increase as the number of dictionary increases. Table 4. Micro F1 for different words per dictionary for Base line method and Dictionary (MDBoW) No: of Words per Dictionary Base Line Method Dic 1 Dic 2 Dic3 Dic 4 Dic 5 80 0.5457 0.5908 0.6003 0.5596 0.6109 0.5711 120 0.5949 0.6488 0.6331 0.6255 0.5864 0.6023 160 0.6038 0.6197 0.6374 0.6531 0.6285 0.6767 200 0.5493 0.6378 0.6021 0.5919 0.6019 0.5881 Table 5. Macro F1 for different words per dictionary for Base line method and Dictionary (MDBoW) No: of Words per Dictionary Base Line Method Dic 1 Dic 2 Dic3 Dic 4 Dic 5 80 0.5631 0.6047 0.6143 0.5868 0.6307 0.5884 120 0.6161 0.6708 0.6479 0.6432 0.61 0.6278 160 0.6149 0.6289 0.6439 0.6635 0.6366 0.6821 200 0.5643 0.6558 0.6121 0.6037 0.6109 0.6015 5. Conclusion In this paper, the performance of fuzzy clustering Multiple Dictionary Bag of Words model using dictionary used for image classification is investigated by varying the words per dictionary and also the number of dictionaries generated and it is compared with the base line method. In this approach, more words are taken from different independent dictionaries where as in base line method more words will be taken from same dictionary. Thus multiple dictionary method has less storage than baseline approach. It is seen that the method works better when the number of topics and the number of images per topics are more. The results obtained indicate that Multiple Dictionary Bag of Words model using fuzzy clustering increases the recognition performance than the baseline method which uses fuzzy codebook in

2206 K.S. Sujatha et al. / Procedia Engineering 38 ( 2012 ) 2196 2206 Bag of Words method. The performance measures used for evaluation increases as the number of dictionary is increased for a particular value of word per dictionary. References [1] C. Harris and M. Stephens. A combined corner and edge detector. Proceedings of the Fourth Alvey Vision Conference, pages 147 151, 1988 [2] T. Tuytelaars and L. V. Gool. Matching widely separated views based on affine invariant regions. International Journal of Computer Vision, 59(1):61 85, 2004. [3] J. Matas, O. Chum, M. Urban, and T. Pajdla. Robust widebaseline stereo from maximally stable extremal regions. Image and Vision Computing, 22(10):761 767, 2004. [4] K. Mikolajczyk and C. Schmid. Scale and affine invariant interest point detectors. International Journal of Computer Vision, 60(1):63 86, 2004. [5] P. Viola and M. Jones. Robust real-time object detection.proc. of IEEE Workshop on Statistical and Computa-tional Theories of Vision, 2001. [6] D. G. Lowe. Distinctive image features from Scale-invariant key-points. International Journal of Computer Vision, 2(60):91 110, 2004. [7] K. Mikolajczyk and C. Schmid. A performance evaluation of local descriptors. IEEE Trans. on Pattern Analysis and Machine Intelligence, 27(10):1615 1630, 2005. [8] S. Lazebnik, C. Schmid, and J. Ponce. A sparse texture representation using local affine regions. Technical Report, Beckman Institute, University of Illinois, 2004. [9] S. Belongie, J. Malik, and J. Puzicha. Shape matching and object recognition using shape contexts. IEEE Trans. On Pattern Analysis and Machine Intelligence, 24(4):509 522 2002. [10] N. Dalal and B. Triggs. Histograms of oriented gradients for human detection. CVPR, 2005. [11] Bay,H,. Tuytelaars, T., &Van Gool, L. Vision. [12] R. L. Cannon, J. V. Dave, and J. C. Bezdek. Efficient implementation of the fuzzy c-means clustering algornthms. Pattern Analysis and Machine Intelligence, IEEE Transactions on, Volume: PAMI-8:248-255, 1986. [13] D. Nister and H. Stewenius. Scalable recognition with a vocabulary tree. CVPR, 2006. [14] M. Muja and D. Lowe. Fast approximate nearest neighbors with automatic algorithm configuration. In VISAPP, 2009. [15] O. Chum, J. Philbin, J. Sivic, M. Isard, and A. Zisserman. Total recall: Automatic query expansion with a generative feature model for object retrieval. In ICCV, 2007. [16] H. Jégou, M. Douze, and C. Schmid. Hamming embedding and weak geometric consistency for large scale image search. In ECCV, 2008. [17] Mohamed Aly, Peter Welinder, Mario Munich, and Pietro Perona. Towards Automated Large Scale Discovery of Image Families. In CVPR Workshop on Internet Vision, June 2009. [18] Mohamed Aly. Online Learning for Parameter Selection in Large Scale Image Search. In CVPR Workshop OLCV, June 2010.