Relational Indexing. Mauro S. Costa 1 and Linda G. Shapiro 1,2 "

Similar documents
High-Level Computer Vision

3D Models and Matching

3D Models and Matching

Recognition. Clark F. Olson. Cornell University. work on separate feature sets can be performed in

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

Computer vision is the analysis of digital images by a computer for such applications as:

Human Activity Recognition Using Multidimensional Indexing

Beyond Bags of Features

A Graph Theoretic Approach to Image Database Retrieval

Power Functions and Their Use In Selecting Distance Functions for. Document Degradation Model Validation. 600 Mountain Avenue, Room 2C-322

Textural Features for Image Database Retrieval

Identifying Layout Classes for Mathematical Symbols Using Layout Context

On Resolving Ambiguities in Arbitrary-Shape extraction by the Hough Transform

Determination of 3-D Image Viewpoint Using Modified Nearest Feature Line Method in Its Eigenspace Domain

Object Classification Using Tripod Operators

Patch-based Object Recognition. Basic Idea

HOUGH TRANSFORM CS 6350 C V

View-Based 3-D Object Recognition using Shock Graphs Diego Macrini Department of Computer Science University of Toronto Sven Dickinson

Robust Shape Retrieval Using Maximum Likelihood Theory

10/03/11. Model Fitting. Computer Vision CS 143, Brown. James Hays. Slides from Silvio Savarese, Svetlana Lazebnik, and Derek Hoiem

CS 534: Computer Vision 3D Model-based recognition

TRANSPARENT OBJECT DETECTION USING REGIONS WITH CONVOLUTIONAL NEURAL NETWORK

Instance-level recognition

Structural Graph Matching With Polynomial Bounds On Memory and on Worst-Case Effort

FACE DETECTION AND LOCALIZATION USING DATASET OF TINY IMAGES

Model Based Perspective Inversion

CS 664 Image Matching and Robust Fitting. Daniel Huttenlocher

Fitting: The Hough transform

COSC160: Detection and Classification. Jeremy Bolton, PhD Assistant Teaching Professor

Scale Space Based Grammar for Hand Detection

Instance-level recognition

Toward Part-based Document Image Decoding

Geometric Hashing. Overview (1) Overview (2) Reference Frames. Recognition under Similarity Transformations. Feb Goo Jun

High Information Rate and Efficient Color Barcode Decoding

Segmentation of Range Data for the Automatic Construction of Models of Articulated Objects

Object recognition in 3D scenes with occlusions and clutter by Hough voting

Progressive Probabilistic Hough Transform

Sketchable Histograms of Oriented Gradients for Object Detection

Graph Matching: Fast Candidate Elimination Using Machine Learning Techniques

Object detection using non-redundant local Binary Patterns

Recognizing Objects by Matching Oriented Points

Instance-level recognition part 2

Indexing Fingerprints using Minutiae Quadruplets

Consistent Line Clusters for Building Recognition in CBIR

Swarm Based Fuzzy Clustering with Partition Validity

Three-Dimensional Face Recognition: A Fishersurface Approach

Visual localization using global visual features and vanishing points

Biometrics is the science of identifying individuals

GLOBAL SAMPLING OF IMAGE EDGES. Demetrios P. Gerogiannis, Christophoros Nikou, Aristidis Likas

COMPUTER AND ROBOT VISION

ADAPTIVE TEXTURE IMAGE RETRIEVAL IN TRANSFORM DOMAIN

New wavelet based ART network for texture classification

Sea Turtle Identification by Matching Their Scale Patterns

Trademark Matching and Retrieval in Sport Video Databases

Previously. Part-based and local feature models for generic object recognition. Bag-of-words model 4/20/2011

A Statistical Consistency Check for the Space Carving Algorithm.

Fitting: The Hough transform

A New Shape Matching Measure for Nonlinear Distorted Object Recognition

A Virtual Laboratory for Study of Algorithms

Characterizing Strategies of Fixing Full Scale Models in Construction Photogrammetric Surveying. Ryan Hough and Fei Dai

The Detection of Faces in Color Images: EE368 Project Report

Semi-Supervised PCA-based Face Recognition Using Self-Training

CS839: Epistemic Net Visualization Challenge

Texture Classification by Combining Local Binary Pattern Features and a Self-Organizing Map

A Minimum Number of Features with Full-Accuracy Iris Recognition

A Neural Network for Real-Time Signal Processing

3D Reconstruction of a Hopkins Landmark

The Novel Approach for 3D Face Recognition Using Simple Preprocessing Method

Instance-level recognition II.

Design and Evaluation of Processes for Fuel Fabrication: Quarterly Progress Report #7

OBJECT RECOGNITION: OBTAINING 2-D RECONSTRUCTIONS FROM COLOR EDGES. G. Bellaire**, K. Talmi*, E. Oezguer *, and A. Koschan*

Detecting Printed and Handwritten Partial Copies of Line Drawings Embedded in Complex Backgrounds

MORPHOLOGICAL BOUNDARY BASED SHAPE REPRESENTATION SCHEMES ON MOMENT INVARIANTS FOR CLASSIFICATION OF TEXTURES

Introduction. Introduction. Related Research. SIFT method. SIFT method. Distinctive Image Features from Scale-Invariant. Scale.

Real-Time Billboard Substitution in a Video Stream

High Performance Memory Read Using Cross-Coupled Pull-up Circuitry

Computer Science Faculty, Bandar Lampung University, Bandar Lampung, Indonesia

Combined Weak Classifiers

A Comparison of Still-Image Compression Standards Using Different Image Quality Metrics and Proposed Methods for Improving Lossy Image Quality

Fast trajectory matching using small binary images

Visual object classification by sparse convolutional neural networks

Introduction. Chapter Overview

Mobile Human Detection Systems based on Sliding Windows Approach-A Review

Mobile Visual Search with Word-HOG Descriptors

Real-Time Model-Based Hand Localization for Unsupervised Palmar Image Acquisition

Feature Selection. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

Image Segmentation for Image Object Extraction

PART-LEVEL OBJECT RECOGNITION

On Road Vehicle Detection using Shadows

RANSAC and some HOUGH transform

Hand Posture Recognition Using Adaboost with SIFT for Human Robot Interaction

Performance Characterization in Computer Vision

Query Learning Based on Boundary Search and Gradient Computation of Trained Multilayer Perceptrons*

Radar Target Identification Using Spatial Matched Filters. L.M. Novak, G.J. Owirka, and C.M. Netishen MIT Lincoln Laboratory

Video Alignment. Final Report. Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin

Beyond bags of features: Adding spatial information. Many slides adapted from Fei-Fei Li, Rob Fergus, and Antonio Torralba

Local Features: Detection, Description & Matching

Object Detection in a Fixed Frame

3D object recognition used by team robotto

Model-based Enhancement of Lighting Conditions in Image Sequences

Transcription:

Relational Indexing Mauro S. Costa 1 and Linda G. Shapiro 1,2 " 1 Department of Electrical Engineering, Box 352500 2 Department of Computer Science & Engineering, Box 352350 University of Washington Seattle WA 98195 U.S.A. Abstract. We address the problem of recognizing 3D objects in scenes containing multiple objects by means of a new indexing technique called relational indexing. Given a database of relational models, we determine those models whose relational descriptions are most similar to subsets of the relational description of an unknown scene. The relational indexing technique has worst-c~e complexity O(m(~)) for relational graphs of up to k nodes and a database of m models. This paper evaluates the performance of the technique using Monte Carlo experiments. 1 Introduction In a model-based object recognition system, the task of matching image features to model features, in the general case, implies searching the space of all possible correspondences. Indexing is one of the techniques that have been largely utilized to reduce this search space. In recent years, several systems have made use of different approaches to indexing ([1], [4], [5], [8]). In this paper we describe relational indezing: a new approach to indexing into a database of models that makes use of features and the spatial relationships among them. In this new matching technique each model in the database is described by a relational graph of all its features, but small relational subgraphs of the image features are utilized to index into the database and retrieve appropriate model hypotheses. For a database of m models and for relational graphs of up to k nodes the algorithm has worst-case complexity O(rn(~)). This paper investigates the use of this new technique in a model-based 3D recognition system; Monte Carlo experiments are used to evaluate its performance as a hypotheses generation mechanism. 2 Relational Indexing Notation An attributed relational description D is a labeled graph D = (N, E) where N is a set of attributed nodes and E is a set of labeled edges. For each attributed * This research was supported by the National Science Foundation under grant number IRI-9023977, by the Boeing Commercial Airplane Group, and by the Washington Technology Center.

131 node n E N, let A(n) denote the attribute vector associated with node n. Each labeled edge e E E will be denoted as e = (ni, nj,li,j) where ni and nj are nodes of N and Li,j is the label associated with the edge between them. Lid is usually a scalar, but it can also be a vector. A relational description D = (N, E) can be broken down into subgraphs, each having a small number of nodes. We will consider subgraphs of two nodes, called t-graph.s. All of our graphs are complete graphs, so a graph of k nodes has (~) 2-graphs, each consisting of a pair of attributed nodes and the labeled relationship between them. The relationship between the two nodes may be a meaningful spatial relationship or the null relationship r~one. We will refer to the set of 2-graphs of a relational description Dl as 7~. Figure 1 illustrates a partial graph representing an object and all the 2-graphs for the given relational graph. MODEL-VIEW ~3 a; RELATIONS: @ncloses b: coaxial FEJ~TURE& 1 : coaxiats-mufti 2: ellipse 3: parallel lines PARTIAL GRAPH 2-GRAPHS a b nodes: leatures Fig. 1. Sample graph and corresponding 2-graphs for the "hexnut" object. 3 Relational Indexing Algorithm Let DB = {Mz, M2,...,Mm} be the database of models, where each Mi = (Ni, Ei) is an attributed relational description. Let D = (N, E) be a relational description that has been extracted from an image and T be the set of all 2- graphs of D. We would like to find the closest models to D. This is accomplished

132 in two steps: an off-line preprocessing step to set up the indexing mechanism and an on-line hypotheses generation step. The off-line step is as follows. Let T M~ be the set of 2-graphs of Mi. Each element G M; in this set is encoded to produce an index I M~, which is used to access a hash table. The bin corresponding to the particular encoded 2-graph G M~ stores information about which model M~ gave rise to that particular index. This encoding and storing of information in the hash table is done off-line and for all models in the database DB. In the on-line step the relational indexing procedure keeps an accumulator Ai for each model Mi in the database (all the accumulators are initialized to zero). Each 2-graph Gz in T is encoded to produce an index Iz. The procedure then uses that index to retrieve from the precomputed hash table all models Mi that contain a 2-graph that is identical to Gl. Identical means that the two nodes have the same attributes and the edge has the same label. For each 2- graph Gz of T, the accumulator A~ of every retrieved model M~ is incremented by one. After the entire voting process, the models whose accumulators have the highest votes are candidates for further consideration. Since the procedure goes through all (~) 2-graphs of T and for each one can retrieve a maximum of m models, the worst-case complexity is O(m(~)). However, the work performed on each model is very small, merely incrementing its accumulator by one. This is very different from methods that perform full relational matching on each model of the database. The relational indexing algorithm is given below. RELATIONAL INDEXING ALGORITHM Preprocessing (off-line) Phase 1. For each model M~ in the database DB do: - Encode each 2-graph G M~ to produce an index. - Store Mi and associated information in the selected bin of the hash table. Matching (on-line) Phase 1. Construct a relational description D for the scene. 2. For each 2-graph Gz of D do: - Encode it, produce and index, and access the hash table. - Cast a vote for each Mi associated with the selected bin. 3. Select Mi's with enough votes as possible hypotheses. Since some models share features and relations, it is expected that some of the hypotheses produced will be incorrect. This indicates that a subsequent verification phase is essential for the method to be successful. It is important to mention that the information stored in the hash table is actually more than just the identity of the model that gave rise to a particular 2-graph index. It also contains information about which specific features (and their attributes) are part of the 2-graph. This information is essential for hypothesis verification and eventual pose estimation.

133 4 Matching with Relational Indexing: an Example In this section we give an example of the experiments we have conducted to demonstrate the use of the relational indexing technique for 3D object recognition with appearance-based features [2]. An appearance-based model of an object is defined as the collection of the features that can be reliably detected from a training set of real images of the object. We chose to use view-class models, in which an object is represented by a small set of characteristic views, each having its own distinct feature set [6]. We have created a database of appearance-based object models for a set of mechanical parts that have both flat and curved surfaces, holes, and threads. The relational descriptions DV, M of all the model-views were derived from a large set of training pairs of real images (280 image pairs of 7 models). In order to illustrate our recognition methodology, we matched nine test images of both single and multiple object scenes to the database of model-views. The nine test images used are shown in figure 2. The database of models was created by encoding all 2-graphs for each of the model-views. For each test scene, features and relations were detected, the relational description was built, and all 2-graphs were encoded. Relational indexing was then performed and the generated hypotheses were normalized by the number of 2-graphs in the original models and ranked in order of strength. Hypotheses that exceeded a preset strength threshold were dubbed "strong hypotheses." These hypotheses are to be passed to the verification procedure for further consideration. In each of the nine tests, the strong hypotheses were classified as type A, type B, or type C. Type A hypotheses are those where the correct model and the correct (closest) view class were identified. Type B hypotheses are those where the correct model was identified, but the chosen view class was not closest to the view in the image. Type B hypotheses can still be verified and used to determine pose if enough corresponding features are found. Type C hypotheses are those where an incorrect model was selected. These incorrect hypotheses should be ruled out in the verification step. The results of the nine tests are as follows: all the objects in the scenes have been correctly recognized (18 type A hypotheses); there were 9 type B hypotheses, and 4 type C hypotheses. Figure 3(a) shows the results for test scene 9, which contains four objects: the "stacked cylinder," the "hexnut," the "wrench," and the "cylinder-block." The system produced five strong hypotheses; four were correct and are overlaid on the image. These hypothesized models were taken through pose computation (affine correspondence of appearance-based model features and scene features) without verification. The fifth strong hypothesis (not shown) matched the object "hexnut" to an incorrect view of the correct object model. The subgraph indices shown in Figure 1 were among those that were used in the matching process. Figure 3(b) illustrates the correct (type A) hypotheses generated for test scene 5. Of the three type B hypotheses generated~ one was for the "cylinderblock" object and two were for the "hexnut" object, both of which are present in the scene. As seen, the method shows promising results. However, a more

134 Fig. 2. The nine test scenes used. thorough characterization of the indexing mechanism performance needs to be assessed. The next section addresses this issue. 5 Monte Carlo Experiments We performed Monte Carlo experiments to investigate how our relational indexing algorithm performs as the following factors are varied: 1. Number of model graphs in the database.

135 Fig. 3. (a) Left image of test scene 5 overlaid with the appearance-based features of the hypothesized model matches. The objects in this scene are the "cylinder-block" and the "hexnut." (b) Right image of test scene O overlaid with the appearance-based features of the hypothesized model matches. The objects in this scene are the "stacked cylinder," the "hexnut," the "wrench," and the "cylinder-block." 2. Average size of model graphs in the database. 3. Size of image graph, measured in terms of the percentages of extra and missing 2-graphs from a given model graph. 4. Size of cluster of similar models in the database. 5. The degree of similarity between models. In order to define similarity between model graphs, let DB = M1, M2,..., M,~ be a database of rrt model graphs. Consider the set of 2-graphs Ti of each model Mi. Let s(i, j) = tti n Tj[ be the measure of similarity between models M~ and Mj. If s(i, j) = 0, then M~ and Mj are fully discriminable by relational indexing. Synthetic data was obtained by generating random databases of models graphs whose nodes and edges were features and relations used in our current system. In order to generate realistic databases, the physical constraints between the features and their relations were taken into account. For each test database, each model was transformed into an "image graph" by randomly removing 2- graphs and adding extra 2-graphs and was then matched against the database of models. A model was retrieved only if at least 50% of its 2-graphs were found in the image. For each such set of image graphs used, we recorded the following quantities: - PCM: The average percentage of correct models retrieved, with respect to the total number of models retrieved - PHV: The average percentage of correct models retrieved wi~h the highest vo~e, with respect to the total number of models retrieved

- M: - C: 136 - PMR: The average percentage of models retrieved, with respect to the total number of models in the database Since we were interested in investigating the effect of similar models in the database on the performance of the technique, the above quantities were measured within and outside a cluster of similar models. For each experiment described above, 100 random replications (R = 100) were performed. Each database generated possessed one cluster of similar models. All the parameters involved in this investigation are listed below, along with the individual values each assumed: - D: Number of models in the database = 50, 100, 200, 500 - G: Average size of model graphs (measured in terms of 2-graphs) = 10, 15, 20, 25, 30. Percentage of missing 2-graphs = 0, 20, 50. - E: Percentage of extra 2-graphs = 0, 20, 50. - S: Similarity among the models in the cluster (measured as a percentage of G) = 20, 30, 40. Cluster size (measured as a percentage of D) : 10, 20, 30 The total number ofexperiments performed was then: R x D x G x M x E x S x C -- 162,000. We generated plots of PCM, PHV, and PMR as a function of the average size model in the database (measured in terms of 2-graphs), for every combination of the following parameters; D, S, C, M, and E. Given the large number of plots obtained, we only illustrate those for which the trends in performance are most significant (for the complete set of results, please see [3]). PCM, the first quantity measured, for both within the cluster and outside the cluster, had value 100% irrespective of the combination of the parameters involved. This means that the correct model was always among the retrieved models. Since relational indexing is a hypotheses generation mechanism, verification has to be performed for each model retrieved. In our current system, we start the verification with the models that obtained the largest number of votes, therefore we are also interested in finding out how often the correct model received the most votes. Quantity PHV reflects this aspect of the performance of the technique. For PHV, measured within the cluster, the plots of Figure 4 are representative of the trends in performance. From 4(a), (b), and (c), it can be seen that the average percentage of correct models retrieved decreases with an increase in the percentage of missing 2-graphs and in the size of the database. This is expected since the total number of models retrieved will increase with the size of the database. Figures 4(d), (e), and (f) show plots of PHV for different values of similarities within the cluster. The trend observed is that performance only changes significantly for small values of the average size models and with the percentage of missing 2-graphs, when the similarity between models is varied. The effects of the size of the cluster of similar models on performance can be seen on the plots

137 of Figures 4(g), (h), and (i). Again, performance decreases with an increase in the cluster size and in the percentage of missing 2-graphs. One important thing to notice is that regardless of the combination of parameters used, the number of extra 2-graphs had a negligible effect on the performance, indicating that the indexing technique is robust with respect to clutter. Quantity PHV measured outside the cluster had a much more stable behavior than for inside the cluster. In fact, the plots obtained are essentially flat, regardless of the combination of parameters used. The only parameters that slightly affect the performance are again the percentage of missing 2-graphs and the size of the database. Since outside the cluster the models are randomly different, the value of PHV is always very close to 100%, except for a larger percentage of missing 2-graphs (50%), for which it decreases to around 85% for the largest size database (500 models). Figure 5 shows some of the results obtained when measuring quantity PMR, within the cluster. It can be seen that the average percentage of total models retrieved remains very low, regardless of the size of the database, according to the plots in Figures 5(a), (b), and (c). The effects of changes in similarities among models in the cluster are depicted in Figures 5(d), (e), and (f). As expected, with an increase in the similarity between models, there is an increase in the percentage of total models retrieved. The larger the cluster size, the larger the average percentage of models retrieved will be. This trend can be observed in Figures 5(g), (h), and (i). As for the case of PHV, the percentage of extra 2-graphs has close to no affect in the performance of the technique in terms of the measured quantity PMR. This also holds for outside cluster measurements. Quantity PMR measured outside clusters, stays fairly constant (ranging from around 1% to 3% of the database size), regardless of the actual values of the parameters involved in the investigation. References 1. A. Califano and R. Mohan. Multidimensional Indexing for Recognition of Visual Shapes. IEEE Transactions on Pattern Analysis and Machine Intelligence, 16(4):373-392, 1994. 2. M. S. Costa and L. G. Shapiro. Scene Analysis Using Appearance-Based Models and Relational Indexing. In International Symposium on Computer Vision, pp. 103-108, Coral Gables, Florida, November 1995. 3. M. S. Costa and L. G. Shapiro. Relational Indexing for Object Recognition. Intelligent Systems Lab Technical Report #ISL-05-96, Department of Electrical Engineering, University of Washington, May 1996. 4. Y. Lamdan and H, J. Wolfson. Geometric Hashing; A general and efficient Model-based Recognition Scheme. In Second International Conference on Computer Vision, pp. 238-249, 1988. 5. C. F. Olson. Probabillstic Indexing for Object Recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 17(5):518-522, 1995. 6. L. G. Shapiro and M. S. Costa. Appearance-Based 3D Object Recognition. In Proc. of the NSF/DARPA Workshop on 3D Object Representation for Computer Vision, New York, NY, December 1994. 7. L. G. Shapiro and R. M. Haralick. A Metric for Comparing Relational Descriptions IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. PAMI-T,pp. 90-94, 1985. 8. F. Stein and G. Medioni. Structural Indexing: Efficient Three Dimensional Object Recognition. IEEE Transactione on Pattern Analysis and Machine Intelligence, 14(2):125-145, 1992.

138 Ioo84 +coo- _- - : >.\ \ i ~ soo ~ 1 7~ 6 4oo ~oo ~oo 13.e,,v~ne 2-;r,~,. 2~.. 00100 150 2oo 250 ~oo average model s~ze (2-graphs) (a) D=S0, S=20, C=10 ~176 ~so ~oo m+o 3OO (b) D=100, S=20, C=10 ~176 is.o 200 2so 3oo average model size (2-9r~ohs) (c) D=2oo, S=2O, C=1o 9 i~ l+oo + i 4o.o I~ +oo & ~ ISO ;,oo ~o ~oo (d) S=20, D=50, C=30 +ooo, i \ ~ ~so ~oo ~o ~oo (e) S=30, D=50, C=30 ~> 1~~ i ~o+o,m,a.gemm. o~ ae! (0 ~176 1so 2oo ~o 3oo le~oq eoo, ]~ eo0 +oo S=40, D=50, C=30 o+ooo &vemge model s~ze (2-graphs) (g) C=10, D=100, S=20 15 ae o~ (h) C=20, D=100, S=20 ;e ~ 8verage model size (2-grllphs) (i) C=30, D=lOO, S=20 Fig. 4. Sample results for quantity PHV (average percentage of correct models retrieved with the highest vote) measured within the cluster of models in the database.

139 B,4o mmr, g =.or t~',l. ~,~ = ~eo "6,o eo oo,;o ~4o ~o =o sver~s model size (2-graphs) (a) D=50, S=20, C=lO ~oo,so =oo ~5o ~oo (b) D=100, S=2O, C=10 oo,5o 2oo 2so 3oo average model size (2-gr~ohs) (c) D=200, S=20, C=]0 200 I 20.0 ~ ~,~,~ ; 9,~, - ~ ~ ~ ~'P~~ 5o~ 160 1 > 160 160 ~e 120 120 "6 r 8.0 v 80 40 40 00,o0 150 200 PSO 300 (d) S=20, D=50, C=30 average model size {2-graphs) (e) S=30, D--50, C=30 00 o!so ~c~ ~-o 300 average model s~ze (2-graphs) (f) S=40, D=50, C=30 zo o } i,o average model size (2~raphs) (g) C=10, D=100, S=20,oo 1~o zoo 2so ~oo (h) C=20, D=100, S=20 average model s~ze (2-graphs) (i) C=30, D=100, S=20 Fig. 5. Sample results for quantity PMR (average percentage of total models retrieved) measured within the cluster of models in the database.