Automatic image content retrieval - are we getting anywhere?

Similar documents
RETRIEVAL OF TRADE MARK IMAGES BY SHAPE FEATURE - THE ARTISAN PROJECT

Retrieval of trade mark images by shape feature. J P Eakins

Searching Image Databases Containing Trademarks

Content Based Image Retrieval with Semantic Features using Object Ontology

Perceptually-driven moment thresholds for shape description in image databases

General Image Database Model

Analysis of Image and Video Using Color, Texture and Shape Features for Object Identification

An Efficient Approach for Color Pattern Matching Using Image Mining

CHAPTER 8 Multimedia Information Retrieval

Character Recognition

INFORMATION RETRIEVAL SYSTEM: CONCEPT AND SCOPE

Image Mining: frameworks and techniques

Category Theory in Ontology Research: Concrete Gain from an Abstract Approach

Holistic Correlation of Color Models, Color Features and Distance Metrics on Content-Based Image Retrieval

For many people it would appear that there is no difference between information and data.

Using surface markings to enhance accuracy and stability of object perception in graphic displays

A Content Based Image Retrieval System Based on Color Features

COMPARISON OF SOME CONTENT-BASED IMAGE RETRIEVAL SYSTEMS WITH ROCK TEXTURE IMAGES

Image retrieval based on bag of images

An Introduction to Content Based Image Retrieval

Journal of Asian Scientific Research FEATURES COMPOSITION FOR PROFICIENT AND REAL TIME RETRIEVAL IN CBIR SYSTEM. Tohid Sedghi

Extending Image Retrieval Systems with a Thesaurus for Shapes

Optimal Grouping of Line Segments into Convex Sets 1

TRANSFORM FEATURES FOR TEXTURE CLASSIFICATION AND DISCRIMINATION IN LARGE IMAGE DATABASES

Multimodal Information Spaces for Content-based Image Retrieval

Content Based Image Retrieval

Invariant-based data model for image databases

Shape Retrieval with Flat Contour Segments

Efficient Content Based Image Retrieval System with Metadata Processing

Chapter 8 Visualization and Optimization

Image retrieval based on region shape similarity

HOW USEFUL ARE COLOUR INVARIANTS FOR IMAGE RETRIEVAL?

Query-Sensitive Similarity Measure for Content-Based Image Retrieval

CE4031 and CZ4031 Database System Principles

Differential Compression and Optimal Caching Methods for Content-Based Image Search Systems

Chapter 4 - Image. Digital Libraries and Content Management

Improving Content Based Image Retrieval Systems with a Thesaurus for Shapes

Detecting Digital Image Forgeries By Multi-illuminant Estimators

IMAGINE Objective. The Future of Feature Extraction, Update & Change Mapping

SOFIA: An Image Database with Image Retrieval by Image Content

Image Matching Using Run-Length Feature

Definition, Detection, and Evaluation of Meeting Events in Airport Surveillance Videos

An Analysis of Image Retrieval Behavior for Metadata Type and Google Image Database

Top-K Retrieval Cbir Systems

Direction-Length Code (DLC) To Represent Binary Objects

Welcome Back to Fundamental of Multimedia (MR412) Fall, ZHU Yongxin, Winson

Water-Filling: A Novel Way for Image Structural Feature Extraction

International Journal of Modern Trends in Engineering and Research e-issn No.: , Date: 2-4 July, 2015

Implementation Techniques

Consistent Line Clusters for Building Recognition in CBIR

Wallpaper Groups and Statistical Geometry

Qualitative Physics and the Shapes of Objects

CE4031 and CZ4031 Database System Principles

A Texture Descriptor for Image Retrieval and Browsing

PERFORMANCE EVALUATION OF ONTOLOGY AND FUZZYBASE CBIR

A Miniature-Based Image Retrieval System

Breaking it Down: The World as Legos Benjamin Savage, Eric Chu

Contextual priming for artificial visual perception

Learning to Match. Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li

On Modeling Variations for Face Authentication

A Method of Annotation Extraction from Paper Documents Using Alignment Based on Local Arrangements of Feature Points

Chapter 7. Conclusions and Future Work

Multimedia Information Retrieval

Self-Organizing Maps of Web Link Information

Drywall state detection in image data for automatic indoor progress monitoring C. Kropp, C. Koch and M. König

Multimedia Database Systems. Retrieval by Content

Sketch Based Image Retrieval Approach Using Gray Level Co-Occurrence Matrix

Content-based Image Retrieval (CBIR)

CHAPTER 1 INTRODUCTION

CONTENT BASED IMAGE RETRIEVAL SYSTEM USING IMAGE CLASSIFICATION

Security and Usability Challenges of Moving-Object CAPTCHAs

Content-Based Image Retrieval Readings: Chapter 8:

Boosting Image Database Retrieval

A NOVEL FEATURE EXTRACTION METHOD BASED ON SEGMENTATION OVER EDGE FIELD FOR MULTIMEDIA INDEXING AND RETRIEVAL

A Top-Down Visual Approach to GUI development

Object perception by primates

Recognizing hand-drawn images using shape context

A Texture Extraction Technique for. Cloth Pattern Identification

Computer Aided Draughting and Design: Graded Unit 1

Comparing the Usability of RoboFlag Interface Alternatives*

A Texture Feature Extraction Technique Using 2D-DFT and Hamming Distance

Content-Based Image Retrieval Readings: Chapter 8:

Domain Specific Search Engine for Students

How to Conduct a Heuristic Evaluation

2. METHODOLOGY 10% (, ) are the ath connected (, ) where

Automated Cognitive Walkthrough for the Web (AutoCWW)

Texture Segmentation by Windowed Projection

FACE DETECTION AND RECOGNITION OF DRAWN CHARACTERS HERMAN CHAU

Image Mining Using Image Feature

A Novel Image Retrieval Method Using Segmentation and Color Moments

Massive Data Analysis

IJESRT. Scientific Journal Impact Factor: (ISRA), Impact Factor: 1.852

Rough Feature Selection for CBIR. Outline

Design. Introduction

Higher National Unit specification: general information. Graded Unit 2

Human Motion Detection and Tracking for Video Surveillance

MULTIVIEW REPRESENTATION OF 3D OBJECTS OF A SCENE USING VIDEO SEQUENCES

Detecting motion by means of 2D and 3D information

Content Based Image Retrieval (CBIR) Using Segmentation Process

Recordkeeping Standards Analysis of HealthConnect

Transcription:

Automatic image content retrieval - are we getting anywhere? John P Eakins Department of Computing, University of Northumbria at Newcastle, Newcastle upon Tyne NE1 8ST, United Kingdom Introduction Recent advances in computer-based storage, transmission and display of multimedia data have been truly impressive. The sight (and sound) of information emanating from all parts of the planet is becoming commonplace. Initially, users are often amazed at the technology that makes it all possible. As they gain familiarity with such systems, though, their expectations can rise rapidly. Difficulties in accessing desired sites, or slow response times when downloading images, are soon recognized as inadequacies in the level of service provided, and lead to demands for improvement. Success in meeting users' needs is not just a matter of improving transmission speeds and image quality, however. Far more fundamental issues need to be considered. Users will want to put the next generation of multimedia systems to all kinds of new and unforeseen uses. In order to remain successful, system providers will have to devise new tools to meet these needs. One requirement that can be predicted with some confidence is the need for far more sophisticated retrieval capabilities than any current system offers. The limitations of current technology are well illustrated in a recent editorial in IEEE Multimedia, which shows just how difficult and timeconsuming it is to perform a simple task such as obtaining a few images of animals from the Web (Jain, 1995). The capability required to meet this need is content-based image retrieval (CBIR) - the ability to retrieve images from a collection on the basis of features extracted or inferred from the images themselves. Computer-based CBIR, the application of image processing and related techniques to derive retrieval features without the need for human intervention, is now a highly active area for research and development (Cawkell (1993), Gudivada & Raghavan (1995)). While most of this work is still experimental, commercial-strength systems such as QBIC (Flickner et al, 1995) and Virage (Gupta et al, 1996) are now beginning to appear. However, the range of facilities offered by present systems is often quite limited. In many cases, retrieval capabilities appear to be governed more by what current technology can offer than by user demand. This is perhaps not surprising, given the lack of maturity of much of the underlying technology and the difficulty of identifying what users really want. However, evidence on user requirements is beginning to accumulate (Enser, 1995), even if its conclusions are somewhat tentative at present. This paper attempts to assess how successful current developments in technology are likely to be in meeting these user needs. The need for retrieval by image content Does a system holding collections of electronic images need to provide retrieval by image content - in other words, some kind of image classification or indexing facility? The experience of generations of picture librarians, backed up by studies of user queries in such libraries (e.g. Enser, 1993), suggests powerfully that it does. Journalists requesting photographs from a newspaper library, designers

looking for materials with a particular colour or texture, and engineers looking for drawings of a particular type of part, all have a clear idea of what they want, even if they prefer to delegate the actual searching to an intermediary. The existence - and continuing use - of detailed classification schemes such as ICONCLASS (Gordon, 1990) for art images, and the Opitz code (Opitz et al, 1969) for machined parts, reinforces this message. Is automatic retrieval by image content (i.e. computerized extraction of index features from images, and subsequent retrieval of images containing the desired features) also worthwhile? Again, the enthusiasm of art librarians using IBM's prototype QBIC system (Holt & Hartwick, 1994) suggests that it is, drastically cutting the time needed both to catalogue an image and to process a user query. Manual techniques of picture classification and indexing are inherently time-consuming, and there is evidence that their effectiveness is often limited. Studies of inter-indexer consistency (reviewed by Enser, 1995) have revealed wide variations in choice of keywords for any given image, and analysis of user queries suggest that a given image may satisfy such a wide range of queries that it is impossible to foresee them all. Enser's rather depressing conclusion is that "if the retrieval utility of an image suffers from low predictability, the subject indexing of that image must have low utility". The case for improved methods of image retrieval is thus a strong one. To what extent can automated systems meet this need? To answer this question, we first need a framework within which we can distinguish between different types - or levels - of image content retrieval. Potentially, one could seek to retrieve an image by many types of attribute, including the following: low-level attributes such as its colour, texture or shape (e.g. an image containing green triangles); derived attributes such as the presence or arrangement of specific types of object (e.g. chairs around a table); inferred abstract attributes such as a particular type of event (e.g. a football match); identifying attributes such as the presence of named individuals, locations, or events (e.g. the Queen opening Parliament);. subjective attributes such as the emotions one might associate with the image (e.g. happiness); non-derivable attributes such as who created the image, where and when. This wide range of potential query types is of course one of the reasons why classification and indexing of images is so difficult. Not all types of query are applicable to all types of image - a Fig 1. Examples of abstract images, taken from the UK Trade Marks registry. Crown Copyright reserved database of abstract images of the type shown in Fig. 1 is unlikely to be worth querying at any but the lowest level, while a natural scene such as that shown in Fig. 2 would be more likely to answer queries at one of the higher levels.

Fig 2. An example of a natural scene Previous authors have proposed a number of possible frameworks within which one might discuss different types of image query, including Panofsky's classification of fine art image attributes into preiconography, iconography, and iconology (Panofsky, 1955), Enser's classification of queries to a picture library into unique, non-unique, unique refined and non-unique refined (Enser, 1993), and my own division of image content retrieval systems into spatial, image paradigm and shape retrieval systems (Eakins, 1992). However, none of these can encompass the entire range of query types listed above. Hence a new framework is proposed, as indicated below. Three levels of content retrieval The types of possible image query listed above range from the highly concrete (find all pictures containing red) to the very abstract (find pictures depicting "suffering"). As one moves towards the abstract end of this continuum, queries become harder to formulate and to execute, and opinions on what constitutes successful retrieval become more diverse. Hence a classification of query types in order of increasing abstraction seems appropriate. Level 1, the lowest level, comprises retrieval by what Gudivada & Raghavan (1995) describe as primitive features such as colour, texture, shape or the spatial location of image elements. This level could perhaps be further divided into the following sub-categories: retrieval by colour feature (find all pictures containing yellow and blue regions); retrieval by texture (find images containing regions with texture similar to that of woven cloth); retrieval by shape (find drawings similar in shape to this sketch) 1 ; retrieval by spatial location (find pictures with objects of interest in the top left-hand corner) 2 ; retrieval by a combination of the above (find images containing yellow stars arranged in a ring); though these distinctions can appear somewhat artificial to the user, since few experimental or commercial systems now limit their retrieval capabilities to a single type of feature. This kind of retrieval is mainly required for specialist applications such as location of drawings in a design archive (Eakins, 1993), registration of trade mark images (Eakins, 1994), or colour matching of fashion 1 corresponding to the shape retrieval category identified in my earlier review (Eakins, 1992). 2 corresponding to the spatial information retrieval category identified in my earlier review.

accessories (Petkovic, 1996). It is the only class of query that can meaningfully be put to abstract (non-representational) images of the type illustrated in Fig. 1. Level 2 comprises retrieval by derived attributes (Gudivada & Raghavan describe these as logical features), involving some degree of logical inference about the identity of the objects depicted in the image. It is also sometimes referred to as retrieval by semantic content, and corresponds both to the image paradigm retrieval distinguished in my earlier review, and to Panofsky's pre-iconographic level of picture description. This level of retrieval can usefully be further divided into: retrieval of objects of a given type (find pictures of a passenger train crossing a bridge); retrieval of individual objects or persons (find a picture of Nelson's column in Trafalgar Square). The first of these sub-categories corresponds in part to Enser's non-unique and non-unique refined categories of pictorial query, the second to his unique and unique refined categories (Enser, 1993). This level of retrieval is probably of more general applicability than level 1 - particularly for representational images of the type shown in Fig. 2. Many of the queries analysed in Enser's study of newspaper picture libraries fall into this overall category, though often qualified by further search criteria which cannot be readily derived from the image - an example being a request for "North London street scenes during the 1950's" (Enser, 1993). Level 3 comprises retrieval by abstract attributes, involving a high degree of abstract - and possibly subjective - reasoning about the meaning and purpose of the objects or scenes depicted. Again, this level of retrieval can usefully be subdivided, as follows: retrieval of named events or types of activity (find pictures of English folk dancing); retrieval of pictures with emotional or symbolic significance (find a picture depicting "atonement"). The difference between these two sub-categories is in some respects similar to Panofsky's distinction between iconography (describing a picture's actual subject matter) and iconology (its deeper artistic or religious significance). Although somewhat specialized, queries at this level can be encountered in both newspaper and art libraries, and pose a major challenge to the ingenuity of the searcher. The current state of the art At level 1, significant progress has been made by exploiting the results of computer vision research. Experimental and even a few commercial systems now exist which can provide content-based retrieval on the basis of primitive features such as colour or shape. Systems are typically example-based, allowing users to submit a query image and retrieve the most similar stored images from a database. Some types of feature - particularly colour - have proved more effective as retrieval cues than others. Retrieval by colour, in which colour histograms are generated for each stored image (or selected regions within the image) and matched with a histogram extracted from the query image, was first introduced by Swain & Ballard (1991). It has since been significantly improved by researchers such as Stricker & Orengo (1995), and can now generate some quite impressive-looking results, particularly when combined with some element of spatial matching (Stricker & Dimai, 1996). Retrieval by texture, comparing values of automatically-extracted features such as contrast. coarseness, directionality and regularity as defined by Tamura et al (1978), has also been achieved with some degree of success, particularly when used in conjunction with colour matching (Niblack et al, 1993). Shape retrieval has proved a considerably greater challenge, despite considerable research into the topic (Niblack et al, 1993). At present, most shape retrieval systems operate by using well-established

image-processing techniques to compute a set of shape features for identified regions of interest within each stored image. Queries are then answered by computing the same set of features for the query image, and retrieving those stored images whose features most closely match those of the query. Two main types of shape feature have been used (see Fig. 3) - global features such as aspect ratio, circularity and moment invariants (Niblack et al, 1993) and local features such as points of high curvature (Ramesh & Sethi, 1995) or sets of consecutive boundary segments (Mehrotra & Gary, 1995). However, the effectiveness of current shape retrieval systems still leaves a lot to be desired. Evaluation results reported with IBM's QBIC system (Faloutsos et al, 1994) show clearly that its retrieval effectiveness is markedly better for colour matching than for shape matching. And there is worrying evidence that few, if any, of the shape feature measures in current use are accurate predictors of human judgements of shape similarity (Mumford, 1991; Scassellati et al, 1994). Current commercial systems such as QBIC and Virage offer retrieval by colour, texture or shape, either singly or in combination. They provide the user with a range of interfaces, to allow query by image example, by menu selection of colour or texture, or by hand-drawn sketch. In order to improve search efficiency, they use multidimensional index structures such as the R*-tree (Beckmann et al, 1990), which avoid the need to match the query with every image in the database. Despite the limitations of the underlying technology, such systems can provide useful results, particularly in the hands of a skilled user. Of all current systems, Virage is perhaps the most interesting. Unlike its competitors, it is designed as a series of independent modules, which can be combined in a variety of ways. This makes it easy to extend the system by building in new types of query interface or additional customized modules to process specialized collections of images such as trademarks. Alternatively, Virage modules can be incorporated within existing database management products to extend their image-handling capabilities. An on-line demonstration of the Virage system is available on the World- Wide Web at http://www.virage.com/online/. The common thread running through all systems supporting level 1 retrieval is their reliance on automatic extraction and matching of primitive features such as boundary shape characteristics or colour histograms. Feature extraction may be performed once and for all when the images are first added to the database. Alternatively, as with MIT's Photobook system (Pentland et al, 1994), features may be computed at search time as required, giving greater flexibility at the expense of speed. Either way, the emphasis is on concrete features which are relatively easy to extract from raw image data - and, more importantly, can be derived purely from information present in the images themselves. The underlying assumption of all current level 1 systems is that there is no need to refer to any external knowledge base when indexing a new image or processing a query. At level 2, progress has been more limited. Retrieval at this level involves matching derived or abstract attributes of the image such as the presence of specified types of building or animal in a scene. This kind of feature matching is qualitatively different, as it requires not just feature analysis, but a complex inferential process, making reference to some stored knowledge base of object paradigms. No amount of colour, texture or shape analysis can tell us on its own whether a particular image represents a dog, since dogs come in many shapes, sizes and colours. The ease with which humans can recognize an image as a dog conceals a complex and as yet incompletely-understood process. For this reason, there are many who consider that the only way to provide level 2 retrieval is to employ human indexers to add suitable keywords (Gudivada & Raghavan, 1995). However, some researchers are attempting to bridge the gap between primitive-level and semanticlevel retrieval. Although elements of this approach were present in the early IMAID system (Chang & Fu, 1981), the first systematic attempt to tackle this problem appears to be that of Rabitti & Stanchev (1989). Their GRIM_DBMS was capable of retrieved line drawings of objects of a given type within a narrow predefined domain, using a schema constructed by the systems designer as its knowledge base.

The system analysed object drawings, labelling each with a set of possible interpretations and their probabilities. These were then used to derive likely interpretations of the scene within which they appeared. Object and scene interpretations could be queried using an SQL-like command language. Despite the fact that the domain of objects chosen to illustrate the system was severely restricted, and that no retrieval effectiveness results were reported, GRIM_DBMS remains important as an indicator of how automatic semantic image retrieval might be achieved.. More recently, Hermes et al (1995) have used similar techniques to derive natural-language descriptors directly from images of outdoor scenes for their IRIS system. Colour, texture, region and spatial information from the image are input to a graph parser which derives the most likely interpretation of each region within the image, and hence of the overall scene, generating text descriptors which can be input to any text retrieval system (see Fig. 3). Like GRIM_DBMS, IRIS uses a complex reasoning process, linked to an external knowledge base, to derive image interpretations; also like GRIM_DBMS, the process of building up the knowledge base requires considerable human effort. However, its image analysis and interpretation capacity is much more highly-developed than GRIM_DBMS. Its success, albeit within a highly restricted domain, reinforces the hypothesis that level 2 retrieval is basically a deductive reasoning process which can in principle be automated. Fig 3. Combining different aspects of an image to generate automatic labelling of a natural scene in IRIS. At level 3, the problems have so far proved insurmountable. Identification of the objects or individuals in a scene is a formidable enough task; attempting to understand the significance of their presence or location in the scene well-nigh impossible without considerable intellectual effort. It might be possible to teach an image understanding system that a combination of small human beings around a table bearing a cake topped with candles can be interpreted as a child's birthday party. But identifying the connotations of a picture of, say, the Jarrow march is clearly not a logical reasoning process, and it is difficult to see how retrieval at this level could ever be automated. Even the former task requires access to a knowledge base of formidable complexity. Prospects for viable image retrieval at this level seem vanishingly small - though this has not deterred Kato (1992) from attempting to index a

collection of paintings by generating subjective descriptors such as "warm" and "romantic" directly from their colour histograms. From the details reported, his approach seems somewhat naive. In summary, significant progress has been made in delivering usable image retrieval at level 1 - though even here there is still considerable scope for improvement in both search efficiency and effectiveness. By contrast, progress at levels 2 and 3 has been extremely slow, though some potentially promising lines of investigation have been identified. Evidence both from informal discussions with picture librarians and from studies such as Enser (1993) makes it quite clear that while retrieval at level 1 might be nice to have, most users' real needs can be met only by providing higher levels of retrieval. There is thus a significant gap between users' needs and the capabilities of even the most advanced of today's experimental systems. What are the prospects of bridging this gap in the foreseeable future? Research issues At level 1, the retrieval process itself appears to be relatively well understood, and the main task is to improve both its efficiency and effectiveness. Better ways need to be found of identifying salient image features for a given range of applications, particularly those characterizing object shape. Improved methods of recognizing key features in noisy or distorted images are also required. Further advances in indexing and matching efficiency are needed to handle ever larger collections of images. All these areas are actively being researched, and significant improvements in the performance of image retrieval systems operating at this level can reasonably be expected. At level 2 and beyond, the main problem is that the retrieval process is not well understood. The main block to progress is the process of interpreting an image in order to generate a sufficiently rich set of logical features for retrieval. As indicated above, many consider the task of automatically generating such interpretations too far beyond the limits of current technology to be worth pursuing (Gudivada & Raghavan, 1995). However, the experiments of Rabbitti & Stanchev (1989) and Hermes et al (1995) suggest otherwise. Given improved mechanisms for building up a knowledge base of image feature interpretations, it should be possible to construct systems that offer level 2 image retrieval within restricted but non-trivial domains, such as photographs of sporting events. A number of potentially fruitful lines of approach are available. One is the use of neural networks for object recognition - a technique that has already been used with some success in the image retrieval field (Rickman & Stonham, 1993). Neural networks certainly have the flexibility to distinguish between images of, say, dogs and sheep - though since their reasoning process cannot be examined, one can never have complete confidence in their conclusions. Another is the use of ideas arising from studies into human visual cognition, such as those of Marr (1982) and Biederman (1987), who attempt to explain the mechanisms by which humans recognize and identify concrete objects such as a chair or a bicycle. Biederman, for example, has outlined a possible process involving edge detection, primitive feature extraction and component identification as intermediate steps. Whether or not this eventually proves valid for human vision, it can still provide a useful starting-point for designers of automatic object recognition systems. This approach offers a real possibility of worthwhile progress at level 2 - though it clearly still leaves a lot of questions unanswered, particularly the structure of the knowledge base needed to support the recognition process, and how it would be populated. Even at level 1, such an approach can prove useful. My colleagues and I (Shields et al, 1995; Eakins et al, 1996) are currently developing a shape retrieval system for trade mark images known as ARTISAN 3, which incorporates elements of 3 Automatic Retrieval of Trademark Images by Shape ANalysis

Fig 4a. Boundaries extracted by ARTISAN from a typical abstract trademark image. a Fig 4b. Boundaries grouped by ARTISAN into families on the basis of proximity and parallelism. b Biederman's hypothesis, grouping image elements into patterns by recognizing non-accidental properties such as shape similarity, symmetry and collinearity. These are used both to "clean up" noise-affected images and to group image components into families from which additional shape features can be extracted (Fig. 4). Early results from this project (Fig. 5) are encouraging. Fig 7a. An ARTISAN query screen, showing a query image (top left) and run-time retrieval options. Fig 7b. Retrieval results from the query illustrated in Fig 7a, showing the ten most similar retrieved shapes in similarity order. The query shape is also included for comparison purposes A further research area we cannot afford to neglect concerns the user. We need to know much more about why people request images, what they do with them, and how they judge the relevance of retrieved images to their needs. More comprehensive studies of the type conducted by Enser (1993) are urgently needed. Without this knowledge, we cannot adequately evaluate image retrieval systems; without systematic evaluation, we cannot expect systems to deliver results which users find acceptable. Even at level 1, where the issue reduces to comparing human and machine judgements of shape or texture similarity, very little work has been reported (a notable exception being Scassellati et al (1994)). While such studies will not in themselves solve the problems of image retrieval, they are a necessary prerequisite to success.

Conclusions There seems little doubt that automatic image content retrieval will be seen as a desirable, if not essential, facility in the next generation of multimedia systems. Given current research trends, it should be feasible to equip such systems with effective low-level automatic image content retrieval facilities. The development of systems capable of matching users' real needs is a much more formidable task. However, the problems involved in meeting these needs, at least up to level 2, do not appear insoluble, and a number of possible ways of tackling these problems have already been identified. They should provide a worthwhile challenge to the research community for some time to come. Acknowledgements Thanks are due to J M Boardman for his development work on the ARTISAN project, to T Hermes for providing images from IRIS, and the British Library and the UK Patent Office for their financial support. References Beckmann, N et al (1990) "The R*-tree: an efficient and robust access method for points and rectangles" ACM SIGMOD Record 19(2), 322-331 Biederman, I (1987) "Recognition-by-components: a theory of human image understanding" Psychological Review 94(2), 115-147 Cawkell, A E (1993) "Indexing collections of electronic images: a review" British Library Research Review 15. British Library, London Chang, N S & Fu, K S (1981) "Picture query languages for pictorial database systems" IEEE Computer 14(11), 23-33 Chang, S K et al (1988) "An intelligent image database system" IEEE Transactions on Software Engineering 14(5), 681-688 Eakins, J P (1992) "Pictorial information systems - prospects and problems" Proc 14th BCS IRSG Research Colloquium on Information Retrieval, Lancaster, 102-123. Eakins, J P (1993) "Design criteria for a shape retrieval system" Computers in Industry 21, 167-184 Eakins, J P (1994) "Retrieval of trade mark images by shape feature" Proc First International Conference on Electronic Library and Visual Information System Research, de Montfort University, Milton Keynes, 101-109 Eakins J P, Shields K, & Boardman J M (1996) "ARTISAN - a shape retrieval system based on boundary family indexing" in Storage and Retrieval for Image and Video Databases IV, (Sethi, I K & Jain, R C, eds), Proc SPIE 2670, pp 17-28 Enser, P G B (1993) "Query analysis in a visual retrieval context" Journal of Document and Text Management 1(1), 25-52 Enser, P G B (1995) "Pictorial information retrieval" Journal of Documentation, 51(2), 126-170 Faloutsos, C et al (1994) "Efficient and effective querying by image content" Journal of Intelligent Information Systems 3, 231-262

Flickner, M et al (1995) "Query by image and video content: the QBIC system" IEEE Computer, 28(9), 23-32 Gordon, C (1990) "An introduction to ICONCLASS" in Terminology for Museums, Proceedings of an International Conference, Cambridge, 1988 (Roberts, D A, ed), 233-244. Museum Documentation Association Grosky, W I & Mehrotra, R (1990) "Index-based object recognition in pictorial database management" Computer Vision Graphics, and Image Processing 52(3), 416-436 Grosky, W I & Mehrotra, R (1992) "Image Database Management" Advances in Computers 34, 237-291 Gudivada, V N & Raghavan, V V, eds (1995) "Content-based image retrieval systems" IEEE Computer, 28(9), 18-22 Gupta, A et al (1996) "The Virage image search engine: an open framework for image management" in Storage and Retrieval for Image and Video Databases IV, (Sethi, I K & Jain, R C, eds), Proc SPIE 2670, pp 76-87 Hermes, T et al (1995) "Image retrieval for information systems" in Storage and Retrieval for Image and Video Databases III (Niblack, W R & Jain, R C, eds), Proc SPIE 2420, pp 394-405 Holt, B and Hartwick, L (1994) "Retrieving art images by image content: the UC Davis QBIC project" Proc First International Conference on Electronic Library and Visual Information System Research, de Montfort University, Milton Keynes, 93-100 Jain, R (1995) "World-wide maze" IEEE Multimedia 2(3), 3 Kato, T (1992) "Database architecture for content-based image retrieval" in Image Storage and Retrieval Systems (Jambardino, A A & Niblack, W R, eds), Proc SPIE 2185, pp 112-123 Marr, D (1982) "Vision" Freeman, San Francisco Mehrotra, R and Gary, J E (1995) "Similar-shape retrieval in shape data management" IEEE Computer 28(9), 57-62 Mumford, D (1991) "Mathematical theories of shape: do they model perception?" in Geometric Methods in Computer Vision, Proc SPIE 1570, pp 2-10 Niblack, W et al (1993) "The QBIC project: querying images by color, texture and shape" IBM Research Report RJ-9203 Opitz, H et al (1969) "Workpiece classification and its industrial application" International Journal of Machine Tool Design Research 9, 39-50 Panofsky, E (1955) Meaning in the visual arts. Doubleday Anchor Books, Garden City, NY Pentland A et al (1994) "Photobook: tools for content-based manipulation of image databases" in Storage and Retrieval for Image and Video Databases II (Niblack, W R & Jain, R C, eds), Proc SPIE 2185, pp 34-47 Petkovic, D (1996) "Query by Image Content" oral presentation to Storage and Retrieval for Image and Video Databases IV, San Jose, California, February 1996 Rabbitti, F & Stanchev, P (1989) "GRIM_DBMS: a graphical image database management system" in Visual Database Systems (Kunii, T, ed), pp 415-430. Elsevier, Amsterdam.

Ramesh, N & Sethi, I K (1995) "Feature identification as an aid to content-based image retrieval" in Storage and Retrieval for Image and Video Databases III (Niblack, W R & Jain, R C, eds), Proc SPIE 2420, pp 2-11 Rickman, R M & Stonham, T J (1993) "Similarity retrieval from image databases - neural networks can deliver" in Storage and Retrieval for Image and Video Databases I (Niblack, W R & Jain, R C, eds), Proc SPIE 1908, pp 85-94 Scassellati, B et al (1994) "Retrieving images by 2-D shape: a comparison of computation methods with human perceptual judgements" in Storage and Retrieval for Image and Video Databases II (Niblack, W R & Jain, R C, eds), Proc SPIE 2185, pp 2-14 Shields K, Eakins J P & Boardman J M (1995) Automatic image retrieval using shape features presented at 17th BCS IRSG Research Colloquium, Crewe, April 1995 Stricker, M & Dimai, A (1996) "Color indexing with weak spatial constraints" in Storage and Retrieval for Image and Video Databases IV, (Sethi, I K & Jain, R C, eds), Proc SPIE 2670, pp 29-40 Stricker, M & Orengo, M (1995) "Similarity of color images" in Storage and Retrieval for Image and Video Databases III (Niblack, W R & Jain, R C, eds), Proc SPIE 2420, pp 381-392 Swain, M J & Ballard, D H (1991) "Color indexing" International Journal of Computer Vision 7(1), 11-32 Tamura, H et al (1978) "Textural features corresponding to visual perception" IEEE Transactions on Systems, Man and Cybernetics 8(6), 460-472