HIERARCHICAL VISUAL DESCRIPTION SCHEMES FOR STILL IMAGES AND VIDEO SEQUENCES

Similar documents
The ToCAI Description Scheme for Indexing and Retrieval of Multimedia Documents 1

IST MPEG-4 Video Compliant Framework

Overview of the MPEG-7 Standard and of Future Challenges for Visual Information Analysis

Searching Video Collections:Part I

DATA and signal modeling for images and video sequences. Region-Based Representations of Image and Video: Segmentation Tools for Multimedia Services

PROJECTION MODELING SIMPLIFICATION MARKER EXTRACTION DECISION. Image #k Partition #k

An Intelligent System for Archiving and Retrieval of Audiovisual Material Based on the MPEG-7 Description Schemes

Lesson 11. Media Retrieval. Information Retrieval. Image Retrieval. Video Retrieval. Audio Retrieval

Offering Access to Personalized Interactive Video

Multimedia Databases. Wolf-Tilo Balke Younès Ghammad Institut für Informationssysteme Technische Universität Braunschweig

Lesson 6. MPEG Standards. MPEG - Moving Picture Experts Group Standards - MPEG-1 - MPEG-2 - MPEG-4 - MPEG-7 - MPEG-21

Tracking of video objects using a backward projection technique

Spatial Scene Level Shape Error Concealment for Segmented Video

Multimedia Databases. 9 Video Retrieval. 9.1 Hidden Markov Model. 9.1 Hidden Markov Model. 9.1 Evaluation. 9.1 HMM Example 12/18/2009

Video Representation. Video Analysis

Tracking of Human Body using Multiple Predictors

Extraction, Description and Application of Multimedia Using MPEG-7

Overview of the MPEG-7 Standard and of Future Challenges for Visual Information Analysis

Binju Bentex *1, Shandry K. K 2. PG Student, Department of Computer Science, College Of Engineering, Kidangoor, Kottayam, Kerala, India

Chapter 3 Image Registration. Chapter 3 Image Registration

Scalable Hierarchical Summarization of News Using Fidelity in MPEG-7 Description Scheme

Objective Evaluation of Video Segmentation Quality

Chapter 11.3 MPEG-2. MPEG-2: For higher quality video at a bit-rate of more than 4 Mbps Defined seven profiles aimed at different applications:

A Geometrical Key-frame Selection Method exploiting Dominant Motion Estimation in Video

1.2. ζ t Result : Shot. Eye truth : Shot Frame number

INTERACTIVE CONTENT-BASED VIDEO INDEXING AND BROWSING

AIIA shot boundary detection at TRECVID 2006

Using the MPEG-7 Audio-Visual Description Profile for 3D Video Content Description

A Content Based Image Retrieval System Based on Color Features

Solid Modelling. Graphics Systems / Computer Graphics and Interfaces COLLEGE OF ENGINEERING UNIVERSITY OF PORTO

XETA: extensible metadata System

VIDEO SUMMARIZATION USING FUZZY DESCRIPTORS AND A TEMPORAL SEGMENTATION. INPG, 45 Avenue Felix Viallet Grenoble Cedex, FRANCE

The MPEG-7 Description Standard 1

ABSTRACT 1. INTRODUCTION

Multimedia Database Systems. Retrieval by Content

Automatic visual recognition for metro surveillance

70 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 6, NO. 1, FEBRUARY ClassView: Hierarchical Video Shot Classification, Indexing, and Accessing

Content based Image Retrieval Using Multichannel Feature Extraction Techniques

SOFTWARE ENGINEERING Prof.N.L.Sarda Computer Science & Engineering IIT Bombay. Lecture #10 Process Modelling DFD, Function Decomp (Part 2)

Bipartite Graph Partitioning and Content-based Image Clustering

Figure 1: Representation of moving images using layers Once a set of ane models has been found, similar models are grouped based in a mean-square dist

Video Alignment. Literature Survey. Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin

Lecture 3 Image and Video (MPEG) Coding

Georgios Tziritas Computer Science Department

MPEG-4: Overview. Multimedia Naresuan University

A Robust Wipe Detection Algorithm

Image Access and Data Mining: An Approach

Object detection using non-redundant local Binary Patterns

AUTOMATIC VIDEO INDEXING

Browsing News and TAlk Video on a Consumer Electronics Platform Using face Detection

A MPEG-4/7 based Internet Video and Still Image Browsing System

Hand-Eye Calibration from Image Derivatives

USING METADATA TO PROVIDE SCALABLE BROADCAST AND INTERNET CONTENT AND SERVICES

Countermeasure for the Protection of Face Recognition Systems Against Mask Attacks

INFORMATION RETRIEVAL SYSTEM: CONCEPT AND SCOPE

VISUAL content compression and indexing have generally

Software Tools for Recording Image Inspection Processes

CS 534: Computer Vision Segmentation and Perceptual Grouping

Introducing MESSIA: A Methodology of Developing Software Architectures Supporting Implementation Independence

Video Key-Frame Extraction using Entropy value as Global and Local Feature

Video Analysis for Browsing and Printing

TEVI: Text Extraction for Video Indexing

ISO/IEC INTERNATIONAL STANDARD. Information technology Multimedia content description interface Part 5: Multimedia description schemes

Automatic video summarization based on MPEG-7 descriptions

Integrating Low-Level and Semantic Visual Cues for Improved Image-to-Video Experiences

Recall precision graph

Initial Partitions. Region. Region Interpolation. 3. Region parametrization. Ordering. Partition Creation. Interpolated Partitions

Analysis of Image and Video Using Color, Texture and Shape Features for Object Identification

Multimedia structuring using trees

This tutorial has been prepared for computer science graduates to help them understand the basic-to-advanced concepts related to data mining.

8.5 Application Examples

Real-time Monitoring System for TV Commercials Using Video Features

MPEG-4 AUTHORING TOOL FOR THE COMPOSITION OF 3D AUDIOVISUAL SCENES

Task analysis based on observing hands and objects by vision

CS229: Action Recognition in Tennis

Still Image Objective Segmentation Evaluation using Ground Truth

Region-based Segmentation

Textural Features for Image Database Retrieval

Title: Automatic event detection for tennis broadcasting. Author: Javier Enebral González. Director: Francesc Tarrés Ruiz. Date: July 8 th, 2011

signal-to-noise ratio (PSNR), 2

Video Search and Retrieval Overview of MPEG-7 Multimedia Content Description Interface

Video Alignment. Final Report. Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin

Lecture 25 of 41. Spatial Sorting: Binary Space Partitioning Quadtrees & Octrees

Semantic Video Indexing

YouTube (and other online video-sharing sites) The audience and message. In-the-field uses. The techniques. Camera movements Panning: left, right

CHAPTER-23 MINING COMPLEX TYPES OF DATA

Patent Image Retrieval

MODELING AND HIERARCHY

Ray Tracing Acceleration Data Structures

Object Detection in Video Streams

Fast Non-Linear Video Synopsis

Multiple-Choice Questionnaire Group C

Lecture 3: Art Gallery Problems and Polygon Triangulation

Action Classification in Soccer Videos with Long Short-Term Memory Recurrent Neural Networks

Image retrieval based on region shape similarity

Executing Evaluations over Semantic Technologies using the SEALS Platform

Interoperable Content-based Access of Multimedia in Digital Libraries

Story Unit Segmentation with Friendly Acoustic Perception *

EXPLORING ON STEGANOGRAPHY FOR LOW BIT RATE WAVELET BASED CODER IN IMAGE RETRIEVAL SYSTEM

A Hierarchical Keyframe User Interface for Browsing Video over the Internet

Transcription:

HIERARCHICAL VISUAL DESCRIPTION SCHEMES FOR STILL IMAGES AND VIDEO SEQUENCES Universitat Politècnica de Catalunya Barcelona, SPAIN philippe@gps.tsc.upc.es P. Salembier, N. O Connor 2, P. Correia 3 and F. Pereira 3 2 Dublin City University Dublin, IRELAND Noel.OConnor@Teltec.DCU.IE 3 Instituto Superior Técnico Lisboa, PORTUGAL {paulo.correia,fernando.pereira}@lx.it.pt ABSTRACT This paper proposes two s to describe the visual information of an AV document. The first one, is devoted to still images. It describes the image visual appearance and its structure with regions as well as its semantic content in terms of objects. The second is devoted to video sequences. It describes the sequence structure as well as its semantic content in terms of events. Features such as motion, camera activity, etc. are included in this. Moreover, it involves static visual representations such as key-frames, background mosaics and keyregions. These elements are considered as still images and are described by the first.. INTRODUCTION Describing Audio-Visual (AV) content is currently becoming an active field of research []. This activity is motivated by the increasing availability of AV documents as well as standardization efforts such as MPEG-7 [2]. Using MPEG-7 terminology, an AV document description involves Descriptors (termed Ds: syntax & semantic of a representation entity for a feature) and Description Schemes (termed s: set of Ds and s plus the structure and semantics of their relationships). Most of the work on AV indexing has focussed on Ds and very few contributions are available for s. This paper proposes two s to describe the visual information of an AV document. The first, called Still Image, is devoted to still images. It describes the visual appearance and structure of the image as well as its semantic content in terms of objects. The visual appearance is described with regions and their associated color and geometrical features. The semantic description is based on objects and defines their type, identity and possible activity. The second, called Video, is devoted to video sequences. It describes the sequence structure as well as its semantic content in terms of events. Features related to sequence properties such as motion, camera activity, etc. are included in this. Moreover, it involves visual representations such as key-frames, background mosaics and key-regions. These elements are considered as still images and are described by the Still Image. The proposed s are not intended to address audio & production features. 2. TOOLS FOR DOCUMENT DESCRIPTION The proposed s are inspired from the classical way of describing books. Let us briefly comment on the two basic tools used in this context: the Table of Contents and the Index. 2.. The Table of Contents The Table of Contents is a hierarchical representation that splits the document into elementary pieces (chapters, sections, subsections, etc). The order in which the items are presented follows the -D structure of the book. Although the section titles may carry semantic information, the main goal of the Table of Contents is not to describe the content itself but to define the document structure. The corresponding s are the page numbers. The role and structure of the Table of Contents are described in the left side of Fig.. 2.2 The Index The goal of the Index is not to define the book structure, but to define a set of potentially interesting items and to provide references to the sections where these items are discussed. The items have been selected because of their semantic value. Obviously, a given item may appear in several sections of the book. This situation is described by multiple references as shown in the right side of Fig. (these multiple references are not present in the case of a Table of Contents). In many cases, the index is also presented in a hierarchical fashion to allow fast access to the item of interest. Finally, the references involved in the Index are generally defined in terms of page number. However, if the Table of Contents has a fine granularity, the index could directly refer to subsections of the Table of contents. Table of Contents Chapter section. section.2 Chapter 2 section 2. section 2.2 subsection 2.2. subsection 2.2.2 Chapter 3 Book Index Topic item item 2 Topic 2 sub-topic item item 2 sub-topic 2 Figure : Description of a book content by Table of Contents & Index As can be seen, the classical approach for book content description relies on a dual strategy: ) define the document structure (the Table of Contents) and 2) define the locations where items carrying a specific semantic meaning appear (the Index). In the following, this dual strategy will be used for both the Still Image and the Video s. 3. STILL IMAGE To describe images, an approach similar to that of section II may be used. It involves two hierarchical structures represented by trees: Region and s: 3.. The The plays the role of the Table of Contents. Its goal is to describe the spatial organization of the image. Note that the

structure of the document is now 2-D and not only -D as for books. The nodes of the tree represent connected components of the space called regions. We use the term region, and not object, because regions themselves need not have a clear semantic meaning (e.g. region R 9 in Fig. 2). The structure of the defines the inclusion relationship between elementary items as in the Table of Contents. Instead of having a section, which can be further decomposed into subsections, we have now regions that can be further decomposed into sub-regions. Note that the tree describes the entire image (similarly, the Table of Contents describes the entire book without leaving any holes). R8 R5 Tree of regions R4 R9 R2 R2 R6 R R3 R0 R3 R7 R Tree of objects Figure 2: Example of Region and. A very simple example is illustrated in the left side of Fig. 2. The entire original Akiyo image can be seen as the root of the tree. The tree defines how this global region can be subdivided into elementary regions. Finally, the s attached to the Region Tree describe visual properties of the regions: color, spatial and geometrical characteristics. The could be an arbitrary tree, that is a tree where each node can have an arbitrary number of children. However, in the sequel, we will only use the inclusion relationship between regions. This relation can be represented by a simple binary tree. As a result, the tree will describe how a region can be decomposed into two components. It is known as a Binary Partition Tree [3]. The binary partition tree can be viewed as a region-based multiscale representation of the visual content. Large regions are represented on higher levels of the tree and fine details can be obtained from lower levels. The involves s related to the signal properties. It may be created automatically or at least semiautomatically. In particular, regions appearing in the lower levels of the tree should be defined by their homogeneity in terms of signal properties (color for example). 3.2. The The is the image Index. It is composed of a list of objects that were judged as being of potential interest during the indexing. This tree is composed of objects with a semantic meaning. The corresponding s define the type, identity and activity of objects. Moreover, as an Index, an important functionality of this tree is to relate its objects to regions of the. Note that the refers to the and not to the original image itself. This choice has been made because it is assumed that the granularity of the is sufficiently fine. An example of an is shown in the right side of Fig. 2. The hierarchical relation is of the type is-made-of. For example, the object Body is made of sub-entities called Hair, Face and Torso. Each object in the has to refer to one or several regions in the. For example in Fig. 2, the object Body refers to region R 3. The object TV refer to regions R 4 and R 8 since two TV screens are visible in the image. Conversely, one region may be referred to by several objects. This is the case for regions involving various semantic meanings. For example in Fig. 2, region R 0 is referred to by both the Jacket and the Torso objects. Finally, note that all objects should refer to at least one region but not all regions have to be referred by an object. In this case, we have an unidentified region, that is a region described only by its visual appearance (e.g., R 9 in Fig. 2). The s involved in the have semantic meaning. Even if some of these s may be instantiated automatically (presence of face for example), the structure of the tree itself as well as a large number of s imply (at least today) human assistance. 3.3. Global view of the Still Image The global view of the Still Image is represented in Fig. 3. In this diagram, a is represented by a gray rectangle and a by a white rectangle. The symbol ( ) represents the aggregation of elements of different (same) types. The numbers close to the rectangles indicate the cardinality with which the respective elements may be present in the. 0,2 0, Color Region Geometry 0, Type Still Image Visual representation * 0, Object * Type Reference to regions 0, 0, Identity Object activity,n Elementary object activity Figure 3: Global view of the. White (gray) rectangles represent s (description schemes) At the higher level, the is decomposed in a visual representation and two sub-s representing the Region and the s respectively. The visual representation is a used for browsing (an icon image, for example). The Region represents the. The first node represents the region of support of the entire image which can be of arbitrary shape as in the case of MPEG-4 VOP. It is described by a color, a geometry and two (sub-)region s. The geometry deals with features such as position, size, orientation and shape. The two region s describe how the image can be split into two components (the is binary). If the region is a leaf node of the tree, it is only described by color and geometry s. The tree structure introduces a

notion of scalability in the description itself. A given region in the can be described by its color and geometry s. This gives a first description level. If necessary, this first description can be improved by accessing the s of the children or even the s of the leaves of the sub-tree starting from the region of interest. The Object represents the. It follows the same hierarchical approach as the Region except that the tree is not binary. Each object is described by a Type and an arbitrary number of references to regions and of (sub-)object s. The Type provides semantic information about the object: type, identity and activity. The exact definition of these s may involve various thesauruses. In this context, at least three kind of thesaurus seem to be interesting:. Thesaurus of types: Typical examples include face, body, specific props, etc. 2. Thesaurus of identity: For some object types, the description can go into more details by defining the identity of the object. A typical example consists of being able to associate an instance of the object type face with a specific person. 3. Thesaurus of activity: Depending on the object type, the activity is an important. The term activity should be understood in a broad sense. Static (such as standing, sitting, etc) as well as dynamic activities (such as entering, taking something, etc) may be included. 4. VIDEO The Video also relies on the dual strategy discussed in section II. It involves two trees. The first one is devoted to the description of the video structure and is called the Sequence Tree whereas the second one, termed the Event Tree, describes what is happening. The similarity between these trees and the Table of contents and the Index used to describe a book is very high since the documents they describe are both -D in essence. 4.. The Sequence Tree The Sequence Tree is the Table of Contents of the video sequence [4]. It defines the structure of the sequence and its visual properties. Each node of the tree represents a connected component in time called sequence. The tree describes how each sequence can be divided in shorter (sub-)sequences. The leaves of the tree are assumed to be shots (or at least part of shots). In particular, they do not involve any editing effect. A very simple example is illustrated in the left side of Fig. 4. The time line of the video sequence is segmented in a hierarchical fashion. Each tree node points to a time segment of the video. The Sequence Tree involves s related to the visual properties of the video signal. The associated to nonleaf nodes of the Sequence Tree (sequences that are not shots) is simply a time reference indicating the beginning and the end of the corresponding sequence. The leaves (the shots) are more precisely described. Their description involves a time reference, a classification of the type of transition (e.g. cut, fade, wipe), a characterization of the camera activity and several visual representations: key-frames, a background mosaic and keyregions. The Sequence Tree can be constructed using tools such as shot transition detection algorithms and shot clustering algorithms. A large part of the process can be automated. Of course, user interaction may be necessary to correct mistakes. 4.2. The Event Tree The Event Tree is the Index of the Video sequence. It defines in a hierarchical fashion a set of events and sub-events and characterizes and relates them with sequences belonging to the Sequence Tree. A simple example is illustrated in the right side of Fig. 4. If we assume for instance that the video sequence corresponds to the beginning of a TV news program, the main events can be classified in categories such as: credits, studio scenes, speaker views, news reports (politic, economy, social life, sport), etc. Each element of this tree points to the occurrences of the corresponding shots or sequences in the Sequence Tree. The Event Tree concentrates the s related to the semantic in the video. Some of these s can be predefined in a thesaurus of events and simply consists of an index defining the event type. Annotation is another way to characterize the semantic value. Finally, to relate events with temporal video segments, a called reference to sequence is assigned to each event. The construction of the Event Tree may very likely rely on supervised techniques and human interaction. Time Axis Segment Sub-segment Sub-segment 2 Sub-segment 3 Sub-segment 4 segment 2 Segment 3 Segment 4 Segment 5 Segment 6 Segment 7 Segment Tree Shot Shot2 Shot3 Event Tree Introduction Summary Program logo Studio Overview News Presenter Figure 4: example of Sequence and Event Trees. News Items International Clinton Case Pope in Cuba National Twins Sports Closing 4.3. Global view of the Video The global view of the Video is presented in Fig. 5. The higher level of the Video is decomposed into three major elements: The Sequence, the Event and the Visual summary. The visual summary is a short and sub-sampled version of the sequence used for browsing. Both the Sequence and the Event are arbitrary trees. A sequence is described by two time references (start & end), by an arbitrary number of (sub-) sequences and by shots. The shot itself is a rather complex and will be more precisely described later on. An event is described by an arbitrary number of (sub-)events and three s: an index defining the type of event, an annotation and references to sequences. As in the case of the Still Image, the reference plays a major role in relating the semantic information contained in the Event Tree with the signal-based characterization of the video contained in the Sequence Tree.

Video Sequence Tree 0, * Sequence Visual summary * Event 2 * * Shot Reference to sequence Key-frame Mosaic Annotation Shot transitions Event Tree Event type 2 * 0, * Warping parameters 0, * Background mosaic 2 Key-regions 0, 0, * Key-region instance Figure 5: Global view of the video Camera motion Elementary cam. motion Let us recall that a shot is a continuous set of frames without editing effects. The goal of the Shot is to characterize the signal properties of this set of frames. The set of s includes two time references (start & end), a characterization of the transition effects at both sides of the shots, a devoted to the camera motion (which is assumed to be continuous) and some visual information. The camera motion can be characterized by parameters such as the camera translation, rotation, focal distance, etc and by typical activity classes such as static, pan, zoom, tilt, etc. As the camera activity may not be constant during the entire shot, homogeneous activities may be segmented in time and are represented by the Elementary camera motion. The visual information can be characterized, as a first example, by a set of key-frames. Each key-frame is a still image extracted from the shot. As a result, the key-frame is composed of the Still Image presented in section III plus a time reference. If the visual content remains stable during the shot, one (or a few) keyframe(s) is(are) sufficient to characterize the entire shot. In the presence of significant changes due to camera motion, an alternative representation consists of a background mosaic [5] plus a set of foreground regions (called key-regions ). In practice, key-regions have been identified as not belonging to the background because they have a different motion or are not in the same depth plane. This information is gathered in the mosaic which is decomposed into one background mosaic and several Key-region s. The background mosaic is also a static representation that is conveniently described by the Still Image. This description can be enhanced by making available the set of warping parameters used to create the mosaic. The key-regions is decomposed into several key-region instances. This decomposition is included here in order to be able to represent various visual instances of the same region within the shot. Typical examples include various instances of regions representing a face or a human body. Finally, each key-region instance is described by a Still Image and a time reference. The global view of the Video is summarized in Fig. 5. Finally, note that even if the Still Image is primarily used to characterize static information such as key-frames, or key-objects, its description accuracy can be improved by describing the time evolution of its composing regions. 5. SUPPORTED SEARCH FUNCTIONALITIES These support a large number of search functionalities related to the visual properties of the image or the sequences. In the sequel, the most important type of queries are discussed. 5.. Semantic queries The Object and Event Trees concentrate the semantic information extracted during the indexing process. The semantic value of possible object types, identities, activities and event types is assumed to be precisely defined in thesauruses. As a result, if the corresponding semantic notion has been recognized during the indexing process, the description is able to answer to semantic queries. The search process is straightforward, since the query has simply to be translated into a set of thesaurus indexes that are matched with the indexes involved in the description. Furthermore, part of the semantic description is also included as free text in the Annotation s. In this case, classical textbased search and retrieval is assumed to be used. 5.2. Signal & Similarity-based queries The Region and Event Tree define the visual appearance and the structure of the signal. As a result, this part of the description supports signal & similarity-based queries. Signal-based queries refer to all queries related to region color or geometry, to time references, to shot transitions, to camera activity. Typical examples include finding regions with a specific color, size, orientation, shape characteristics or sequence with a specific camera motion or any combination of these properties. 5.3. Combined semantic and signal-based queries A typical example of collaboration between the semantic and the signal-based description deals with queries referring to composition of objects or sequences. For instance, it is not realistic to describe in the all possible spatial relations between objects. Queries of the type: find images where object A is placed just below object B which is in turn on the right side of object C have to be solved by a search involving both the Object and the. Indeed, the describes all possible spatial relationships between objects that refer to regions. 5.4. Flexibility to deal with unexpected queries The main interest of the Object and Event Trees is their efficiency to deal with high level queries. However, they have limited flexibility. Indeed, during the indexing process, one has defined all the objects and events of interest and their corresponding s. In practice, the user has made a selection of the objects and events he/she thinks might be of interest for the retrieval phase. In the example of Fig. 2, the user has assumed that the most important objects are instances of body, face, hair, torso, TV and jacket. Objects such as microphone, eye, mouth, shirt were discarded and no information about them is available in the. What can be done if the query is unexpected, that is if it addresses an object, an event or a property that has not been considered as being relevant during the indexing process? A partial answer is given by the Region and Segment

Trees and their relation with the Object and Event Trees. Let us consider two examples dealing with still images: ) Unexpected query involving an unknown property of a known object: Considering the example of Fig. 6, assume the following query: find images of a woman with dark hair. The search process may involve 3 steps and would naturally start with the in order to select images with a woman:. Analyze the s to use as much as possible the semantic s and efficiently pre-select candidates. Find images with a of type woman. 2. Look for the hair that is related to the woman. At this stage, we assume that the hair has been considered as an object of interest during the indexing process. 3. Since no information about the hair color is available in the (this object property has not been considered as being of interest during the indexing process), the search has to look for information in the. The region describing the visual properties of the hair object is found by following the reference from the hair object to the. The color of the region (here R 6 ) gives the final answer. 4: Analyze the color of the referenced region R 2 R 4 R R 3 R 5 R 6 R 7 R 8 R 9 R 0 R R 2 R 3 2: Find woman 3: Find the related hair : Analyze Object tree Figure 6: example of the search process for a query involving semantic and signal properties: find images of a woman with dark hair. 2) Unexpected query involving an unknown object: Consider the example of Fig. 7 and assume that the query is: find images of a woman speaker with a microphone. The problem with this query is that the object microphone has not been considered as being of interest during the indexing process. In this case, two solutions are possible: - either to say that the second part of the query cannot be processed and that only the first part will be dealt with, or - to open the door to intelligent search engines that could try to figure out whether the unknown object is present or not (Note that the following example implies a fair amount of intelligence in the search engine side. The goal is not to claim that this solution is easily implementable with today's technology, but to show the flexibility and how it could cope with future progresses. In particular, this solution allows using more intelligent search engines as technology evolves). Let us describe a possible search scenario involving 4 steps as described by Fig. 7:. Analyze the s to use as much as possible the semantic s and pre-select candidates. Find images with a of type woman with activity speaking. 2. Look for one (or several) sub-component(s) of the body where the microphone might be located, for example the Torso. 3. Go to the following the reference from the Torso object. Analyze the corresponding region R 0 and its sub-regions. 4. The search engine finds region R 3 that is small, located in the center of region R 0 and that is dark. Taking into account all these pieces of information, the search engine may come to the conclusion that the probability of R 3 being a microphone is high. As a result, the image matches quite well the query. R R 2 R 3 4: Analyze the subtree corresponding to R 4 the torso region R 6 R 7 R 5 2: Find woman with the corresponding speaking activity R 9 R 0 R 8 R R R 3 5: A dark region of small 2 size has been found on the torso region. It might be a microphone. : Analyze Object tree 3: Find a body part where the microphone might be Figure7: example of search process for a query involving an unknown object: find images of a woman speaker with a microphone 6. CONCLUSIONS This paper proposes two s to describe the visual information of an AV document. The Still Image describes the visual appearance and structure of an image as well as its semantic content in terms of objects. The Video describes the sequence structure as well as its semantic content in terms of events. Features related to sequence properties such as motion, camera activity, etc. are included in this. ACKNOWLEDGEMENT This work was carried out within the ACTS project DICEMAN in collaboration with France-Telecom CCETT. The work was partfunded by the European Commission. The views expressed are those of the authors and should not be taken to represent the views of the European Commission or its services. REFERENCES [] Y. Rui, T. Huang, S.F. Chang, Image retrieval: Past, Present and Future, accepted for publication in Jour. of Visual Communication and Image Representation, 999 [2] MPEG-7: context and objectives, Tech. report ISO/IEC JTC/SC29/WG/w2460, Oct. 98, Atlantic City. [3] P. Salembier, L. Garrido, Binary partition tree as an efficient representation for filtering, segmentation and information retrieval, IEEE ICIP'98, Chicago (IL), USA, Oct. 4-7, 98. [4] Y. Rui, T. Huang, S. Mehrotra, Exploring video structure beyond the shots, Proc. of IEEE Int. Conf. on Multimedia Computing and Systems (ICMCS), pp237-240, June 28-July, 998, Austin, Texas USA [5] H. Sawhney, S. Ayer. Compact representations of videos through dominant and multiple motion estimation. IEEE Trans. on Pattern Analysis and Machine Intelligence, 8:84-830, August 996.