CIMWOS: A MULTIMEDIA ARCHIVING AND INDEXING SYSTEM

Size: px
Start display at page:

Download "CIMWOS: A MULTIMEDIA ARCHIVING AND INDEXING SYSTEM"

Transcription

1 CIMWOS: A MULTIMEDIA ARCHIVING AND INDEXING SYSTEM Nick Hatzigeorgiu, Nikolaos Sidiropoulos and Harris Papageorgiu Institute for Language and Speech Processing Epidavrou & Artemidos 6, Maroussi, Greece nikos@xanthi.ilsp.gr, nsidir@xanthi.ilsp.gr, xaris@ilsp.gr Abstract This paper describes a multimedia, multilingual and multimodal research system called CIMWOS (Combined IMage and WOrd Spotting). CIMWOS incorporates an extensive set of multimedia technologies, integrating three major subsystems (text, speech, and image processing). It produces a rich collection of XML metadata annotations following the MPEG-7 standard. These XML annotations are further merged and loaded into the CIMWOS Multimedia Database. An ergonomic and user-friendly Web-based interface allows the user to efficiently retrieve video segments by a combination of media description, content metadata and natural language text. Currently the system is under evaluation and contains broadcast news and documentaries in three languages (English, Greek, and French), while the open architecture allows for more languages to be incorporated in the future. Key Words: Multimedia Databases, Electronic Libraries, User Interfaces, Applications recognition, video text detection and character recognition. All outputs converge to a textual XML metadata annotation scheme following the MPEG-7 standard. These XML annotations are further merged and loaded into the CIMWOS Multimedia Database. Additionally, they can be dynamically transformed for interchanging semantic-based information into RDF and Topic Maps documents via XSL stylesheets. The CIMWOS retrieval engine is based on a weighted boolean model with intelligent indexing components. An ergonomic and user-friendly Web-based interface allows the user to efficiently retrieve video segments by a combination of media description, content metadata and natural language text. The database is a large collection of broadcast news and documentaries in three languages (English, Greek, and French), while the open architecture allows for more languages to be incorporated in the future. The system is currently under evaluation. 1. Introduction There are several projects aiming at developing advanced technologies and systems to tackle the problems encountered in multimedia archiving and indexing [1-3]. CIMWOS (Combined IMage and WOrd Spotting) [4] incorporates an extensive set of multimedia technologies, combining the areas of expertise of eight participating institutions with the goal of creating an integrated system that can retrieve information from video sources. CIMWOS supports content-based indexing, archiving, retrieval and on-demand delivery of audiovisual content. The system uses state-of-the-art algorithms for text, speech and image processing. The audio processing operations employ robust continuous speech recognition, speech/non-speech classification, speaker clustering and speaker identification. Text processing tools operate on the text stream produced by the speech recognizer and perform named entity detection, term recognition, topic detection, and story segmentation. Image processing includes video segmentation and key frame extraction, face detection and face identification, object and scene Figure 1: CIMWOS modules 2. System modules Each video is processed by the various CIMWOS information extraction modules (Fig. 1). These modules can be grouped in three main categories that comprise the speech, text and image processing subsystems

2 2.1. Speech Processing Subsystem Broadcast news exhibit a wide variety of audio characteristics, including clean speech, telephone speech, conference speech, music, and speech corrupted by music or noise. Transcribing the audio, i.e., producing a (raw) transcript of what is being said, determining who is speaking when, what topic a segment is about, or which organizations are mentioned, are all challenging problems. Adverse background conditions can lead to significant degradation in performance. Consequently, adaptation to the varied acoustic properties of the signal or to a particular speaker, and enhancements to the segmentation process, are generally acknowledged to be key areas for research and improvement to render indexing systems usable. This is reflected in the effort dedicated to advance the current state-of-the-art in these areas. In CIMWOS, these tasks are handled by the speech processing subsystem [5], which comprises speaker change detection (SCD), automatic speech recognition (ASR), speaker identification (SID) and speaker clustering (SC). The ASR engine is a real-time, large vocabulary, speaker-independent, gender-independent, continuous speech recognizer trained in a wide variety of noise conditions encountered in the broadcast news domain Text Processing Subsystem Text processing tools, operating on the text stream produced by the speech processing subsystem, perform detection, identification, and classification tasks Named Entity Detection The NED module [6] identifies all named locations, persons and organizations, as well as dates, percentages and monetary amounts in video transcriptions. CIMWOS uses a series of basic language technology building blocks, which are modular and combined in a pipeline. An initial finite state preprocessor performs tokenization and sentence boundary identification on the output of the speech recognizer. A part-of-speech tagger trained on a manually annotated corpus and a lexicon-based lemmatizer carry out morphological analysis and lemmatization. A lookup module matches name lists and trigger-words against the text, and, eventually, a finite state parser recognizes named entities on the basis of a pattern grammar. Training the NED module consists in populating the gazetteer lists and semi-automatically extracting the pattern rules. A corpus of words per language already tagged with the NED classes was used to guide system training and development Term Recognition The term recognizer identifies possible single or multiword terms, using both linguistic and statistical modeling. Linguistic processing is based on an augmented term grammar, the results of which are statistically filtered using frequency-based scores Story Detection & Topic Classification Story detection (SD) and topic classification (TC) use a set of models trained on an annotated corpus of stories and their associated topics. The basis of SD and TC is a generative, mixture-based HMM including one state per topic and one state modeling general language, i.e., words not specific to any topic. After emitting a single word, the model re-enters the beginning state and the next word is generated. At the end of the story, a final state is reached. In SD, a sliding window of fixed size is used to note the change in topic-specific words, resulting in a set of stable regions in which topics change only slightly or not at all. TC then classifies the text sections according to the set of topic models. The modeled inventory of topics is a flat, Reuters-derived structure, containing a few main categories and several sub-categories, a structure shallow enough to provide reasonable amounts of training data per topic but still fine-grained enough to allow for flexibility and detail in queries. All technologies used are inherently language independent and of a statistical nature Image Processing Subsystem A video sequence consists of many individual images which are called frames and are generally considered as the smallest unit we are concerned with, when segmenting a video. An uninterrupted video stream generated by one camera is called a shot. The image processing subsystem consists of Video segmentation and Key frame extraction, Face Detection and Face Identification spotting persons at any location, scale and orientation, Object Recognition marking an object in a 'recognition view', given knowledge cumulated from a set of previously seen 'learning views' and Video Text recognition detecting and recognizing close captions in images and video. This work has been reported in [7-10]. 3. Integration Architecture All processing modules in the corresponding three modalities (Audio, Image and Text) converge to a textual XML metadata annotation scheme [11] following the MPEG-7 descriptors. These XML metadata annotations are further processed, merged and loaded into the CIMWOS Multimedia Database. The merging Component of CIMWOS is responsible for the amalgamation of the various XML annotations and the creation of one self-contained object that is compliant with the CIMWOS Database. Additionally, the resulting object can be dynamically transformed for interchanging 227

3 semantic-based information into RDF and Topic Maps documents via XSL stylesheets. The CIMWOS Database is a relational database that contains all the information of each one of the XML metadata files and also all the relevant indexes needed to facilitate search and retrieval of information based on user-specified search terms. 4. Retrieval and Presentation Because of its size, a video clip can take a long time to move from one location to another, such as from the digital video library to the user. Likewise, if a library consists of only 30-minute clips, when users check one out, it may take 30 minutes to determine whether the clip meets users needs. To make the retrieval and viewing of information faster, the digital video library will need to support partitioning video into small-sized clips. In CIMWOS, the retrieval engine is based on a weighted Boolean model equipped with intelligent indexing components. The basic retrieval unit is the passage. The passage, which plays the role of a document of a standard retrieval system, is indexed on a set of textual features: words, terms, named entities, speakers and topics. Each passage is linked to one or multiple shots, and each shot is indexed on another set of textual features: faces, objects and video text (caption or scene). Via the linking of passages and shots, each passage is finally assigned a broader set (union) of features that will be used for retrieval. All the features are textual, no image or audio conceptual feature is supported (i.e. motion parameters or similarity between histograms). Each passage is represented as a set of features and retrieval of passages is performed based on computing similarity in the feature space. The results are ranked on the basis of the computed similarity values. The retrieval procedure is the following: first, passages are filtered on the basis of the advanced features in order to reduce the search space of the free query. If no such features have been selected, the search space remains unchanged. Next, a boolean-based matching operation takes place in order to compute the similarity between the query and each passage of the possibly reduced space. Finally, passages are ranked based on this similarity. The result set can be visualized by summarizing the relevant passages. While skimming a passage and its associated metadata, the end user can select a passage and view its associated metadata, read a summary containing the first words of the transcribed speech and a sequence of thumbnails, or play the passage using intelligent streaming. 5. The CIMWOS User Interface The CIMWOS user interface is a Web interface that is based on well-known Web technologies, like HTML and ASP. The user views standard HTML Web pages and does not need to download specific software. The user can browse the CIMWOS database or submit queries. The system processes user-submitted queries using both ASP code and dll modules, formulates an appropriate SQL query, retrieves the matching information from the CIMWOS multimedia database and formulates the results according to user selections. Table 1: Selection steps for the Browse Interface Language Selection Output Format Categories English French Greek All Lang. HTML XML Creators Faces Locations Named Entities Objects Video Text Speakers Terms Topics Figure 2: Time search form Figure 3: Standard search form 228

4 Figure 4: Advanced Search Figure 5: Search results 229

5 The Web interface is accessible through the following address: Some of the video material included in the digital library has copyright restrictions (e.g. for research purposes only), so anonymous access is not yet available; however, a limited-time password can be obtained by the webmaster or the authors of this paper. The interface offers four different modes of access: Browse, Time Search, Standard Search and Advanced Search. The Browse mode permits the user to look trough the videos using three selection steps (Language selection, Output format, Categories) with the choices of table 1. Time search (Fig. 2) allows the user to choose a specific video from the database and the view a particular video segment. Standard search (Fig. 3) is a simplified search interface offering a subset of the advanced search capabilities. Figure 6: Search results in XML Advanced Search (Fig. 4) allows the user to access the full power of the search engine. It permits any Boolean combination of multiple search criteria that include text (extracted by the speech processing subsystem), named entities, topics, speakers, terms, faces, objects, locations and video text. The user can also choose the output format for the results (HTML or XML), whether the ranking engine will be used on the results and the number of thumbnails that will be returned for each result. The results from any kind of search can be viewed as a Web page with a list of video segments that satisfy the search criteria (Fig. 5). In this list the user can view the video title, its duration and its format (currently streaming video in MS WMP format is the only available format), as well as the ranking for each result. However, the user can also request the results to be presented in XML format. In that case, the user receives an XML document (Fig. 6) with a limitation of 10 results per query. Figure 7: Details of a video segment For each of the resulting segments, it is possible to view the full details of the available metadata (Fig. 7) including information for the parent video and the transcription of speech extracted from the segment. Finally, the user can also view the video segment itself together with the transcription (Fig. 8), as well as adjacent video segments. Figure 8: Viewing a video segment 230

6 6. Conclusions CIMWOS has developed an integrated environment for accessing multimedia digital content providing advanced searching and browsing capabilities. Using the results from research efforts on emerging technologies for multimedia data processing, indexing and retrieval, and embedding them in a video library system that covers broadcast news and documentaries in English, French and Greek, has produced a system that can index and retrieve video segments based on a variety of search criteria. The CIMWOS Digital Video Library uses intelligent, automatic mechanisms that provide full-content search and retrieval from a large (scaling to several hours) online digital video library. The tools that have been exploited during the project can automatically populate the library and support access to it. The adopted approach uses combined speech, language and image understanding technology to automatically transcribe, segment and index the video. Retrieval is text-based, performed on the material that results from image analysis, speech recognition and natural language processing of the transcribed text. The information that is produced by the speech, image, and text components is suitable for intelligent indexing and retrieval so that a user can find a video segment of interest within the digital library. The user interface for the resulting indexed multimedia database is a Web-form based interface. It is a user-friendly interface that offers a number of different modes of access to the data contained in the library. The results of a search are video segments, and the user can view them using streaming video, or save the detailed information for each segment in XML format. 7. Acknowledgements This material is based on work supported by cost reimbursement contract number IST for research and technological development projects from the Commission of the European Communities, Directorate- General Information Society. 8. References [1] Wactlar H., Olligschlaeger A., Hauptmann A., Christel M., Complementary Video and Audio Analysis for Broadcast News Archives, Communications of the ACM, 43(2), February 2000, [2] Michael R. Lyu, Edward Yau, Sam Sze., Video and multimedia digital libraries: A multilingual, multimodal digital video library system, Proc. 2nd ACM/IEEE-CS joint conf. On Digital Libraries, July 2002, [3] Sankar A., Gadde R.R. and Weng F., SRI s broadcast news system Toward faster, smaller and better speech recognition, Proc. DARPA Broadcast News Workshop, 1999, [4] EU IST project CIMWOS ( /cimwos/) [5] F. Kubala, J. Davenport, et al., The 1997 BBN BYBLOS System applied to Broadcast News Transcription, Proc. DARPA Broadcast News Transcription and Understanding Workshop, Lansdowne VA, Feb [6] Demiros, I., Boutsis, S., Giouli, V., Liakata, M., Papageorgiou, H., Piperidis, S., Named Entity Recognition in Greek Texts, Proc. 2nd Int. Conf. on LREC 2000, Athens 2000, [7] Fabien Cardinaux and Sebastien Marcel, Face Verification using MLP and SVM, XI Journees NeuroSciences et sciences pour l'ingenieur (NSI 2002), [8] T. Tuytelaars, A. Zaatri, L. Van Gool and H. Van Brussel, Automatic Object Recognition as part of an Integrated Supervisory Control System, Proc. IEEE Conference on Robotics and Automation, ICRA 2000, [9] D. Chen, H. Bourlard and J-Ph. Thiran, Text identification in complex background using SVM, Proc. Int. Conf. on computer vision and pattern recognition, Dec. 2001, [10] J.M. Odobez and D. Chen (2002), Robust Video Text Segmentation and Recognition with Multiple Hypotheses, Proc. ICIP, Sep [11] Papageorgiou H., Prokopidis P., Demiros I., Giouli G., Constantinidis A., and Piperidis S. Multi-level XMLbased Corpus Annotation. Proc. 3rd LREC Conference, Las Palmas 2002,

CIMWOS: A MULTIMEDIA, MULTIMODAL AND MULTILINGUAL INDEXING AND RETRIEVAL SYSTEM

CIMWOS: A MULTIMEDIA, MULTIMODAL AND MULTILINGUAL INDEXING AND RETRIEVAL SYSTEM Ebroul Izquierdo, editor. Digital Media Processing for Multimedia Interactive Services. Proceedings of the 4 th European Workshop on Image Analysis for Multimedia Interactive Services. World Scientific

More information

clarin:el an infrastructure for documenting, sharing and processing language data

clarin:el an infrastructure for documenting, sharing and processing language data clarin:el an infrastructure for documenting, sharing and processing language data Stelios Piperidis, Penny Labropoulou, Maria Gavrilidou (Athena RC / ILSP) the problem 19/9/2015 ICGL12, FU-Berlin 2 use

More information

Semantic Video Indexing

Semantic Video Indexing Semantic Video Indexing T-61.6030 Multimedia Retrieval Stevan Keraudy stevan.keraudy@tkk.fi Helsinki University of Technology March 14, 2008 What is it? Query by keyword or tag is common Semantic Video

More information

Multi-level XML-based Corpus Annotation

Multi-level XML-based Corpus Annotation Multi-level XML-based Corpus Annotation Harris Papageorgiou, Prokopis Prokopidis, Voula Giouli, Iason Demiros, Alexis Konstantinidis, Stelios Piperidis Institute for Language and Speech Processing Epidavrou

More information

Lecture Video Indexing and Retrieval Using Topic Keywords

Lecture Video Indexing and Retrieval Using Topic Keywords Lecture Video Indexing and Retrieval Using Topic Keywords B. J. Sandesh, Saurabha Jirgi, S. Vidya, Prakash Eljer, Gowri Srinivasa International Science Index, Computer and Information Engineering waset.org/publication/10007915

More information

Department of Computer Science & Engineering. The Chinese University of Hong Kong Final Year Project LYU0102

Department of Computer Science & Engineering. The Chinese University of Hong Kong Final Year Project LYU0102 Department of Computer Science & Engineering The Chinese University of Hong Kong LYU0102 Supervised by Prof. LYU, Rung Tsong Michael Group Members: Chan Pik Wah Ngai Cheuk Han Prepared by Chan Pik Wah

More information

CHAPTER 8 Multimedia Information Retrieval

CHAPTER 8 Multimedia Information Retrieval CHAPTER 8 Multimedia Information Retrieval Introduction Text has been the predominant medium for the communication of information. With the availability of better computing capabilities such as availability

More information

Digital Newsletter. Editorial. Second Review Meeting in Brussels

Digital Newsletter. Editorial. Second Review Meeting in Brussels Editorial The aim of this newsletter is to inform scientists, industry as well as older people in general about the achievements reached within the HERMES project. The newsletter appears approximately

More information

Spoken Document Retrieval (SDR) for Broadcast News in Indian Languages

Spoken Document Retrieval (SDR) for Broadcast News in Indian Languages Spoken Document Retrieval (SDR) for Broadcast News in Indian Languages Chirag Shah Dept. of CSE IIT Madras Chennai - 600036 Tamilnadu, India. chirag@speech.iitm.ernet.in A. Nayeemulla Khan Dept. of CSE

More information

From Multimedia Retrieval to Knowledge Management. Pedro J. Moreno JM Van Thong Beth Logan

From Multimedia Retrieval to Knowledge Management. Pedro J. Moreno JM Van Thong Beth Logan From Multimedia Retrieval to Knowledge Management Pedro J. Moreno JM Van Thong Beth Logan CRL 2002/02 March 2002 From Multimedia Retrieval to Knowledge Management Pedro J. Moreno JM Van Thong Beth Logan

More information

Multimedia Indexing and Retrieval Using Natural Language, Speech and Image Processing Methods UNCORRECTED PROOF

Multimedia Indexing and Retrieval Using Natural Language, Speech and Image Processing Methods UNCORRECTED PROOF JWBK0- JWBK0-Stamou January, 0 : Char Count= 0 Multimedia Indexing and Retrieval Using Natural Language, Speech and Image Processing Methods Harris Papageorgiou, Prokopis Prokopidis, Athanassios Protopapas

More information

Lecture 12: Video Representation, Summarisation, and Query

Lecture 12: Video Representation, Summarisation, and Query Lecture 12: Video Representation, Summarisation, and Query Dr Jing Chen NICTA & CSE UNSW CS9519 Multimedia Systems S2 2006 jchen@cse.unsw.edu.au Last week Structure of video Frame Shot Scene Story Why

More information

The Stanford/Technicolor/Fraunhofer HHI Video Semantic Indexing System

The Stanford/Technicolor/Fraunhofer HHI Video Semantic Indexing System The Stanford/Technicolor/Fraunhofer HHI Video Semantic Indexing System Our first participation on the TRECVID workshop A. F. de Araujo 1, F. Silveira 2, H. Lakshman 3, J. Zepeda 2, A. Sheth 2, P. Perez

More information

Real-time large-scale analysis of audiovisual data

Real-time large-scale analysis of audiovisual data Finnish Center of Excellence in Computational Inference Real-time large-scale analysis of audiovisual data Department of Signal Processing and Acoustics Aalto University School of Electrical Engineering

More information

TIPSTER Text Phase II Architecture Requirements

TIPSTER Text Phase II Architecture Requirements 1.0 INTRODUCTION TIPSTER Text Phase II Architecture Requirements 1.1 Requirements Traceability Version 2.0p 3 June 1996 Architecture Commitee tipster @ tipster.org The requirements herein are derived from

More information

Enterprise Multimedia Integration and Search

Enterprise Multimedia Integration and Search Enterprise Multimedia Integration and Search José-Manuel López-Cobo 1 and Katharina Siorpaes 1,2 1 playence, Austria, 2 STI Innsbruck, University of Innsbruck, Austria {ozelin.lopez, katharina.siorpaes}@playence.com

More information

Information Retrieval

Information Retrieval Multimedia Computing: Algorithms, Systems, and Applications: Information Retrieval and Search Engine By Dr. Yu Cao Department of Computer Science The University of Massachusetts Lowell Lowell, MA 01854,

More information

The ToCAI Description Scheme for Indexing and Retrieval of Multimedia Documents 1

The ToCAI Description Scheme for Indexing and Retrieval of Multimedia Documents 1 The ToCAI Description Scheme for Indexing and Retrieval of Multimedia Documents 1 N. Adami, A. Bugatti, A. Corghi, R. Leonardi, P. Migliorati, Lorenzo A. Rossi, C. Saraceno 2 Department of Electronics

More information

Offering Access to Personalized Interactive Video

Offering Access to Personalized Interactive Video Offering Access to Personalized Interactive Video 1 Offering Access to Personalized Interactive Video Giorgos Andreou, Phivos Mylonas, Manolis Wallace and Stefanos Kollias Image, Video and Multimedia Systems

More information

3 Publishing Technique

3 Publishing Technique Publishing Tool 32 3 Publishing Technique As discussed in Chapter 2, annotations can be extracted from audio, text, and visual features. The extraction of text features from the audio layer is the approach

More information

Hands On: Multimedia Methods for Large Scale Video Analysis (Lecture) Dr. Gerald Friedland,

Hands On: Multimedia Methods for Large Scale Video Analysis (Lecture) Dr. Gerald Friedland, Hands On: Multimedia Methods for Large Scale Video Analysis (Lecture) Dr. Gerald Friedland, fractor@icsi.berkeley.edu 1 Today Recap: Some more Machine Learning Multimedia Systems An example Multimedia

More information

Summary of Bird and Simons Best Practices

Summary of Bird and Simons Best Practices Summary of Bird and Simons Best Practices 6.1. CONTENT (1) COVERAGE Coverage addresses the comprehensiveness of the language documentation and the comprehensiveness of one s documentation of one s methodology.

More information

TRENTINOMEDIA: Exploiting NLP and Background Knowledge to Browse a Large Multimedia News Store

TRENTINOMEDIA: Exploiting NLP and Background Knowledge to Browse a Large Multimedia News Store TRENTINOMEDIA: Exploiting NLP and Background Knowledge to Browse a Large Multimedia News Store Roldano Cattoni 1, Francesco Corcoglioniti 1,2, Christian Girardi 1, Bernardo Magnini 1, Luciano Serafini

More information

Adaptive Multimedia Messaging based on MPEG-7 The M 3 -Box

Adaptive Multimedia Messaging based on MPEG-7 The M 3 -Box Adaptive Multimedia Messaging based on MPEG-7 The M 3 -Box Abstract Jörg Heuer José Luis Casas André Kaup {Joerg.Heuer, Jose.Casas, Andre.Kaup}@mchp.siemens.de Siemens Corporate Technology, Information

More information

Searching Video Collections:Part I

Searching Video Collections:Part I Searching Video Collections:Part I Introduction to Multimedia Information Retrieval Multimedia Representation Visual Features (Still Images and Image Sequences) Color Texture Shape Edges Objects, Motion

More information

Joint UNECE/Eurostat/OECD work session on statistical metadata (METIS) (Geneva, 3-5 April 2006)

Joint UNECE/Eurostat/OECD work session on statistical metadata (METIS) (Geneva, 3-5 April 2006) WP. 20 ENGLISH ONLY UNITED NATIONS STATISTICAL COMMISSION and ECONOMIC COMMISSION FOR EUROPE CONFERENCE OF EUROPEAN STATISTICIANS EUROPEAN COMMISSION STATISTICAL OFFICE OF THE EUROPEAN COMMUNITIES (EUROSTAT)

More information

QMUL-ACTIVA: Person Runs detection for the TRECVID Surveillance Event Detection task

QMUL-ACTIVA: Person Runs detection for the TRECVID Surveillance Event Detection task QMUL-ACTIVA: Person Runs detection for the TRECVID Surveillance Event Detection task Fahad Daniyal and Andrea Cavallaro Queen Mary University of London Mile End Road, London E1 4NS (United Kingdom) {fahad.daniyal,andrea.cavallaro}@eecs.qmul.ac.uk

More information

Speech Recognition Systems for Automatic Transcription, Voice Command & Dialog applications. Frédéric Beaugendre

Speech Recognition Systems for Automatic Transcription, Voice Command & Dialog applications. Frédéric Beaugendre Speech Recognition Systems for Automatic Transcription, Voice Command & Dialog applications Frédéric Beaugendre www.seekiotech.com SeekioTech Start-up hosted at the multimedia incubator «la Belle de Mai»,

More information

Video Summarization Using MPEG-7 Motion Activity and Audio Descriptors

Video Summarization Using MPEG-7 Motion Activity and Audio Descriptors Video Summarization Using MPEG-7 Motion Activity and Audio Descriptors Ajay Divakaran, Kadir A. Peker, Regunathan Radhakrishnan, Ziyou Xiong and Romain Cabasson Presented by Giulia Fanti 1 Overview Motivation

More information

STS Infrastructural considerations. Christian Chiarcos

STS Infrastructural considerations. Christian Chiarcos STS Infrastructural considerations Christian Chiarcos chiarcos@uni-potsdam.de Infrastructure Requirements Candidates standoff-based architecture (Stede et al. 2006, 2010) UiMA (Ferrucci and Lally 2004)

More information

Search Engines. Information Retrieval in Practice

Search Engines. Information Retrieval in Practice Search Engines Information Retrieval in Practice All slides Addison Wesley, 2008 Beyond Bag of Words Bag of Words a document is considered to be an unordered collection of words with no relationships Extending

More information

A Hierarchical Keyframe User Interface for Browsing Video over the Internet

A Hierarchical Keyframe User Interface for Browsing Video over the Internet A Hierarchical Keyframe User Interface for Browsing Video over the Internet Maël Guillemot, Pierre Wellner, Daniel Gatica-Pérez & Jean-Marc Odobez IDIAP, Rue du Simplon 4, Martigny, Switzerland {guillemo,

More information

Lesson 11. Media Retrieval. Information Retrieval. Image Retrieval. Video Retrieval. Audio Retrieval

Lesson 11. Media Retrieval. Information Retrieval. Image Retrieval. Video Retrieval. Audio Retrieval Lesson 11 Media Retrieval Information Retrieval Image Retrieval Video Retrieval Audio Retrieval Information Retrieval Retrieval = Query + Search Informational Retrieval: Get required information from database/web

More information

Available online at ScienceDirect. Procedia Computer Science 87 (2016 ) 12 17

Available online at  ScienceDirect. Procedia Computer Science 87 (2016 ) 12 17 Available online at www.sciencedirect.com ScienceDirect Procedia Computer Science 87 (2016 ) 12 17 4th International Conference on Recent Trends in Computer Science & Engineering Segment Based Indexing

More information

Optimal Video Adaptation and Skimming Using a Utility-Based Framework

Optimal Video Adaptation and Skimming Using a Utility-Based Framework Optimal Video Adaptation and Skimming Using a Utility-Based Framework Shih-Fu Chang Digital Video and Multimedia Lab ADVENT University-Industry Consortium Columbia University Sept. 9th 2002 http://www.ee.columbia.edu/dvmm

More information

An Intelligent System for Archiving and Retrieval of Audiovisual Material Based on the MPEG-7 Description Schemes

An Intelligent System for Archiving and Retrieval of Audiovisual Material Based on the MPEG-7 Description Schemes An Intelligent System for Archiving and Retrieval of Audiovisual Material Based on the MPEG-7 Description Schemes GIORGOS AKRIVAS, SPIROS IOANNOU, ELIAS KARAKOULAKIS, KOSTAS KARPOUZIS, YANNIS AVRITHIS

More information

Multimedia Integration for Cooking Video Indexing

Multimedia Integration for Cooking Video Indexing Multimedia Integration for Cooking Video Indexing Reiko Hamada 1, Koichi Miura 1, Ichiro Ide 2, Shin ichi Satoh 3, Shuichi Sakai 1, and Hidehiko Tanaka 4 1 The University of Tokyo, 7-3-1 Hongo, Bunkyo-ku,

More information

Baseball Game Highlight & Event Detection

Baseball Game Highlight & Event Detection Baseball Game Highlight & Event Detection Student: Harry Chao Course Adviser: Winston Hu 1 Outline 1. Goal 2. Previous methods 3. My flowchart 4. My methods 5. Experimental result 6. Conclusion & Future

More information

Speaker Diarization System Based on GMM and BIC

Speaker Diarization System Based on GMM and BIC Speaer Diarization System Based on GMM and BIC Tantan Liu 1, Xiaoxing Liu 1, Yonghong Yan 1 1 ThinIT Speech Lab, Institute of Acoustics, Chinese Academy of Sciences Beijing 100080 {tliu, xliu,yyan}@hccl.ioa.ac.cn

More information

Using Multimedia Metadata

Using Multimedia Metadata Using Multimedia Metadata Semantic Gap Manual textual annotation Subjective process Meta data generation MP3 EXIF Dublin Core MPEG-7 MPEG-7 and the semantic gap MPEG-7 Annotation MPEG future, MPEG-21 MPEG-7

More information

Story Unit Segmentation with Friendly Acoustic Perception *

Story Unit Segmentation with Friendly Acoustic Perception * Story Unit Segmentation with Friendly Acoustic Perception * Longchuan Yan 1,3, Jun Du 2, Qingming Huang 3, and Shuqiang Jiang 1 1 Institute of Computing Technology, Chinese Academy of Sciences, Beijing,

More information

BUAA AUDR at ImageCLEF 2012 Photo Annotation Task

BUAA AUDR at ImageCLEF 2012 Photo Annotation Task BUAA AUDR at ImageCLEF 2012 Photo Annotation Task Lei Huang, Yang Liu State Key Laboratory of Software Development Enviroment, Beihang University, 100191 Beijing, China huanglei@nlsde.buaa.edu.cn liuyang@nlsde.buaa.edu.cn

More information

XVIP: An XML-Based Video Information Processing System

XVIP: An XML-Based Video Information Processing System XVIP: An XML-Based Video Information Processing System C.H. Ngai, P.W. Chan, Edward Yau, Michael R. Lyu Department of Computer Science and Engineering The Chinese University of Hong Kong Shatin, Hong Kong

More information

USING METADATA TO PROVIDE SCALABLE BROADCAST AND INTERNET CONTENT AND SERVICES

USING METADATA TO PROVIDE SCALABLE BROADCAST AND INTERNET CONTENT AND SERVICES USING METADATA TO PROVIDE SCALABLE BROADCAST AND INTERNET CONTENT AND SERVICES GABRIELLA KAZAI 1,2, MOUNIA LALMAS 1, MARIE-LUCE BOURGUET 1 AND ALAN PEARMAIN 2 Department of Computer Science 1 and Department

More information

Video search requires efficient annotation of video content To some extent this can be done automatically

Video search requires efficient annotation of video content To some extent this can be done automatically VIDEO ANNOTATION Market Trends Broadband doubling over next 3-5 years Video enabled devices are emerging rapidly Emergence of mass internet audience Mainstream media moving to the Web What do we search

More information

Text, Speech, and Vision for Video Segmentation: The Informedia TM Project

Text, Speech, and Vision for Video Segmentation: The Informedia TM Project Text, Speech, and Vision for Video Segmentation: The Informedia TM Project Alexander G. Hauptmann Michael A. Smith School Computer Science Dept. Electrical and Computer Engineering Carnegie Mellon University

More information

Ontology driven voice-based interaction in mobile environment

Ontology driven voice-based interaction in mobile environment Ontology driven voice-based interaction in mobile environment Jiri Kopsa 1, Zdenek Mikovec 1, Pavel Slavik 1 1 Czech Technical University in Prague Karlovo namesti 13, Prague 2, Czech Republic j.kopsa@fee.ctup.cz,

More information

Workshop W14 - Audio Gets Smart: Semantic Audio Analysis & Metadata Standards

Workshop W14 - Audio Gets Smart: Semantic Audio Analysis & Metadata Standards Workshop W14 - Audio Gets Smart: Semantic Audio Analysis & Metadata Standards Jürgen Herre for Integrated Circuits (FhG-IIS) Erlangen, Germany Jürgen Herre, hrr@iis.fhg.de Page 1 Overview Extracting meaning

More information

Content-Based Multimedia Information Retrieval

Content-Based Multimedia Information Retrieval Content-Based Multimedia Information Retrieval Ishwar K. Sethi Intelligent Information Engineering Laboratory Oakland University Rochester, MI 48309 Email: isethi@oakland.edu URL: www.cse.secs.oakland.edu/isethi

More information

Noisy Text Clustering

Noisy Text Clustering R E S E A R C H R E P O R T Noisy Text Clustering David Grangier 1 Alessandro Vinciarelli 2 IDIAP RR 04-31 I D I A P December 2004 1 IDIAP, CP 592, 1920 Martigny, Switzerland, grangier@idiap.ch 2 IDIAP,

More information

Deep Learning for Broadcast Videos and Multimedia

Deep Learning for Broadcast Videos and Multimedia Deep Learning for Broadcast Videos and Multimedia Lorenzo Baraldi lorenzo.baraldi@unimore.it University of Modena and Reggio Emilia Deep Learning State-of-the-art in image classification, object detection,

More information

Natural Language Processing Pipelines to Annotate BioC Collections with an Application to the NCBI Disease Corpus

Natural Language Processing Pipelines to Annotate BioC Collections with an Application to the NCBI Disease Corpus Natural Language Processing Pipelines to Annotate BioC Collections with an Application to the NCBI Disease Corpus Donald C. Comeau *, Haibin Liu, Rezarta Islamaj Doğan and W. John Wilbur National Center

More information

MATRIX BASED SEQUENTIAL INDEXING TECHNIQUE FOR VIDEO DATA MINING

MATRIX BASED SEQUENTIAL INDEXING TECHNIQUE FOR VIDEO DATA MINING MATRIX BASED SEQUENTIAL INDEXING TECHNIQUE FOR VIDEO DATA MINING 1 D.SARAVANAN 2 V.SOMASUNDARAM Assistant Professor, Faculty of Computing, Sathyabama University Chennai 600 119, Tamil Nadu, India Email

More information

!!!!!! Portfolio Summary!! for more information July, C o n c e r t T e c h n o l o g y

!!!!!! Portfolio Summary!! for more information  July, C o n c e r t T e c h n o l o g y Portfolio Summary July, 2014 for more information www.concerttechnology.com bizdev@concerttechnology.com C o n c e r t T e c h n o l o g y Overview The screenplay project covers emerging trends in social

More information

MESH. Multimedia Semantic Syndication for Enhanced News Services. Project Overview

MESH. Multimedia Semantic Syndication for Enhanced News Services. Project Overview MESH Multimedia Semantic Syndication for Enhanced News Services Project Overview Presentation Structure 2 Project Summary Project Motivation Problem Description Work Description Expected Result The MESH

More information

AUTOMATIC VIDEO INDEXING

AUTOMATIC VIDEO INDEXING AUTOMATIC VIDEO INDEXING Itxaso Bustos Maite Frutos TABLE OF CONTENTS Introduction Methods Key-frame extraction Automatic visual indexing Shot boundary detection Video OCR Index in motion Image processing

More information

Confidence Measures: how much we can trust our speech recognizers

Confidence Measures: how much we can trust our speech recognizers Confidence Measures: how much we can trust our speech recognizers Prof. Hui Jiang Department of Computer Science York University, Toronto, Ontario, Canada Email: hj@cs.yorku.ca Outline Speech recognition

More information

A Digital Library Framework for Reusing e-learning Video Documents

A Digital Library Framework for Reusing e-learning Video Documents A Digital Library Framework for Reusing e-learning Video Documents Paolo Bolettieri, Fabrizio Falchi, Claudio Gennaro, and Fausto Rabitti ISTI-CNR, via G. Moruzzi 1, 56124 Pisa, Italy paolo.bolettieri,fabrizio.falchi,claudio.gennaro,

More information

Annual Public Report - Project Year 2 November 2012

Annual Public Report - Project Year 2 November 2012 November 2012 Grant Agreement number: 247762 Project acronym: FAUST Project title: Feedback Analysis for User Adaptive Statistical Translation Funding Scheme: FP7-ICT-2009-4 STREP Period covered: from

More information

Automatic Metadata Extraction for Archival Description and Access

Automatic Metadata Extraction for Archival Description and Access Automatic Metadata Extraction for Archival Description and Access WILLIAM UNDERWOOD Georgia Tech Research Institute Abstract: The objective of the research reported is this paper is to develop techniques

More information

MPEG-7. Multimedia Content Description Standard

MPEG-7. Multimedia Content Description Standard MPEG-7 Multimedia Content Description Standard Abstract The purpose of this presentation is to provide a better understanding of the objectives & components of the MPEG-7, "Multimedia Content Description

More information

Towards Corpus Annotation Standards The MATE Workbench 1

Towards Corpus Annotation Standards The MATE Workbench 1 Towards Corpus Annotation Standards The MATE Workbench 1 Laila Dybkjær, Niels Ole Bernsen Natural Interactive Systems Laboratory Science Park 10, 5230 Odense M, Denmark E-post: laila@nis.sdu.dk, nob@nis.sdu.dk

More information

INFORMEDIA TM : NEWS-ON-DEMAND EXPERIMENTS IN SPEECH RECOGNITION

INFORMEDIA TM : NEWS-ON-DEMAND EXPERIMENTS IN SPEECH RECOGNITION INFORMEDIA TM : NEWS-ON-DEMAND EXPERIMENTS IN SPEECH RECOGNITION Howard D. Wactlar, Alexander G. Hauptmann and Michael J. Witbrock ABSTRACT In theory, speech recognition technology can make any spoken

More information

Deliverable D1.4 Report Describing Integration Strategies and Experiments

Deliverable D1.4 Report Describing Integration Strategies and Experiments DEEPTHOUGHT Hybrid Deep and Shallow Methods for Knowledge-Intensive Information Extraction Deliverable D1.4 Report Describing Integration Strategies and Experiments The Consortium October 2004 Report Describing

More information

Multimedia Database Systems. Retrieval by Content

Multimedia Database Systems. Retrieval by Content Multimedia Database Systems Retrieval by Content MIR Motivation Large volumes of data world-wide are not only based on text: Satellite images (oil spill), deep space images (NASA) Medical images (X-rays,

More information

CORRELATION BASED CAR NUMBER PLATE EXTRACTION SYSTEM

CORRELATION BASED CAR NUMBER PLATE EXTRACTION SYSTEM CORRELATION BASED CAR NUMBER PLATE EXTRACTION SYSTEM 1 PHYO THET KHIN, 2 LAI LAI WIN KYI 1,2 Department of Information Technology, Mandalay Technological University The Republic of the Union of Myanmar

More information

Columbia University High-Level Feature Detection: Parts-based Concept Detectors

Columbia University High-Level Feature Detection: Parts-based Concept Detectors TRECVID 2005 Workshop Columbia University High-Level Feature Detection: Parts-based Concept Detectors Dong-Qing Zhang, Shih-Fu Chang, Winston Hsu, Lexin Xie, Eric Zavesky Digital Video and Multimedia Lab

More information

Information Extraction from News Video using Global Rule Induction Technique

Information Extraction from News Video using Global Rule Induction Technique Information Extraction from News Video using Global Rule Induction Technique Lekha Chaisorn and 2 Tat-Seng Chua Media Semantics Department, Media Division, Institute for Infocomm Research (I 2 R), Singapore

More information

Video Processing for Judicial Applications

Video Processing for Judicial Applications Video Processing for Judicial Applications Konstantinos Avgerinakis, Alexia Briassouli, Ioannis Kompatsiaris Informatics and Telematics Institute, Centre for Research and Technology, Hellas Thessaloniki,

More information

Multimedia Event Detection for Large Scale Video. Benjamin Elizalde

Multimedia Event Detection for Large Scale Video. Benjamin Elizalde Multimedia Event Detection for Large Scale Video Benjamin Elizalde Outline Motivation TrecVID task Related work Our approach (System, TF/IDF) Results & Processing time Conclusion & Future work Agenda 2

More information

CS229: Action Recognition in Tennis

CS229: Action Recognition in Tennis CS229: Action Recognition in Tennis Aman Sikka Stanford University Stanford, CA 94305 Rajbir Kataria Stanford University Stanford, CA 94305 asikka@stanford.edu rkataria@stanford.edu 1. Motivation As active

More information

A Miniature-Based Image Retrieval System

A Miniature-Based Image Retrieval System A Miniature-Based Image Retrieval System Md. Saiful Islam 1 and Md. Haider Ali 2 Institute of Information Technology 1, Dept. of Computer Science and Engineering 2, University of Dhaka 1, 2, Dhaka-1000,

More information

Informedia News-On Demand: Using Speech Recognition to Create a Digital Video Library

Informedia News-On Demand: Using Speech Recognition to Create a Digital Video Library Informedia News-On Demand: Using Speech Recognition to Create a Digital Video Library Howard D. Wactlar 1, Alexander G. Hauptmann 1 and Michael J. Witbrock 2,3 March 19 th, 1998 CMU-CS-98-109 1 School

More information

WEB SEARCH, FILTERING, AND TEXT MINING: TECHNOLOGY FOR A NEW ERA OF INFORMATION ACCESS

WEB SEARCH, FILTERING, AND TEXT MINING: TECHNOLOGY FOR A NEW ERA OF INFORMATION ACCESS 1 WEB SEARCH, FILTERING, AND TEXT MINING: TECHNOLOGY FOR A NEW ERA OF INFORMATION ACCESS BRUCE CROFT NSF Center for Intelligent Information Retrieval, Computer Science Department, University of Massachusetts,

More information

A CASE STUDY: Structure learning for Part-of-Speech Tagging. Danilo Croce WMR 2011/2012

A CASE STUDY: Structure learning for Part-of-Speech Tagging. Danilo Croce WMR 2011/2012 A CAS STUDY: Structure learning for Part-of-Speech Tagging Danilo Croce WM 2011/2012 27 gennaio 2012 TASK definition One of the tasks of VALITA 2009 VALITA is an initiative devoted to the evaluation of

More information

Patent Terminlogy Analysis: Passage Retrieval Experiments for the Intellecutal Property Track at CLEF

Patent Terminlogy Analysis: Passage Retrieval Experiments for the Intellecutal Property Track at CLEF Patent Terminlogy Analysis: Passage Retrieval Experiments for the Intellecutal Property Track at CLEF Julia Jürgens, Sebastian Kastner, Christa Womser-Hacker, and Thomas Mandl University of Hildesheim,

More information

A Survey Of Different Text Mining Techniques Varsha C. Pande 1 and Dr. A.S. Khandelwal 2

A Survey Of Different Text Mining Techniques Varsha C. Pande 1 and Dr. A.S. Khandelwal 2 A Survey Of Different Text Mining Techniques Varsha C. Pande 1 and Dr. A.S. Khandelwal 2 1 Department of Electronics & Comp. Sc, RTMNU, Nagpur, India 2 Department of Computer Science, Hislop College, Nagpur,

More information

COALA: CONTENT-ORIENTED AUDIOVISUAL LIBRARY ACCESS

COALA: CONTENT-ORIENTED AUDIOVISUAL LIBRARY ACCESS COALA: CONTENT-ORIENTED AUDIOVISUAL LIBRARY ACCESS NASTARAN FATEMI Swiss Federal Institute of Technology (EPFL) E-mail: Nastaran.Fatemi@epfl.ch OMAR ABOU KHALED University of Applied Sciences of Western

More information

Automated Tagging to Enable Fine-Grained Browsing of Lecture Videos

Automated Tagging to Enable Fine-Grained Browsing of Lecture Videos Automated Tagging to Enable Fine-Grained Browsing of Lecture Videos K.Vijaya Kumar (09305081) under the guidance of Prof. Sridhar Iyer June 28, 2011 1 / 66 Outline Outline 1 Introduction 2 Motivation 3

More information

MultiMatch. D1.4 Functional Specification of the Second Prototype

MultiMatch. D1.4 Functional Specification of the Second Prototype Project no. 033104 MultiMatch Technology-enhanced Learning and Access to Cultural Heritage Instrument: Specific Targeted Research Project FP6-2005-IST-5 D1.4 Functional Specification of the Second Prototype

More information

Shrey Patel B.E. Computer Engineering, Gujarat Technological University, Ahmedabad, Gujarat, India

Shrey Patel B.E. Computer Engineering, Gujarat Technological University, Ahmedabad, Gujarat, India International Journal of Scientific Research in Computer Science, Engineering and Information Technology 2018 IJSRCSEIT Volume 3 Issue 3 ISSN : 2456-3307 Some Issues in Application of NLP to Intelligent

More information

Writer Identification In Music Score Documents Without Staff-Line Removal

Writer Identification In Music Score Documents Without Staff-Line Removal Writer Identification In Music Score Documents Without Staff-Line Removal Anirban Hati, Partha P. Roy and Umapada Pal Computer Vision and Pattern Recognition Unit Indian Statistical Institute Kolkata,

More information

Easy Ed: An Integration of Technologies for Multimedia Education 1

Easy Ed: An Integration of Technologies for Multimedia Education 1 Easy Ed: An Integration of Technologies for Multimedia Education 1 G. Ahanger and T.D.C. Little Multimedia Communications Laboratory Department of Electrical and Computer Engineering Boston University,

More information

Efficient Indexing and Searching Framework for Unstructured Data

Efficient Indexing and Searching Framework for Unstructured Data Efficient Indexing and Searching Framework for Unstructured Data Kyar Nyo Aye, Ni Lar Thein University of Computer Studies, Yangon kyarnyoaye@gmail.com, nilarthein@gmail.com ABSTRACT The proliferation

More information

ISO/IEC INTERNATIONAL STANDARD. Information technology Multimedia content description interface Part 5: Multimedia description schemes

ISO/IEC INTERNATIONAL STANDARD. Information technology Multimedia content description interface Part 5: Multimedia description schemes INTERNATIONAL STANDARD ISO/IEC 15938-5 First edition 2003-05-15 Information technology Multimedia content description interface Part 5: Multimedia description schemes Technologies de l'information Interface

More information

Class 5: Attributes and Semantic Features

Class 5: Attributes and Semantic Features Class 5: Attributes and Semantic Features Rogerio Feris, Feb 21, 2013 EECS 6890 Topics in Information Processing Spring 2013, Columbia University http://rogerioferis.com/visualrecognitionandsearch Project

More information

Annotation Graphs, Annotation Servers and Multi-Modal Resources

Annotation Graphs, Annotation Servers and Multi-Modal Resources Annotation Graphs, Annotation Servers and Multi-Modal Resources Infrastructure for Interdisciplinary Education, Research and Development Christopher Cieri and Steven Bird University of Pennsylvania Linguistic

More information

Integrating Visual and Textual Cues for Query-by-String Word Spotting

Integrating Visual and Textual Cues for Query-by-String Word Spotting Integrating Visual and Textual Cues for D. Aldavert, M. Rusiñol, R. Toledo and J. Lladós Computer Vision Center, Dept. Ciències de la Computació Edifici O, Univ. Autònoma de Barcelona, Bellaterra(Barcelona),

More information

Chapter 2. Architecture of a Search Engine

Chapter 2. Architecture of a Search Engine Chapter 2 Architecture of a Search Engine Search Engine Architecture A software architecture consists of software components, the interfaces provided by those components and the relationships between them

More information

Language Resources. Khalid Choukri ELRA/ELDA 55 Rue Brillat-Savarin, F Paris, France Tel Fax.

Language Resources. Khalid Choukri ELRA/ELDA 55 Rue Brillat-Savarin, F Paris, France Tel Fax. Language Resources By the Other Data Center over 15 years fruitful partnership Khalid Choukri ELRA/ELDA 55 Rue Brillat-Savarin, F-75013 Paris, France Tel. +33 1 43 13 33 33 -- Fax. +33 1 43 13 33 30 choukri@elda.org

More information

Annotation by category - ELAN and ISO DCR

Annotation by category - ELAN and ISO DCR Annotation by category - ELAN and ISO DCR Han Sloetjes, Peter Wittenburg Max Planck Institute for Psycholinguistics P.O. Box 310, 6500 AH Nijmegen, The Netherlands E-mail: Han.Sloetjes@mpi.nl, Peter.Wittenburg@mpi.nl

More information

ANC2Go: A Web Application for Customized Corpus Creation

ANC2Go: A Web Application for Customized Corpus Creation ANC2Go: A Web Application for Customized Corpus Creation Nancy Ide, Keith Suderman, Brian Simms Department of Computer Science, Vassar College Poughkeepsie, New York 12604 USA {ide, suderman, brsimms}@cs.vassar.edu

More information

MIRACLE at ImageCLEFmed 2008: Evaluating Strategies for Automatic Topic Expansion

MIRACLE at ImageCLEFmed 2008: Evaluating Strategies for Automatic Topic Expansion MIRACLE at ImageCLEFmed 2008: Evaluating Strategies for Automatic Topic Expansion Sara Lana-Serrano 1,3, Julio Villena-Román 2,3, José C. González-Cristóbal 1,3 1 Universidad Politécnica de Madrid 2 Universidad

More information

A cocktail approach to the VideoCLEF 09 linking task

A cocktail approach to the VideoCLEF 09 linking task A cocktail approach to the VideoCLEF 09 linking task Stephan Raaijmakers Corné Versloot Joost de Wit TNO Information and Communication Technology Delft, The Netherlands {stephan.raaijmakers,corne.versloot,

More information

CHIST-ERA Projects Seminar Topic IUI

CHIST-ERA Projects Seminar Topic IUI CHIST-ERA Projects Seminar Topic IUI Heiko Schuldt, Alexey Andrushevich, Laurence Devillers (based on slides from S. Dupont) Brussels, March 21-23, 2017 Introduction: Projects of the topic eglasses: The

More information

Managing Learning Objects in Large Scale Courseware Authoring Studio 1

Managing Learning Objects in Large Scale Courseware Authoring Studio 1 Managing Learning Objects in Large Scale Courseware Authoring Studio 1 Ivo Marinchev, Ivo Hristov Institute of Information Technologies Bulgarian Academy of Sciences, Acad. G. Bonchev Str. Block 29A, Sofia

More information

Clustering Methods for Video Browsing and Annotation

Clustering Methods for Video Browsing and Annotation Clustering Methods for Video Browsing and Annotation Di Zhong, HongJiang Zhang 2 and Shih-Fu Chang* Institute of System Science, National University of Singapore Kent Ridge, Singapore 05 *Center for Telecommunication

More information

Fish species recognition from video using SVM classifier

Fish species recognition from video using SVM classifier Fish species recognition from video using SVM classifier Katy Blanc, Diane Lingrand, Frédéric Precioso Univ. Nice Sophia Antipolis, I3S, UMR 7271, 06900 Sophia Antipolis, France CNRS, I3S, UMR 7271, 06900

More information

Browsing News and TAlk Video on a Consumer Electronics Platform Using face Detection

Browsing News and TAlk Video on a Consumer Electronics Platform Using face Detection MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Browsing News and TAlk Video on a Consumer Electronics Platform Using face Detection Kadir A. Peker, Ajay Divakaran, Tom Lanning TR2005-155

More information

Scalable Hierarchical Summarization of News Using Fidelity in MPEG-7 Description Scheme

Scalable Hierarchical Summarization of News Using Fidelity in MPEG-7 Description Scheme Scalable Hierarchical Summarization of News Using Fidelity in MPEG-7 Description Scheme Jung-Rim Kim, Seong Soo Chun, Seok-jin Oh, and Sanghoon Sull School of Electrical Engineering, Korea University,

More information