CONTENT MODEL FOR MOBILE ADAPTATION OF MULTIMEDIA INFORMATION

Similar documents
MULTIMEDIA ADAPTATION FOR DYNAMIC ENVIRONMENTS

Authoring Adaptive Multimedia Services

Mobile Multimedia Services Content Adaptation

A Reference Model for Flexible Content Development

Data Synchronization in Mobile Computing Systems Lesson 12 Synchronized Multimedia Markup Language (SMIL)

Optimal Video Adaptation and Skimming Using a Utility-Based Framework

Types and Methods of Content Adaptation. Anna-Kaisa Pietiläinen

RTT TECHNOLOGY TOPIC October The wireless web

Accessing SMIL-based Dynamically Adaptable Multimedia Presentations from Mobile Devices

An Output Schema for Multimedia Data in Multimedia Database Systems

MPEG-21: The 21st Century Multimedia Framework

M.SARAVANA KARTHIKEYAN

IMS Client Framework for All IP-Based Communication Networks

From Mobility Management to Connectivity Management

MPEG-4 AUTHORING TOOL FOR THE COMPOSITION OF 3D AUDIOVISUAL SCENES

Glossary 1. ARPU or Average Revenue per User A method of measuring revenue associated with the delivery of mobile commerce services by MNOs.

Interchange formats. Introduction Application areas Requirements Track and object model Real-time transfer Different interchange formats Comparison

MPEG-4. Today we'll talk about...

Interactive Authoring Tool for Extensible MPEG-4 Textual Format (XMT)

USING METADATA TO PROVIDE SCALABLE BROADCAST AND INTERNET CONTENT AND SERVICES

MPEG: It s Need, Evolution and Processing Methods

3GPP TS V5.2.0 ( )

Oracle Data Modelling & Database Design Course Content:35-40hours

09. Mobile Commerce. Contents. Mobile Computing and Commerce

BLM6196 COMPUTER NETWORKS AND COMMUNICATION PROTOCOLS

Mobile Commerce. Electronic Commerce

Apple Inc. November 2007

On Design of 3D and Multimedia Extension of Information System Using VRML

WAP-Speech: Deriving Synergy between WAP and the Spoken Dialog Interface

Voice/Data Integration in Wireless Communication Networks

A Hierarchical Keyframe User Interface for Browsing Video over the Internet

This tutorial is designed for those who would like to understand the basics of i-mode in simple and easy steps.

Internet Application Developer

A video streaming technique, in which video content is played back while the video data is being downloaded from an origin server, is useful for the

Dolby Vision. Streams within the MPEG-DASH format

Local area network (LAN) Wide area networks (WANs) Circuit. Circuit switching. Packets. Based on Chapter 2 of Gary Schneider.

Adaptive Multimedia Messaging based on MPEG-7 The M 3 -Box

Multi-modal Web IBM Position

A Design of Index Library using Management Information Modeling in Moving Picture Contents

DIF8914 Distributed Information Systems. Issues for Development of Mobile Multimedia Systems. Carl-Fredrik Sørensen

MPML: A Multimodal Presentation Markup Language with Character Agent Control Functions

9/8/2016. Characteristics of multimedia Various media types

Please view notes for further information on later slides

3 Publishing Technique

A WAY TO ACCESS WIRELESS NETWORK

An Intelligent System for Archiving and Retrieval of Audiovisual Material Based on the MPEG-7 Description Schemes

Comprehensive Structured Context Profiles (CSCP): Design and Experiences

Design of the Data-Retrieving Engine for Distributed Multimedia Presentations

Bridging and Switching Basics

A MULTIPOINT VIDEOCONFERENCE RECEIVER BASED ON MPEG-4 OBJECT VIDEO. Chih-Kai Chien, Chen-Yu Tsai, and David W. Lin

CPET 499/ITC 250 Web Systems. Topics

Introduction to LAN/WAN. Application Layer 4

Scalable Video Coding

A SMIL Editor and Rendering Tool for Multimedia Synchronization and Integration

Examining potentials for future of mobile Internet

Future Directions in Miniaturization inemi, January 16, 2008, Shanghai

A Concept for an Electronic Magazine

Hardware Assisted Recursive Packet Classification Module for IPv6 etworks ABSTRACT

A NOVEL MECHANISM FOR MEDIA RESOURCE CONTROL IN SIP MOBILE NETWORKS

E-Commerce. Infrastructure I: Computer Networks

A Plexos International Network Operating Technology May 2006

User Interaction: XML and JSON

RECOMMENDATION ITU-R BT.1720 *

Scalable Hierarchical Summarization of News Using Fidelity in MPEG-7 Description Scheme

MPEG-21 SESSION MOBILITY FOR HETEROGENEOUS DEVICES

Engineering Aspects of Web Hypermedia:

T Mobile Multimedia. Sakari Luukkainen Helsinki University of Technology Telecommunications Software and Multimedia Laboratory

Dolby Vision. Streams within the HTTP Live Streaming format

Delay Constrained ARQ Mechanism for MPEG Media Transport Protocol Based Video Streaming over Internet

A Study on Transmission System for Realistic Media Effect Representation

Multimedia CTI Services for Telecommunication Systems

Research on Construction of Road Network Database Based on Video Retrieval Technology

Mobile Devices: Server and Management Lesson 04 Gateways and Portals

Network protocols and. network systems INTRODUCTION CHAPTER

ACCESSIBLE DESIGN THEMES

WEB SEARCH, FILTERING, AND TEXT MINING: TECHNOLOGY FOR A NEW ERA OF INFORMATION ACCESS

File: SiteExecutive 2013 Content Intelligence Modules User Guide.docx Printed January 20, Page i

INTRODUCING THE UNIFIED E-BOOK FORMAT AND A HYBRID LIBRARY 2.0 APPLICATION MODEL BASED ON IT. 1. Introduction

Abstractions in Multimedia Authoring: The MAVA Approach

Easy Ed: An Integration of Technologies for Multimedia Education 1

Multimedia Communications

The Transport Layer: User Datagram Protocol

SMIL 2.0. Synchronized Multimedia Integration Language. Espen Nilsen

Research on 3G Terminal-Based Agricultural Information Service

Increazing interactivity in IPTV using MPEG-21 descriptors

«Computer Science» Requirements for applicants by Innopolis University

Intermediary Systems: A Survey

Extraction, Description and Application of Multimedia Using MPEG-7

Binju Bentex *1, Shandry K. K 2. PG Student, Department of Computer Science, College Of Engineering, Kidangoor, Kottayam, Kerala, India

Style Sheet Languages for Hypertext

Chapter 2 The Internet and World Wide Web

INTERNATIONAL ORGANISATION FOR STANDARISATION ORGANISATION INTERNATIONALE DE NORMALISATION ISO/IEC JTC1/SC29/WG11 CODING OF MOVING PICTURES AND AUDIO

MM Message Assembly Mode

Abstract. Proceedings of the IEEE Sixth International Symposium on Multimedia Software Engineering (ISMSE 04) /04 $20.

8. Markup Languages. Characteristics of Computer-Use Markup Languages. Markup. Markup language- History

Distributed Systems. Overview. Distributed Systems September A distributed system is a piece of software that ensures that:

A Study on Metadata Extraction, Retrieval and 3D Visualization Technologies for Multimedia Data and Its Application to e-learning

Streaming Technologies Delivering Multimedia into the Future. May 2014

Multimedia-Systems: Applications

Differential Compression and Optimal Caching Methods for Content-Based Image Search Systems

Transcription:

CONTENT MODEL FOR MOBILE ADAPTATION OF MULTIMEDIA INFORMATION Maija Metso, Antti Koivisto and Jaakko Sauvola MediaTeam, MVMP Unit Infotech Oulu, University of Oulu e-mail: {maija.metso, antti.koivisto, jaakko.sauvola}@ee.oulu.fi Abstract - The penetration of Internet services has introduced a set of transmission and presentation standards. At the same time, limited bandwidth resources of cellular networks and terminals set strict requirements to the transmission and presentation of multimedia. This paper proposes a new, abstract model to represent and adapt multimedia to hybrid environments. The model includes a layered mapping of semantic and physical entities and is combined under the taxonomy of multimedia adaptation to optimize end-to-end service. We conclude with an example case that utilizes the proposed techniques. Keywords: Multimedia adaptation, content model, multimedia processing, presentation INTRODUCTION Many of the presentations found in popular services can not be used with low capability terminals or with low bandwidth networks due to their increasing physical size and complexity. Alternatively, they can be provided via tailored, often proprietary access methods requiring additional effort and cost from providers and additional adaptation from the client. Existing WWW or intranet services have remained practically unavailable for mobile users, although efforts such the Wireless Application Protocol (WAP) [1] are finally laying a foundation for true mobile multimedia. Emerging markup-languages such as the XML (Extensible Markup Language) [2] and its subsets like SMIL (Synchronized Multimedia Integration Language) [3] developed by the World Wide Web Consortium (W3C) can be applied in modeling multimedia presentations. Video-related standards such as MPEG-4 will introduce an object-oriented approach to construct dynamic presentations [4]. Nevertheless, these approaches are mostly concerned with document layout, structure and synchronization where as model enabling flexible adaptation of the content using both semantics and physical constraints has not been yet explored. Our multimedia adaptation taxonomy [5] introduced a two-tiered approach to adapt multimedia presentations. In second chapter we present a new media content model for intelligent multimedia adaptation for low capability (e.g. wireless) terminals and networks. The usage of the model and media adaptation taxonomy together for semantic and physical adaptation is presented in third chapter. The semantic adaptation discussed in first subsection addresses some additional requirements for the structure and usage of the content model. The physical adaptation, guided by the semantic adaptation is also presented. Finally, we conclude with an example, where the content model is applied for mobile multimedia newspaper -application.

A CONTENT MODEL FOR MEDIA ADAPTATION A prerequisite for efficient adaptation of multimedia information is a careful analysis of the properties of different media types. Video, voice, images and text require different adaptation algorithms. The complex nature of multimedia makes the adaptation difficult to design and implement. The proposed media content model encapsulates the properties of multimedia presentation and supports the manipulation of entire service in a layered manner. The purpose of adaptation is to alter dynamically the content, e.g. size, format and formation of multimedia presentations. This is done with respect to the Quality of Service (QoS) defined by the transmission and presentation environment of the end user, and by the semantic properties of the presentation itself. The process comprises semantic and physical entities, which are simultaneously used to create, optimize and scale the presentation into the desired size and format. The semantic process first defines the content of the presentation, which the physical adaptation process then transforms into the final format. In the media content model depicted in Figure 1 the presentation is divided into four semantic layers, where the topmost layer describes the presentation as an entity. The presentation layer can be divided into two or more subscripts, which describe the logical entities (indexes) within the presentation, e.g. pages or files in multimedia presentation stream. The object layer consists of media objects, which are defined as logical entities of the subscript layer, e.g. images, headlines or audio clips. Finally, the primitive layer defines the properties of the actual physical data in form of headers and bodies. Figure 1. Media Content Model. Several relationships are defined between and within the layers of the model. The physical data is mostly stored in the primitive layer while the semantics are encapsulated in the abstracted layers. The model is used to describe temporal and spatial relationships between parts of the presentation, and to categorize and priorize the components. Different components of a multimedia presentation are adapted based on the semantic and physical requirements of the entire service event (required by user). Different layers provide alternatives for adaptation, i.e. it can be applied to the targeted media primitives, media objects, subscripts or the entire presentation, including methods of transmission.

APPLYING THE MEDIA CONTENT MODEL WITH ADAPTATION TAXONOMY The taxonomy of multimedia adaptation in [5] presented a two-tiered approach for adapting multimedia presentations. It combined the processing requirements set by the application and content domains, all of which influence the entire lifecycle of a presentation: preparation, adaptation and transmission. The adaptation taxonomy consists of two parts: optimization (semantic adaptation) and scaling (physical adaptation). Figure 2. Adaptation Taxonomy versus Media Content Model. The relationship between the media content model and the multimedia adaptation taxonomy is illustrated in Figure 2. The presentation and subscript layers are used for optimization purposes while the media object and primitive layers are utilized in the physical adaptation. Applying the Semantics of the Presentation In principle, semantic adaptation is a selective process. Based on available and requested information, multiple pieces of information are combined into a complete presentation or service. This process can be improved, if we are able to understand the content, identify the requester and the user preferences and classify the components required by the presentation into priorized categories. Semantic adaptation is based on users and service providers choices and it is affected by the semantic content of a presentation these three parameter classes are referred as the semantic QoS. They determine which part of the data is relevant for final presentation. The optimization process has to manage a number of constraints: the users needs must be fulfilled, and service identifiers extracted. Secondly, as the semantic processing must guide and collaborate with the process of physical adaptation to convert the framework into the final presentation environment. Scaling the Physical Data Physical adaptation of media is defined as the combination of (1) conversion, (2) scaling and (3) distillation processes guided by the characteristics of media formats and physical QoS. Here the QoS defines the bounds for example size, type, duration, transmission and other parameters. For instance, we have a multimedia presentation containing multiple video- and audio sequences with captions and formatted text that are selected and generated during the content preparation phase, Figure 3. In the decomposition and QoS determination phase, the presentation is constructed from separate media objects. They represent individual video- and audio sequences and text sections. Then, the bodies and headers are scaled and converted according to the physical QoS. The adaptation phase can include the replacement of audio and video sequences with

captions or synthesized speech (conversion) or by substituting a high-quality video clip with a low-quality preview version (scaling) or even further with a selected set of keyframes. Adapted media objects are merged in spatial (layout) and temporal domains in the composition phase. The distillation process is used to create a presentation script that describes the actual form and outlook of the presentation. In the case of mixed audio- and video, the composition phase would include synchronization of streams into a multimedia presentation format. Figure 3. The process of physical adaptation is guided by physical QoS. Integrated Adaptation The final adaptation process can be seen as the integration of physical and semantic phases. They complement each other by addressing different properties of the presentation. The semantic adaptation determines relevant information but does not specify the presentation. In the physical domain, the actual information content of the presentation remains constant while the actual physical format, such as file format or size is altered. The physical adaptation can be made completely automatic, taking data as input, and transforming it according to given parameters. However, the semantic adaptation requires collection of parameters as input from the content provider or author (or a consumer process). The high-level structure and content of the presentation can be formed without any knowledge of specific markup-languages or media formats, whereas the low-level processing can be done with a little or no knowledge of the context. APPLICATION EXAMPLE In the following we demonstrate the usage of our media content model with an example application, mobile multimedia newspaper. It is used in conjunction with adaptation taxonomy to create and adapt multimedia presentations into two different transmission and presentation environments. Beginning with the same semantic content, we prepare, adapt and transmit a multimedia presentation to both

a WAP (Wireless Application Protocol) compliant cellular phone and a portable computer with wireless LAN connection. The application uses an indexed video database, which contains raw video streams, associated captions and textual information about news headlines. First, the user with the cellular phone asks for recent news from the service (e.g. date parameters in terminal GUI). Based on the information stored in the database, a presentation is created with most recent information. Figure 4. Adaptation layers and QoS news service Here, the specific model used in adaptation is presented in Figure 4. The presentation is decomposed into subscripts, encapsulating a single headline, and further into media objects which map to text sections and video clips. The semantic (vertical) QoS is defined by the user request and the semantic information about the news headline describing the priorization and relationships between different parts of the content. The physical (horizontal) QoS is determined from the information on user s terminal and network connection. The adaptation process begins with the selection of headlines, which is primarily based on priorization specified by the news author. Then, at the subscript layer, objects carrying the most important content are selected. Having the content for the presentation, the media primitives are scaled according to the physical QoS. As the user with a WAP-phone is unable to view video sequences on his terminal, they are removed and the text is stripped to minimum based on the priorization of the content. The remaining media objects are scaled, converted and distilled to form a WML-document [1], which is transmitted to the user. At the same time the laptop user requests the same service. Because the physical QoS does not limit the text length or the formatting of the presentation the headline can be presented in more detail using high quality graphics and audio. Based on the available bandwidth, the video can either be sent as high quality or preview version. If network has only limited bandwidth, the user is provided with a keyframe-presentation of the sequence, accompanied with captions and the same text as with the WML-version. The scaling of video and the generation and scaling of keyframes are both computationally expensive operations. Due to the dynamic nature of a news service, the adaptation is performed on-demand thus affecting latency suffered by the service consumer.

By using a hierarchically built semantic content model together with adaptation, we are able to separate presentation content from its representation and classify the content of the presentation and its parts according to their semantic meaning. This makes the progression very straightforward: it is easy to adjust the content and provide it in different formats to different platforms. Flexibility extends also to the client-side: if required, the initially seen minimum content can be easily extended. The effort to provide the same content to various platforms is reduced significantly, when the adaptation can be executed automatically. Expensive and limited bandwidth is saved as only necessary information is transmitted and the presentations are received with shorter latencies. CONCLUSIONS AND FUTURE WORK In this paper we proposed a new technique to model multimedia data adaptively for multiple heterogeneous networks and terminals. By mixing intelligence that combines the requirements and semantic (content) information with low-level processing, we presented an entirely new approach to create, provide and access multimedia information. Armed with the possibilities to convert, scale and represent the media, and having acquired a high-level view to entire presentation, including primitive components, we are able to development the adaptation model to support end-toend multimedia services and associated media retrieval and processing algorithms. ACKNOWLEDGMENTS The financial support from the Academy of Finland, Technology Development Centre of Finland, Graduate School in Electronics, Telecommunications and Automation (GETA) and Princess project consortium is gratefully acknowledged. REFERENCES [1] Wireless Application Protocol Forum, Wireless Application Protocol Specification 1.1 & Wireless Markup Language Specification 1.0, 1998. [2] World Wide Web Consortium, The Extensible Markup Language 1.0 Recommendation, 1998. [3] Bouthillier L., Synchronized Multimedia on the Web. Web Techniques Magazine, September 1998, Volume 3, Issue 9. [4] Motion Pictures Experts Group, Overview of the MPEG-4 Standard, ISO/IEC JTC1/SC29/WG11 N2459, 1998. [5] Metso M., Koivisto A. & Sauvola J. Multimedia Adaptatation for Dynamic Environments. Proc. 2nd IEEE Workshop on Multimedia Signal Processing, 1998, p. 203-208.