A MULTIPOINT VIDEOCONFERENCE RECEIVER BASED ON MPEG-4 OBJECT VIDEO. Chih-Kai Chien, Chen-Yu Tsai, and David W. Lin

Size: px
Start display at page:

Download "A MULTIPOINT VIDEOCONFERENCE RECEIVER BASED ON MPEG-4 OBJECT VIDEO. Chih-Kai Chien, Chen-Yu Tsai, and David W. Lin"

Transcription

1 A MULTIPOINT VIDEOCONFERENCE RECEIVER BASED ON MPEG-4 OBJECT VIDEO Chih-Kai Chien, Chen-Yu Tsai, and David W. Lin Dept. of Electronics Engineering and Center for Telecommunications Research National Chiao Tung University Hsinchu, Taiwan 30010, R.O.C. s: ABSTRACT We consider the design and implementation of a novel type of software-based multipoint videoconference receiver on a personal computer (PC), where some distinguishing features are that MPEG-4 object-based coding is used in encoding each video stream and that the decoded videos are composed into one for display. The resulting receiver includes an RTP-based network interface, a set of MPEG-4 video and AAC audio decoders (whose number depends on number of source sites), and a unit to compose the decoded media streams for display. We develop a graphical user interface using the Windows SDK for convenience in system control and monitoring as well as display of results. With an AMD CPU running at 2.1-GHz with 512-MB RAM, the current un-optimized implementation yields a speed on the order of 10 frames per second for CIF ( ) video when receiving from only one source site. The frame rate reduces approximately in proportion to number of sources. 1. INTRODUCTION In a typical multipoint videoconference system, the receiver places the decoded videos from different sources in different windows. We consider constructing a different kind of system in which the decoded videos are composed into a virtual conference room scene. For this, a natural and simplest approach is to segment and encode the source videos separately at their respective transmitter sites, and let each receiver decode all received videos and compose and display the result. The MPEG-4 standards, with their provision for object-based video coding, appear naturally fitting for this use. In this work, we consider the design and implementation of the receiver on a personal computer (PC), in software. There are four major components in the receiver: the network interface, the video decoder, the audio decoder, and the composition unit, as illustrated in Fig. 1. As will be Work supported by National Science Council of R.O.C. under Grant NSC E Fig. 1. Structure of the proposed videoconference receiver. explained below, we decide to use the Real-time Transport Protocol (RTP) for the network interface and develop our own composition method, leaving the video and the audio encoded and decoded according to MPEG-4 specifications. In what follows, Section 2 gives an overview of the (earliest) MPEG-4 standards and comments on the usefulness of each part in our work. Section 3 discusses how we decode and compose multiple videos into one scene. Section 4 describes the integration of the receiver system. Some experimental results are described in Section 5. Finally, Section 6 contains the conclusion. 2. THE MPEG-4 STANDARDS AND THEIR APPLICATION IN THIS WORK The original MPEG-4 standards are divided into four basic parts: Systems (ISO/IEC ), Visual (ISO/IEC ), Audio (ISO/IEC ), and Delivery Multimedia Integration Framework, or DMIF in short (ISO/IEC ) [1], [2]. The MPEG-4 Systems part specifies how audio-visual

2 Fig. 2. Relation between the two decoders in time domain. scenes can be created from individual objects. To actually create the audio-visual scenes for a particular application, one may employ a suitable authoring tool designed to the MPEG-4 specifications. Unfortunately, no suitable authoring tools are found for our application in the course of this work. Hence we decide to develop a simple composition method and write our own program for it. Aside from other innovations, a major novelty in the MPEG-4 Visual part is the provision for coding of arbitraryshaped objects. In fact, each picture is considered as composed of a number of video objects. For each object, a socalled alpha-plane image sequence defines the support of the object in each video frame as well as the transparency of each pixel therein. Pixels belonging to the object are encoded largely employing typical motion-compensated block- DCT techniques; and so are the alpha-plane images. These features are used in our implementation. The MPEG-4 Audio part is rather generic in that it considers different kinds of audio signal (e.g., speech, natural sound, and synthetic sound) and different kinds of signal manipulation (e.g., compression and synthesis). As a result, it facilitates object-based audio coding and manipulation. Our system merely makes use of its compression functionality. We employ a publicly available MPEG-4 AAC (Advanced Audio Coding) decoder in this work. The MPEG-4 DMIF part specifies the interface between the application and the transport, so that the transport can be relatively transparent to the application and the application can be developed largely free from concerns over the transport. Despite the intent, recent MPEG activities show waned interest in DMIF and existing DMIF software, such as [3], has suffered functionality and maintenance problems. Therefore, we decide to use RTP [4] for the network interface. 3. VIDEO DECODING AND COMPOSITION As discussed previously, we develop a simple video composition method in this work. This involves not only composition of multiple videos into one scene, but also synchronization among the multiple decoders. For simplicity, we Fig. 3. Different spatial relations between two images. (a) Case 0, no overlap. (b) (e) Cases 1 4, respectively, with overlaps. assume that all videos have the same frame rate and relegate the situation with disparate frame rates to potential future research. Note, however, that in the latter situation, we may consider skipping some frames in the videos with higher frame rates, which should not cause much problem in videoconferencing when the frame rates are high enough. In this case, only minor modifications are needed in the present program. We first consider the situation with two videos and two decoders. The two decoders are placed in two threads on the PC. To synchronize the decoded videos, for each frame we let the first decoder wait until the second decoder completes. Then it starts working on the next frame. Composition and display of the videos are done in the second thread. Figure 2 shows the temporal relation between the two decoders. To compose the two images decoded by the two decoders, note that there are several possible spatial relations between them, as illustrated in Fig. 3. Case 0 is where there is no overlap between the two images. If there is some overlap between the two images, we let the first image occlude the second in the overlapped area. In the integrated receiver system, we let the user determine and specify where each decoded video is to be placed in the display window.

3 Fig. 4. Composition of four (or three) videos. Now consider the situation with four (or three) videos. Figure 4 explains our method of composition. The outputs of the first and the second decoders are composed together first, and so are the outputs of the third and the fourth decoders. Then the two composed images are composed to yield the final displayed image. The padding referred in the figure is to pad the image into the form of a box for display purpose. For more videos, we simply extend the composition tree. In any case, it is always the highest-indexed video that controls the composition and display operation. 4. RECEIVER SYSTEM INTEGRATION 4.1. Overall System Structure Figure 5 shows how the integrated receiver program works. The GUI block creates a window and the user can input the ports and the positions of the different videos. Since there may be multiple video and audio streams to be handled, we use the multi-threading technique to manage the video and the audio decoders. The multi-threading technique lets the operating system handle the scheduling of the threads. Good scheduling can made efficient use of the PC s available computing, storage, and communication resources. So, after the user has specified the ports and the positions of the videos, the system creates the decoder threads. It also obtains information on which decoders are active at present, so as to determine the highest-indexed video decoder and pass the control responsibility (including video composition and display) to it. Now the video and the audio decoders can begin their work in decoding the data received through the RTP network interface. The composed video is displayed in a window and the composed audio is played. Fig. 5. Flow diagram of the integrated receiver. In the following subsections we give some further details of the system components The RTP Network Interface The RTP facilitates end-to-end delivery of data with realtime characteristics, such as interactive audio and video, and is thus suitable for our application [4]. It supports sequence numbering, timestamping, and delivery monitoring. It can also support multicasting if the underlying network has such capability, which makes it naturally fitted to support multiparty multimedia conferences. However, our implementation does not make use of this feature, neither do we assume that the network has multicasting capability. The RTP is usually run on top of UDP. One should note that RTP itself does not provide quality-of-service guarantees (such as orderly and timely delivery of packets); it is up to the higher layers (such as the user) to provide them, aided by the information provided by the RTP and as far as it is within the ability of the lower-layer services. We use the jrtplib software [5] developed at the Expertise Centre for Digital Media (EDM) to realize the RTP network interface. One RTP session is created for each video or audio stream. The transmitter [6] handles the required RTP packetization. Two important parameters that need to be set for each session are the timestamp and the portbase. The timestamp parameter is set to 1 section per second for a video stream and to 1 section per 4 seconds for an audio stream. (These parameter values are somewhat in-

4 appropriate for real-time applications, but are chosen based on experience for smooth running of the program. Further work is needed to determine the underlying problem and potential solution.) For the portbase, since each transmitter sends two streams (video and audio), the video stream uses a user-specified port number and the audio stream uses that number plus 100. In our system, after the setting of these parameters, we put the receiving of the RTP packets in a main loop. We use the function StartRTP to enable the main loop to receive RTP data and the function ResetRTP to disable the loop to stop receiving RTP data. For some reason yet unclear, the software would lose the first two packets and receive the third and later packets successfully Video Decoding and Display Our video decoder is from the Microsoft MPEG-4 Video Reference Software [7], which is a public source for encoding and decoding of video in MPEG-4 format. We did optimization of the encoder for Intel processors [8], but not the decoder. For our application, we use the binary shape coding feature belonging to main profile of MPEG-4 video. To integrate the video decoder into the overall system, we modify the original decoder program into a function called MPEG4VDecoder and give it two parameters: handle (of the display window) and thread index. The handle can give the decoder some information to control the window and display the video stream. To display the video, we convert the decoder output from the original 4:2:0 format to the 4:4:4 format. Then we calculate the RGB values of each pixel from the luminance and the chrominance values. And then we use the SetPixelV function provided by the Windows SDK library to display the RGB values pixel by pixel. However, experience shows that the SetPixelV function is very slow and can significantly slow down the overall speed of the receiver system. Hence, to reduce its use, instead of using it on all pixels in the display window, we only use it to update the pixels in the object areas of two successive frames Audio Decoding and Composition For audio decoding, we use the Freeware Audio Decoder (FAAD2) [9] written by M. Bakker. The decoder can handle HE, LC, MAIN and LTP profiles, but we only make use of the MAIN profile. After decoding the audio stream received from the RTP network interface, the result is saved as a temporal audio file in the WAV format, which is a common format used in PC audio. The Media Control Interface (MCI), a high level open interface, provides two ways to play WAV-format audio. Since each audio section is four seconds, the decoder Fig. 6. Some typical composed scenes with two videos. would wait for that long as the decoder output is played by the MCI. For audio composition, two intuitive methods are (1) to sum all audio streams and (2) to play only one stream. The first method suffers an overflow problem which can be solved by proper scaling. This is left to potential future work. For simplicity in the final system we only play the audio from the first transmitter site. 5. EXPERIMENTAL RESULTS We present some experimental results in this section. For convenience, for the video part we use the common CIF ( ) test sequence Bream with its associated binary shape information. Multiple instances of the sequence are considered separate video transmissions. Figure 6 shows some typical composed scenes with two videos. Table 1 shows some performance data from employing a PC with an AMD CPU running at 2.1-GHz with 512-MB RAM. The program is not yet optimized. We see that if only one original video decoder is present, the rate can get up to 53.4 frames per second (fps). The processing speed decreases roughly in proportion to the number of decoders, but somewhat slower. This presumably is because the computing resources are used a little more efficiently when there are multiple decoder threads. As explained previously, we pad the decoded video to the 4:4:4 format and convert the result to RGB values for display. As shown in Table 1, the frame rate decreases by about 20% in each case when we add the padding function. Adding video composition and displaying of the result, where the SetPixelV function is used over the whole video frame, reduces the frame rates exceedingly (by one order of magnitude), as can be seen in Table 1 for the cases with

5 Table 1. Processing Rate (in Frames per Second) Under Various Configurations No. of Decoders Configuration Original With Padding With Compos.+Display Only Composition 17.6 Reduced SetPixel With Audio We considered the design and implementation of a novel type of software-based multipoint videoconference receiver on a PC, where some distinguishing features were the use of MPEG-4 object-based coding and the composition of decoded videos into one scene. The resulting receiver included an RTP-based network interface, a set of MPEG-4 video and AAC audio decoders (whose number depends on number of source sites), and a unit to compose the decoded media streams for display. Some topics for potential future research are as follows. 1. Optimization of software components to increase speed and reduce transmission latency: Top on the list should be a faster method to display the composed video. Other components, such as the video and the audio decoders, can also be improved in speed. 2. More elegant audio composition. 3. More sophisticated handling of videos with different frame rates. 4. More elaborate ways of video composition: This may include scaling and three-dimensional rotation of the decoded videos, as well as integration with proper background and foreground. 7. REFERENCES [1] International Committee for Information Technology Standards, Fig. 7. Time analysis of the receiver system. one and two videos. With only video composition but not the display of the result, the frame rate can be much higher, as shown in Table 1 for the case with two videos. Time analysis shows that image display alone takes over 70% of the overall processing time, as illustrated in Fig. 7. Hence, unless we can find more efficient ways to set the display pixels, we should minimize the use of the time-consuming SetPixelV function. After limiting SetPixelV function calls to the video object areas as mentioned briefly previously, we increase the frame rates by about 140%, as shown in Table 1. Finally, we add the audio decoder into the system. From Table 1, we see that the frame rate decreases only by roughly 10% in order of magnitude in each case. Hence the audio decoder has a relatively low complexity. 6. CONCLUSION AND FUTURE WORK [2] MPEG-4 Video Group, MPEG-4 overview (V.21 Jeju Version), doc. no. ISO/IEC JTC1/SC29/WG11 N4668, Mar [3] DMIF software, [4] H. Schulzrine, S. Casner, R. Frederick, and V. Jacobson, RTP: a transport protocol for real-time application, RFC 3550, Network Working Group, July [5] JRTPLIB, [6] C.-Y. Tsai, Integration of videoconference transmitter with MPEG-4 object-based video encoding, M.S. thesis, Dept. of Electronics Engineering, National Chaio Tung University, Hsinchu, Taiwan, R.O.C., June [7] Microsoft, ISO/IEC (MPEG-4) Video Reference Software User Manual, Oct [8] M.-Y. Liu, Real-time implementation of MPEG-4 video encoder using SIMD-enhanced Intel processor, M.S. thesis, Degree Program of Electrical Engineering and Computer Science, National Chaio Tung University, Hsinchu, Taiwan, R.O.C., July [9] AudioCoding.com,

Outline Introduction MPEG-2 MPEG-4. Video Compression. Introduction to MPEG. Prof. Pratikgiri Goswami

Outline Introduction MPEG-2 MPEG-4. Video Compression. Introduction to MPEG. Prof. Pratikgiri Goswami to MPEG Prof. Pratikgiri Goswami Electronics & Communication Department, Shree Swami Atmanand Saraswati Institute of Technology, Surat. Outline of Topics 1 2 Coding 3 Video Object Representation Outline

More information

MISB EG Motion Imagery Standards Board Engineering Guideline. 24 April Delivery of Low Bandwidth Motion Imagery. 1 Scope.

MISB EG Motion Imagery Standards Board Engineering Guideline. 24 April Delivery of Low Bandwidth Motion Imagery. 1 Scope. Motion Imagery Standards Board Engineering Guideline Delivery of Low Bandwidth Motion Imagery MISB EG 0803 24 April 2008 1 Scope This Motion Imagery Standards Board (MISB) Engineering Guideline (EG) provides

More information

Chapter 11.3 MPEG-2. MPEG-2: For higher quality video at a bit-rate of more than 4 Mbps Defined seven profiles aimed at different applications:

Chapter 11.3 MPEG-2. MPEG-2: For higher quality video at a bit-rate of more than 4 Mbps Defined seven profiles aimed at different applications: Chapter 11.3 MPEG-2 MPEG-2: For higher quality video at a bit-rate of more than 4 Mbps Defined seven profiles aimed at different applications: Simple, Main, SNR scalable, Spatially scalable, High, 4:2:2,

More information

EE Multimedia Signal Processing. Scope & Features. Scope & Features. Multimedia Signal Compression VI (MPEG-4, 7)

EE Multimedia Signal Processing. Scope & Features. Scope & Features. Multimedia Signal Compression VI (MPEG-4, 7) EE799 -- Multimedia Signal Processing Multimedia Signal Compression VI (MPEG-4, 7) References: 1. http://www.mpeg.org 2. http://drogo.cselt.stet.it/mpeg/ 3. T. Berahimi and M.Kunt, Visual data compression

More information

Scalable Video Coding

Scalable Video Coding Introduction to Multimedia Computing Scalable Video Coding 1 Topics Video On Demand Requirements Video Transcoding Scalable Video Coding Spatial Scalability Temporal Scalability Signal to Noise Scalability

More information

MPEG-4 AUTHORING TOOL FOR THE COMPOSITION OF 3D AUDIOVISUAL SCENES

MPEG-4 AUTHORING TOOL FOR THE COMPOSITION OF 3D AUDIOVISUAL SCENES MPEG-4 AUTHORING TOOL FOR THE COMPOSITION OF 3D AUDIOVISUAL SCENES P. Daras I. Kompatsiaris T. Raptis M. G. Strintzis Informatics and Telematics Institute 1,Kyvernidou str. 546 39 Thessaloniki, GREECE

More information

MPEG-4. Today we'll talk about...

MPEG-4. Today we'll talk about... INF5081 Multimedia Coding and Applications Vårsemester 2007, Ifi, UiO MPEG-4 Wolfgang Leister Knut Holmqvist Today we'll talk about... MPEG-4 / ISO/IEC 14496...... is more than a new audio-/video-codec...

More information

CMPT 365 Multimedia Systems. Media Compression - Video Coding Standards

CMPT 365 Multimedia Systems. Media Compression - Video Coding Standards CMPT 365 Multimedia Systems Media Compression - Video Coding Standards Spring 2017 Edited from slides by Dr. Jiangchuan Liu CMPT365 Multimedia Systems 1 Video Coding Standards H.264/AVC CMPT365 Multimedia

More information

Georgios Tziritas Computer Science Department

Georgios Tziritas Computer Science Department New Video Coding standards MPEG-4, HEVC Georgios Tziritas Computer Science Department http://www.csd.uoc.gr/~tziritas 1 MPEG-4 : introduction Motion Picture Expert Group Publication 1998 (Intern. Standardization

More information

Streaming and Recording Capabilities

Streaming and Recording Capabilities Streaming and Recording Capabilities PCS-G50/G50P All PCS-G70/G70P All PCS-XG55S All PCS-XG80S All Introduction Sony visual communication system PCS-XG55S/XG80S and PCS-G50/G50P/G70/ G70P (here after referred

More information

The Virtual Meeting Room

The Virtual Meeting Room Contact Details of Presenting Authors Stefan Rauthenberg (rauthenberg@hhi.de), Peter Kauff (kauff@hhi.de) Tel: +49-30-31002 266, +49-30-31002 615 Fax: +49-30-3927200 Summation Brief explaination of the

More information

Module 7 VIDEO CODING AND MOTION ESTIMATION

Module 7 VIDEO CODING AND MOTION ESTIMATION Module 7 VIDEO CODING AND MOTION ESTIMATION Lesson 20 Basic Building Blocks & Temporal Redundancy Instructional Objectives At the end of this lesson, the students should be able to: 1. Name at least five

More information

Introduction to LAN/WAN. Application Layer 4

Introduction to LAN/WAN. Application Layer 4 Introduction to LAN/WAN Application Layer 4 Multimedia Multimedia: Audio + video Human ear: 20Hz 20kHz, Dogs hear higher freqs DAC converts audio waves to digital E.g PCM uses 8-bit samples 8000 times

More information

A Hybrid Architecture for Video Transmission

A Hybrid Architecture for Video Transmission 2017 Asia-Pacific Engineering and Technology Conference (APETC 2017) ISBN: 978-1-60595-443-1 A Hybrid Architecture for Video Transmission Qian Huang, Xiaoqi Wang, Xiaodan Du and Feng Ye ABSTRACT With the

More information

MPEG-4: Overview. Multimedia Naresuan University

MPEG-4: Overview. Multimedia Naresuan University MPEG-4: Overview Multimedia Naresuan University Sources - Chapters 1 and 2, The MPEG-4 Book, F. Pereira and T. Ebrahimi - Some slides are adapted from NTNU, Odd Inge Hillestad. MPEG-1 and MPEG-2 MPEG-1

More information

Interworking Between SIP and MPEG-4 DMIF For Heterogeneous IP Video Conferencing

Interworking Between SIP and MPEG-4 DMIF For Heterogeneous IP Video Conferencing Interworking Between SIP and DMIF For Heterogeneous IP Video Conferencing Toufik Ahmed 1, Ahmed Mehaoua 1 and Raouf Boutaba 2 1 University of Versailles, CNRS-PRiSM Lab. 45 av. des Etats-Unis, 78000, Versailles,

More information

Fast Decision of Block size, Prediction Mode and Intra Block for H.264 Intra Prediction EE Gaurav Hansda

Fast Decision of Block size, Prediction Mode and Intra Block for H.264 Intra Prediction EE Gaurav Hansda Fast Decision of Block size, Prediction Mode and Intra Block for H.264 Intra Prediction EE 5359 Gaurav Hansda 1000721849 gaurav.hansda@mavs.uta.edu Outline Introduction to H.264 Current algorithms for

More information

The Video Z-buffer: A Concept for Facilitating Monoscopic Image Compression by exploiting the 3-D Stereoscopic Depth map

The Video Z-buffer: A Concept for Facilitating Monoscopic Image Compression by exploiting the 3-D Stereoscopic Depth map The Video Z-buffer: A Concept for Facilitating Monoscopic Image Compression by exploiting the 3-D Stereoscopic Depth map Sriram Sethuraman 1 and M. W. Siegel 2 1 David Sarnoff Research Center, Princeton,

More information

ERROR-ROBUST INTER/INTRA MACROBLOCK MODE SELECTION USING ISOLATED REGIONS

ERROR-ROBUST INTER/INTRA MACROBLOCK MODE SELECTION USING ISOLATED REGIONS ERROR-ROBUST INTER/INTRA MACROBLOCK MODE SELECTION USING ISOLATED REGIONS Ye-Kui Wang 1, Miska M. Hannuksela 2 and Moncef Gabbouj 3 1 Tampere International Center for Signal Processing (TICSP), Tampere,

More information

Advances of MPEG Scalable Video Coding Standard

Advances of MPEG Scalable Video Coding Standard Advances of MPEG Scalable Video Coding Standard Wen-Hsiao Peng, Chia-Yang Tsai, Tihao Chiang, and Hsueh-Ming Hang National Chiao-Tung University 1001 Ta-Hsueh Rd., HsinChu 30010, Taiwan pawn@mail.si2lab.org,

More information

RECOMMENDATION ITU-R BT.1720 *

RECOMMENDATION ITU-R BT.1720 * Rec. ITU-R BT.1720 1 RECOMMENDATION ITU-R BT.1720 * Quality of service ranking and measurement methods for digital video broadcasting services delivered over broadband Internet protocol networks (Question

More information

Interframe coding A video scene captured as a sequence of frames can be efficiently coded by estimating and compensating for motion between frames pri

Interframe coding A video scene captured as a sequence of frames can be efficiently coded by estimating and compensating for motion between frames pri MPEG MPEG video is broken up into a hierarchy of layer From the top level, the first layer is known as the video sequence layer, and is any self contained bitstream, for example a coded movie. The second

More information

Digital Image Stabilization and Its Integration with Video Encoder

Digital Image Stabilization and Its Integration with Video Encoder Digital Image Stabilization and Its Integration with Video Encoder Yu-Chun Peng, Hung-An Chang, Homer H. Chen Graduate Institute of Communication Engineering National Taiwan University Taipei, Taiwan {b889189,

More information

9/8/2016. Characteristics of multimedia Various media types

9/8/2016. Characteristics of multimedia Various media types Chapter 1 Introduction to Multimedia Networking CLO1: Define fundamentals of multimedia networking Upon completion of this chapter students should be able to define: 1- Multimedia 2- Multimedia types and

More information

OPTIMIZATION OF LOW DELAY WAVELET VIDEO CODECS

OPTIMIZATION OF LOW DELAY WAVELET VIDEO CODECS OPTIMIZATION OF LOW DELAY WAVELET VIDEO CODECS Andrzej Popławski, Marek Domański 2 Uniwersity of Zielona Góra, Institute of Computer Engineering and Electronics, Poland 2 Poznań University of Technology,

More information

Module 10 MULTIMEDIA SYNCHRONIZATION

Module 10 MULTIMEDIA SYNCHRONIZATION Module 10 MULTIMEDIA SYNCHRONIZATION Lesson 36 Packet architectures and audio-video interleaving Instructional objectives At the end of this lesson, the students should be able to: 1. Show the packet architecture

More information

Video Transcoding Architectures and Techniques: An Overview. IEEE Signal Processing Magazine March 2003 Present by Chen-hsiu Huang

Video Transcoding Architectures and Techniques: An Overview. IEEE Signal Processing Magazine March 2003 Present by Chen-hsiu Huang Video Transcoding Architectures and Techniques: An Overview IEEE Signal Processing Magazine March 2003 Present by Chen-hsiu Huang Outline Background & Introduction Bit-rate Reduction Spatial Resolution

More information

MPEG: It s Need, Evolution and Processing Methods

MPEG: It s Need, Evolution and Processing Methods MPEG: It s Need, Evolution and Processing Methods Ankit Agarwal, Prateeksha Suwalka, Manohar Prajapati ECE DEPARTMENT, Baldev Ram mirdha institute of technology (EC) ITS- 3,EPIP SItapura, Jaipur-302022(India)

More information

Streaming Technologies Glossary

Streaming Technologies Glossary Streaming Technologies Glossary A ACELP.net ADSL AIFF Artifact Asf Asx Avi B Bitrate C C Capture card Codec CIF (QCIF) Client Industry standard audio compression technology developed by VoiceAge Corporation.

More information

Chapter 11: Understanding the H.323 Standard

Chapter 11: Understanding the H.323 Standard Página 1 de 7 Chapter 11: Understanding the H.323 Standard This chapter contains information about the H.323 standard and its architecture, and discusses how Microsoft Windows NetMeeting supports H.323

More information

Bluray (

Bluray ( Bluray (http://www.blu-ray.com/faq) MPEG-2 - enhanced for HD, also used for playback of DVDs and HDTV recordings MPEG-4 AVC - part of the MPEG-4 standard also known as H.264 (High Profile and Main Profile)

More information

Video Redundancy Coding in H.263+ Stephan Wenger Technische Universität Berlin

Video Redundancy Coding in H.263+ Stephan Wenger Technische Universität Berlin Video Redundancy Coding in H.263+ Stephan Wenger Technische Universität Berlin stewe@cs.tu-berlin.de ABSTRACT: The forthcoming new version of ITU- T s advanced video compression recommendation H.263 [1]

More information

DIGITAL TELEVISION 1. DIGITAL VIDEO FUNDAMENTALS

DIGITAL TELEVISION 1. DIGITAL VIDEO FUNDAMENTALS DIGITAL TELEVISION 1. DIGITAL VIDEO FUNDAMENTALS Television services in Europe currently broadcast video at a frame rate of 25 Hz. Each frame consists of two interlaced fields, giving a field rate of 50

More information

Fast Wavelet-based Macro-block Selection Algorithm for H.264 Video Codec

Fast Wavelet-based Macro-block Selection Algorithm for H.264 Video Codec Proceedings of the International MultiConference of Engineers and Computer Scientists 8 Vol I IMECS 8, 19-1 March, 8, Hong Kong Fast Wavelet-based Macro-block Selection Algorithm for H.64 Video Codec Shi-Huang

More information

2 RTP Encapsulation and its Application in NS-2 Simulation

2 RTP Encapsulation and its Application in NS-2 Simulation 3rd International Conference on Multimedia Technology(ICMT 2013) RTP Encapsulation for Scalable Video Stream and its Application in NS-2 Simulation Zhou Ying, Zhang Jihong, Liu Wei Abstract. Real-time

More information

The RTP Encapsulation based on Frame Type Method for AVS Video

The RTP Encapsulation based on Frame Type Method for AVS Video Applied Mechanics and Materials Online: 2012-12-27 ISSN: 1662-7482, Vols. 263-266, pp 1803-1808 doi:10.4028/www.scientific.net/amm.263-266.1803 2013 Trans Tech Publications, Switzerland The RTP Encapsulation

More information

Video coding. Concepts and notations.

Video coding. Concepts and notations. TSBK06 video coding p.1/47 Video coding Concepts and notations. A video signal consists of a time sequence of images. Typical frame rates are 24, 25, 30, 50 and 60 images per seconds. Each image is either

More information

CODING METHOD FOR EMBEDDING AUDIO IN VIDEO STREAM. Harri Sorokin, Jari Koivusaari, Moncef Gabbouj, and Jarmo Takala

CODING METHOD FOR EMBEDDING AUDIO IN VIDEO STREAM. Harri Sorokin, Jari Koivusaari, Moncef Gabbouj, and Jarmo Takala CODING METHOD FOR EMBEDDING AUDIO IN VIDEO STREAM Harri Sorokin, Jari Koivusaari, Moncef Gabbouj, and Jarmo Takala Tampere University of Technology Korkeakoulunkatu 1, 720 Tampere, Finland ABSTRACT In

More information

Automatic Video Caption Detection and Extraction in the DCT Compressed Domain

Automatic Video Caption Detection and Extraction in the DCT Compressed Domain Automatic Video Caption Detection and Extraction in the DCT Compressed Domain Chin-Fu Tsao 1, Yu-Hao Chen 1, Jin-Hau Kuo 1, Chia-wei Lin 1, and Ja-Ling Wu 1,2 1 Communication and Multimedia Laboratory,

More information

Locating 1-D Bar Codes in DCT-Domain

Locating 1-D Bar Codes in DCT-Domain Edith Cowan University Research Online ECU Publications Pre. 2011 2006 Locating 1-D Bar Codes in DCT-Domain Alexander Tropf Edith Cowan University Douglas Chai Edith Cowan University 10.1109/ICASSP.2006.1660449

More information

RTP. Prof. C. Noronha RTP. Real-Time Transport Protocol RFC 1889

RTP. Prof. C. Noronha RTP. Real-Time Transport Protocol RFC 1889 RTP Real-Time Transport Protocol RFC 1889 1 What is RTP? Primary objective: stream continuous media over a best-effort packet-switched network in an interoperable way. Protocol requirements: Payload Type

More information

Copyright Notice. Springer papers: Springer. Pre-prints are provided only for personal use. The final publication is available at link.springer.

Copyright Notice. Springer papers: Springer. Pre-prints are provided only for personal use. The final publication is available at link.springer. Copyright Notice The document is provided by the contributing author(s) as a means to ensure timely dissemination of scholarly and technical work on a non-commercial basis. This is the author s version

More information

ISO/IEC INTERNATIONAL STANDARD. Information technology Coding of audio-visual objects Part 18: Font compression and streaming

ISO/IEC INTERNATIONAL STANDARD. Information technology Coding of audio-visual objects Part 18: Font compression and streaming INTERNATIONAL STANDARD ISO/IEC 14496-18 First edition 2004-07-01 Information technology Coding of audio-visual objects Part 18: Font compression and streaming Technologies de l'information Codage des objets

More information

Alkit Reflex RTP reflector/mixer

Alkit Reflex RTP reflector/mixer Alkit Reflex RTP reflector/mixer Mathias Johanson, Ph.D. Alkit Communications Introduction Real time audio and video communication over IP networks is attracting a lot of interest for applications like

More information

JPEG 2000 vs. JPEG in MPEG Encoding

JPEG 2000 vs. JPEG in MPEG Encoding JPEG 2000 vs. JPEG in MPEG Encoding V.G. Ruiz, M.F. López, I. García and E.M.T. Hendrix Dept. Computer Architecture and Electronics University of Almería. 04120 Almería. Spain. E-mail: vruiz@ual.es, mflopez@ace.ual.es,

More information

INTERNATIONAL ORGANISATION FOR STANDARDISATION ORGANISATION INTERNATIONALE DE NORMALISATION ISO/IEC JTC1/SC29/WG11 CODING OF MOVING PICTURES AND AUDIO

INTERNATIONAL ORGANISATION FOR STANDARDISATION ORGANISATION INTERNATIONALE DE NORMALISATION ISO/IEC JTC1/SC29/WG11 CODING OF MOVING PICTURES AND AUDIO INTERNATIONAL ORGANISATION FOR STANDARDISATION ORGANISATION INTERNATIONALE DE NORMALISATION ISO/IEC JTC1/SC29/WG11 CODING OF MOVING PICTURES AND AUDIO ISO/IEC JTC1/SC29 WG11 N15073 February 2015, Geneva,

More information

Optical Storage Technology. MPEG Data Compression

Optical Storage Technology. MPEG Data Compression Optical Storage Technology MPEG Data Compression MPEG-1 1 Audio Standard Moving Pictures Expert Group (MPEG) was formed in 1988 to devise compression techniques for audio and video. It first devised the

More information

Multimedia Applications over Packet Networks

Multimedia Applications over Packet Networks Multimedia Networking and Quality of Service Mario Baldi Technical Univeristy of Torino Computer Engineering Department mario.baldi@polito.it +39 011 564 7067 staff.polito.it/mario.baldi Nota di Copyright

More information

Visionair. TNA - Collaborative remote visualization ENSAM URJC. Miguel Ramos García. Universidad Rey Juan Carlos, Móstoles (Spain)

Visionair. TNA - Collaborative remote visualization ENSAM URJC. Miguel Ramos García. Universidad Rey Juan Carlos, Móstoles (Spain) Visionair TNA - Collaborative remote visualization ENSAM URJC Miguel Ramos García Universidad Rey Juan Carlos, Móstoles (Spain) Index Introduction... 3 1 Framework, the first contact.... 4 1.1 Information

More information

RTP/RTCP protocols. Introduction: What are RTP and RTCP?

RTP/RTCP protocols. Introduction: What are RTP and RTCP? RTP/RTCP protocols Introduction: What are RTP and RTCP? The spread of computers, added to the availability of cheap audio/video computer hardware, and the availability of higher connection speeds have

More information

MPEG-4 Structured Audio Systems

MPEG-4 Structured Audio Systems MPEG-4 Structured Audio Systems Mihir Anandpara The University of Texas at Austin anandpar@ece.utexas.edu 1 Abstract The MPEG-4 standard has been proposed to provide high quality audio and video content

More information

In the name of Allah. the compassionate, the merciful

In the name of Allah. the compassionate, the merciful In the name of Allah the compassionate, the merciful Digital Video Systems S. Kasaei Room: CE 315 Department of Computer Engineering Sharif University of Technology E-Mail: skasaei@sharif.edu Webpage:

More information

over the Internet Tihao Chiang { Ya-Qin Zhang k enormous interests from both industry and academia.

over the Internet Tihao Chiang { Ya-Qin Zhang k enormous interests from both industry and academia. An End-to-End Architecture for MPEG-4 Video Streaming over the Internet Y. Thomas Hou Dapeng Wu y Wenwu Zhu z Hung-Ju Lee x Tihao Chiang { Ya-Qin Zhang k Abstract It is a challenging problem to design

More information

INTERNATIONAL ORGANISATION FOR STANDARDISATION ORGANISATION INTERNATIONALE DE NORMALISATION ISO/IEC JTC1/SC29/WG11 CODING OF MOVING PICTURES AND AUDIO

INTERNATIONAL ORGANISATION FOR STANDARDISATION ORGANISATION INTERNATIONALE DE NORMALISATION ISO/IEC JTC1/SC29/WG11 CODING OF MOVING PICTURES AND AUDIO INTERNATIONAL ORGANISATION FOR STANDARDISATION ORGANISATION INTERNATIONALE DE NORMALISATION ISO/IEC JTC1/SC29/WG11 CODING OF MOVING PICTURES AND AUDIO ISO/IEC JTC1/SC29/WG11 N15071 February 2015, Geneva,

More information

Optimal Estimation for Error Concealment in Scalable Video Coding

Optimal Estimation for Error Concealment in Scalable Video Coding Optimal Estimation for Error Concealment in Scalable Video Coding Rui Zhang, Shankar L. Regunathan and Kenneth Rose Department of Electrical and Computer Engineering University of California Santa Barbara,

More information

ISO/IEC TR TECHNICAL REPORT. Information technology Coding of audio-visual objects Part 24: Audio and systems interaction

ISO/IEC TR TECHNICAL REPORT. Information technology Coding of audio-visual objects Part 24: Audio and systems interaction TECHNICAL REPORT ISO/IEC TR 14496-24 First edition 2008-01-15 Information technology Coding of audio-visual objects Part 24: Audio and systems interaction Technologies de l'information Codage d'objets

More information

FRAME-RATE UP-CONVERSION USING TRANSMITTED TRUE MOTION VECTORS

FRAME-RATE UP-CONVERSION USING TRANSMITTED TRUE MOTION VECTORS FRAME-RATE UP-CONVERSION USING TRANSMITTED TRUE MOTION VECTORS Yen-Kuang Chen 1, Anthony Vetro 2, Huifang Sun 3, and S. Y. Kung 4 Intel Corp. 1, Mitsubishi Electric ITA 2 3, and Princeton University 1

More information

MPEG-2. ISO/IEC (or ITU-T H.262)

MPEG-2. ISO/IEC (or ITU-T H.262) MPEG-2 1 MPEG-2 ISO/IEC 13818-2 (or ITU-T H.262) High quality encoding of interlaced video at 4-15 Mbps for digital video broadcast TV and digital storage media Applications Broadcast TV, Satellite TV,

More information

A Transport Infrastructure Supporting Real Time Interactive MPEG-4 Client-Server Applications over IP Networks

A Transport Infrastructure Supporting Real Time Interactive MPEG-4 Client-Server Applications over IP Networks A Transport Infrastructure Supporting Real Time Interactive MPEG-4 Client-Server Applications over IP Networks Haining Liu, Xiaoping Wei, and Magda El Zarki Department of Information and Computer Science

More information

Format-Independent Multimedia Streaming

Format-Independent Multimedia Streaming University of Wollongong Research Online Faculty of Informatics - Papers Faculty of Informatics 2006 Format-Independent Multimedia Streaming J. Thomas-Kerr University of Wollongong I. Burnett University

More information

Digital video coding systems MPEG-1/2 Video

Digital video coding systems MPEG-1/2 Video Digital video coding systems MPEG-1/2 Video Introduction What is MPEG? Moving Picture Experts Group Standard body for delivery of video and audio. Part of ISO/IEC/JTC1/SC29/WG11 150 companies & research

More information

MITIGATING THE EFFECT OF PACKET LOSSES ON REAL-TIME VIDEO STREAMING USING PSNR AS VIDEO QUALITY ASSESSMENT METRIC ABSTRACT

MITIGATING THE EFFECT OF PACKET LOSSES ON REAL-TIME VIDEO STREAMING USING PSNR AS VIDEO QUALITY ASSESSMENT METRIC ABSTRACT MITIGATING THE EFFECT OF PACKET LOSSES ON REAL-TIME VIDEO STREAMING USING PSNR AS VIDEO QUALITY ASSESSMENT METRIC Anietie Bassey, Kufre M. Udofia & Mfonobong C. Uko Department of Electrical/Electronic

More information

Multimedia Systems Video II (Video Coding) Mahdi Amiri April 2012 Sharif University of Technology

Multimedia Systems Video II (Video Coding) Mahdi Amiri April 2012 Sharif University of Technology Course Presentation Multimedia Systems Video II (Video Coding) Mahdi Amiri April 2012 Sharif University of Technology Video Coding Correlation in Video Sequence Spatial correlation Similar pixels seem

More information

Spatial Scene Level Shape Error Concealment for Segmented Video

Spatial Scene Level Shape Error Concealment for Segmented Video Spatial Scene Level Shape Error Concealment for Segmented Video Luis Ducla Soares 1, Fernando Pereira 2 1 Instituto Superior de Ciências do Trabalho e da Empresa Instituto de Telecomunicações, Lisboa,

More information

TRIBHUVAN UNIVERSITY Institute of Engineering Pulchowk Campus Department of Electronics and Computer Engineering

TRIBHUVAN UNIVERSITY Institute of Engineering Pulchowk Campus Department of Electronics and Computer Engineering TRIBHUVAN UNIVERSITY Institute of Engineering Pulchowk Campus Department of Electronics and Computer Engineering A Final project Report ON Minor Project Java Media Player Submitted By Bisharjan Pokharel(061bct512)

More information

Lecture 3 Image and Video (MPEG) Coding

Lecture 3 Image and Video (MPEG) Coding CS 598KN Advanced Multimedia Systems Design Lecture 3 Image and Video (MPEG) Coding Klara Nahrstedt Fall 2017 Overview JPEG Compression MPEG Basics MPEG-4 MPEG-7 JPEG COMPRESSION JPEG Compression 8x8 blocks

More information

Tema 0: Transmisión de Datos Multimedia

Tema 0: Transmisión de Datos Multimedia Tema 0: Transmisión de Datos Multimedia Clases de aplicaciones multimedia Redes basadas en IP y QoS Computer Networking: A Top Down Approach Featuring the Internet, 3 rd edition. Jim Kurose, Keith Ross

More information

White paper: Video Coding A Timeline

White paper: Video Coding A Timeline White paper: Video Coding A Timeline Abharana Bhat and Iain Richardson June 2014 Iain Richardson / Vcodex.com 2007-2014 About Vcodex Vcodex are world experts in video compression. We provide essential

More information

Improving student feedback in virtual classrooms

Improving student feedback in virtual classrooms Improving student feedback in virtual classrooms M. van der Schyff, H.C. Ferreira Department of Electrical and Electronic Engineering Rand Afrikaans University, South Africa mvds@ing.rau.ac.za, hcf@ing.rau.ac.za

More information

Parallel Implementation of Arbitrary-Shaped MPEG-4 Decoder for Multiprocessor Systems

Parallel Implementation of Arbitrary-Shaped MPEG-4 Decoder for Multiprocessor Systems Parallel Implementation of Arbitrary-Shaped MPEG-4 oder for Multiprocessor Systems Milan Pastrnak *,a,c, Peter H.N. de With a,c, Sander Stuijk c and Jef van Meerbergen b,c a LogicaCMG Nederland B.V., RTSE

More information

ADAPTIVE PICTURE SLICING FOR DISTORTION-BASED CLASSIFICATION OF VIDEO PACKETS

ADAPTIVE PICTURE SLICING FOR DISTORTION-BASED CLASSIFICATION OF VIDEO PACKETS ADAPTIVE PICTURE SLICING FOR DISTORTION-BASED CLASSIFICATION OF VIDEO PACKETS E. Masala, D. Quaglia, J.C. De Martin Λ Dipartimento di Automatica e Informatica/ Λ IRITI-CNR Politecnico di Torino, Italy

More information

Multimedia Networking and Quality of Service

Multimedia Networking and Quality of Service Multimedia Networking and Quality of Service Mario Baldi Politecnico di Torino (Technical Univeristy of Torino) Department of Computer Engineering mario.baldi [at] polito.it +39 011 564 7067 staff.polito.it/mario.baldi

More information

QUANTIZER DESIGN FOR EXPLOITING COMMON INFORMATION IN LAYERED CODING. Mehdi Salehifar, Tejaswi Nanjundaswamy, and Kenneth Rose

QUANTIZER DESIGN FOR EXPLOITING COMMON INFORMATION IN LAYERED CODING. Mehdi Salehifar, Tejaswi Nanjundaswamy, and Kenneth Rose QUANTIZER DESIGN FOR EXPLOITING COMMON INFORMATION IN LAYERED CODING Mehdi Salehifar, Tejaswi Nanjundaswamy, and Kenneth Rose Department of Electrical and Computer Engineering University of California,

More information

Application and Desktop Sharing. Omer Boyaci November 1, 2007

Application and Desktop Sharing. Omer Boyaci November 1, 2007 Application and Desktop Sharing Omer Boyaci November 1, 2007 Overview Introduction Demo Architecture Challenges Features Conclusion Application Sharing Models Application specific + Efficient - Participants

More information

Figure 1: Representation of moving images using layers Once a set of ane models has been found, similar models are grouped based in a mean-square dist

Figure 1: Representation of moving images using layers Once a set of ane models has been found, similar models are grouped based in a mean-square dist ON THE USE OF LAYERS FOR VIDEO CODING AND OBJECT MANIPULATION Luis Torres, David Garca and Anna Mates Dept. of Signal Theory and Communications Universitat Politecnica de Catalunya Gran Capita s/n, D5

More information

implementation using GPU architecture is implemented only from the viewpoint of frame level parallel encoding [6]. However, it is obvious that the mot

implementation using GPU architecture is implemented only from the viewpoint of frame level parallel encoding [6]. However, it is obvious that the mot Parallel Implementation Algorithm of Motion Estimation for GPU Applications by Tian Song 1,2*, Masashi Koshino 2, Yuya Matsunohana 2 and Takashi Shimamoto 1,2 Abstract The video coding standard H.264/AVC

More information

A COST-EFFICIENT RESIDUAL PREDICTION VLSI ARCHITECTURE FOR H.264/AVC SCALABLE EXTENSION

A COST-EFFICIENT RESIDUAL PREDICTION VLSI ARCHITECTURE FOR H.264/AVC SCALABLE EXTENSION A COST-EFFICIENT RESIDUAL PREDICTION VLSI ARCHITECTURE FOR H.264/AVC SCALABLE EXTENSION Yi-Hau Chen, Tzu-Der Chuang, Chuan-Yung Tsai, Yu-Jen Chen, and Liang-Gee Chen DSP/IC Design Lab., Graduate Institute

More information

International Journal of Emerging Technology and Advanced Engineering Website: (ISSN , Volume 2, Issue 4, April 2012)

International Journal of Emerging Technology and Advanced Engineering Website:   (ISSN , Volume 2, Issue 4, April 2012) A Technical Analysis Towards Digital Video Compression Rutika Joshi 1, Rajesh Rai 2, Rajesh Nema 3 1 Student, Electronics and Communication Department, NIIST College, Bhopal, 2,3 Prof., Electronics and

More information

QoE Characterization for Video-On-Demand Services in 4G WiMAX Networks

QoE Characterization for Video-On-Demand Services in 4G WiMAX Networks QoE Characterization for Video-On-Demand Services in 4G WiMAX Networks Amitabha Ghosh IBM India Research Laboratory Department of Electrical Engineering University of Southern California, Los Angeles http://anrg.usc.edu/~amitabhg

More information

Module 10 MULTIMEDIA SYNCHRONIZATION

Module 10 MULTIMEDIA SYNCHRONIZATION Module 10 MULTIMEDIA SYNCHRONIZATION Lesson 33 Basic definitions and requirements Instructional objectives At the end of this lesson, the students should be able to: 1. Define synchronization between media

More information

IO [io] MAYAH. IO [io] Audio Video Codec Systems

IO [io] MAYAH. IO [io] Audio Video Codec Systems IO [io] MAYAH IO [io] Audio Video Codec Systems MPEG 4 Audio Video Embedded 24/7 Real-Time Solution MPEG 4 Audio Video Production and Streaming Solution ISMA compliant 24/7 Audio Video Realtime Solution

More information

Rate Distortion Optimization in Video Compression

Rate Distortion Optimization in Video Compression Rate Distortion Optimization in Video Compression Xue Tu Dept. of Electrical and Computer Engineering State University of New York at Stony Brook 1. Introduction From Shannon s classic rate distortion

More information

Lecture 7, Video Coding, Motion Compensation Accuracy

Lecture 7, Video Coding, Motion Compensation Accuracy Lecture 7, Video Coding, Motion Compensation Accuracy Last time we saw several methods to obtain a good motion estimation, with reduced complexity (efficient search), and with the possibility of sub-pixel

More information

Video Compression MPEG-4. Market s requirements for Video compression standard

Video Compression MPEG-4. Market s requirements for Video compression standard Video Compression MPEG-4 Catania 10/04/2008 Arcangelo Bruna Market s requirements for Video compression standard Application s dependent Set Top Boxes (High bit rate) Digital Still Cameras (High / mid

More information

Advanced Video Coding: The new H.264 video compression standard

Advanced Video Coding: The new H.264 video compression standard Advanced Video Coding: The new H.264 video compression standard August 2003 1. Introduction Video compression ( video coding ), the process of compressing moving images to save storage space and transmission

More information

IST MPEG-4 Video Compliant Framework

IST MPEG-4 Video Compliant Framework IST MPEG-4 Video Compliant Framework João Valentim, Paulo Nunes, Fernando Pereira Instituto de Telecomunicações, Instituto Superior Técnico, Av. Rovisco Pais, 1049-001 Lisboa, Portugal Abstract This paper

More information

3GPP TS V5.2.0 ( )

3GPP TS V5.2.0 ( ) TS 26.140 V5.2.0 (2002-12) Technical Specification 3rd Generation Partnership Project; Technical Specification Group Services and System Aspects; Multimedia Messaging Service (MMS); Media formats and codecs

More information

Comparison of Shaping and Buffering for Video Transmission

Comparison of Shaping and Buffering for Video Transmission Comparison of Shaping and Buffering for Video Transmission György Dán and Viktória Fodor Royal Institute of Technology, Department of Microelectronics and Information Technology P.O.Box Electrum 229, SE-16440

More information

Information technology - Coding of audiovisual objects - Part 2: Visual

Information technology - Coding of audiovisual objects - Part 2: Visual This is a preview of "INCITS/ISO/IEC 14496...". Click here to purchase the full version from the ANSI store. INCITS/ISO/IEC 14496-2:2004[R2012] (ISO/IEC 14496-2:2004, IDT) Information technology - Coding

More information

VIDEO streaming applications over the Internet are gaining. Brief Papers

VIDEO streaming applications over the Internet are gaining. Brief Papers 412 IEEE TRANSACTIONS ON BROADCASTING, VOL. 54, NO. 3, SEPTEMBER 2008 Brief Papers Redundancy Reduction Technique for Dual-Bitstream MPEG Video Streaming With VCR Functionalities Tak-Piu Ip, Yui-Lam Chan,

More information

Content Based Image Retrieval Using Color Quantizes, EDBTC and LBP Features

Content Based Image Retrieval Using Color Quantizes, EDBTC and LBP Features Content Based Image Retrieval Using Color Quantizes, EDBTC and LBP Features 1 Kum Sharanamma, 2 Krishnapriya Sharma 1,2 SIR MVIT Abstract- To describe the image features the Local binary pattern (LBP)

More information

Request for Comments: 2032 Category: Standards Track Bellcore October 1996

Request for Comments: 2032 Category: Standards Track Bellcore October 1996 Network Working Group Request for Comments: 2032 Category: Standards Track T. Turletti MIT C. Huitema Bellcore October 1996 Status of this Memo RTP Payload Format for H.261 Video Streams This document

More information

signal-to-noise ratio (PSNR), 2

signal-to-noise ratio (PSNR), 2 u m " The Integration in Optics, Mechanics, and Electronics of Digital Versatile Disc Systems (1/3) ---(IV) Digital Video and Audio Signal Processing ƒf NSC87-2218-E-009-036 86 8 1 --- 87 7 31 p m o This

More information

STACK ROBUST FINE GRANULARITY SCALABLE VIDEO CODING

STACK ROBUST FINE GRANULARITY SCALABLE VIDEO CODING Journal of the Chinese Institute of Engineers, Vol. 29, No. 7, pp. 1203-1214 (2006) 1203 STACK ROBUST FINE GRANULARITY SCALABLE VIDEO CODING Hsiang-Chun Huang and Tihao Chiang* ABSTRACT A novel scalable

More information

DATA HIDING IN PDF FILES AND APPLICATIONS BY IMPERCEIVABLE MODIFICATIONS OF PDF OBJECT PARAMETERS

DATA HIDING IN PDF FILES AND APPLICATIONS BY IMPERCEIVABLE MODIFICATIONS OF PDF OBJECT PARAMETERS DATA HIDING IN PDF FILES AND APPLICATIONS BY IMPERCEIVABLE MODIFICATIONS OF PDF OBJECT PARAMETERS 1 Jiun-Tsung Wang ( 王竣聰 ) and 2 Wen-Hsiang Tsai ( 蔡文祥 ) 1 Institute of Multimedia Eng., National Chiao

More information

MpegRepair Software Encoding and Repair Utility

MpegRepair Software Encoding and Repair Utility PixelTools MpegRepair Software Encoding and Repair Utility MpegRepair integrates fully featured encoding, analysis, decoding, demuxing, transcoding and stream manipulations into one powerful application.

More information

Networking Applications

Networking Applications Networking Dr. Ayman A. Abdel-Hamid College of Computing and Information Technology Arab Academy for Science & Technology and Maritime Transport Multimedia Multimedia 1 Outline Audio and Video Services

More information

Stereo Vision Image Processing Strategy for Moving Object Detecting

Stereo Vision Image Processing Strategy for Moving Object Detecting Stereo Vision Image Processing Strategy for Moving Object Detecting SHIUH-JER HUANG, FU-REN YING Department of Mechanical Engineering National Taiwan University of Science and Technology No. 43, Keelung

More information

Comparative Study of Partial Closed-loop Versus Open-loop Motion Estimation for Coding of HDTV

Comparative Study of Partial Closed-loop Versus Open-loop Motion Estimation for Coding of HDTV Comparative Study of Partial Closed-loop Versus Open-loop Motion Estimation for Coding of HDTV Jeffrey S. McVeigh 1 and Siu-Wai Wu 2 1 Carnegie Mellon University Department of Electrical and Computer Engineering

More information

RTP for Application Sharing Payload Format Extensions

RTP for Application Sharing Payload Format Extensions [MS-RTASPF]: Intellectual Property Rights Notice for Open Specifications Documentation Technical Documentation. Microsoft publishes Open Specifications documentation for protocols, file formats, languages,

More information