A MULTIPOINT VIDEOCONFERENCE RECEIVER BASED ON MPEG-4 OBJECT VIDEO. Chih-Kai Chien, Chen-Yu Tsai, and David W. Lin
|
|
- Lesley Brooks
- 5 years ago
- Views:
Transcription
1 A MULTIPOINT VIDEOCONFERENCE RECEIVER BASED ON MPEG-4 OBJECT VIDEO Chih-Kai Chien, Chen-Yu Tsai, and David W. Lin Dept. of Electronics Engineering and Center for Telecommunications Research National Chiao Tung University Hsinchu, Taiwan 30010, R.O.C. s: ABSTRACT We consider the design and implementation of a novel type of software-based multipoint videoconference receiver on a personal computer (PC), where some distinguishing features are that MPEG-4 object-based coding is used in encoding each video stream and that the decoded videos are composed into one for display. The resulting receiver includes an RTP-based network interface, a set of MPEG-4 video and AAC audio decoders (whose number depends on number of source sites), and a unit to compose the decoded media streams for display. We develop a graphical user interface using the Windows SDK for convenience in system control and monitoring as well as display of results. With an AMD CPU running at 2.1-GHz with 512-MB RAM, the current un-optimized implementation yields a speed on the order of 10 frames per second for CIF ( ) video when receiving from only one source site. The frame rate reduces approximately in proportion to number of sources. 1. INTRODUCTION In a typical multipoint videoconference system, the receiver places the decoded videos from different sources in different windows. We consider constructing a different kind of system in which the decoded videos are composed into a virtual conference room scene. For this, a natural and simplest approach is to segment and encode the source videos separately at their respective transmitter sites, and let each receiver decode all received videos and compose and display the result. The MPEG-4 standards, with their provision for object-based video coding, appear naturally fitting for this use. In this work, we consider the design and implementation of the receiver on a personal computer (PC), in software. There are four major components in the receiver: the network interface, the video decoder, the audio decoder, and the composition unit, as illustrated in Fig. 1. As will be Work supported by National Science Council of R.O.C. under Grant NSC E Fig. 1. Structure of the proposed videoconference receiver. explained below, we decide to use the Real-time Transport Protocol (RTP) for the network interface and develop our own composition method, leaving the video and the audio encoded and decoded according to MPEG-4 specifications. In what follows, Section 2 gives an overview of the (earliest) MPEG-4 standards and comments on the usefulness of each part in our work. Section 3 discusses how we decode and compose multiple videos into one scene. Section 4 describes the integration of the receiver system. Some experimental results are described in Section 5. Finally, Section 6 contains the conclusion. 2. THE MPEG-4 STANDARDS AND THEIR APPLICATION IN THIS WORK The original MPEG-4 standards are divided into four basic parts: Systems (ISO/IEC ), Visual (ISO/IEC ), Audio (ISO/IEC ), and Delivery Multimedia Integration Framework, or DMIF in short (ISO/IEC ) [1], [2]. The MPEG-4 Systems part specifies how audio-visual
2 Fig. 2. Relation between the two decoders in time domain. scenes can be created from individual objects. To actually create the audio-visual scenes for a particular application, one may employ a suitable authoring tool designed to the MPEG-4 specifications. Unfortunately, no suitable authoring tools are found for our application in the course of this work. Hence we decide to develop a simple composition method and write our own program for it. Aside from other innovations, a major novelty in the MPEG-4 Visual part is the provision for coding of arbitraryshaped objects. In fact, each picture is considered as composed of a number of video objects. For each object, a socalled alpha-plane image sequence defines the support of the object in each video frame as well as the transparency of each pixel therein. Pixels belonging to the object are encoded largely employing typical motion-compensated block- DCT techniques; and so are the alpha-plane images. These features are used in our implementation. The MPEG-4 Audio part is rather generic in that it considers different kinds of audio signal (e.g., speech, natural sound, and synthetic sound) and different kinds of signal manipulation (e.g., compression and synthesis). As a result, it facilitates object-based audio coding and manipulation. Our system merely makes use of its compression functionality. We employ a publicly available MPEG-4 AAC (Advanced Audio Coding) decoder in this work. The MPEG-4 DMIF part specifies the interface between the application and the transport, so that the transport can be relatively transparent to the application and the application can be developed largely free from concerns over the transport. Despite the intent, recent MPEG activities show waned interest in DMIF and existing DMIF software, such as [3], has suffered functionality and maintenance problems. Therefore, we decide to use RTP [4] for the network interface. 3. VIDEO DECODING AND COMPOSITION As discussed previously, we develop a simple video composition method in this work. This involves not only composition of multiple videos into one scene, but also synchronization among the multiple decoders. For simplicity, we Fig. 3. Different spatial relations between two images. (a) Case 0, no overlap. (b) (e) Cases 1 4, respectively, with overlaps. assume that all videos have the same frame rate and relegate the situation with disparate frame rates to potential future research. Note, however, that in the latter situation, we may consider skipping some frames in the videos with higher frame rates, which should not cause much problem in videoconferencing when the frame rates are high enough. In this case, only minor modifications are needed in the present program. We first consider the situation with two videos and two decoders. The two decoders are placed in two threads on the PC. To synchronize the decoded videos, for each frame we let the first decoder wait until the second decoder completes. Then it starts working on the next frame. Composition and display of the videos are done in the second thread. Figure 2 shows the temporal relation between the two decoders. To compose the two images decoded by the two decoders, note that there are several possible spatial relations between them, as illustrated in Fig. 3. Case 0 is where there is no overlap between the two images. If there is some overlap between the two images, we let the first image occlude the second in the overlapped area. In the integrated receiver system, we let the user determine and specify where each decoded video is to be placed in the display window.
3 Fig. 4. Composition of four (or three) videos. Now consider the situation with four (or three) videos. Figure 4 explains our method of composition. The outputs of the first and the second decoders are composed together first, and so are the outputs of the third and the fourth decoders. Then the two composed images are composed to yield the final displayed image. The padding referred in the figure is to pad the image into the form of a box for display purpose. For more videos, we simply extend the composition tree. In any case, it is always the highest-indexed video that controls the composition and display operation. 4. RECEIVER SYSTEM INTEGRATION 4.1. Overall System Structure Figure 5 shows how the integrated receiver program works. The GUI block creates a window and the user can input the ports and the positions of the different videos. Since there may be multiple video and audio streams to be handled, we use the multi-threading technique to manage the video and the audio decoders. The multi-threading technique lets the operating system handle the scheduling of the threads. Good scheduling can made efficient use of the PC s available computing, storage, and communication resources. So, after the user has specified the ports and the positions of the videos, the system creates the decoder threads. It also obtains information on which decoders are active at present, so as to determine the highest-indexed video decoder and pass the control responsibility (including video composition and display) to it. Now the video and the audio decoders can begin their work in decoding the data received through the RTP network interface. The composed video is displayed in a window and the composed audio is played. Fig. 5. Flow diagram of the integrated receiver. In the following subsections we give some further details of the system components The RTP Network Interface The RTP facilitates end-to-end delivery of data with realtime characteristics, such as interactive audio and video, and is thus suitable for our application [4]. It supports sequence numbering, timestamping, and delivery monitoring. It can also support multicasting if the underlying network has such capability, which makes it naturally fitted to support multiparty multimedia conferences. However, our implementation does not make use of this feature, neither do we assume that the network has multicasting capability. The RTP is usually run on top of UDP. One should note that RTP itself does not provide quality-of-service guarantees (such as orderly and timely delivery of packets); it is up to the higher layers (such as the user) to provide them, aided by the information provided by the RTP and as far as it is within the ability of the lower-layer services. We use the jrtplib software [5] developed at the Expertise Centre for Digital Media (EDM) to realize the RTP network interface. One RTP session is created for each video or audio stream. The transmitter [6] handles the required RTP packetization. Two important parameters that need to be set for each session are the timestamp and the portbase. The timestamp parameter is set to 1 section per second for a video stream and to 1 section per 4 seconds for an audio stream. (These parameter values are somewhat in-
4 appropriate for real-time applications, but are chosen based on experience for smooth running of the program. Further work is needed to determine the underlying problem and potential solution.) For the portbase, since each transmitter sends two streams (video and audio), the video stream uses a user-specified port number and the audio stream uses that number plus 100. In our system, after the setting of these parameters, we put the receiving of the RTP packets in a main loop. We use the function StartRTP to enable the main loop to receive RTP data and the function ResetRTP to disable the loop to stop receiving RTP data. For some reason yet unclear, the software would lose the first two packets and receive the third and later packets successfully Video Decoding and Display Our video decoder is from the Microsoft MPEG-4 Video Reference Software [7], which is a public source for encoding and decoding of video in MPEG-4 format. We did optimization of the encoder for Intel processors [8], but not the decoder. For our application, we use the binary shape coding feature belonging to main profile of MPEG-4 video. To integrate the video decoder into the overall system, we modify the original decoder program into a function called MPEG4VDecoder and give it two parameters: handle (of the display window) and thread index. The handle can give the decoder some information to control the window and display the video stream. To display the video, we convert the decoder output from the original 4:2:0 format to the 4:4:4 format. Then we calculate the RGB values of each pixel from the luminance and the chrominance values. And then we use the SetPixelV function provided by the Windows SDK library to display the RGB values pixel by pixel. However, experience shows that the SetPixelV function is very slow and can significantly slow down the overall speed of the receiver system. Hence, to reduce its use, instead of using it on all pixels in the display window, we only use it to update the pixels in the object areas of two successive frames Audio Decoding and Composition For audio decoding, we use the Freeware Audio Decoder (FAAD2) [9] written by M. Bakker. The decoder can handle HE, LC, MAIN and LTP profiles, but we only make use of the MAIN profile. After decoding the audio stream received from the RTP network interface, the result is saved as a temporal audio file in the WAV format, which is a common format used in PC audio. The Media Control Interface (MCI), a high level open interface, provides two ways to play WAV-format audio. Since each audio section is four seconds, the decoder Fig. 6. Some typical composed scenes with two videos. would wait for that long as the decoder output is played by the MCI. For audio composition, two intuitive methods are (1) to sum all audio streams and (2) to play only one stream. The first method suffers an overflow problem which can be solved by proper scaling. This is left to potential future work. For simplicity in the final system we only play the audio from the first transmitter site. 5. EXPERIMENTAL RESULTS We present some experimental results in this section. For convenience, for the video part we use the common CIF ( ) test sequence Bream with its associated binary shape information. Multiple instances of the sequence are considered separate video transmissions. Figure 6 shows some typical composed scenes with two videos. Table 1 shows some performance data from employing a PC with an AMD CPU running at 2.1-GHz with 512-MB RAM. The program is not yet optimized. We see that if only one original video decoder is present, the rate can get up to 53.4 frames per second (fps). The processing speed decreases roughly in proportion to the number of decoders, but somewhat slower. This presumably is because the computing resources are used a little more efficiently when there are multiple decoder threads. As explained previously, we pad the decoded video to the 4:4:4 format and convert the result to RGB values for display. As shown in Table 1, the frame rate decreases by about 20% in each case when we add the padding function. Adding video composition and displaying of the result, where the SetPixelV function is used over the whole video frame, reduces the frame rates exceedingly (by one order of magnitude), as can be seen in Table 1 for the cases with
5 Table 1. Processing Rate (in Frames per Second) Under Various Configurations No. of Decoders Configuration Original With Padding With Compos.+Display Only Composition 17.6 Reduced SetPixel With Audio We considered the design and implementation of a novel type of software-based multipoint videoconference receiver on a PC, where some distinguishing features were the use of MPEG-4 object-based coding and the composition of decoded videos into one scene. The resulting receiver included an RTP-based network interface, a set of MPEG-4 video and AAC audio decoders (whose number depends on number of source sites), and a unit to compose the decoded media streams for display. Some topics for potential future research are as follows. 1. Optimization of software components to increase speed and reduce transmission latency: Top on the list should be a faster method to display the composed video. Other components, such as the video and the audio decoders, can also be improved in speed. 2. More elegant audio composition. 3. More sophisticated handling of videos with different frame rates. 4. More elaborate ways of video composition: This may include scaling and three-dimensional rotation of the decoded videos, as well as integration with proper background and foreground. 7. REFERENCES [1] International Committee for Information Technology Standards, Fig. 7. Time analysis of the receiver system. one and two videos. With only video composition but not the display of the result, the frame rate can be much higher, as shown in Table 1 for the case with two videos. Time analysis shows that image display alone takes over 70% of the overall processing time, as illustrated in Fig. 7. Hence, unless we can find more efficient ways to set the display pixels, we should minimize the use of the time-consuming SetPixelV function. After limiting SetPixelV function calls to the video object areas as mentioned briefly previously, we increase the frame rates by about 140%, as shown in Table 1. Finally, we add the audio decoder into the system. From Table 1, we see that the frame rate decreases only by roughly 10% in order of magnitude in each case. Hence the audio decoder has a relatively low complexity. 6. CONCLUSION AND FUTURE WORK [2] MPEG-4 Video Group, MPEG-4 overview (V.21 Jeju Version), doc. no. ISO/IEC JTC1/SC29/WG11 N4668, Mar [3] DMIF software, [4] H. Schulzrine, S. Casner, R. Frederick, and V. Jacobson, RTP: a transport protocol for real-time application, RFC 3550, Network Working Group, July [5] JRTPLIB, [6] C.-Y. Tsai, Integration of videoconference transmitter with MPEG-4 object-based video encoding, M.S. thesis, Dept. of Electronics Engineering, National Chaio Tung University, Hsinchu, Taiwan, R.O.C., June [7] Microsoft, ISO/IEC (MPEG-4) Video Reference Software User Manual, Oct [8] M.-Y. Liu, Real-time implementation of MPEG-4 video encoder using SIMD-enhanced Intel processor, M.S. thesis, Degree Program of Electrical Engineering and Computer Science, National Chaio Tung University, Hsinchu, Taiwan, R.O.C., July [9] AudioCoding.com,
Outline Introduction MPEG-2 MPEG-4. Video Compression. Introduction to MPEG. Prof. Pratikgiri Goswami
to MPEG Prof. Pratikgiri Goswami Electronics & Communication Department, Shree Swami Atmanand Saraswati Institute of Technology, Surat. Outline of Topics 1 2 Coding 3 Video Object Representation Outline
More informationMISB EG Motion Imagery Standards Board Engineering Guideline. 24 April Delivery of Low Bandwidth Motion Imagery. 1 Scope.
Motion Imagery Standards Board Engineering Guideline Delivery of Low Bandwidth Motion Imagery MISB EG 0803 24 April 2008 1 Scope This Motion Imagery Standards Board (MISB) Engineering Guideline (EG) provides
More informationChapter 11.3 MPEG-2. MPEG-2: For higher quality video at a bit-rate of more than 4 Mbps Defined seven profiles aimed at different applications:
Chapter 11.3 MPEG-2 MPEG-2: For higher quality video at a bit-rate of more than 4 Mbps Defined seven profiles aimed at different applications: Simple, Main, SNR scalable, Spatially scalable, High, 4:2:2,
More informationEE Multimedia Signal Processing. Scope & Features. Scope & Features. Multimedia Signal Compression VI (MPEG-4, 7)
EE799 -- Multimedia Signal Processing Multimedia Signal Compression VI (MPEG-4, 7) References: 1. http://www.mpeg.org 2. http://drogo.cselt.stet.it/mpeg/ 3. T. Berahimi and M.Kunt, Visual data compression
More informationScalable Video Coding
Introduction to Multimedia Computing Scalable Video Coding 1 Topics Video On Demand Requirements Video Transcoding Scalable Video Coding Spatial Scalability Temporal Scalability Signal to Noise Scalability
More informationMPEG-4 AUTHORING TOOL FOR THE COMPOSITION OF 3D AUDIOVISUAL SCENES
MPEG-4 AUTHORING TOOL FOR THE COMPOSITION OF 3D AUDIOVISUAL SCENES P. Daras I. Kompatsiaris T. Raptis M. G. Strintzis Informatics and Telematics Institute 1,Kyvernidou str. 546 39 Thessaloniki, GREECE
More informationMPEG-4. Today we'll talk about...
INF5081 Multimedia Coding and Applications Vårsemester 2007, Ifi, UiO MPEG-4 Wolfgang Leister Knut Holmqvist Today we'll talk about... MPEG-4 / ISO/IEC 14496...... is more than a new audio-/video-codec...
More informationCMPT 365 Multimedia Systems. Media Compression - Video Coding Standards
CMPT 365 Multimedia Systems Media Compression - Video Coding Standards Spring 2017 Edited from slides by Dr. Jiangchuan Liu CMPT365 Multimedia Systems 1 Video Coding Standards H.264/AVC CMPT365 Multimedia
More informationGeorgios Tziritas Computer Science Department
New Video Coding standards MPEG-4, HEVC Georgios Tziritas Computer Science Department http://www.csd.uoc.gr/~tziritas 1 MPEG-4 : introduction Motion Picture Expert Group Publication 1998 (Intern. Standardization
More informationStreaming and Recording Capabilities
Streaming and Recording Capabilities PCS-G50/G50P All PCS-G70/G70P All PCS-XG55S All PCS-XG80S All Introduction Sony visual communication system PCS-XG55S/XG80S and PCS-G50/G50P/G70/ G70P (here after referred
More informationThe Virtual Meeting Room
Contact Details of Presenting Authors Stefan Rauthenberg (rauthenberg@hhi.de), Peter Kauff (kauff@hhi.de) Tel: +49-30-31002 266, +49-30-31002 615 Fax: +49-30-3927200 Summation Brief explaination of the
More informationModule 7 VIDEO CODING AND MOTION ESTIMATION
Module 7 VIDEO CODING AND MOTION ESTIMATION Lesson 20 Basic Building Blocks & Temporal Redundancy Instructional Objectives At the end of this lesson, the students should be able to: 1. Name at least five
More informationIntroduction to LAN/WAN. Application Layer 4
Introduction to LAN/WAN Application Layer 4 Multimedia Multimedia: Audio + video Human ear: 20Hz 20kHz, Dogs hear higher freqs DAC converts audio waves to digital E.g PCM uses 8-bit samples 8000 times
More informationA Hybrid Architecture for Video Transmission
2017 Asia-Pacific Engineering and Technology Conference (APETC 2017) ISBN: 978-1-60595-443-1 A Hybrid Architecture for Video Transmission Qian Huang, Xiaoqi Wang, Xiaodan Du and Feng Ye ABSTRACT With the
More informationMPEG-4: Overview. Multimedia Naresuan University
MPEG-4: Overview Multimedia Naresuan University Sources - Chapters 1 and 2, The MPEG-4 Book, F. Pereira and T. Ebrahimi - Some slides are adapted from NTNU, Odd Inge Hillestad. MPEG-1 and MPEG-2 MPEG-1
More informationInterworking Between SIP and MPEG-4 DMIF For Heterogeneous IP Video Conferencing
Interworking Between SIP and DMIF For Heterogeneous IP Video Conferencing Toufik Ahmed 1, Ahmed Mehaoua 1 and Raouf Boutaba 2 1 University of Versailles, CNRS-PRiSM Lab. 45 av. des Etats-Unis, 78000, Versailles,
More informationFast Decision of Block size, Prediction Mode and Intra Block for H.264 Intra Prediction EE Gaurav Hansda
Fast Decision of Block size, Prediction Mode and Intra Block for H.264 Intra Prediction EE 5359 Gaurav Hansda 1000721849 gaurav.hansda@mavs.uta.edu Outline Introduction to H.264 Current algorithms for
More informationThe Video Z-buffer: A Concept for Facilitating Monoscopic Image Compression by exploiting the 3-D Stereoscopic Depth map
The Video Z-buffer: A Concept for Facilitating Monoscopic Image Compression by exploiting the 3-D Stereoscopic Depth map Sriram Sethuraman 1 and M. W. Siegel 2 1 David Sarnoff Research Center, Princeton,
More informationERROR-ROBUST INTER/INTRA MACROBLOCK MODE SELECTION USING ISOLATED REGIONS
ERROR-ROBUST INTER/INTRA MACROBLOCK MODE SELECTION USING ISOLATED REGIONS Ye-Kui Wang 1, Miska M. Hannuksela 2 and Moncef Gabbouj 3 1 Tampere International Center for Signal Processing (TICSP), Tampere,
More informationAdvances of MPEG Scalable Video Coding Standard
Advances of MPEG Scalable Video Coding Standard Wen-Hsiao Peng, Chia-Yang Tsai, Tihao Chiang, and Hsueh-Ming Hang National Chiao-Tung University 1001 Ta-Hsueh Rd., HsinChu 30010, Taiwan pawn@mail.si2lab.org,
More informationRECOMMENDATION ITU-R BT.1720 *
Rec. ITU-R BT.1720 1 RECOMMENDATION ITU-R BT.1720 * Quality of service ranking and measurement methods for digital video broadcasting services delivered over broadband Internet protocol networks (Question
More informationInterframe coding A video scene captured as a sequence of frames can be efficiently coded by estimating and compensating for motion between frames pri
MPEG MPEG video is broken up into a hierarchy of layer From the top level, the first layer is known as the video sequence layer, and is any self contained bitstream, for example a coded movie. The second
More informationDigital Image Stabilization and Its Integration with Video Encoder
Digital Image Stabilization and Its Integration with Video Encoder Yu-Chun Peng, Hung-An Chang, Homer H. Chen Graduate Institute of Communication Engineering National Taiwan University Taipei, Taiwan {b889189,
More information9/8/2016. Characteristics of multimedia Various media types
Chapter 1 Introduction to Multimedia Networking CLO1: Define fundamentals of multimedia networking Upon completion of this chapter students should be able to define: 1- Multimedia 2- Multimedia types and
More informationOPTIMIZATION OF LOW DELAY WAVELET VIDEO CODECS
OPTIMIZATION OF LOW DELAY WAVELET VIDEO CODECS Andrzej Popławski, Marek Domański 2 Uniwersity of Zielona Góra, Institute of Computer Engineering and Electronics, Poland 2 Poznań University of Technology,
More informationModule 10 MULTIMEDIA SYNCHRONIZATION
Module 10 MULTIMEDIA SYNCHRONIZATION Lesson 36 Packet architectures and audio-video interleaving Instructional objectives At the end of this lesson, the students should be able to: 1. Show the packet architecture
More informationVideo Transcoding Architectures and Techniques: An Overview. IEEE Signal Processing Magazine March 2003 Present by Chen-hsiu Huang
Video Transcoding Architectures and Techniques: An Overview IEEE Signal Processing Magazine March 2003 Present by Chen-hsiu Huang Outline Background & Introduction Bit-rate Reduction Spatial Resolution
More informationMPEG: It s Need, Evolution and Processing Methods
MPEG: It s Need, Evolution and Processing Methods Ankit Agarwal, Prateeksha Suwalka, Manohar Prajapati ECE DEPARTMENT, Baldev Ram mirdha institute of technology (EC) ITS- 3,EPIP SItapura, Jaipur-302022(India)
More informationStreaming Technologies Glossary
Streaming Technologies Glossary A ACELP.net ADSL AIFF Artifact Asf Asx Avi B Bitrate C C Capture card Codec CIF (QCIF) Client Industry standard audio compression technology developed by VoiceAge Corporation.
More informationChapter 11: Understanding the H.323 Standard
Página 1 de 7 Chapter 11: Understanding the H.323 Standard This chapter contains information about the H.323 standard and its architecture, and discusses how Microsoft Windows NetMeeting supports H.323
More informationBluray (
Bluray (http://www.blu-ray.com/faq) MPEG-2 - enhanced for HD, also used for playback of DVDs and HDTV recordings MPEG-4 AVC - part of the MPEG-4 standard also known as H.264 (High Profile and Main Profile)
More informationVideo Redundancy Coding in H.263+ Stephan Wenger Technische Universität Berlin
Video Redundancy Coding in H.263+ Stephan Wenger Technische Universität Berlin stewe@cs.tu-berlin.de ABSTRACT: The forthcoming new version of ITU- T s advanced video compression recommendation H.263 [1]
More informationDIGITAL TELEVISION 1. DIGITAL VIDEO FUNDAMENTALS
DIGITAL TELEVISION 1. DIGITAL VIDEO FUNDAMENTALS Television services in Europe currently broadcast video at a frame rate of 25 Hz. Each frame consists of two interlaced fields, giving a field rate of 50
More informationFast Wavelet-based Macro-block Selection Algorithm for H.264 Video Codec
Proceedings of the International MultiConference of Engineers and Computer Scientists 8 Vol I IMECS 8, 19-1 March, 8, Hong Kong Fast Wavelet-based Macro-block Selection Algorithm for H.64 Video Codec Shi-Huang
More information2 RTP Encapsulation and its Application in NS-2 Simulation
3rd International Conference on Multimedia Technology(ICMT 2013) RTP Encapsulation for Scalable Video Stream and its Application in NS-2 Simulation Zhou Ying, Zhang Jihong, Liu Wei Abstract. Real-time
More informationThe RTP Encapsulation based on Frame Type Method for AVS Video
Applied Mechanics and Materials Online: 2012-12-27 ISSN: 1662-7482, Vols. 263-266, pp 1803-1808 doi:10.4028/www.scientific.net/amm.263-266.1803 2013 Trans Tech Publications, Switzerland The RTP Encapsulation
More informationVideo coding. Concepts and notations.
TSBK06 video coding p.1/47 Video coding Concepts and notations. A video signal consists of a time sequence of images. Typical frame rates are 24, 25, 30, 50 and 60 images per seconds. Each image is either
More informationCODING METHOD FOR EMBEDDING AUDIO IN VIDEO STREAM. Harri Sorokin, Jari Koivusaari, Moncef Gabbouj, and Jarmo Takala
CODING METHOD FOR EMBEDDING AUDIO IN VIDEO STREAM Harri Sorokin, Jari Koivusaari, Moncef Gabbouj, and Jarmo Takala Tampere University of Technology Korkeakoulunkatu 1, 720 Tampere, Finland ABSTRACT In
More informationAutomatic Video Caption Detection and Extraction in the DCT Compressed Domain
Automatic Video Caption Detection and Extraction in the DCT Compressed Domain Chin-Fu Tsao 1, Yu-Hao Chen 1, Jin-Hau Kuo 1, Chia-wei Lin 1, and Ja-Ling Wu 1,2 1 Communication and Multimedia Laboratory,
More informationLocating 1-D Bar Codes in DCT-Domain
Edith Cowan University Research Online ECU Publications Pre. 2011 2006 Locating 1-D Bar Codes in DCT-Domain Alexander Tropf Edith Cowan University Douglas Chai Edith Cowan University 10.1109/ICASSP.2006.1660449
More informationRTP. Prof. C. Noronha RTP. Real-Time Transport Protocol RFC 1889
RTP Real-Time Transport Protocol RFC 1889 1 What is RTP? Primary objective: stream continuous media over a best-effort packet-switched network in an interoperable way. Protocol requirements: Payload Type
More informationCopyright Notice. Springer papers: Springer. Pre-prints are provided only for personal use. The final publication is available at link.springer.
Copyright Notice The document is provided by the contributing author(s) as a means to ensure timely dissemination of scholarly and technical work on a non-commercial basis. This is the author s version
More informationISO/IEC INTERNATIONAL STANDARD. Information technology Coding of audio-visual objects Part 18: Font compression and streaming
INTERNATIONAL STANDARD ISO/IEC 14496-18 First edition 2004-07-01 Information technology Coding of audio-visual objects Part 18: Font compression and streaming Technologies de l'information Codage des objets
More informationAlkit Reflex RTP reflector/mixer
Alkit Reflex RTP reflector/mixer Mathias Johanson, Ph.D. Alkit Communications Introduction Real time audio and video communication over IP networks is attracting a lot of interest for applications like
More informationJPEG 2000 vs. JPEG in MPEG Encoding
JPEG 2000 vs. JPEG in MPEG Encoding V.G. Ruiz, M.F. López, I. García and E.M.T. Hendrix Dept. Computer Architecture and Electronics University of Almería. 04120 Almería. Spain. E-mail: vruiz@ual.es, mflopez@ace.ual.es,
More informationINTERNATIONAL ORGANISATION FOR STANDARDISATION ORGANISATION INTERNATIONALE DE NORMALISATION ISO/IEC JTC1/SC29/WG11 CODING OF MOVING PICTURES AND AUDIO
INTERNATIONAL ORGANISATION FOR STANDARDISATION ORGANISATION INTERNATIONALE DE NORMALISATION ISO/IEC JTC1/SC29/WG11 CODING OF MOVING PICTURES AND AUDIO ISO/IEC JTC1/SC29 WG11 N15073 February 2015, Geneva,
More informationOptical Storage Technology. MPEG Data Compression
Optical Storage Technology MPEG Data Compression MPEG-1 1 Audio Standard Moving Pictures Expert Group (MPEG) was formed in 1988 to devise compression techniques for audio and video. It first devised the
More informationMultimedia Applications over Packet Networks
Multimedia Networking and Quality of Service Mario Baldi Technical Univeristy of Torino Computer Engineering Department mario.baldi@polito.it +39 011 564 7067 staff.polito.it/mario.baldi Nota di Copyright
More informationVisionair. TNA - Collaborative remote visualization ENSAM URJC. Miguel Ramos García. Universidad Rey Juan Carlos, Móstoles (Spain)
Visionair TNA - Collaborative remote visualization ENSAM URJC Miguel Ramos García Universidad Rey Juan Carlos, Móstoles (Spain) Index Introduction... 3 1 Framework, the first contact.... 4 1.1 Information
More informationRTP/RTCP protocols. Introduction: What are RTP and RTCP?
RTP/RTCP protocols Introduction: What are RTP and RTCP? The spread of computers, added to the availability of cheap audio/video computer hardware, and the availability of higher connection speeds have
More informationMPEG-4 Structured Audio Systems
MPEG-4 Structured Audio Systems Mihir Anandpara The University of Texas at Austin anandpar@ece.utexas.edu 1 Abstract The MPEG-4 standard has been proposed to provide high quality audio and video content
More informationIn the name of Allah. the compassionate, the merciful
In the name of Allah the compassionate, the merciful Digital Video Systems S. Kasaei Room: CE 315 Department of Computer Engineering Sharif University of Technology E-Mail: skasaei@sharif.edu Webpage:
More informationover the Internet Tihao Chiang { Ya-Qin Zhang k enormous interests from both industry and academia.
An End-to-End Architecture for MPEG-4 Video Streaming over the Internet Y. Thomas Hou Dapeng Wu y Wenwu Zhu z Hung-Ju Lee x Tihao Chiang { Ya-Qin Zhang k Abstract It is a challenging problem to design
More informationINTERNATIONAL ORGANISATION FOR STANDARDISATION ORGANISATION INTERNATIONALE DE NORMALISATION ISO/IEC JTC1/SC29/WG11 CODING OF MOVING PICTURES AND AUDIO
INTERNATIONAL ORGANISATION FOR STANDARDISATION ORGANISATION INTERNATIONALE DE NORMALISATION ISO/IEC JTC1/SC29/WG11 CODING OF MOVING PICTURES AND AUDIO ISO/IEC JTC1/SC29/WG11 N15071 February 2015, Geneva,
More informationOptimal Estimation for Error Concealment in Scalable Video Coding
Optimal Estimation for Error Concealment in Scalable Video Coding Rui Zhang, Shankar L. Regunathan and Kenneth Rose Department of Electrical and Computer Engineering University of California Santa Barbara,
More informationISO/IEC TR TECHNICAL REPORT. Information technology Coding of audio-visual objects Part 24: Audio and systems interaction
TECHNICAL REPORT ISO/IEC TR 14496-24 First edition 2008-01-15 Information technology Coding of audio-visual objects Part 24: Audio and systems interaction Technologies de l'information Codage d'objets
More informationFRAME-RATE UP-CONVERSION USING TRANSMITTED TRUE MOTION VECTORS
FRAME-RATE UP-CONVERSION USING TRANSMITTED TRUE MOTION VECTORS Yen-Kuang Chen 1, Anthony Vetro 2, Huifang Sun 3, and S. Y. Kung 4 Intel Corp. 1, Mitsubishi Electric ITA 2 3, and Princeton University 1
More informationMPEG-2. ISO/IEC (or ITU-T H.262)
MPEG-2 1 MPEG-2 ISO/IEC 13818-2 (or ITU-T H.262) High quality encoding of interlaced video at 4-15 Mbps for digital video broadcast TV and digital storage media Applications Broadcast TV, Satellite TV,
More informationA Transport Infrastructure Supporting Real Time Interactive MPEG-4 Client-Server Applications over IP Networks
A Transport Infrastructure Supporting Real Time Interactive MPEG-4 Client-Server Applications over IP Networks Haining Liu, Xiaoping Wei, and Magda El Zarki Department of Information and Computer Science
More informationFormat-Independent Multimedia Streaming
University of Wollongong Research Online Faculty of Informatics - Papers Faculty of Informatics 2006 Format-Independent Multimedia Streaming J. Thomas-Kerr University of Wollongong I. Burnett University
More informationDigital video coding systems MPEG-1/2 Video
Digital video coding systems MPEG-1/2 Video Introduction What is MPEG? Moving Picture Experts Group Standard body for delivery of video and audio. Part of ISO/IEC/JTC1/SC29/WG11 150 companies & research
More informationMITIGATING THE EFFECT OF PACKET LOSSES ON REAL-TIME VIDEO STREAMING USING PSNR AS VIDEO QUALITY ASSESSMENT METRIC ABSTRACT
MITIGATING THE EFFECT OF PACKET LOSSES ON REAL-TIME VIDEO STREAMING USING PSNR AS VIDEO QUALITY ASSESSMENT METRIC Anietie Bassey, Kufre M. Udofia & Mfonobong C. Uko Department of Electrical/Electronic
More informationMultimedia Systems Video II (Video Coding) Mahdi Amiri April 2012 Sharif University of Technology
Course Presentation Multimedia Systems Video II (Video Coding) Mahdi Amiri April 2012 Sharif University of Technology Video Coding Correlation in Video Sequence Spatial correlation Similar pixels seem
More informationSpatial Scene Level Shape Error Concealment for Segmented Video
Spatial Scene Level Shape Error Concealment for Segmented Video Luis Ducla Soares 1, Fernando Pereira 2 1 Instituto Superior de Ciências do Trabalho e da Empresa Instituto de Telecomunicações, Lisboa,
More informationTRIBHUVAN UNIVERSITY Institute of Engineering Pulchowk Campus Department of Electronics and Computer Engineering
TRIBHUVAN UNIVERSITY Institute of Engineering Pulchowk Campus Department of Electronics and Computer Engineering A Final project Report ON Minor Project Java Media Player Submitted By Bisharjan Pokharel(061bct512)
More informationLecture 3 Image and Video (MPEG) Coding
CS 598KN Advanced Multimedia Systems Design Lecture 3 Image and Video (MPEG) Coding Klara Nahrstedt Fall 2017 Overview JPEG Compression MPEG Basics MPEG-4 MPEG-7 JPEG COMPRESSION JPEG Compression 8x8 blocks
More informationTema 0: Transmisión de Datos Multimedia
Tema 0: Transmisión de Datos Multimedia Clases de aplicaciones multimedia Redes basadas en IP y QoS Computer Networking: A Top Down Approach Featuring the Internet, 3 rd edition. Jim Kurose, Keith Ross
More informationWhite paper: Video Coding A Timeline
White paper: Video Coding A Timeline Abharana Bhat and Iain Richardson June 2014 Iain Richardson / Vcodex.com 2007-2014 About Vcodex Vcodex are world experts in video compression. We provide essential
More informationImproving student feedback in virtual classrooms
Improving student feedback in virtual classrooms M. van der Schyff, H.C. Ferreira Department of Electrical and Electronic Engineering Rand Afrikaans University, South Africa mvds@ing.rau.ac.za, hcf@ing.rau.ac.za
More informationParallel Implementation of Arbitrary-Shaped MPEG-4 Decoder for Multiprocessor Systems
Parallel Implementation of Arbitrary-Shaped MPEG-4 oder for Multiprocessor Systems Milan Pastrnak *,a,c, Peter H.N. de With a,c, Sander Stuijk c and Jef van Meerbergen b,c a LogicaCMG Nederland B.V., RTSE
More informationADAPTIVE PICTURE SLICING FOR DISTORTION-BASED CLASSIFICATION OF VIDEO PACKETS
ADAPTIVE PICTURE SLICING FOR DISTORTION-BASED CLASSIFICATION OF VIDEO PACKETS E. Masala, D. Quaglia, J.C. De Martin Λ Dipartimento di Automatica e Informatica/ Λ IRITI-CNR Politecnico di Torino, Italy
More informationMultimedia Networking and Quality of Service
Multimedia Networking and Quality of Service Mario Baldi Politecnico di Torino (Technical Univeristy of Torino) Department of Computer Engineering mario.baldi [at] polito.it +39 011 564 7067 staff.polito.it/mario.baldi
More informationQUANTIZER DESIGN FOR EXPLOITING COMMON INFORMATION IN LAYERED CODING. Mehdi Salehifar, Tejaswi Nanjundaswamy, and Kenneth Rose
QUANTIZER DESIGN FOR EXPLOITING COMMON INFORMATION IN LAYERED CODING Mehdi Salehifar, Tejaswi Nanjundaswamy, and Kenneth Rose Department of Electrical and Computer Engineering University of California,
More informationApplication and Desktop Sharing. Omer Boyaci November 1, 2007
Application and Desktop Sharing Omer Boyaci November 1, 2007 Overview Introduction Demo Architecture Challenges Features Conclusion Application Sharing Models Application specific + Efficient - Participants
More informationFigure 1: Representation of moving images using layers Once a set of ane models has been found, similar models are grouped based in a mean-square dist
ON THE USE OF LAYERS FOR VIDEO CODING AND OBJECT MANIPULATION Luis Torres, David Garca and Anna Mates Dept. of Signal Theory and Communications Universitat Politecnica de Catalunya Gran Capita s/n, D5
More informationimplementation using GPU architecture is implemented only from the viewpoint of frame level parallel encoding [6]. However, it is obvious that the mot
Parallel Implementation Algorithm of Motion Estimation for GPU Applications by Tian Song 1,2*, Masashi Koshino 2, Yuya Matsunohana 2 and Takashi Shimamoto 1,2 Abstract The video coding standard H.264/AVC
More informationA COST-EFFICIENT RESIDUAL PREDICTION VLSI ARCHITECTURE FOR H.264/AVC SCALABLE EXTENSION
A COST-EFFICIENT RESIDUAL PREDICTION VLSI ARCHITECTURE FOR H.264/AVC SCALABLE EXTENSION Yi-Hau Chen, Tzu-Der Chuang, Chuan-Yung Tsai, Yu-Jen Chen, and Liang-Gee Chen DSP/IC Design Lab., Graduate Institute
More informationInternational Journal of Emerging Technology and Advanced Engineering Website: (ISSN , Volume 2, Issue 4, April 2012)
A Technical Analysis Towards Digital Video Compression Rutika Joshi 1, Rajesh Rai 2, Rajesh Nema 3 1 Student, Electronics and Communication Department, NIIST College, Bhopal, 2,3 Prof., Electronics and
More informationQoE Characterization for Video-On-Demand Services in 4G WiMAX Networks
QoE Characterization for Video-On-Demand Services in 4G WiMAX Networks Amitabha Ghosh IBM India Research Laboratory Department of Electrical Engineering University of Southern California, Los Angeles http://anrg.usc.edu/~amitabhg
More informationModule 10 MULTIMEDIA SYNCHRONIZATION
Module 10 MULTIMEDIA SYNCHRONIZATION Lesson 33 Basic definitions and requirements Instructional objectives At the end of this lesson, the students should be able to: 1. Define synchronization between media
More informationIO [io] MAYAH. IO [io] Audio Video Codec Systems
IO [io] MAYAH IO [io] Audio Video Codec Systems MPEG 4 Audio Video Embedded 24/7 Real-Time Solution MPEG 4 Audio Video Production and Streaming Solution ISMA compliant 24/7 Audio Video Realtime Solution
More informationRate Distortion Optimization in Video Compression
Rate Distortion Optimization in Video Compression Xue Tu Dept. of Electrical and Computer Engineering State University of New York at Stony Brook 1. Introduction From Shannon s classic rate distortion
More informationLecture 7, Video Coding, Motion Compensation Accuracy
Lecture 7, Video Coding, Motion Compensation Accuracy Last time we saw several methods to obtain a good motion estimation, with reduced complexity (efficient search), and with the possibility of sub-pixel
More informationVideo Compression MPEG-4. Market s requirements for Video compression standard
Video Compression MPEG-4 Catania 10/04/2008 Arcangelo Bruna Market s requirements for Video compression standard Application s dependent Set Top Boxes (High bit rate) Digital Still Cameras (High / mid
More informationAdvanced Video Coding: The new H.264 video compression standard
Advanced Video Coding: The new H.264 video compression standard August 2003 1. Introduction Video compression ( video coding ), the process of compressing moving images to save storage space and transmission
More informationIST MPEG-4 Video Compliant Framework
IST MPEG-4 Video Compliant Framework João Valentim, Paulo Nunes, Fernando Pereira Instituto de Telecomunicações, Instituto Superior Técnico, Av. Rovisco Pais, 1049-001 Lisboa, Portugal Abstract This paper
More information3GPP TS V5.2.0 ( )
TS 26.140 V5.2.0 (2002-12) Technical Specification 3rd Generation Partnership Project; Technical Specification Group Services and System Aspects; Multimedia Messaging Service (MMS); Media formats and codecs
More informationComparison of Shaping and Buffering for Video Transmission
Comparison of Shaping and Buffering for Video Transmission György Dán and Viktória Fodor Royal Institute of Technology, Department of Microelectronics and Information Technology P.O.Box Electrum 229, SE-16440
More informationInformation technology - Coding of audiovisual objects - Part 2: Visual
This is a preview of "INCITS/ISO/IEC 14496...". Click here to purchase the full version from the ANSI store. INCITS/ISO/IEC 14496-2:2004[R2012] (ISO/IEC 14496-2:2004, IDT) Information technology - Coding
More informationVIDEO streaming applications over the Internet are gaining. Brief Papers
412 IEEE TRANSACTIONS ON BROADCASTING, VOL. 54, NO. 3, SEPTEMBER 2008 Brief Papers Redundancy Reduction Technique for Dual-Bitstream MPEG Video Streaming With VCR Functionalities Tak-Piu Ip, Yui-Lam Chan,
More informationContent Based Image Retrieval Using Color Quantizes, EDBTC and LBP Features
Content Based Image Retrieval Using Color Quantizes, EDBTC and LBP Features 1 Kum Sharanamma, 2 Krishnapriya Sharma 1,2 SIR MVIT Abstract- To describe the image features the Local binary pattern (LBP)
More informationRequest for Comments: 2032 Category: Standards Track Bellcore October 1996
Network Working Group Request for Comments: 2032 Category: Standards Track T. Turletti MIT C. Huitema Bellcore October 1996 Status of this Memo RTP Payload Format for H.261 Video Streams This document
More informationsignal-to-noise ratio (PSNR), 2
u m " The Integration in Optics, Mechanics, and Electronics of Digital Versatile Disc Systems (1/3) ---(IV) Digital Video and Audio Signal Processing ƒf NSC87-2218-E-009-036 86 8 1 --- 87 7 31 p m o This
More informationSTACK ROBUST FINE GRANULARITY SCALABLE VIDEO CODING
Journal of the Chinese Institute of Engineers, Vol. 29, No. 7, pp. 1203-1214 (2006) 1203 STACK ROBUST FINE GRANULARITY SCALABLE VIDEO CODING Hsiang-Chun Huang and Tihao Chiang* ABSTRACT A novel scalable
More informationDATA HIDING IN PDF FILES AND APPLICATIONS BY IMPERCEIVABLE MODIFICATIONS OF PDF OBJECT PARAMETERS
DATA HIDING IN PDF FILES AND APPLICATIONS BY IMPERCEIVABLE MODIFICATIONS OF PDF OBJECT PARAMETERS 1 Jiun-Tsung Wang ( 王竣聰 ) and 2 Wen-Hsiang Tsai ( 蔡文祥 ) 1 Institute of Multimedia Eng., National Chiao
More informationMpegRepair Software Encoding and Repair Utility
PixelTools MpegRepair Software Encoding and Repair Utility MpegRepair integrates fully featured encoding, analysis, decoding, demuxing, transcoding and stream manipulations into one powerful application.
More informationNetworking Applications
Networking Dr. Ayman A. Abdel-Hamid College of Computing and Information Technology Arab Academy for Science & Technology and Maritime Transport Multimedia Multimedia 1 Outline Audio and Video Services
More informationStereo Vision Image Processing Strategy for Moving Object Detecting
Stereo Vision Image Processing Strategy for Moving Object Detecting SHIUH-JER HUANG, FU-REN YING Department of Mechanical Engineering National Taiwan University of Science and Technology No. 43, Keelung
More informationComparative Study of Partial Closed-loop Versus Open-loop Motion Estimation for Coding of HDTV
Comparative Study of Partial Closed-loop Versus Open-loop Motion Estimation for Coding of HDTV Jeffrey S. McVeigh 1 and Siu-Wai Wu 2 1 Carnegie Mellon University Department of Electrical and Computer Engineering
More informationRTP for Application Sharing Payload Format Extensions
[MS-RTASPF]: Intellectual Property Rights Notice for Open Specifications Documentation Technical Documentation. Microsoft publishes Open Specifications documentation for protocols, file formats, languages,
More information