Error Concealment Used for P-Frame on Video Stream over the Internet

Similar documents
MPEG-4: Simple Profile (SP)

over the Internet Tihao Chiang { Ya-Qin Zhang k enormous interests from both industry and academia.

Fast Mode Decision for H.264/AVC Using Mode Prediction

Error Control Techniques for Interactive Low-bit Rate Video Transmission over the Internet.

Module 6 STILL IMAGE COMPRESSION STANDARDS

System Modeling and Implementation of MPEG-4. Encoder under Fine-Granular-Scalability Framework

Zonal MPEG-2. Cheng-Hsiung Hsieh *, Chen-Wei Fu and Wei-Lung Hung

Rate-Distortion Optimized Layered Coding with Unequal Error Protection for Robust Internet Video

One-pass bitrate control for MPEG-4 Scalable Video Coding using ρ-domain

Optimal Estimation for Error Concealment in Scalable Video Coding

SINGLE PASS DEPENDENT BIT ALLOCATION FOR SPATIAL SCALABILITY CODING OF H.264/SVC

Improving Interactive Video in Wireless Networks Using Path Diversity 1

Multi-path Transport of FGS Video

Week 14. Video Compression. Ref: Fundamentals of Multimedia

Unequal Error Recovery Scheme for Multimedia Streaming in Application-Level Multicast

Fast Decision of Block size, Prediction Mode and Intra Block for H.264 Intra Prediction EE Gaurav Hansda

SMART: An Efficient, Scalable and Robust Streaming Video System

streams cannot be decoded properly, which would severely affect the quality of reconstructed video. Therefore, error resilience is utilized to solve

CODING METHOD FOR EMBEDDING AUDIO IN VIDEO STREAM. Harri Sorokin, Jari Koivusaari, Moncef Gabbouj, and Jarmo Takala

Adaptation of Scalable Video Coding to Packet Loss and its Performance Analysis

Lecture 5: Error Resilience & Scalability

Unit-level Optimization for SVC Extractor

Comparison of Shaping and Buffering for Video Transmission

ERROR-ROBUST INTER/INTRA MACROBLOCK MODE SELECTION USING ISOLATED REGIONS

IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 19, NO. 9, SEPTEMBER

Digital Video Processing

Fast frame memory access method for H.264/AVC

A Robust Error Resilient Approach for Data Hiding in MPEG Video Files Using Multivariate Regression and Flexible Macroblock Ordering

H.264/AVC Baseline Profile to MPEG-4 Visual Simple Profile Transcoding to Reduce the Spatial Resolution

Motion Estimation. Original. enhancement layers. Motion Compensation. Baselayer. Scan-Specific Entropy Coding. Prediction Error.

ADAPTIVE JOINT H.263-CHANNEL CODING FOR MEMORYLESS BINARY CHANNELS

A 3-D Virtual SPIHT for Scalable Very Low Bit-Rate Embedded Video Compression

Recommended Readings

Motion Estimation for Video Coding Standards

H.264 to MPEG-4 Transcoding Using Block Type Information

A deblocking filter with two separate modes in block-based video coding

A Novel Deblocking Filter Algorithm In H.264 for Real Time Implementation

SNR Scalable Transcoding for Video over Wireless Channels

FAST SPATIAL LAYER MODE DECISION BASED ON TEMPORAL LEVELS IN H.264/AVC SCALABLE EXTENSION

JPEG 2000 vs. JPEG in MPEG Encoding

The new Hybrid approach to protect MPEG-2 video header

H.264 Video Transmission with High Quality and Low Bitrate over Wireless Network

Lecture 5: Video Compression Standards (Part2) Tutorial 3 : Introduction to Histogram

Reduced Frame Quantization in Video Coding

FRAME-RATE UP-CONVERSION USING TRANSMITTED TRUE MOTION VECTORS

Bit Error Recovery in MMR Coded Bitstreams Using Error Detection Points

A COST-EFFICIENT RESIDUAL PREDICTION VLSI ARCHITECTURE FOR H.264/AVC SCALABLE EXTENSION

Video Redundancy Coding in H.263+ Stephan Wenger Technische Universität Berlin

Hybrid Video Compression Using Selective Keyframe Identification and Patch-Based Super-Resolution

2 RTP Encapsulation and its Application in NS-2 Simulation

MODELING AND SIMULATION OF MPEG-2 VIDEO TRANSPORT OVER ATM NETWOR.KS CONSIDERING THE JITTER EFFECT

Streaming Video Based on Temporal Frame Transcoding.

Unbalanced Multiple Description Video Communication Using Path Diversity

VIDEO COMPRESSION STANDARDS

High Efficient Intra Coding Algorithm for H.265/HVC

To address these challenges, extensive research has been conducted and have introduced six key areas of streaming video, namely: video compression,

Semi-Hierarchical Based Motion Estimation Algorithm for the Dirac Video Encoder

Low complexity H.264 list decoder for enhanced quality real-time video over IP

Linköping University Post Print. A Minimum Mean Square Error Estimation and Mixture-Based Approach to Packet Video Error Concealment

Improving Interactive Video in Ad-hoc Networks Using Path Diversity

EE Multimedia Signal Processing. Scope & Features. Scope & Features. Multimedia Signal Compression VI (MPEG-4, 7)

A NOVEL SCANNING SCHEME FOR DIRECTIONAL SPATIAL PREDICTION OF AVS INTRA CODING

Combined Copyright Protection and Error Detection Scheme for H.264/AVC

Network-based model for video packet importance considering both compression artifacts and packet losses

Complexity Reduced Mode Selection of H.264/AVC Intra Coding

The Performance of MANET Routing Protocols for Scalable Video Communication

2014 Summer School on MPEG/VCEG Video. Video Coding Concept

Fast Implementation of VC-1 with Modified Motion Estimation and Adaptive Block Transform

Video Compression Standards (II) A/Prof. Jian Zhang

Multimedia networked applications: standards, protocols and research trends

Enhanced Hexagon with Early Termination Algorithm for Motion estimation

International Journal of Emerging Technology and Advanced Engineering Website: (ISSN , Volume 2, Issue 4, April 2012)

Homogeneous Transcoding of HEVC for bit rate reduction

Experimental Evaluation of H.264/Multiview Video Coding over IP Networks

Multiframe Blocking-Artifact Reduction for Transform-Coded Video

Welcome Back to Fundamentals of Multimedia (MR412) Fall, 2012 Chapter 10 ZHU Yongxin, Winson

OVERVIEW OF IEEE 1857 VIDEO CODING STANDARD

Real-Time Course. Video Streaming Over network. June Peter van der TU/e Computer Science, System Architecture and Networking

Interframe coding A video scene captured as a sequence of frames can be efficiently coded by estimating and compensating for motion between frames pri

Fast Motion Estimation for Shape Coding in MPEG-4

Robust MPEG-2 SNR Scalable Coding Using Variable End-of-Block

Real-time and smooth scalable video streaming system with bitstream extractor intellectual property implementation

A High Quality/Low Computational Cost Technique for Block Matching Motion Estimation

A Hybrid Temporal-SNR Fine-Granular Scalability for Internet Video

VIDEO streaming applications over the Internet are gaining. Brief Papers

SMART: An Efficient, Scalable, and Robust Streaming Video System

Distributed Video Coding

Professor, CSE Department, Nirma University, Ahmedabad, India

Image Error Concealment Based on Watermarking

Performance of a Multiple Description Streaming Media Content Delivery Network

ADAPTIVE PICTURE SLICING FOR DISTORTION-BASED CLASSIFICATION OF VIDEO PACKETS

MITIGATING THE EFFECT OF PACKET LOSSES ON REAL-TIME VIDEO STREAMING USING PSNR AS VIDEO QUALITY ASSESSMENT METRIC ABSTRACT

Standard Codecs. Image compression to advanced video coding. Mohammed Ghanbari. 3rd Edition. The Institution of Engineering and Technology

2 Framework of The Proposed Voice Quality Assessment System

System Modeling and Implementation of MPEG-4. Encoder under Fine-Granular-Scalability Framework

An Implementation of Multiple Region-Of-Interest Models in H.264/AVC

Bit error recovery in Internet facsimile without retransmission

Network-Adaptive Video Coding and Transmission

10.2 Video Compression with Motion Compensation 10.4 H H.263

5LSE0 - Mod 10 Part 1. MPEG Motion Compensation and Video Coding. MPEG Video / Temporal Prediction (1)

Transcription:

Error Concealment Used for P-Frame on Video Stream over the Internet MA RAN, ZHANG ZHAO-YANG, AN PING Key Laboratory of Advanced Displays and System Application, Ministry of Education School of Communication and Information Engineering Shanghai University No. 114,YanChang Road, Shanghai CNINA Abstract: Error concealment on the end of decoder, which makes use of the correctly received video data to conceal the lost video, does not need any redundancy data and has no impact to bandwidth of network. A new concealment method based on multiple motion predictions is proposed here. In this paper, motion vectors in the past and/or future frames are used to achieve the current motion vectors when the current frame is suffered, so the lost parts can be reconstructed. Experimental results show that the performance of the new method is superior to traditional single or dual forward prediction error concealment. Key-Words: motion vector;error concealment;video stream;internet 1 Introduction Compressed video stream is very sensitive to packet loss in the Internet, due to variable length coding(vlc) and interframe prediction coding, which cause spatial and temporal error propagation. As a consequence, the research on error concealment is an important challenging task. On the end of decoder, error concealment technique makes use of the correctly received video data to conceal the lost video. In the late years, the investigated concealment methods can be categorized into two main areas: intraframe and interframe[1]-[5]. In intraframe concealment, the corrupted parts of the frame are interpolated from either the surrounding pixels or their transform coefficients, of the same frame. To interframe concealment, the lost parts are reconstructed from either the previous uncorrupted frames or their motion vectors. In this paper, a new interframe concealment method based on multiple predictions is presented to conceal the missing motion vectors for P-frame. The paper organization is as follows. In Section 2, channel characteristics of Internet are reviewed. The new concealment method is introduced in Section 3. Experimental results are presented in Section 4. Finally, the conclusion is stated in Section 5. 2 Characteristics of Internet 2.1 Characteristics Of Internet and Gilbert Channel Model Some conclusions about the packet loss characteristics over the Internet are drawn in Ref.[6] as follows: no bit errors in IP/UDP packets; no relationship between the packet loss rate, the packet size and the bitrate; packet losses randomly or in very short bursts like two or three packets; packet loss rates dependent on the connection and the time of day. Channel characteristics of Internet determine the probabilities that packets can be received correctly.

The packet loss behavior is always explained well with a 2-state Markov Chain, i.e., the Gilbert model shown in Fig.1[7]~[8]. State 1 represents a packet loss and State 0 represents the case where packets are correctly transmitted. Two parameters, p and q, are used to characterize the model, where p is the probability of going form State 0 to State 1, and q is the probability of going from State 1 to State 0. Here, p+q does not necessarily equal 1. Different values of p and q define different packet loss conditions that can occur on the Internet. The probability that n consecutive packets are lost is p(1-q) n-1. If (1-q)>p, the probability of losing a packet is greater after having already lost a packet than after having successfully received a packet: this is generally the case on Internet video transmission where packet losses occur as bursts. When p and q are fixed, the mean number of consecutive packets lost can be easily calculated and is equal to p/q 2. 1-p p 0 1 q 1-q Fig.1 Gilbert model 2.2 Structures of IP Packet and Video Packet The structures of IP packet and video packet are shown in Fig.2. The size of video packet is limited to prevent the packet fragmentation by the default maximum transmission unit (MTU), i.e., 576 bytes, where the overhead is 50-bytes long. It is clear that the use of large packet size will reduce the total number of generated packets and overhead. So, the video packet approach adopted by MPEG-4 uses periodic resynchronization markers throughout the bitstream[9]~[10]. The length of the video packet is not based on the number of MBs(Macroblocks), instead of on the number of bits contained in that packet. IP UDP Resync marker RTP Macroblock number FlexMux SyncLayer Quant_scale HEC MPEG-4 Video Packet Macroblock Data Fig.2 Structures of IP packet and video packet 3 Concealment used for P-Frame A concealment method, called as Dual forward prediction concealment, is proposed in Ref.[5]. The final prediction frame is obtained by averaging the two previous prediction frames when the two frames are not corrupted. When either of previous prediction frames is corrupted, the final frame copies directly with the one correct prediction frame. Obviously, the method only considers the past reference frames, and ignores the same important relativity between the current frame and the future frames. Furthermore, when transmission conditions in the Internet are too bad, it is possible that the two previous prediction frames are lost; so the above method is invalid. MV 01 MV 12 MV 23 MV 45 I 0 P 1 P 2 P 3 P 4 P 5 Fig.3 Directions of motion vectors Fig.3 shows the directions of motion vectors in video sequence. Referring to the previous I-frame or predicted P-frame, the current P-frame is reconstructed by using motion compensation and prediction. Firstly, there are strongly relativities among most of video sequences, and the motion vectors of neighbor frames are similar. Therefore, not only the past prediction frames but also the future prediction frames can be used to conceal the discard parts of the current frame. Secondly, the trend of motions of objects in video sequences can be better estimated by using the motion vectors of multiple frames than of single one. Thirdly, due to the motions of objects in video sequence are coherent, the closer the distance of the two prediction frames is, the less the difference between the corresponding vectors is. Lastly, in packet switched network, video packets are out of order when arriving on the end of decoder. Consequently, the decoder always has capability of buffering and reordering, which makes possible that several prediction frames can be stored.

As stated above, a new interframe concealment used for P-frame is proposed on condition that I-frame is uncorrupted. According to the different relativities between the current prediction frame and the past and/or future prediction frames, several appropriate motion vectors are selected from the received past and/or future candidate vectors based on certain strategies. The new concealment is described by taking MV34 in Fig.3 for example. Here, assuming MV34 is lost, i.e. Frame P4 is corrupted. Taking storing capacity and computing complexity of the decoder into account, four motion vectors are regarded as candidate vectors: MV01, MV12, MV23 and. Due to random packet loss, either all or some MV 45 of candidate vectors are received. The principle of choosing vectors is to select the ones that are closer as possible. As a consequence, the to MV34 following strategies are used to determine candidate vectors: Strategy 1: choosing MV 23 and MV45 when all candidate vectors are received 1/2( MV 23 + MV 45 Strategy 2: choosing MV12 not received ) (1) and MV45 when MV23 is in the received ones, to replace MV34 in the other cases. 4 Experiments directly 4.1 Choosing Strategies Testing Segments of Sequence Akiyo are randomly extracted to test the validity of choosing strategies, where the first frames are intra-encoded and the others are inter-encoded. Assuming the fifth frames are suffered, the discarded parts of the fifth frames are concealed by the above strategies, respectively. Figure 4. shows the test results of one of video segment ( Frame 0 ~ Frame 5), where some part of Frame 4 is lost. Comparing Figure 4. (c) ~ (h), the corresponding PSNRs are decreasing, which validate the priority of choosing candidate vectors is right. Note, the PSNR of (g) is lowest because in this given video segment, the eyes part in the first four frames is slow-motion while there is quick-break from Frame 4 to Frame 5. Only the future motion vector is considered in (g), so the effect of Strategy 5 is worse than the one of Strategy 6 against. 1/3( MV12 +2 MV45 Strategy 3: choosing MV12 not received 2 MV23 - MV 12 Strategy 4: choosing MV 01 and MV45 are not received ) (2) 1/2(3 MV23 - MV 01 and MV23 and MV23 when MV45 (3) is when MV12 ) (4) (a) Original frame (b) No concealment PSNR=19.07dB Strategy 5: choosing 23 MV is received MV 23 or 45 MV 23 MV or MV45 or MV34 MV 45 Strategy 6: choosing MV 01 MV 23 nor MV45 is received 3 MV12-2 MV01 and MV12 when only (5) when neither (6) (c) Strategy 1 (d) Strategy 2 PSNR=32.51dB PSNR=32.17dB Strategy 7: choosing the vector, which is closest to

(e) Strategy 3 (f) Strategy 4 PSNR=31.26dB PSNR=31.09dB (g) Strategy 5 ( MV45 ) (h) Strategy 6 PSNR=29.77dB PSNR=31.05dB Fig.4 Reconstructed frames by using different candidate vectors The average PSRNs gained by testing a great deal sequence segments are shown in Table 1: 1) the proposed multiple prediction method is effective on error concealment; 2) different candidate vectors have different abilities to conceal. It must be pointed out that Strategy 3 just is the improvement to dual forward prediction and Strategy 5 is single prediction actually. Choosing strategies PSRN(dB) No concealment 19.02 Strategy 1 33.56 Strategy 2 31.66 Strategy 3 31.10 Strategy 4 31.04 Strategy 5 29.31 Strategy 6 28.41 Table 1 PSNRs of reconstructed frames 4.2 Internet Testing The MPEG-4 Verification Model (without B frames) is used for coding the standard raw video sequence Akiyo in QCIF format under the Gilbert channel model. In order to avoid error propagation into the subsequent frames, the packetization algorithm based on FlexMux layer is adopted according to Ref.[7]. The goal of the packetization is that the data of the same frame is put into a packet as possible, so dependency among the packets can be removed. In the experiment, each packet contains one P-frame. Shown as Table 2, 1) there are the different PSNRs under the different packet loss rates; 2) the higher is the loss rate, the stronger should be the degradation of the PSNRs; 3) the quality of reconstructed video is still acceptable in the case of 2 or 3 packets lost continuously. values of p and q probabilities of n continuous lost packets n=1 n=2 n=3 PSNR(dB) p=0 0 0 0 37.59 p=0.05 q=0.85 p=0.67 q=0.07 p=0.10 q=0.70 p=0.50 q=0.20 0.050 0.008 0.001 35.45 0.070 0.023 0.008 31.51 0.100 0.030 0.009 29.52 0.200 0.100 0.050 24.99 Table 2 PSNRs under different loss rates 5 Conclusion In this paper, the new concealment method for P-frame is described in enough view of the relativities in the motion vectors of the interframes, and the experimental results are satisfied. It is pointed out that this paper only analyses temporal concealment, and in fact the neighbor motion vectors around the discarded parts of the same frame also be used to conceal if the appropriate interleaving technique is introduced. This will be our research in the future. Acknowledgements This project supported by the Key Disciplinary Development Program of Shanghai (2001-44) and Shanghai Leading Academic Disciplines (T0102).

References: [1] Tamer Shanableh, Mohammed Ghanbai, Loss Concealment Using B-Pictures Motion Information, IEEE Trans. Multimedia, VOL.5, NO.2, 2003, pp. 257-266 [2] Y. Wang, Q. F. Zhu and L. Shaw, et al. Maximally Smooth Image Recovery in Transform Coding, IEEE Trans. Communication, VOL.41, NO.10, 1993, pp.1544-1551 [3] H. Sun, W. Kwok, Concealment of Damaged Block Transform Coded Image Using Projections onto Convex Sets, IEEE Trans. Image Processing, VOL.4, NO.4, 1995, pp. 470-477 [4] C. P. Lim, E. A. W. Tan, M. Ghanbari and S. Ghanbari, Cell Loss Concealment and Packetization in packet video, Int. J. Imag. Syst. Technol., Special Issue on Image and Video Compression, VOL. 10, NO. 1, 1999, pp.54-58 [5] K. Imurra, Y. Machida, Error Resilient Video Schemes for Real-Time and Low-Bitrate Mobile Communications[J], Signal Processing: Image Commnun., VOL.14, NO.5, 1999, pp.519-530 [6] ITU-TSS Q.15/SG 16., The accompanying document Appendix 11.pdf about Proposed Error Patterns for Internet Experiments. Red Bank,New Jersey,1999 [7] P. Meyorya, R. Lamy and L. Besacier, Recovering of Packet Loss for Distributed Speech Recognition, Proc. Eusipco 2002, France, 2002 [8] M. Yajnik, S. Moon, J. Kurose and D. Towsley, Measurement and Modelling of the temporal Dependence in Packet Loss, IEEE Inform 99, New York, 1999. [9] Jae-Young Pyun, Jae-Jeong Shim, Sung-Jea Ko and Sang-Hyun Park, Packet Loss Resilience for Video Stream over the Internet, IEEE Trans. Consumer Electronics, VOL.48, NO.3, 2002, pp.556~563 [10] Dapeng Wu, Yiwei Thomas, Wenwu Zhu, Hung-Ju Lee, Tihao Chiang, Ya_Qin Zhang and H. Jonathan Chao, On End-to-End Architecture for Transporting MPEG-4 Video Over the Internet, IEEE Trans. Circuits & System For Video Technology, VOL.10, NO.6, 2000, pp.923-941