Spatial Sparsity Induced Temporal Prediction for Hybrid Video Compression

Similar documents
BLOCK MATCHING-BASED MOTION COMPENSATION WITH ARBITRARY ACCURACY USING ADAPTIVE INTERPOLATION FILTERS

Coding of Coefficients of two-dimensional non-separable Adaptive Wiener Interpolation Filter

Reduced Frame Quantization in Video Coding

Iterated Denoising for Image Recovery

Optimal Estimation for Error Concealment in Scalable Video Coding

Deblocking Filter Algorithm with Low Complexity for H.264 Video Coding

An Efficient Mode Selection Algorithm for H.264

Data Hiding in Video

Fast Decision of Block size, Prediction Mode and Intra Block for H.264 Intra Prediction EE Gaurav Hansda

Pre- and Post-Processing for Video Compression

Optimizing the Deblocking Algorithm for. H.264 Decoder Implementation

Performance Comparison between DWT-based and DCT-based Encoders

Multi-View Image Coding in 3-D Space Based on 3-D Reconstruction

Frequency Band Coding Mode Selection for Key Frames of Wyner-Ziv Video Coding

Overview: motion-compensated coding

New Techniques for Improved Video Coding

Rate Distortion Optimization in Video Compression

RICE UNIVERSITY. Spatial and Temporal Image Prediction with Magnitude and Phase Representations. Gang Hua

An Optimized Template Matching Approach to Intra Coding in Video/Image Compression

Video compression with 1-D directional transforms in H.264/AVC

Image Segmentation Techniques for Object-Based Coding

Wavelet-Based Video Compression Using Long-Term Memory Motion-Compensated Prediction and Context-Based Adaptive Arithmetic Coding

H.264/AVC BASED NEAR LOSSLESS INTRA CODEC USING LINE-BASED PREDICTION AND MODIFIED CABAC. Jung-Ah Choi, Jin Heo, and Yo-Sung Ho

A Quantized Transform-Domain Motion Estimation Technique for H.264 Secondary SP-frames

Comparative Study of Partial Closed-loop Versus Open-loop Motion Estimation for Coding of HDTV

Module 7 VIDEO CODING AND MOTION ESTIMATION

Fast frame memory access method for H.264/AVC

ERROR-ROBUST INTER/INTRA MACROBLOCK MODE SELECTION USING ISOLATED REGIONS

Reduced 4x4 Block Intra Prediction Modes using Directional Similarity in H.264/AVC

EE 5359 Low Complexity H.264 encoder for mobile applications. Thejaswini Purushotham Student I.D.: Date: February 18,2010

Fraunhofer Institute for Telecommunications - Heinrich Hertz Institute (HHI)

IBM Research Report. Inter Mode Selection for H.264/AVC Using Time-Efficient Learning-Theoretic Algorithms

AN ALGORITHM FOR BLIND RESTORATION OF BLURRED AND NOISY IMAGES

International Journal of Emerging Technology and Advanced Engineering Website: (ISSN , Volume 2, Issue 4, April 2012)

Adaptive Up-Sampling Method Using DCT for Spatial Scalability of Scalable Video Coding IlHong Shin and Hyun Wook Park, Senior Member, IEEE

Complexity Reduced Mode Selection of H.264/AVC Intra Coding

A deblocking filter with two separate modes in block-based video coding

Optimum Quantization Parameters for Mode Decision in Scalable Extension of H.264/AVC Video Codec

A Novel Deblocking Filter Algorithm In H.264 for Real Time Implementation

Pattern based Residual Coding for H.264 Encoder *

Intra-Mode Indexed Nonuniform Quantization Parameter Matrices in AVC/H.264

An Improved H.26L Coder Using Lagrangian Coder Control. Summary

Optimized Progressive Coding of Stereo Images Using Discrete Wavelet Transform

Implementation and analysis of Directional DCT in H.264

EE Low Complexity H.264 encoder for mobile applications

Efficient Halving and Doubling 4 4 DCT Resizing Algorithm

Block-based Watermarking Using Random Position Key

Bit Allocation for Spatial Scalability in H.264/SVC

Mesh Based Interpolative Coding (MBIC)

Network Image Coding for Multicast

Title Adaptive Lagrange Multiplier for Low Bit Rates in H.264.

STUDY AND IMPLEMENTATION OF VIDEO COMPRESSION STANDARDS (H.264/AVC, DIRAC)

Rate-distortion Optimized Streaming of Compressed Light Fields with Multiple Representations

Video Coding Using Spatially Varying Transform

Research on Distributed Video Compression Coding Algorithm for Wireless Sensor Networks

DIGITAL TELEVISION 1. DIGITAL VIDEO FUNDAMENTALS

Efficient MPEG-2 to H.264/AVC Intra Transcoding in Transform-domain

Stereo Image Compression

Rate-distortion Optimized Streaming of Compressed Light Fields with Multiple Representations

A LOW-COMPLEXITY AND LOSSLESS REFERENCE FRAME ENCODER ALGORITHM FOR VIDEO CODING

Comparative and performance analysis of HEVC and H.264 Intra frame coding and JPEG2000

Decoded. Frame. Decoded. Frame. Warped. Frame. Warped. Frame. current frame

FAST AND EFFICIENT SPATIAL SCALABLE IMAGE COMPRESSION USING WAVELET LOWER TREES

VHDL Implementation of H.264 Video Coding Standard

Digital Video Processing

An Efficient Table Prediction Scheme for CAVLC

SINGLE PASS DEPENDENT BIT ALLOCATION FOR SPATIAL SCALABILITY CODING OF H.264/SVC

EE 5359 MULTIMEDIA PROCESSING SPRING Final Report IMPLEMENTATION AND ANALYSIS OF DIRECTIONAL DISCRETE COSINE TRANSFORM IN H.

A 3-D Virtual SPIHT for Scalable Very Low Bit-Rate Embedded Video Compression

An Approach for Reduction of Rain Streaks from a Single Image

IMAGE DENOISING TO ESTIMATE THE GRADIENT HISTOGRAM PRESERVATION USING VARIOUS ALGORITHMS

Efficient Dictionary Based Video Coding with Reduced Side Information

An Optimized Pixel-Wise Weighting Approach For Patch-Based Image Denoising

2014 Summer School on MPEG/VCEG Video. Video Coding Concept

Model-based Enhancement of Lighting Conditions in Image Sequences

Video Transcoding Architectures and Techniques: An Overview. IEEE Signal Processing Magazine March 2003 Present by Chen-hsiu Huang

Bit-Depth Scalable Coding Using a Perfect Picture and Adaptive Neighboring Filter *

A COST-EFFICIENT RESIDUAL PREDICTION VLSI ARCHITECTURE FOR H.264/AVC SCALABLE EXTENSION

Outline Introduction MPEG-2 MPEG-4. Video Compression. Introduction to MPEG. Prof. Pratikgiri Goswami

Homogeneous Transcoding of HEVC for bit rate reduction

IN the early 1980 s, video compression made the leap from

MCTF and Scalability Extension of H.264/AVC and its Application to Video Transmission, Storage, and Surveillance

Structure-adaptive Image Denoising with 3D Collaborative Filtering

Patch-Based Color Image Denoising using efficient Pixel-Wise Weighting Techniques

DCT image denoising: a simple and effective image denoising algorithm

NEW CAVLC ENCODING ALGORITHM FOR LOSSLESS INTRA CODING IN H.264/AVC. Jin Heo, Seung-Hwan Kim, and Yo-Sung Ho

Mode-Dependent Pixel-Based Weighted Intra Prediction for HEVC Scalable Extension

MOTION estimation is one of the major techniques for

Performance Analysis of DIRAC PRO with H.264 Intra frame coding

A Low Bit-Rate Video Codec Based on Two-Dimensional Mesh Motion Compensation with Adaptive Interpolation

Advanced Video Coding: The new H.264 video compression standard

Advanced Encoding Features of the Sencore TXS Transcoder

FAST MOTION ESTIMATION WITH DUAL SEARCH WINDOW FOR STEREO 3D VIDEO ENCODING

Template based illumination compensation algorithm for multiview video coding

Review and Implementation of DWT based Scalable Video Coding with Scalable Motion Coding.

Compression of Stereo Images using a Huffman-Zip Scheme

Video Compression An Introduction

Complexity Reduction Tools for MPEG-2 to H.264 Video Transcoding

One-pass bitrate control for MPEG-4 Scalable Video Coding using ρ-domain

A Novel Statistical Distortion Model Based on Mixed Laplacian and Uniform Distribution of Mpeg-4 FGS

Transcription:

Spatial Sparsity Induced Temporal Prediction for Hybrid Video Compression Gang Hua and Onur G. Guleryuz Department of Electrical and Computer Engineering, Rice University, Houston, TX 77005 ganghua@rice.edu DoCoMo USA Laboratories, Inc. 3240 Hillview Avenue, Palo Alto, CA 94304 guleryuz@docomolabs-usa.com January 10, 2007 Abstract In this paper we propose a new motion compensated prediction technique that enables successful predictive encoding during fades, blended scenes, temporally decorrelated noise, and many other temporal evolutions which force predictors used in traditional hybrid video coders to fail. We model reference frame blocks to be used in motion compensated prediction as consisting of two superimposed parts: One part that is relevant for prediction and another part that is not relevant. By performing prediction in a domain where the video frames are spatially sparse, our work allows the automatic isolation of the prediction-relevant parts. These are then used to enable better prediction than would be possible otherwise. Our sparsity induced prediction algorithm (SIP) generates successful predictors by exploiting the non-convex structure of the sets that natural images and video frames lie in. Correctly determining this non-convexity through sparse representations allows better performance in hybrid video codecs equipped with the proposed work. 1 Introduction As video compression matures, temporal prediction techniques that try to yield significant performance improvements must concentrate on providing gains over ever more sophisticated evolutions in video. In this paper we propose a technique that targets temporal evolutions over which motion compensated prediction employed by stateof-the-art codecs [6, 11] results in failed predictors, culminating in non-differentially

Temporal Evolution Reference frame Frame to be Coded Required processing for each predicted block Prediction Prediction Accuracy (PSNR) Scene transition from a blend of two scenes (peppers & barbara). Denoise, find peppers out of the blend of peppers & barbara, amplify peppers. 28.954 db (a) Scene transition from a blend of three scenes (peppers, barbara & boat). Denoise, find peppers out of the blend of peppers, barbara & boat, amplify peppers. 26.874 db (b) Scene transition with a cross-fade. One scene (barbara) fades out, the other (peppers) fades in. Denoise, find barbara, reduce barbara, find peppers, amplify peppers. 34.952 db (c) Brightness change due to a lightmap. Denoise, find lightmap, invert lightmap. 31.697 db (d) Scene transition from a blend with a brightness change. (e) Denoise, find lightmap, invert lightmap, find peppers out of the blend of peppers & barbara, amplify peppers. 27.274 db Figure 1: Example temporal evolutions that can be targeted using the prediction algorithm proposed in this paper. The frame to be coded is predicted using the reference frame in a hybrid video compression setting. Both frames have additive Gaussian noise of standard deviation σ w = 5. The fourth column provides a summary of the required processing for successful prediction (our algorithm accomplishes these results using simple low-level prediction). Traditional motion compensated prediction results in significant prediction errors that result in non-differential encoding of the frame to be coded. The prediction accuracy of our technique is shown in the last column. Observe that the prediction is successful even under complicated scenarios that involve brightness changes and sophisticated fades. The algorithm manages to fish-out scenes, recombine them, correct lighting, etc., to form these predictors. encoded (INTRA) frames and macroblocks. Figure 1 uses example video frames composed of standard test images to illustrate a simple subset of the temporal evolutions that can be effectively targeted using our work. Traditional motion compensated prediction relies on translated blocks from reference frames to directly match blocks in the frame to be coded. However, translated reference frame blocks may consist of two superimposed parts: One part that is rele-

vant for prediction and another part that is not relevant. In many types of temporal evolutions in video, such as fades from one scene to another, blended scenes, temporally decorrelated noise, etc., the prediction-irrelevant part can become severe and significantly hurt prediction accuracy, resulting in the encoding of expensive INTRA macroblocks. By performing prediction in a domain where video frames are expected to be sparse, the technique we propose allows the isolation of the prediction-relevant parts of predicting blocks. These isolated parts are then used to enable better prediction than would be possible with conventional techniques. Conceptually, our work processes reference frame blocks that are about to be used in prediction so that they become much better predictors of the frame to be coded. Figure 2 shows the positioning of the proposed technique inside a basic hybrid video encoder. Frame to be coded + + - Transform Coder (causal information) Coded differential Transform Decoder + + + Sparsity Induced Prediction Motion Compensation 1 z Previously decoded frame, to be used as reference Figure 2: The proposed work inside a basic hybrid video encoder. Sparsity induced prediction is incorporated as a module that processes motion compensated prediction estimates so that they become better predictors of the frame to be coded. Our work formulates predictors in spatial transform domain where we assume that the frame to be coded and the reference frames to be used in its prediction are sparse. By using causal information we estimate temporal correlations in transform domain, formulate predictions of the transform coefficients of the frame to be coded, and finally perform an inverse transform to obtain the equivalent pixel domain prediction. In contrast, temporal prediction research in video compression has mainly concentrated on simple pixel domain parameterizations that are geared toward accounting for interpolation errors, aliasing noise, and brightness changes that may be present in the reference frames (see for e.g., [10, 9, 7]). While these techniques do form better predictors, the temporal evolution models they are targeted toward are very limited compared to the work proposed here 1. Our work may also be related to multihypothesis prediction approaches [3]. Note that filter-based techniques typically result in a small set of filters having low-pass characteristics which are not applicable on frames rich in spatial frequencies and textures (see Section 2 and Figure 3). As we will show, by only using causal 1 Tools like weighted prediction [6] can target fades but require blending scenes to be available among previously decoded frames, in isolated form.

information over limited spatial neighborhoods, our technique can result in effective prediction over widely varying spatial statistics. Earlier research on the other hand needs to signal per-block choice of filters with overhead bits and typically requires much larger neighborhoods for the design of filters. One of the strengths of our work is its effective transform domain parameterization which allows adaptation and accurate prediction with little training data. This is a desirable characteristic when operating over nonstationary frame statistics with localized singularities. Being close to the temporal prediction techniques cited above in terms of application area, our work is conceptually closer to denoising and inverse blending techniques that use sparse representations and exploit the non-convex structure of the sets that natural images lie in (see for e.g., [1, 2, 8]). The inverse blending techniques typically operate by using two images that are different blends of two target natural images. Note that our framework is very different as we predictively encode one image based on the other. Our work is also more powerful as it can handle noise, blends involving more than two images, cross-fades (i.e., a transition from one blend to a different blend so that the target image is a blend itself), brightness changes, and many more evolutions as well as their combinations. In comparison to l p regularized denoising setups, note that the noise that we remove from the reference frame is highly structured and cannot be dealt with using simple denoising techniques. Section 2 introduces the basic ideas and motivates the proposed technique. The main algorithm is considered in Section 3 and simulation results are provided in Section 4. The paper concludes in Section 5 with concluding remarks. 2 Basic Ideas Let x n (N 1) denote the frame to be coded and let y be the reference frame to be used in its prediction. For notational convenience assume that only one reference frame will be used in prediction and that motion compensation has taken place, i.e., if x n 1 denotes the decoded reference frame, y = MC( x n 1 ), where MC denotes the motion compensation operation. Let H 1,...,H M (N N) denote an overcomplete set of linear transforms. For issues of implementation simplicity and computational speed, in this paper H i, i = 1,...,M, are given by a p p block DCT transform and its p 2 1 shifts with M = p 2. We note however that it is straightforward to utilize our work with different transforms and different redundancy factors. We would like to implement M temporal predictors of x n which we denote as ˆx i n, i = 1,...,M. Each predictor is simple and is composed via c i = H T i y (1) d i (j) = γ i (j)c i (j), j = 1,...,N (2) ˆx i n = H i d i, (3)

where γ i (j) are linear prediction weights that are determined to minimize mean squared prediction error. The final prediction of x n is obtained as an average of the M predictors via ˆx n = 1 M ˆx i M n, (4) though weighted averaging techniques similar to [5] can also be used. i=1 The rationale for forming predictors using (1) - (4) can be summarized as follows: Localized linear transforms have good statistical decorrelation properties so that on most video sequences the prediction model of Equation 2 remains valid, i.e., a transform coefficient of x n is mostly correlated to the same coefficient of y. The simple transform domain parameterization of (1) - (4) thus obviates the need to determine complex cross correlations over large neighborhoods as would be required by general pixel domain formulations with many parameters. Images and video frames are sparse in transform domain with many transform coefficients zero or close to zero. The spatial structures of the remaining large coefficients form delineating signatures for images, and together with non-convex image models arising from sparse representations [4], lead to intuitive cartoons as in Figure 3. These properties allow our algorithms to easily fish-out one scene from a blend of two or more, predict during cross-fades, etc. Using translation invariant decompositions allow high performance around localized image singularities so that a portion of the transforms provide accurate predictions even when others cannot due to model failures [1] 2. Of course, if desired, it is straightforward to augment our basic prediction setup in (1) through (4) to account for further correlations among coefficients. 3 Main Algorithm Modern high-performance video compression involves many steps and optimizations that are geared toward satisfying several important requirements. Macroblock structure utilized in codecs and the order in which macroblocks are coded present practical issues for any temporal prediction algorithm. In this section we discuss our main algorithm that is geared toward operation inside a high performance video codec (h264/avc reference software - JM10.2 [6]) in macroblock units. Our main concern is the design of an algorithm that obtains the prediction weights in (2) using available information during the decoding of each macroblock. Since the utilized transforms are block based, the transform domain prediction operation can be thought of as predicting p p blocks in x n using the corresponding blocks in y and then averaging the predictions suitably to satisfy (4). In order to help 2 Translation invariant decompositions also reduce blocking effects when using block based transforms.

Suppress when predicting h k x 2 Suppress when predicting h k x (amplify when predicting h k x 1 ) (amplify when predicting h k x 1 ) 2 h k x 1 x2 h k ( x 1 x2) / 2 overlap Figure 3: Band-pass filtered versions of two toy images x 1 and x 2 and of their blend (the filter corresponds to the k th block DCT basis, 0 < k p 2 ). The cartoons depict the values that would be obtained through a translational invariant decomposition at the k th frequency. The resulting values tend to be significant over islands that usually correspond to uniform patches, objects, and edges in typical scenes (shown shaded - unshaded regions denote insignificant values). Except for overlaps, it is easy to obtain h k x 1 or h k x 2 from h k (x 1 + x 2 )/2 using the prediction model of Equation 2 with the aid of temporal correlations. On overlaps the undesired portions act as noise and the prediction model will provide predictions of reduced accuracy. As indicated in the figure, correct prediction will amplify or suppress the same frequencies in a spatially varying fashion depending on the target image. notation, let us adopt a block viewpoint and deviate slightly from the frame-wide notation of Section 2. Suppose b x (p 2 1) represents a p p block in x n that is to be predicted and let b y be the corresponding block in y. We know the transform coefficients of b y which we will use via (4) to determine predictions of the transform coefficients of b x. We need to determine the per coefficient prediction weights 3. Figure 4 (a) illustrates b x within the frame to be coded (current frame), inside the macroblock to be coded (current macroblock). In determining the prediction weights we would like to utilize the information provided by previously decoded macroblocks inside the current frame 4. Let Λ bx denote a L L spatial neighborhood around b x. Consider all blocks formed by shifting a p p mask inside Λ bx. Denote those blocks that completely overlap known data (training data) by t 1,...,t Q. Let u 1,...,u Q be the corresponding blocks in y. Block prediction: Suppose we are performing the prediction for the k th DCT coefficient of b x. Let h k (p 2 1) denote the k th DCT basis function. We derive the prediction weight γ k as the weight that minimizes the mean squared prediction error on the training data, i.e., 3 For the moment we leave issues related to motion estimation to Section 4 and keep assuming that y is such that translations are properly accounted for. 4 If previously decoded macroblocks are not available one can set the prediction weights to an appropriate value (such as 1) or determine optimal weights and send them using overhead bits. h k

Q γ k = argmin h T k t q γh T k u q 2. (5) γ q=1 The prediction for the k th DCT coefficient in b x is then set as ˆ h T k b x = γ k h T k b y. (6) Once the prediction is carried out for all DCT coefficients of b x, an inverse block DCT yields the pixel domain prediction ˆb x. Observe that the solution of (5) is straightforward and the computations for its calculation can be reused when predicting other blocks so that complexity remains contained. Given our desired predictor in (4), it is clear that we need to predict all p p blocks in the current macroblock, i.e., all shifts of a p p block that overlap the current macroblock. h.264/avc utilizes a 4 4 transform (compression transform) to encode prediction errors. In forming predictors for our blocks, we try to utilize available data as much as possible so that compression transform coefficients of the prediction error are used to augment the known data regions of the current macroblock as soon as they become available. Note however that the exact pattern in which compression transform coefficients are sent depends on the motion modes determined at the motion estimation stage [6]. For example, the motion mode where motion blocks are 8 8 sends coefficients in a different order compared to the 16 16 mode. As such, our technique orders block predictions based on the motion mode in order to utilize prediction error updates as much as possible. In order to save space, we avoid further discussion of these detailed but conceptually straightforward implementation issues in the presentation. Our implementation also utilizes predictions provided for previous blocks to augment the training data by scanning the current macroblock in layers (see Figure 4 (b) for an example scan). The procedure below outlines sparsity induced prediction of a macroblock as carried out by a hybrid video decoder: Scan all p p blocks overlapping the current macroblock (Figure 4 (b)). Predict the p p block. Update current prediction. Store prediction for final averaging. Expand and update the training region. Combine stored predictions for an overall prediction inside the macroblock. 4 Simulation Results We incorporated the SIP algorithm inside h264/avc reference software, JM10.2 [6]. In our simulation results we used p = 4 and L = 12 (corresponding to +/ 1 block around each p p block). For each macroblock, we used a 1 bit overhead to determine if the sparsity induced prediction is used or not. This bit was set within the ratedistortion optimization loop of the JM software. Only luminance macroblocks were predicted with the new technique. All test sequences are QCIF (144 176) resolution.

Neighborhood b x Block b x Previously decoded macroblocks p p L Current macroblock Macroblocks to be coded L (a) Figure 4: Implementing the prediction algorithm for macroblock based operation. (a) Block b x and neighborhood Λ bx. (b) Example prediction scan inside a macroblock in layers of horizontal, one pixel thick strips. Previous predictions are incorporated into the training data of later ones. At certain points in the scan, prediction of compression transform blocks are completed and corresponding prediction error updates are incorporated. Beyond the traditional motion search, we implemented a motion search on an integer grid to find the optimal integer motion vectors for SIP. Figure 5 shows a set of five-frame transitions from the video sequences car (Figure 5 (a), (b), and (d)) and glasgow (Figure 5 (c)). In Figure 6 we provide corresponding rate-distortion results using QP = 22, 26, 30. Each five-frame sequence was encoded in the IPPPP pattern. We report rate at 30 frames/sec. Both rate and distortion are averages for the four P frames since both h.264/avc and SIP utilize the same INTRA frame. Figure 5 (a)-(c) correspond to fades, and (d) depicts a special effect with localized blurring, fades, and brightness changes. Figure 5 (a) has mostly rotational motion, (b) and (c) have translations, and (d) is stationary. As seen in Figure 6, SIP obtains improvements for all cases even for scenes rich in spatial frequencies. Allowing the use of fractional motion is expected to improve these results. Over entire sequences the gains provided by the proposed technique varies depending on the density of sophisticated temporal evolutions in the sequence. On simple sequences (such as container ), gains are on the order of %2 5 improvements in rate at constant distortion, whereas on more complicated sequences (such as movie trailers), gains can go up to %10 15. The bulk of the per-frame decoding complexity of the proposed work can be summed as 3p 2 multiplies +p 2 divides +4p 2 additions per-pixel (in order to solve (5) and apply (6)), and a translation invariant DCT decomposition. Note however that this complexity can be reduced significantly by reducing the size of the training set. Encoder complexity is more cumbersome due to the motion search. Complexity can be alleviated by restricting the technique to macroblocks where traditional prediction fails or is deemed inefficient in a ratedistortion sense. Encoder complexity can be reduced by using fast motion search (b)

(a) (b) (c) (d) Figure 5: Test set of five-frame transitions. 42 42 41 41 40 Distortion (PSNR) Distortion (PSNR) 40 39 38 37 38 37 36 36 35 SIP h.264/avc 35 150 39 200 250 300 350 400 Rate (kbits/sec) 450 500 550 34 100 600 SIP h.264/avc 200 300 (a) 42 39 41 600 700 40 Distortion (PSNR) 38 Distortion (PSNR) 500 (b) 40 37 36 35 39 38 37 36 34 SIP h.264/avc 33 300 400 Rate (kbits/sec) 400 500 600 700 800 900 Rate (kbits/sec) (c) 1000 1100 1200 1300 35 34 100 SIP h.264/avc 150 200 250 300 Rate (kbits/sec) 350 400 450 (d) Figure 6: Rate-distortion results corresponding to each of the transitions in Figure 5. The average utilization of the new mode in the predicted frames in (a), (b), (c), and (d) is about 44%, 78%, 60%, 14% of the macroblocks, respectively.

algorithms. 5 Conclusion We introduced a temporal prediction algorithm that enables successful differential encoding in hybrid video coders during sophisticated temporal evolutions where traditional motion compensated prediction fails. Our algorithm exploits temporal correlations in spatial transform domain where video frames are assumed to be sparse. Simulation results demonstrate the efficacy of a simple implementation of our formulation inside a modern, high performance video codec. Even with a simple transform and statistics determined over spatially limited neighborhoods, our technique can accomplish high performance predictions. Our work can be augmented by sending more overhead information so that prediction accuracy can be improved in cases where causal statistics lead to inaccurate results. References [1] R. R. Coifman and D. L. Donoho, Translation invariant denoising, in Wavelets and Statistics, Springer Lecture Notes in Statistics 103, pp. 125-150, New York:Springer-Verlag. [2] M. Elad and M. Aharon, Image Denoising Via Sparse and Redundant representations over Learned Dictionaries, IEEE Trans. on Image Processing, Vol. 15, no. 12, pp. 3736-3745, December 2006. [3] B. Girod, Efficiency Analysis of Multihypothesis Motion-Compensated Prediction for Video Coding, IEEE Trans. on Image Processing, Vol. 9, no. 2, pp. 173-183, Feb. 2000. [4] O. G. Guleryuz, Nonlinear Approximation Based Image Recovery Using Adaptive Sparse Reconstructions and Iterated Denoising: Part I - Theory, IEEE Transactions on Image Processing, vol. 15, No. 3, pp. 539-554, March, 2006. [5] O. G. Guleryuz, Weighted Overcomplete Denoising, Proc. Asilomar Conference on Signals and Systems, Pacific Grove, CA, Nov. 2003. [6] Joint Video Team of ITU-T and ISO/IEC JTC 1, Draft ITU T Recommendation and Final Draft International Standard of Joint Video Specification (ITU-T Rec. H.264 ISO/IEC 14496-10 AVC), Joint Video Team (JVT) of ISO/IEC MPEG and ITU-T VCEG, JVT-G050, March 2003. [7] K. Kamikura, H. Watanabe, H Jozawa, H. Kotera, and S. Ichinose, Global Brightness- Variation Compensation for Video Coding, IEEE Transactions on Circuits and Systems for Video technology, Dec. 1998. [8] P. Kisilev, M. Zibulevshy, and Y. Y. Zeevi, Blind separation of mixed images using multiscale transforms, Proc. IEEE Conf. on Image Proc., ICIP2003, Barcelona, Spain, 2003. [9] J. Vatis, B. Edler, I. Wassermann, D. T. Nguyen, and J. Ostermann, Coding of Coefficients of two-dimensional non-separable Adaptive Wiener Interpolation Filter Proc. VCIP 2005, SPIE Visual Communication and Image Processing, Bejing,China. 2005. [10] T. Wedi, Adaptive Interpolation Filter for Motion and Aliasing Compensated Prediction, Proc. Electronic Imaging 2002: Visual Communications and Image Processing (VCIP 2002), San Jose, California USA, January 2002. [11] T. Wiegand, G. J. Sullivan, G. Bjontegaard, and A. Luthra, Overview of the H.264/AVC Video Coding Standard, IEEE Trans. on CSVT, vol. 13, no.7, pp. 560-576, July 2003.