Citation for the original published paper (version of record):

Similar documents
Frequency Band Coding Mode Selection for Key Frames of Wyner-Ziv Video Coding

Research on Distributed Video Compression Coding Algorithm for Wireless Sensor Networks

Robust Video Coding. Heechan Park. Signal and Image Processing Group Computer Science Department University of Warwick. for CS403

MOTION ESTIMATION AT THE DECODER USING MAXIMUM LIKELIHOOD TECHNIQUES FOR DISTRIBUTED VIDEO CODING. Ivy H. Tseng and Antonio Ortega

Distributed Video Coding

LOW DELAY DISTRIBUTED VIDEO CODING. António Tomé. Instituto Superior Técnico Av. Rovisco Pais, Lisboa, Portugal

A Novel Statistical Distortion Model Based on Mixed Laplacian and Uniform Distribution of Mpeg-4 FGS

Region-based Fusion Strategy for Side Information Generation in DMVC

Rate Distortion Optimization in Video Compression

Distributed Grayscale Stereo Image Coding with Unsupervised Learning of Disparity

Distributed Source Coding for Image and Video Applications. Acknowledgements

2014 Summer School on MPEG/VCEG Video. Video Coding Concept

Module 7 VIDEO CODING AND MOTION ESTIMATION

STUDY AND IMPLEMENTATION OF VIDEO COMPRESSION STANDARDS (H.264/AVC, DIRAC)

Reduced Frame Quantization in Video Coding

Fast Decision of Block size, Prediction Mode and Intra Block for H.264 Intra Prediction EE Gaurav Hansda

A Image Comparative Study using DCT, Fast Fourier, Wavelet Transforms and Huffman Algorithm

Compression Algorithms for Flexible Video Decoding

Topics in AI (CPSC 532L): Multimodal Learning with Vision, Language and Sound. Lecture 12: Deep Reinforcement Learning

ENCODER POWER CONSUMPTION COMPARISON OF DISTRIBUTED VIDEO CODEC AND H.264/AVC IN LOW-COMPLEXITY MODE

Compression of Stereo Images using a Huffman-Zip Scheme

A LOW-COMPLEXITY AND LOSSLESS REFERENCE FRAME ENCODER ALGORITHM FOR VIDEO CODING

arxiv: v1 [cs.cv] 2 Sep 2018

CS 260: Seminar in Computer Science: Multimedia Networking

Intra-Mode Indexed Nonuniform Quantization Parameter Matrices in AVC/H.264

VC 12/13 T16 Video Compression

JPEG compression of monochrome 2D-barcode images using DCT coefficient distributions

Performance Comparison between DWT-based and DCT-based Encoders

Review and Implementation of DWT based Scalable Video Coding with Scalable Motion Coding.

Fundamentals of Video Compression. Video Compression

Deblocking Filter Algorithm with Low Complexity for H.264 Video Coding

Compression of Light Field Images using Projective 2-D Warping method and Block matching

Complexity Reduction Tools for MPEG-2 to H.264 Video Transcoding

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition

WZS: WYNER-ZIV SCALABLE PREDICTIVE VIDEO CODING. Huisheng Wang, Ngai-Man Cheung and Antonio Ortega

System Modeling and Implementation of MPEG-4. Encoder under Fine-Granular-Scalability Framework

Introduction to Video Compression

Measurements and Bits: Compressed Sensing meets Information Theory. Dror Baron ECE Department Rice University dsp.rice.edu/cs

Video Compression An Introduction

A Video Coding Framework with Spatial Scalability

International Journal of Emerging Technology and Advanced Engineering Website: (ISSN , Volume 2, Issue 4, April 2012)

Video compression with 1-D directional transforms in H.264/AVC

Multimedia Communications. Transform Coding

Partial Video Encryption Using Random Permutation Based on Modification on Dct Based Transformation

ECE 533 Digital Image Processing- Fall Group Project Embedded Image coding using zero-trees of Wavelet Transform

10-701/15-781, Fall 2006, Final

Image Coding and Data Compression

SINGLE PASS DEPENDENT BIT ALLOCATION FOR SPATIAL SCALABILITY CODING OF H.264/SVC

A Quantized Transform-Domain Motion Estimation Technique for H.264 Secondary SP-frames

IBM Research Report. Inter Mode Selection for H.264/AVC Using Time-Efficient Learning-Theoretic Algorithms

Efficient and Low Complexity Surveillance Video Compression using Distributed Scalable Video Coding

Vidhya.N.S. Murthy Student I.D Project report for Multimedia Processing course (EE5359) under Dr. K.R. Rao

H.264 STANDARD BASED SIDE INFORMATION GENERATION IN WYNER-ZIV CODING

Context based optimal shape coding

An Optimized Template Matching Approach to Intra Coding in Video/Image Compression

CODING METHOD FOR EMBEDDING AUDIO IN VIDEO STREAM. Harri Sorokin, Jari Koivusaari, Moncef Gabbouj, and Jarmo Takala

Digital Video Processing

Optimizing the Deblocking Algorithm for. H.264 Decoder Implementation

Index. 1. Motivation 2. Background 3. JPEG Compression The Discrete Cosine Transformation Quantization Coding 4. MPEG 5.

Advanced Video Coding: The new H.264 video compression standard

System Modeling and Implementation of MPEG-4. Encoder under Fine-Granular-Scalability Framework

Multi-View Image Coding in 3-D Space Based on 3-D Reconstruction

An Imperceptible and Blind Watermarking Scheme Based on Wyner-Ziv Video Coding for Wireless Video Sensor Networks

CS 335 Graphics and Multimedia. Image Compression

Decoded. Frame. Decoded. Frame. Warped. Frame. Warped. Frame. current frame

Part 1 of 4. MARCH

EE Low Complexity H.264 encoder for mobile applications

A NOVEL SCANNING SCHEME FOR DIRECTIONAL SPATIAL PREDICTION OF AVS INTRA CODING

Compressive Sensing for Multimedia. Communications in Wireless Sensor Networks

Stereo Image Compression

Image Compression Algorithm and JPEG Standard

Navigarea Autonomă utilizând Codarea Video Distribuită Free Navigation with Distributed Video Coding (DVC)

UNIFIED PLATFORM-INDEPENDENT AIRBORNE NETWORKING ARCHITECTURE FOR VIDEO COMPRESSION

Lossless Image Compression with Lossy Image Using Adaptive Prediction and Arithmetic Coding

MRT based Fixed Block size Transform Coding

Distributed Grayscale Stereo Image Coding with Improved Disparity and Noise Estimation

Video Compression MPEG-4. Market s requirements for Video compression standard

Optimal Estimation for Error Concealment in Scalable Video Coding

Research on the New Image De-Noising Methodology Based on Neural Network and HMM-Hidden Markov Models

Block-based Watermarking Using Random Position Key

Study and Development of Image Authentication using Slepian Wolf Coding

ADVANCED IMAGE PROCESSING METHODS FOR ULTRASONIC NDE RESEARCH C. H. Chen, University of Massachusetts Dartmouth, N.

DIGITAL IMAGE PROCESSING WRITTEN REPORT ADAPTIVE IMAGE COMPRESSION TECHNIQUES FOR WIRELESS MULTIMEDIA APPLICATIONS

Chapter 11.3 MPEG-2. MPEG-2: For higher quality video at a bit-rate of more than 4 Mbps Defined seven profiles aimed at different applications:

STACK ROBUST FINE GRANULARITY SCALABLE VIDEO CODING

Lecture 5: Compression I. This Week s Schedule

When Network Embedding meets Reinforcement Learning?

AN ANALYTICAL STUDY OF LOSSY COMPRESSION TECHINIQUES ON CONTINUOUS TONE GRAPHICAL IMAGES

QUANTIZER DESIGN FOR EXPLOITING COMMON INFORMATION IN LAYERED CODING. Mehdi Salehifar, Tejaswi Nanjundaswamy, and Kenneth Rose

ARCHITECTURES OF INCORPORATING MPEG-4 AVC INTO THREE-DIMENSIONAL WAVELET VIDEO CODING

Wyner Ziv-Based Multiview Video Coding Xun Guo, Yan Lu, Member, IEEE, Feng Wu, Senior Member, IEEE, Debin Zhao, and Wen Gao, Senior Member, IEEE

15 Data Compression 2014/9/21. Objectives After studying this chapter, the student should be able to: 15-1 LOSSLESS COMPRESSION

IMAGE COMPRESSION TECHNIQUES

Lossy Coding 2 JPEG. Perceptual Image Coding. Discrete Cosine Transform JPEG. CS559 Lecture 9 JPEG, Raster Algorithms

Homogeneous Transcoding of HEVC for bit rate reduction

Lecture 8 JPEG Compression (Part 3)

ECE521: Week 11, Lecture March 2017: HMM learning/inference. With thanks to Russ Salakhutdinov

Pre- and Post-Processing for Video Compression

Implementation and analysis of Directional DCT in H.264

HYBRID TRANSFORMATION TECHNIQUE FOR IMAGE COMPRESSION

Transcription:

http://www.diva-portal.org This is the published version of a paper published in International Journal of Advanced Computer Sciences and Applications. Citation for the original published paper (version of record): Kouma, J-P., Söderström, U. (2016) Wyner-Ziv Video Coding using Hadamard Transform and Deep Learning. International Journal of Advanced Computer Sciences and Applications, 7(7): 582-589 http://dx.doi.org/10.14569/ijacsa.2016.070779 Access to the published version may require subscription. N.B. When citing this work, cite the original published paper. Permanent link to this version: http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-126345

Wyner-Ziv Video Coding using Hadamard Transform and Deep Learning Jean-Paul Kouma Ulrik Söderström Department of Applied Physics and Electronics Umea University, Sweden Abstract Predictive schemes are current standards of video coding. Unfortunately they do not apply well for lightweight devices such as mobile phones. The high encoding complexity is the bottleneck of the Quality of Experience (QoE) of a video conversation between mobile phones. A considerable amount of research has been conducted towards tackling that bottleneck. Most of the schemes use the so-called Wyner-Ziv Video Coding Paradigm, with results still not comparable to those of predictive coding. This paper shows a novel approach for Wyner-Ziv video compression. It is based on the Reinforcement Learning and Hadamard Transform. Our Scheme shows very promising results. Keywords Wyner-Ziv; video coding; rate distortion; Hadamard transform; Deep learning; Expectation Maximization I. INTRODUCTION Video compression schemes such as MPEG4 and H.264 [1] are the current state-of-art, where correlation between or among frames are exploited at the encoder side. Such schemes usually achieve high compression with a fairly low complexity at the very expense of a high complexity encoder. Compression schemes like MPEG4 or H.264 are suitable for scenarios where the encoder has enough power computation, like video-ondemand servers. Mobile phones are today the de facto device for communication. People want to do more and more with their mobile phone. They want to be able to have a real-time video communication experience comparable to that of computers. Unfortunately, current video compression technologies [1] barely permit it: encoder complexity is the bottleneck. Either comparable frame rate can not be achieved or conversation cannot last long because of battery is scarce. In either case quality of experience will be dropped. The video community is aware of that issue as a great deal of research has been conducted since the emergence of camerabased mobile devices. The common insight toward tackle the issue is the so-called Wyner-Ziv Video Coding (WZVC) or Distributed video coding. WZVC is the consequence of information-theoretic bounds established in the 1970s. First by Slepian and Wolf for distributed lossless coding [2], and then Wyner and Ziv for lossy coding with decoder side information [3]. Let {(X k, Y k )} k=1 be a sequence of independent drawing of a pair of dependent variables (X, Y ) taking values in the finite sets X and Y, respectively. The decoder has access to the side information Y. Illustration is shown in figure 1. Wyner and Ziv suggest that whether or not the side information Y is available at the encoder, X can be compressed - to Z and decoded to ˆX - at a rate R X Y (D) where D = E[d(X, ˆX)] is an acceptable distortion. In WZCV, unlike that of predictive coding paradigms (i.e. H.264), individual frames are encoded separately but decoded conditionally. According to [3], the compression effectiveness of WZVC schemes should be comparable to that of predictive coding. A typical WZVC setup is shown in figure 2 where both terminals are lightweight modern mobile phone capable of decoding MPEG4 frames for example. The corresponding Wyner-Ziv decoder is thought to be powerful computer capable of exploiting statistics between frames and output MPEG4 streams in real-time, using much more complex algorithms. Most of the conducted studies in the area of WZCV have been using binary codes. Major contributions come from Standford University [4] and UC Berkeley [5]. Both methods followed a common pattern; those methods were first developed to perform in pixel domain and later in transform domain (namely Discrete Cosine Transform). Those methods suffered from three major drawbacks: The overhead of working in binary domain - since DCT pixels or alternatively transform coefficients have to be converted back and forth from and to bit planes during decoding process. Rate control - All the pixels values, alternatively transform coefficients need to be converted to binary with the same amount of bits, making the rate control difficult. The decoding algorithms used - either generative or discriminative - were somewhat too simplistic and did not work well in practice. We propose a new practical compression scheme, based on Hadamard transform and reinforcement learning. In contrast to previous works, our method deals with non-binary codes. The encoding is relatively of low complexity with an inherent rate control. We also show that our algorithm outperforms that of the state of art [4] and [5] and really is comparable to predictive coding schemes. II. LOW COMPLEXITY VIDEO ENCODING The encoding challenge is to implement an encoder with lower encoding complexity than that of predictive video coding methods [1] and still achieve comparable codec effectiveness. 582 P a g e

Fig. 1. Source coding with side information available at the decoder Fig. 2. Wyner-Ziv setup of video compression A. Problem setup Let Y and X be two consecutive video frames. X is to be encoded without the knowledge of Y. The challenge is to compress X at the bit rate R bit/pixel, where R = H(Y X) =. H(Y X) is the entropy or the amount of information (in bits) needed to represent the frame X conditioned on that of Y. Usually, Y and X are highly correlated after motion compensation [6], H(X Y ) < H(X). In practice the challenge is to encode X at a rate even lower than R ɛ < H(Y X) - leading to lossy compression - and still achieve reconstruction with satisfying fidelity, as suggested by Ziv et al [3]. C. The Hadamard Transform The Hadamard transform is an orthogonal transform that has been used in numerous image coding applications [7], [8]. The transform matrix of dimension 2 k for k N is given by the following recursive formula [ ] H(2 k H(2 ) = k 1) H(2 k 1) H(2 k 1) H(2 k 1) and H(2) = [ ] 1 1 1 1 B. The encoder Let Z be the compressed frame or Wyner-Ziv frame from X. In our case, compression with compression ratio n : m is achieved simply projecting the row version of frame X R 1 n onto the n : th first dimensions of the orthogonal Hadamard vector basis G R n m, where m < n. Z = X G. The final stream is Z = [Z σ 2 ] where σ 2 is the variance between the current frame X and the previous frame X 1. Its given by σ 2 = N X[n] X 1 [n 1] (1) n=0 A. Problem setup III. DECODING The aim here is to reconstruct the encoded frame Z to X. At time t, the decoder has knowledge of the incoming Wyner- Ziv frame Z and previously decoded frame Y. To be able to reconstruct X, Z has somehow to contain enough information about Z, possibly together with Y ; that is the case in predictive coding schemes, where Z carries information about Y, usually motion vectors (pixel configuration for that matter) and Huffman or Arithmetic coded residuals of the motion-compensated frames [1]. Since we do not really 583 P a g e

know the real pixel setup X we will be left to guessing it out of the pixel space Λ. Mathematically, the problem could be formulated as follow: at a time t, the decoder observing Z and Y, aims to find the best pixel configuration ˆX.. The term best used here actually points out that X is unobserved. This formation leaves us to a Maximum Likelihood problem ˆX arg max Λ P ( ˆX, Z, Y ) (2) the output of our Q-learning scheme - Q-values - is two dimensional, as opposed to other schemes [11] outputting one-dimensional. An episode ends at either n iterations or p db of PSNR. Whichever comes first. For example n = 20 or p = 45. The reward is delayed, i.e. until episode ends B. Maximum Likelihood Decoding X by simply estimating doing X = Z G 1 is obviously not optimal, but it may be worth mentioning why at this point of the study. As X grows in size, pixels in it decrease in term of correlation. Consequently, the coefficients in Z won t explain X well. Estimating X become thus equivalent to solving an under-determined system of equations - fewer equations than unknowns. To find the best estimate of X we have to model an optimal Maximum Likelihood Estimator (MLE). That is, designing the Maximum Likelihood estimator to capture as much decoding information in Y and Z as possible, that is capable of estimating the following joint quantity MLE( ˆX) = P ( ˆX, Z, Y ; Θ) (3) where Θ is the generative model. Hidden Markov Mode (HMM)l [9] and Reinforcement Learning models (such as Q- Learnining) [10] are two good candidates for such problem. But modelling such MLE problem could be rather complex if applying either HMM or Q-learning due to the dimensionality of the tuples. Fortunately Q-learning has a variant, using function approximators and experience replay [11], that has shown to deal well with high dimensions. C. Q-learning Q-learning is a deep learning technique. Generally spoken, the learning model tries to learn the optimal so-called actionselection policy. We are given an agent, states S and a set of actions per state A. At a time t, the agent receives a reward r t by executing an action a t being in state s t. The goal of the agent is to maximize its total reward by learning optimal action for each state; that is the cumulative discounted longterm reward Q(a, s), starting from the current state. During learning process, the Q(a, s) value is updated as follow [ α Q t+1 (a t, s t ) Q t (a t, s t )+ (4) ] (5) r t+1 + λ arg max Q t (s t+1, a) Q(s t, a t ) a Due to high dimensionality equation (4) cannot be applied directly to our ML problem. Instead we use a variant called Q-learning with experience replay [11]. Q-learning with experience replay, the agent s state-action pair is stored in a data set and re-sampled, with respect to some significance criteria, in some later episodes in conjunction with other selected, usually randomly, data set of state-action pairs. Our Q-learning scheme is endorsing the same intuition and motivation albeit somewhat different in design; The following setup is adopted Each column of the Q-values corresponds to the probability distributions - that are assumed to be Gaussian - over the colocated pixels candidates of X; at every position X[i] in X, we consider 2σ pixel candidates X j [i], σ j σ, j Z Each pixel candidate X j [i] has an initial probability Pij 0 = 1 σ 2 2π e (j) /2σ 2 Our Q-learning design is illustrated in figure 3. D. Maximum Likelihood through Experience Replay Recall that the idea behind this whole Q-learning business is to fit a likelihood function. We aim to find out how to capture information out of the previously reconstructed frame, an encoded stream (that is syndrome and mean-squared error) so as decoding is as effective as possible. A Maximum Likelihood function does permit us to estimate the (degree of) truthfulness of a pixel combination - possibly with some measure of confidence of interval. Even though likelihood could be measured for every pixel configuration of X in Λ estimating the best configuration still remains intractable as X is of high dimension. We use Expectation-Maximization (EM) [9] to estimate the Maximum Likelihood. In our setup, the E-step correspond to the estimation of the Q-values, while the M-step choose the pixel combination that maximizes their probabilities: 1) E-step: In the E-step the derived Q-values at each iteration are used to perform a probability update of the side information as follow P t+1 ij (x i j) = q ijpij t I q ij i=0 where i is the ith side information and j is the jth distribution from the ith side information 2) M-step: The best pixel combination with respect to recent probability distribution update is selected according to ˆx i arg max P ij (x ij ) The Learning process is depicted in figure 4 584 P a g e

Fig. 3. Setup for Neural Network based Q-learning IV. EXPERIMENT RESULTS AND DISCUSSION As mentioned in section III-C, the size of the Function Approximator s output is related the variance information. Thus a side information can in theory be up to n = 256 pixel away from its counterpart. This means that the size of our Q- values is M N 256, where M is the frame height and N is the width. That is, for grayscale QCIF video sequences with M = 144 and N = 176, as used in our experiments, the size of our Q-values will be 144 176 256 = 6 488 064. That means the Q-values alone require a more than 50 Gigabytes of RAM memory! As were running the experiment on a personal computer with 8 GB (4 2GB) of RAM, we figured that only around 1024 20 Q-values could fit at a time. We performed therefore a cherry-picking procedure for the sake of assessing the effectiveness of our novel algorithm. Recall that we aim to decode a frame X given its side information Y. To fit the likelihood function, a set of QCIF video sequences were used as training samples. X and Y were divided in blocks and paired up x 1, x 2, x 3,..., x K and y 1, y 2, y 3,..., y K respectively. The pair (x i,y i ) was selected as a training sample if their variance was less than 100. That means pixels in x i is are at most 10 intensities away from pixels in y i. The length of the block were chosen to 1024 as the width of the Hadamard matrix has to be a power of 2. Recall that the encoder sends z i and the decoder only has access to y i and z i and tries to estimate x i. Thus [x i z i ] will be the input to our Function Approximator. During training phase, we used minibatches of size 1000, while adopting a constant ɛ greedy algorithm of 0.1. The input was scaled between -1 and 1 prior entering the Function Approximator. We used 5 hidden layers - with tanh activation function - with 200 nodes per layer. Figures 5 and 6 show the learning ability in terms of rate distortion of the Function Approximator though iterations/episodes. We notice the increase of the PSNR at each episode. The same cherry-picking procedure was used for testing purpose, since we were computationally limited. We tested the algorithm on QCIF video frames for the sequences Salesman 585 P a g e

Compression ratio TABLE I. Compression ratio TABLE II. Variance information σ 2 = 20 σ 2 = 50 σ 2 = 100 σ 2 = 150 σ 2 = 200 1024:3 38,60 32,60 31,88 29,82 24,73 1024:10 43,85 37,47 36,50 33,81 29,26 1024:20 62,66 54,92 45,13 40,12 34,64 TABLE SHOWING THE RATE DISTORTION PERFORMANCE FOR SALESMAN VIDEO SEQUENCE Variance information σ 2 = 20 σ 2 = 50 σ 2 = 100 σ 2 = 150 σ 2 = 200 1024:3 38,60 32,60 31,88 29,82 24,73 1024:10 43,85 37,47 36,50 33,81 29,26 1024:20 62,66 54,92 45,13 40,12 34,64 TABLE SHOWING THE RATE DISTORTION PERFORMANCE FOR HALL VIDEO SEQUENCE and Hall Monitor. Frames blocks with variance information ranging from around 10 to around 200 were selected. The rate distortion performances are given in tables I and II for the Salesman and Hall video sequences. It is important to notice that even though the Function Approximator is trained on variance information less than 100, we tested our algorithm on variance greater than 100. The reconstruction quality is still good to very good for compression ration 1024:10 and 1024:20, respectively. Compression ratio 1024:3 was also to assess the compression limit. The idea was to check if we could still achieve reasonable distortion by minimizing the number of bits to send. For each frame block, the encoder generates 3 syndrome coefficients (3 integers = 3 bits) and 1 variance information (1 double =1 bits), 25 blocks and 12 fps. A full frame has 25 blocks. Arguably, compression ration 1024:3, 1024:10 and 1024:20 could thus be comparable to 33, 100 and 200 kbps respectively. This insight shows the high potential of our scheme for low complexity, low bitrate and low distortion video coding. V. CONCLUSION We have presented a new and practical Video compression scheme based on the Wyner-Ziv framework [3]. The novelty in our scheme lies mostly in the integration of Q-learning in the decoding process. The Wyner-Ziv coding problem has been subject to a great deal of research for at least the past 15 years. The mainstream of in Wyner-Ziv Video Coding has been based on binary codes [4], [12], [13]. Our algorithm is the first really dealing with non-binary codes. A second advantage is its inherent scalability. Previous schemes, such as punctured codes [12] have used different methods for rate control. Nonbinary codes have he advantage of reducing the computation complexity at the both at the encoded and decoder, since calculations do not have to be performed at a bit level as in [14] for example. We also showed that the Wyner-Ziv problem - at least in our case - can be solved using Q-learning algorithm as Likelihood Estimator with a inherent embedding of the EM framework. However, due our computational limitation, we assessed the algorithm in a cherry-picking manner. The results shown are very good. We arguably showed that our Video Coding scheme was that of low complexity, low bitrate and low distortion Low complexity, low bitrate and low distortion is specially meaningful for lightweight devices such as surveillance cameras, mobile phones or probably Google T M Watches or Google T M Glasses in the near future when provided with cameras. Our encoding scheme is of a very low complexity compared to that of motion estimation based video encoders. The bulk of computation is shifted from the encoder to decoder. The decoder is thought to be powerful server station. This is of a great advantage, especially on lightweight devices, such as mobile phones. REFERENCES [1] T. Wiegand, G. J. Sullivan, G. Bjontegaard, and A. Luthra, Overview of the h. 264/avc video coding standard, Circuits and Systems for Video Technology, IEEE Transactions on, vol. 13, no. 7, pp. 560 576, 2003. [2] D. Slepian and J. K. Wolf, Noiseless coding of correlated information sources, Information Theory, IEEE Transactions on, vol. 19, no. 4, pp. 471 480, 1973. [Online]. Available: http: //ieeexplore.ieee.org/xpls/abs all.jsp?arnumber=1055037 [3] A. Wyner and J. Ziv, The rate-distortion function for source coding with side information at the decoder, IEEE Transactions on Information Theory, vol. 22, no. 1, pp. 1 10, 1976. [Online]. Available: http://ieeexplore.ieee.org/xpls/abs all.jsp?arnumber=1055508 [4] A. Aaron, R. Zhang, and B. Girod, Wyner-ziv coding of motion video, pp. 240 244, 2002. [5] R. Puri, A. Majumdar, and K. Ramchandran, Prism: a video coding paradigm with motion estimation at the decoder, Image Processing, IEEE Transactions on, vol. 16, no. 10, pp. 2436 2448, 2007. [6] J.-P. Kouma and H. Li, Large-Scale Face Image Retrieval: A Wyner-Ziv Coding Approach. InTech, 2011, ch. 2, p. 29. [Online]. Available: http://www.intechopen.com/ books/new-approaches-to-characterization-and-recognition-of-faces/ large-scale-face-image-retrieval-a-wyner-ziv-coding-approach [7] C. Kim, H.-H. Shih, and C.-C. J. Kuo, Fast h. 264 intra-prediction mode selection using joint spatial and transform domain features, Journal of Visual Communication and Image Representation, vol. 17, no. 2, pp. 291 310, 2006. [8] Z. Liu, L. Li, Y. Song, S. Li, S. Goto, and T. Ikenaga, Motion feature and hadamard coefficient-based fast multiple reference frame motion estimation for h. 264, IEEE Transactions on Circuits and Systems for Video Technology, vol. 18, no. 5, pp. 620 632, 2008. [9] A. P. Dempster, N. M. Laird, and D. B. Rdin, Maximum likelihood from incomplete data via the em algorithm, Journal of the Royal Statistical Society, Series B, vol. 39, pp. 1 38, 1977. [10] C. J. C. H. Watkins, Learning from delayed rewards, Ph.D. dissertation, University of Cambridge England, 1989. [11] V. Mnih, K. Kavukcuoglu, D. Silver, A. Graves, I. Antonoglou, D. Wierstra, and M. Riedmiller, Playing atari with deep reinforcement learning, arxiv preprint arxiv:1312.5602, 2013. [12] J. Sun and H. Li, Trellis-based reconstruction in wyner-ziv codec for video, in IASTED International Conference on Internet and Multimedia Systems and Applications, 2005. [13] A. Aaron and B. Girod, Wyner-ziv video coding with low-encoder complexity, in Proc. Picture Coding Symposium, 2004. [14] A. Aaron, S. Rane, E. Setton, and B. Girod, Transform-domain wynerziv codec for video, in in Proc. SPIE Visual Communications and Image Processing, 2004, pp. 520 528. 586 P a g e

Fig. 4. Learning process of our video codec 587 P a g e

Fig. 5. Average Learning ability of our Function Approximator through episodes: Compression ratio 1024:3 588 P a g e

Fig. 6. Average Learning ability of our Function Approximator through episodes: Compression ratio 1024:5 589 P a g e