Optimized Analog Mappings for Distributed Source-Channel Coding

Similar documents
QUANTIZER DESIGN FOR EXPLOITING COMMON INFORMATION IN LAYERED CODING. Mehdi Salehifar, Tejaswi Nanjundaswamy, and Kenneth Rose

1722 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 58, NO. 3, MARCH X/$ IEEE

Frequency Band Coding Mode Selection for Key Frames of Wyner-Ziv Video Coding

Source Coding with Distortion through Graph Coloring

Compression of a Binary Source with Side Information Using Parallelly Concatenated Convolutional Codes

Payload Length and Rate Adaptation for Throughput Optimization in Wireless LANs

MOTION ESTIMATION AT THE DECODER USING MAXIMUM LIKELIHOOD TECHNIQUES FOR DISTRIBUTED VIDEO CODING. Ivy H. Tseng and Antonio Ortega

Network Image Coding for Multicast

Delay-minimal Transmission for Energy Constrained Wireless Communications

Robust Video Coding. Heechan Park. Signal and Image Processing Group Computer Science Department University of Warwick. for CS403

Optimal Estimation for Error Concealment in Scalable Video Coding

AN ALGORITHM FOR BLIND RESTORATION OF BLURRED AND NOISY IMAGES

Analytic Performance Models for Bounded Queueing Systems

Distributed Detection in Sensor Networks: Connectivity Graph and Small World Networks

Distributed Grayscale Stereo Image Coding with Improved Disparity and Noise Estimation

Digital Image Processing Laboratory: MAP Image Restoration

Effects of Weight Approximation Methods on Performance of Digital Beamforming Using Least Mean Squares Algorithm

MULTIMODE TREE CODING OF SPEECH WITH PERCEPTUAL PRE-WEIGHTING AND POST-WEIGHTING

Image Segmentation Techniques for Object-Based Coding

Jointly Optimized Transform Domain Temporal Prediction (TDTP) and Sub-pixel Interpolation

Data Hiding in Video

Compression of Stereo Images using a Huffman-Zip Scheme

Minimum Delay Packet-sizing for Linear Multi-hop Networks with Cooperative Transmissions

Topic 6: Calculus Integration Volume of Revolution Paper 2

Measurements and Bits: Compressed Sensing meets Information Theory. Dror Baron ECE Department Rice University dsp.rice.edu/cs

Joint PHY/MAC Based Link Adaptation for Wireless LANs with Multipath Fading

Adaptive Filtering using Steepest Descent and LMS Algorithm

An Iterative Joint Codebook and Classifier Improvement Algorithm for Finite- State Vector Quantization

Estimating the Information Rate of Noisy Two-Dimensional Constrained Channels

MATH3016: OPTIMIZATION

JPEG compression of monochrome 2D-barcode images using DCT coefficient distributions

Obtaining Feature Correspondences

Coding for the Network: Scalable and Multiple description coding Marco Cagnazzo

Computing over Multiple-Access Channels with Connections to Wireless Network Coding

Error/Erasure-Resilient and Complexity-Constrained Zero-Delay Distributed Coding for Large-Scale Sensor Networks

Asynchronous Distributed Optimization With Event-Driven Communication Minyi Zhong, Student Member, IEEE, and Christos G. Cassandras, Fellow, IEEE

Rate Distortion Optimization in Video Compression

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition

Region-based Fusion Strategy for Side Information Generation in DMVC

Chapter 1. Introduction

3D Surface Recovery via Deterministic Annealing based Piecewise Linear Surface Fitting Algorithm

Error Correction and Detection using Cyclic Redundancy Check

Motion Estimation for Video Coding Standards

Quickest Search Over Multiple Sequences with Mixed Observations

A Probabilistic Approach to the Hough Transform

Adaptive Metric Nearest Neighbor Classification

Programming, numerics and optimization

ELEG Compressive Sensing and Sparse Signal Representations

Blur Space Iterative De-blurring

Complexity-Optimized Low-Density Parity-Check Codes

Memory Placement in Network Compression: Line and Grid Topologies

Image denoising in the wavelet domain using Improved Neigh-shrink

Complexity Reduced Mode Selection of H.264/AVC Intra Coding

Aerial Platform Placement Algorithm to Satisfy Connectivity and Capacity Constraints in Wireless Ad-hoc Networks

THREE DESCRIPTIONS OF SCALAR QUANTIZATION SYSTEM FOR EFFICIENT DATA TRANSMISSION

A Non-Iterative Approach to Frequency Estimation of a Complex Exponential in Noise by Interpolation of Fourier Coefficients

Joint Coding/Routing Optimization for Correlated Sources in Wireless Visual Sensor Networks

EECS 442 Computer vision. Fitting methods

ARELAY network consists of a pair of source and destination

Chapter 4: Implicit Error Detection

Theoretical Concepts of Machine Learning

Hybrid PSO-SA algorithm for training a Neural Network for Classification

Multi-View Image Coding in 3-D Space Based on 3-D Reconstruction

On the construction of Tanner graphs

Convexization in Markov Chain Monte Carlo

Distributed Video Coding

University of Notre Dame Department of Electrical Engineering. EE-87005: Advanced Topics in Multiuser Communications

Spatial Outlier Detection

CONLIN & MMA solvers. Pierre DUYSINX LTAS Automotive Engineering Academic year

JBEAM: Coding Lines and Curves via Digital Beamlets

TRACKING PERFORMANCE OF THE MMAX CONJUGATE GRADIENT ALGORITHM. Bei Xie and Tamal Bose

COURSE: NUMERICAL ANALYSIS. LESSON: Methods for Solving Non-Linear Equations

1. Introduction. 2. Motivation and Problem Definition. Volume 8 Issue 2, February Susmita Mohapatra

Compression of VQM Features for Low Bit-Rate Video Quality Monitoring

Cooperative Visual Monitoring in Energy- Constrained Wireless Sensor Networks

CHAPTER 7. PAPER 3: EFFICIENT HIERARCHICAL CLUSTERING OF LARGE DATA SETS USING P-TREES

Quantized Iterative Message Passing Decoders with Low Error Floor for LDPC Codes

Inclusion of Aleatory and Epistemic Uncertainty in Design Optimization

Convex combination of adaptive filters for a variable tap-length LMS algorithm

PERFORMANCE OF THE DISTRIBUTED KLT AND ITS APPROXIMATE IMPLEMENTATION

A Variable Length Distributed Source Coding Algorithm for WSNs

CSE 573: Artificial Intelligence Autumn 2010

Optimization Methods for Machine Learning (OMML)

Geometry-Driven Distributed Compression of the Plenoptic Function: Performance Bounds and Constructive Algorithms

Lecture 8 Fitting and Matching

Realization of Adaptive NEXT canceller for ADSL on DSP kit

Lagrangian Relaxation: An overview

Context based optimal shape coding

Automated fmri Feature Abstraction using Neural Network Clustering Techniques

Fitting. Instructor: Jason Corso (jjcorso)! web.eecs.umich.edu/~jjcorso/t/598f14!! EECS Fall 2014! Foundations of Computer Vision!

CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, :59pm, PDF to Canvas [100 points]

ADAPTIVE INTERPOLATED MOTION COMPENSATED PREDICTION. Wei-Ting Lin, Tejaswi Nanjundaswamy, Kenneth Rose

Convex Clustering with Exemplar-Based Models

Sketchable Histograms of Oriented Gradients for Object Detection

A Comparison of Still-Image Compression Standards Using Different Image Quality Metrics and Proposed Methods for Improving Lossy Image Quality

Pierre A. Humblet* Abstract

Chapter 14 Global Search Algorithms

Convex Clustering with Exemplar-Based Models

A Robust Wavelet-Based Watermarking Algorithm Using Edge Detection

Efficient Tuning of SVM Hyperparameters Using Radius/Margin Bound and Iterative Algorithms

Transcription:

21 Data Compression Conference Optimized Analog Mappings for Distributed Source-Channel Coding Emrah Akyol, Kenneth Rose Dept. of Electrical & Computer Engineering University of California, Santa Barbara, CA 9316, USA Email:{eakyol, rose}@ece.ucsb.edu Tor Ramstad Dept. of Electronics and Telecommunications Norwegian University of Science and Tech. Trondheim, NO7491, Norway Email: ramstad@iet.ntnu.no Abstract This paper focuses on optimal analog mappings for zero-delay, distributed source-channel coding. The objective is to obtain the optimal vector transformations that map between m-dimensional source spaces and k-dimensional channel spaces, subject to a prescribed power constraint and assuming the mean square error distortion measure. Closed-form necessary conditions for optimality of encoding and decoding mappings are derived. An iterative design algorithm is proposed, which updates encoder and decoder mappings by sequentially enforcing the complementary optimality conditions at each iteration. The obtained encoding functions are shown to be a continuous relative of, and in fact subsume as a special case, the Wyner-Ziv mappings encountered in digital distributed source coding systems, by mapping multiple source intervals to the same channel interval. Example mappings and performance results are presented for Gaussian sources and channels. 1 Introduction Analog mappings for joint source channel coding have been studied since Shannon s seminal work [1] where the use of space filling curves was suggested for transmission of Gaussian sources over higher dimensional Gaussian channels (i.e., with bandwidth expansion) [2, 3, 4, 5]. Although analog mappings with finite delay will not, in general, The work is supported in part by the NSF under grant CCF-728986 168-314/1 $26. 21 IEEE DOI 1.119/DCC.21.92 159

reach the asymptotic coding performance bounds, except for a few special cases (such as Gaussian source and channel of the same dimension [6]), they do possess desirable properties: low complexity, zero delay, and better robustness to varying channel SNR. In our prior work [7], we presented the optimality conditions for such analog mappings for point to point scenarios, along with an algorithm to obtain the locally optimal mappings. In this paper, we extend the work to distributed coding scenarios. We focus on two settings: The first one involves source-channel coding when side information is available to the decoder. This setting has been studied extensively both theoretically [8, 9] and in practice [1]. The second setting involves distributed source-channel coding where multiple correlated sources are encoded separately and transmitted over different channels. An important practical motivation for this setup is sensor networks where sensor measurements are correlated but are not encoded jointly due to physical constraints. This problem has also been studied extensively, especially for Gaussian sources [11]. The derivation of the optimality conditions is a direct extension of our prior work, but the distributed nature of the setting results in complex mappings that are highly nontrivial. Numerical optimization of such mappings heavily depends on initial conditions due to numerous local minima of the cost functional. Note in particular that in the case of Gaussian sources and channels, linear encoders and decoder satisfy the necessary conditions of optimality while, as we will see, careful optimization obtains better mappings that are far from linear. In Section II, we formulate the problem, present the necessary conditions for optimality and proposed an iterative design algorithm. We provide example mappings and comparative performance results in Section III. We conclude in Section IV. 2 Problem and Proposed Solution 2.1 Problem Formulation The first setup of source-channel coding with side information is shown in Figure 1, and the second setup of distributed coding is depicted in Figure 2. In both cases, there are two correlated vector sources X 1 R m 1 and X 2 R m 2. The noise variables, N, N 1, N 2 are assumed to be independent of the sources X 1, X 2. The m 1, m 2 -fold source density is denoted as f X1,X 2 (x 1, x 2 ). In the side information setup, X 2 is available to the decoder, and X 1 is transformed into Y R k by a function g : R m 1 R k and transmitted over an additive noise channel where N R k, with k-fold noise density is f N (n), is independent of X 1, X 2. The received output Ŷ = Y +N is transformed to the estimate ˆX through a function h : R k R m 2 R m 1. The m 1 -fold source density is denoted as f X (x) and k-fold noise density is f N (n). The problem is to find optimal mapping functions g, h that minimize distortion E( X 1 ˆX 1 2 ), (1) 16

Noise Source 1 X1 R m1 Encoder g:r m1 R k Y Rk Source 2 N Rk Y Rk Decoder k h : (R, Rm2 ) Rm1 Reconstruction X 1 Rm1 X2 Rm2 Figure 1: Block diagram: mapping for source-channel coding with side information Noise 1 Source 1 X1 Rm1 Source 2 X2 R m2 N1 Rk1 Encoder 1 g1 : R m1 R k1 Encoder 2 g2 : R m2 R k2 Y 1 R k1 Y 2 R k2 Y 1 Rk1 Y 2 Rk2 Noise 2 Decoder 1 Reconstruction 1 h1 : (Rk1, Rk2 ) Rm1 X 1 Rm1 Decoder 2 Reconstruction 2 h2 : (Rk1, Rk2 ) Rm2 X 2 Rm2 N2 Rk2 Figure 2: Block diagram: mapping for distributed source-channel coding subject to average power constraint Z P [g] = g(x)t g(x)fx (x)dx P. (2) R m1 In the second case, both sources are transformed through g1 : Rm1 Rk1 and g2 : Rm2 Rk2 and transmitted over the noisy channel with k1 -fold noise density, fn 1 (n1 ) and k2 -fold noise density fn 2 (n2 ). At the decoder, X 1 and X 2 are generated by h1 : Rk1 Rk2 Rm1 and h2 : Rk1 Rk2 Rm2. The problem is to find optimal mapping functions g1, g2, h1, h2 that minimize distortion E( X1 X 1 2 + X2 X 2 2 ), (3) subject to the average power constraint Z P [g1, g2 ] = (g1 (x)t g1 (x) + g2 (x)t g2 (x))fx1,x2 (x1, x2 )dx1 dx2 P. (4) Rm1,Rm2 Bandwidth compression-expansion is determined by the setting of the source and channel dimensions, k/m1 for case-1 and k1 /m1, k2 /m2 for case-2. 161

2.2 Optimal Decoding Given Encoder(s) Case-1: Assume that the encoder g is fixed. Then the optimal decoder is the minimum mean square error estimator (MMSE) of X 1 given Ŷ and X 2, i.e., h(ŷ, x 2 ) = E[X 1 ŷ, x 2 ]. (5) Plugging the expressions for expectation, applying Bayes rule and noting that fŷ X1 (ŷ x 1 ) = f N [ŷ g(x 1 )], the optimal decoder can be written, in terms of known quantities, as h(ŷ) = x1 f X1,X 2 (x 1, x 2 ) f N [ŷ g(x 1 )] dx 1 fx1,x 2 (x 1, x 2 ) f N [ŷ g(x 1 )] dx 1. (6) Case-2: Here we assume that the two encoders are fixed. Then, the optimal decoders are h 1 (ŷ 1, ŷ 2 ) = E[X 1 ŷ 1, ŷ 2 ] and h 2 (ŷ 1, ŷ 2 ) = E[X 2 ŷ 1, ŷ 2 ] and can be written as h 1 (ŷ 1, ŷ 2 ) = x1 f X1,X 2 (x 1, x 2 )f N1,N 2 [ŷ 1 g(x 1 ), ŷ 2 g(x 2 )]dx 1 dx 2 fx1,x 2 (x 1, x 2 )f N1,N 2 [ŷ 1 g(x 1 ), ŷ 2 g(x 2 )] dx 1 dx 2 (7) h 2 (ŷ 1, ŷ 2 ) = x2 f X1,X 2 (x 1, x 2 )f N1,N 2 [ŷ 1 g(x 1 ), ŷ 2 g(x 2 )]dx 1 dx 2 fx1,x 2 (x 1, x 2 )f N1,N 2 [ŷ 1 g(x 1 ), ŷ 2 g(x 2 )] dx 1 dx 2. (8) 2.3 Optimal Encoding Given Decoder(s) Case-1: Assume that the decoder h is fixed. Our goal is to minimize the distortion subject to the average power constraint. The distortion is expressed as a functional of g: D[g]= [x 1 h(g(x 1 )+n, x 2 )] T [x 1 h(g(x 1 )+n, x 2 )]f X1,X 2 (x 1, x 2 )f N (n) dx 1 dx 2 dn. We construct the Lagrangian cost functional to minimize over the mapping g, (9) J[g] = D[g] + λ {P [g] P }. (1) To obtain necessary conditions we apply the standard method in variational calculus: The Frechet derivative must vanish at an extremum point of the functional J [12]. This gives the necessary condition for optimality as J[g](x 1, x 2 ) =, x 1, x 2 (11) where J[g](x 1, x 2 )=λf X1,X 2 (x 1, x 2 )g(x 1 ) h (g(x 1 )+n, x 2 )[x h(g(x)+n, x 2 )] f N (n)f X1,X 2 (x 1, x 2 )dn. (12) 162

Case-2: Here we assume the decoders are fixed and we find the encoders g 1, g 2 that minimize the total cost J = D[g 1, g 2 ] + λ {P [g 1, g 2 ] P } where D[g 1, g 2 ]= [x 1 h 1 (g 1 (x 1 )+n 1, g 2 (x 2 )+n 2 )] T [x 1 h 1 (g 1 (x 1 )+n 1, g 2 (x 2 )+n 2 )] +[x 2 h 2 (g 1 (x 1 )+n 1, g 2 (x 2 )+n 2 )] T [x 2 h 2 (g 1 (x 1 )+n 1, g 2 (x 2 )+n 2 )] The necessary conditions are derived by requiring f X1,X 2 (x 1, x 2 )f N1,N 2 (n 1, n 2 ) dx 1 dx 2 dn 1 dn 2. (13) J[g 1 ](x 1, x 2 ) = J[g 2 ](x 1, x 2 ) = x 1, x 2, (14) where J[g 1 ](x 1, x 2 ) = J g 1 and J[g 2 ](x 1, x 2 ) = J g 2 [12]. (Explicit expressions are omitted for brevity, but can be derived similar to Case-1). Note that, unlike the decoder result, the optimal encoder is not specified in closed form. 2.4 Algorithm Sketch The basic idea is to alternate between enforcing the complementary necessary conditions for optimality. Iterations are continued until the algorithm reaches a stationary point where the total cost does not decrease anymore (in practice, until incremental improvements fall below a convergence threshold). Since the encoder(s) condition is not in closed form, we perform steepest descent search on the direction of the Frechet derivative of the Lagrangian with respect to the encoder mapping(s) g for Case-1 and g 1, g 2 for Case-2. By design, the Lagrangian cost decreases monotonically as the algorithm proceeds iteratively. The various encoders are updated as given generically in (15), where i is the iteration index and µ is the step size. g i+1 (x) = g i (x) µ J[g] (15) Note that there is no guarantee that an iterative descent algorithms of this type will converge to the globally optimal solution. The algorithm will converge to a local minimum, which may not be unique. To avoid poor local minima, one can run the algorithm several times with different initial conditions, and in the case of highly complex cost surfaces it may be necessary to apply more structured solutions such as deterministic annealing [13]. The scope of this paper and the simulations are restricted to use a more modest means by applying a noisy channel relaxation variant to avoid local minima following the proposal in [14, 15]. 3 Results In this section, we demonstrate the use of the proposed algorithm by focusing on specific scenarios. Note the algorithm is general and directly applicable to any choice 163

of dimensions as well as source and noise distributions, and also allows for correlations across noise dimensions. However. for conciseness of the results section we assume that sources are jointly Gaussian scalars with correlation coefficient ρ. and are identically distributed. We also assume the noise is scalar and Gaussian. 3.1 Initial Conditions An important observation is that the linear encoder-decoder pair satisfies the necessary conditions of optimality, although, as we will illustrate, there are other mappings that perform better. Hence, initial conditions have paramount importance in obtaining better results. In an attempt to circumvent the poor local minima problem, we embed in the solution the noisy relaxation method of [14, 15]. We initialize the encoding mapping(s) with random initial conditions and run the algorithm for a very noisy channel (high Lagrangian parameter λ). Then, we gradually increase channel SNR (CSNR) (decrease λ) while using the prior mapping as initial condition. We implemented the above algorithm by numerically calculating the derived integrals. For that purpose, we sampled the distribution on a uniform grid. We also imposed bounded support ( 3σ to 3σ) i.e., neglected tails of the infinite support distributions in the examples. 3.2 Case-1: JSCC with Side information Figure 3 presents an example of encoding mapping obtained for correlation coefficient ρ =.97. Interestingly, the analog mapping captures the central characteristic observed in digital Wyner-Ziv mappings, in the sense of many-to-one mappings, or multiple source intervals are mapped to the same channel interval, which will potentially be resolved by the decoder given the side information. Within each bin, there is a mapping function which is approximately linear in this case (scalar Gaussian sources and channel). To see the effect of correlation on the encoding mappings, we present another example in Figure 4. In this case ρ =.9 and, as intuitively expected, the side information is less reliable and source points mapped to the same channel representation grow further apart from each other. Comparative performance results are shown in Figure 5. The proposed mapping outperforms the linear mapping for the entire range of CSNR values significantly. 3.3 Case-2: Distributed JSCC Here we consider jointly Gaussian sources that are transmitted separately over independent channels. Note that our algorithm and derived necessary conditions allow the channels to be correlated, but for the sake of simplicity we assume they are independent, additive Gaussian channels. Figure 6 presents an example of encoding mappings for correlation coefficient ρ =.95. The comparative performance results are shown in Figure 7. The proposed mapping outperforms the linear mapping for the entire range of CSNR values. Note that encoders are highly asymmetric in the sense that one is Wyner-Ziv like and 164

6 Encoder mapping 2 Encoder mapping 4 15 1 2 5 2 5 1 4 15 6 3 2 1 1 2 3 (a) CSNR = 1 2 3 2 1 1 2 3 (b) CSNR = 22 Figure 3: Example encoder mappings at correlation coefficient ρ =.97, Gaussian scalar source, channel and side information 6 Encoder mapping 4 Encoder mapping 4 3 2 2 1 2 1 4 2 6 3 2 1 1 2 3 (a) CSNR = 1 3 3 2 1 1 2 3 (b) CSNR = 23 Figure 4: Example encoder mappings at correlation coefficient ρ =.9, Gaussian scalar source, channel and side information 165

22 2 Linear Encoder with Optimal Decoder Proposed mapping 18 16 SNR (db) 14 12 1 8 5 1 15 2 CSNR (db) Figure 5: Comparative results for correlation coefficient ρ =.9, Gaussian scalar source, channel and side information maps several source intervals to the same channel interval, whereas other encoder is almost a monotonic function. While we can offer tentative explanation to justify the form of this solution, we above all observe that it does outperform the symmetric linear solution. This demonstrates that by breaking from the linear solution and re-optimizing we do obtain an improved solution. Moreover, it suggests that some symmetric highly non-linear solution may exist whose discovery would need more powerful optimization tools a direction we are currently pursuing. 4 Discussion and Future Work In this paper, we derived the necessary conditions of optimality of the distributed analog mappings. Based on these necessary conditions, we derived an iterative algorithm to generate locally optimal analog mappings. Comparative results and example mappings are provided and it is shown that the proposed method provides significant gains over simple linear encoding. As expected the obtained mappings resembles Wyzer-Ziv mappings in the sense that multiple source intervals are mapped to the same channel interval. The algorithm does not guarantee a globally optimal solution. This problem can be largely mitigated by using more powerful optimization, in particular a deterministic annealing approach, which is left as future work. References [1] CE Shannon, Communication in the presence of noise, Proceedings of the IRE, vol. 37, no. 1, pp. 1 21, 1949. 166

6 Encoder 1 mapping 3 Encoder 2 mapping 4 2 2 1 1 2 2 4 3 6 3 2 1 1 2 3 (a) g 1 (x) 4 3 2 1 1 2 3 (b) g 2 (x) Figure 6: Example encoder mappings for correlation coefficient ρ =.95, Gaussian scalar sources and channels, CSNR = 1 28 26 Distributed mapping Proposed mapping Linear Encoder with Optimal Decoder 24 22 SNR (db) 2 18 16 14 12 1 1 12 14 16 18 2 22 24 26 28 3 CSNR (db) Figure 7: Comparative results for correlation coefficient ρ =.95, Gaussian scalar sources and channels 167

[2] VA Kotelnikov, The theory of optimum noise immunity, New York: McGraw- Hill, 1959. [3] S.Y. Chung, On the construction of some capacity-approaching coding schemes, Ph.D. thesis, Massachusetts Institute of Technology, 2. [4] T.A. Ramstad, Shannon mappings for robust communication, Telektronikk, vol. 98, no. 1, pp. 114 128, 22. [5] F. Hekland, GE Oien, and TA Ramstad, Using 2: 1 Shannon mapping for joint source-channel coding, in IEEE Data Compression Conference Proceedings, 25, pp. 223 232. [6] T.M. Cover and J.A. Thomas, Elements of Information Theory, J.Wiley New York, 1991. [7] E. Akyol, K. Rose, and TA Ramstad, Optimal mappings for joint source channel coding, in IEEE Information Theory Workshop, ITW 21, to appear. [8] D. Slepian and J. Wolf, Noiseless coding of correlated information sources, IEEE Transactions on Information Theory, vol. 19, no. 4, pp. 471 48, 1973. [9] A. Wyner and J. Ziv, The rate-distortion function for source coding with side information at the decoder, IEEE Transactions on Information Theory, vol. 22, no. 1, pp. 1 1, 1976. [1] SS Pradhan and K. Ramchandran, Distributed source coding using syndromes (DISCUS): design and construction, IEEE Transactions on Information Theory, vol. 49, no. 3, pp. 626 643, 23. [11] A.B. Wagner, S. Tavildar, and P. Viswanath, Rate region of the quadratic Gaussian two-encoder source-coding problem, IEEE Transactions on Information Theory, vol. 54, no. 5, 28. [12] D.G. Luenberger, Optimization by Vector Space Methods, John Wiley & Sons Inc, 1969. [13] K. Rose, Deterministic annealing for clustering, compression, classification, regression, and related optimization problems, Proceedings of the IEEE, vol. 86, no. 11, pp. 221 2239, 1998. [14] S. Gadkari and K. Rose, Robust vector quantizer design by noisy channel relaxation, IEEE Transactions on Communications, vol. 47, no. 8, pp. 1113 1116, 1999. [15] P. Knagenhjelm, A recursive design method for robust vector quantization, in Proc. Int. Conf. Signal Processing Applications and Technology, pp. 948 954. 168