Signal and Image Recovery from Random Linear Measurements in Compressive Sampling

Similar documents
Signal Reconstruction from Sparse Representations: An Introdu. Sensing

Compressive Sensing for Multimedia. Communications in Wireless Sensor Networks

COMPRESSIVE VIDEO SAMPLING

Compressive Sensing Based Image Reconstruction using Wavelet Transform

Introduction to Topics in Machine Learning

Sparse Reconstruction / Compressive Sensing

Compressed Sensing Algorithm for Real-Time Doppler Ultrasound Image Reconstruction

RECONSTRUCTION ALGORITHMS FOR COMPRESSIVE VIDEO SENSING USING BASIS PURSUIT

Image Reconstruction from Multiple Sparse Representations

Compressed Sensing and Applications by using Dictionaries in Image Processing

TERM PAPER ON The Compressive Sensing Based on Biorthogonal Wavelet Basis

Detection Performance of Radar Compressive Sensing in Noisy Environments

Structurally Random Matrices

Collaborative Sparsity and Compressive MRI

ELEG Compressive Sensing and Sparse Signal Representations

Open Access Reconstruction Technique Based on the Theory of Compressed Sensing Satellite Images

Randomized sampling strategies

Compressive Sensing based image processing in TrapView pest monitoring system

SPARSITY ADAPTIVE MATCHING PURSUIT ALGORITHM FOR PRACTICAL COMPRESSED SENSING

Compressive. Graphical Models. Volkan Cevher. Rice University ELEC 633 / STAT 631 Class

KSVD - Gradient Descent Method For Compressive Sensing Optimization

Measurements and Bits: Compressed Sensing meets Information Theory. Dror Baron ECE Department Rice University dsp.rice.edu/cs

Modified Iterative Method for Recovery of Sparse Multiple Measurement Problems

IMA Preprint Series # 2211

Image Restoration and Background Separation Using Sparse Representation Framework

A Relationship between the Robust Statistics Theory and Sparse Compressive Sensed Signals Reconstruction

The Fundamentals of Compressive Sensing

A* Orthogonal Matching Pursuit: Best-First Search for Compressed Sensing Signal Recovery

Face Recognition via Sparse Representation

arxiv: v1 [stat.ml] 14 Jan 2017

A parallel patch based algorithm for CT image denoising on the Cell Broadband Engine

Lecture 17 Sparse Convex Optimization

BSIK-SVD: A DICTIONARY-LEARNING ALGORITHM FOR BLOCK-SPARSE REPRESENTATIONS. Yongqin Zhang, Jiaying Liu, Mading Li, Zongming Guo

Compressive Cooperative Sensing and Mapping in Mobile Networks

Adaptive step forward-backward matching pursuit algorithm

Reconstruction Improvements on Compressive Sensing

Sparse Watermark Embedding and Recovery using Compressed Sensing Framework for Audio Signals

The Viterbi Algorithm for Subset Selection

An Iteratively Reweighted Least Square Implementation for Face Recognition

Weighted-CS for reconstruction of highly under-sampled dynamic MRI sequences

Department of Electronics and Communication KMP College of Engineering, Perumbavoor, Kerala, India 1 2

SUMMARY. In combination with compressive sensing, a successful reconstruction

Outline Introduction Problem Formulation Proposed Solution Applications Conclusion. Compressed Sensing. David L Donoho Presented by: Nitesh Shroff

Compressive Sensing-based Target Tracking for Wireless Visual Sensor Networks

Sparse Signal Reconstruction using Weight Point Algorithm

COMPRESSED SENSING: A NOVEL POLYNOMIAL COMPLEXITY SOLUTION TO NASH EQUILIBRIA IN DYNAMICAL GAMES. Jing Huang, Liming Wang and Dan Schonfeld

The Benefit of Tree Sparsity in Accelerated MRI

Using. Adaptive. Fourth. Department of Graduate Tohoku University Sendai, Japan jp. the. is adopting. was proposed in. and

Main Menu. Summary. sampled) f has a sparse representation transform domain S with. in certain. f S x, the relation becomes

EE123 Digital Signal Processing

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference

G Practical Magnetic Resonance Imaging II Sackler Institute of Biomedical Sciences New York University School of Medicine. Compressed Sensing

Compressive Topology Identification of Interconnected Dynamic Systems via Clustered Orthogonal Matching Pursuit

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

A Study on Compressive Sensing and Reconstruction Approach

Blind Compressed Sensing Using Sparsifying Transforms

Image reconstruction based on back propagation learning in Compressed Sensing theory

Sparse Component Analysis (SCA) in Random-valued and Salt and Pepper Noise Removal

MAKING HEALTH MANAGEMENT MORE CONCISE AND EFFICIENT: AIRCRAFT CONDITION MONITORING BASED ON COMPRESSED SENSING

ADAPTIVE LOW RANK AND SPARSE DECOMPOSITION OF VIDEO USING COMPRESSIVE SENSING

COMPRESSED SENSING WITH UNKNOWN SENSOR PERMUTATION. Valentin Emiya, Antoine Bonnefoy, Laurent Daudet, Rémi Gribonval

Image Reconstruction based on Block-based Compressive Sensing

Algorithms for sparse X-ray CT image reconstruction of objects with known contour

Hydraulic pump fault diagnosis with compressed signals based on stagewise orthogonal matching pursuit

SPARSE COMPONENT ANALYSIS FOR BLIND SOURCE SEPARATION WITH LESS SENSORS THAN SOURCES. Yuanqing Li, Andrzej Cichocki and Shun-ichi Amari

Image Deblurring Using Adaptive Sparse Domain Selection and Adaptive Regularization

2D and 3D Far-Field Radiation Patterns Reconstruction Based on Compressive Sensing

Sparse Solutions to Linear Inverse Problems. Yuzhe Jin

Image Denoising Using Sparse Representations

Progress In Electromagnetics Research, Vol. 125, , 2012

Compressive Sensing: Theory and Practice

Image Reconstruction Algorithm of Compressed Sensing based on Neural Dynamics

VARIABLE DENSITY COMPRESSED IMAGE SAMPLING

Lecture 19. Lecturer: Aleksander Mądry Scribes: Chidambaram Annamalai and Carsten Moldenhauer

An adaptive random compressive partial sampling method with TV recovery

SPARSE SIGNAL RECONSTRUCTION FROM NOISY COMPRESSIVE MEASUREMENTS USING CROSS VALIDATION. Petros Boufounos, Marco F. Duarte, Richard G.

Compressive Sensing. A New Framework for Sparse Signal Acquisition and Processing. Richard Baraniuk. Rice University

Compressed Sensing for Electron Tomography

Robust Image Watermarking Based on Compressed Sensing Techniques

Image Denoising Based on Hybrid Fourier and Neighborhood Wavelet Coefficients Jun Cheng, Songli Lei

Hyperspectral Data Classification via Sparse Representation in Homotopy

Combinatorial Selection and Least Absolute Shrinkage via The CLASH Operator

Compressive Parameter Estimation with Earth Mover s Distance via K-means Clustering. Dian Mo and Marco F. Duarte

P257 Transform-domain Sparsity Regularization in Reconstruction of Channelized Facies

Compressive Mobile Sensing for Robotic Mapping

An Optimized Pixel-Wise Weighting Approach For Patch-Based Image Denoising

Introduction. Wavelets, Curvelets [4], Surfacelets [5].

Patch-Based Color Image Denoising using efficient Pixel-Wise Weighting Techniques

Constrained adaptive sensing

SEQUENTIAL IMAGE COMPLETION FOR HIGH-SPEED LARGE-PIXEL NUMBER SENSING

Compressive Rendering of Multidimensional Scenes

Bilevel Sparse Coding

Bipartite Graph based Construction of Compressed Sensing Matrices

COMPRESSIVE POINT CLOUD SUPER RESOLUTION

A Compressed Sensing Approach for Biological Microscopy Image Denoising

Optimal Sampling Geometries for TV-Norm Reconstruction of fmri Data

IMAGE COMPRESSION USING HYBRID QUANTIZATION METHOD IN JPEG

University of Alberta. Parallel Sampling and Reconstruction with Permutation in Multidimensional Compressed Sensing. Hao Fang

INCOHERENT DICTIONARY LEARNING FOR SPARSE REPRESENTATION BASED IMAGE DENOISING

Image Inpainting Using Sparsity of the Transform Domain

Transcription:

Signal and Image Recovery from Random Linear Measurements in Compressive Sampling Sarah Mazari and Kamel Belloulata Abstract In many applications, including audio or digital imaging, the Nyquist rate is so high that we get too many samples, and designing systems that allow the acquisition of sample at required rate is costly and sometimes physically impossible. This abundance of samples has led to the development of compression techniques that aim to find the most concise representation of a signal or image making them a necessary tool for the storage or transmission of data. Therefore, to process and transmit or store digital information efficiently, we must acquire it respecting the Nyquist rate and then compress it using suitable algorithm. This process leads to a double wastage: we acquired the entire signal and we process the entire signal to retain only a few coefficients. Recent work on Compressive Sampling can overcome these disadvantages by changing the course of data acquisition. We will define in this paper, the theory of Compressive sampling, and we will show that it is possible using sparse reconstruction algorithms to recover or reconstruct a sparse signal or image with surprisingly fewer samples than those predicted by the Shannon /Nyquist theorem. Keywords Compressive sampling, Incoherence, l minimization, orthogonal matching pursuit, Sparsity. C I. INTRODUCTION OMPRESSIVE sampling proposes a new framework on how to efficiently sample signals with a reduced number of sensors. This new concept retains only the significant part of the signal and rejects from the outset all that is superfluous, therefore, it does not compress, it is reconstruct a global signal by closing the missing information. This new theory exploits the fact that many signals have a sparse representation when expressed in the appropriate base. So a set of non-adaptive linear measurements of a sparse signal contains enough measurements to reconstruction. The success of that reconstruction is guaranteed with high probability if the signal or the image is sparse in an appropriate basis and if the sampling basis and sparsifying basis are incoherent. II. SPARSE AND COMPRESSIBLE SIGNALS In signal processing, we are interested in signals generated by physical systems, they can be modeled by linear systems, hence the need to represent the signals as vectors. A signal is sparse when most of its coefficients are approximately zero. A signal x of size N is k-sparse if x contains k nonzero components with k <<N. The sparse representations consist of the decomposition of the signal on a dictionary containing a number of elements (atoms) higher than the signal dimension. This decomposition will introduce in the new representation of the signal a lot of null values []. The signal x has a sparse representation if, in the base Ψ each vector x can be approximated by: N x = i= s i Ψ i () s i is column vector, s i = x, Ψ i, =Ψ T i x, T denote the transpose. The signals that we encounter in practice are not always exactly sparse, for example, we meet the compressible signals, which are approximately sparse signals, and their coefficients are weakened following a power law decreasing : x i R i ( q ) (2) Where x is a non-increasing rearrangement of x, R is some positive constant, and < q <. Note that in particular, sparse signals are compressible. III. INCOHERENCE Suppose we have a pair of bases (Ф, Ψ), the first base is used to sample the signal x and the second to represent x sparsely. The coherence between the two bases is defined as [2]: µ(φ, Ψ) = N max Ф k T Ψ j (3) The coherence between the bases measures the highest correlation between two elements of the two bases. Sarah Mazari. is Djillali Liabes University of Sidi Bel Abbes BP 89, 22, Algeria, Phone: 23556543; e-mail: mazari_sarah@yahoo.fr. Kamel Belloulata. is Djillali Liabes University of Sidi Bel Abbes BP 89,22, Algeria email: k_belloula@yahoo.fr. Where Ф T k and Ψ j are the kth row and jth column of Φ and Ψ, respectivel. CS is mainly interested in the basis of low coherence. For example, the canonical basis Φ k (t) = δ (t k) and the Fourier basis Ψ j = n 2 e i2πjt/n are perfectly incoherent. 58

We also get a low coherence by taking Ψ as any fixed orthonormal basis and Φ a random matrix with Gaussian entries uniformly distributed. IV. COMPRESSIVE SAMPLING Consider a signal x R N that cannot be sparse in its original base but sparse or approximately sparse in appropriate base Ψ. x= Ψs (4) The sampling of the signal x of length N is performed by measuring M projections of x on testing vectorsy i = x, Ф i, i=,2 M. y = Фx = ФΨs (5) The measurements y(m) are stacked in a vector of T dimension M and the test functions Φ M 3Tas a line in M N matrix Ф. Given a sparse signal in a large space, our goal is to reconstruct the signal efficiently from a number of linear measurements smaller than the dimension of the current signal. It is clear that it is theoretically possible; however, the difficulty lies in the construction of algorithms that performs this task efficiently. There are two major approaches for the sparse reconstruction. The first is based on solving a linear optimization problem minimizing a penalty under constraints of sparsity.we focus in our case on the convex penalties (typically l ) hoping that the choice among all the solutions of underdetermined linear system, the standard l minimum gives a solution with the least possible nonzero components. These methods provide good security and stability but they use linear programming. The second approach uses the greedy methods; these methods calculate the support of the signal iteratively. They are generally faster than optimization methods but so far, they do not perform as well in terms of stability. A. Linear Optimization: Find the sparse solution to underdetermined system of linear equations Θs = y can be summarized as follows: [4], [5] s = argmin s subject to Θ s = y (7) Fig. Measurement process (k<m<n). [3] Figure describes the measurement process with a random measurement matrix Ф, the vector s is k-sparse with k = 4. The measurement vector y is a linear combination of k columns of Ф. This new data acquisition mechanism marks a fundamental departure from the conventional data acquisition compression-transmission-decompression framework: the conventional framework collects a vast amount of data for acquiring a high-resolution signal or image, then essentially discard most of data collected in the compression stage, however, in CS, data are measured in a compressed manner, and the much reduced amount of measurements are transmitted or stored economically, and each measurement is then utilized to recover the signal. V. SIGNAL AND IMAGE RECONSTRUCTION The specificity of CS lies in the fact that a sparse signal or image x can be reconstructed from a signal y which contains considerably less information. At first sight, it seems impossible to reconstruct the signal x from the measurements y as the system (4) is underdetermined and has infinitely many solutions, but if the vector x is sparse then the situation will radically change. The problem is to recover the coefficients by solving, Θs=y, with Θ=ΦΨ (6) At first glance, this task looks difficult since (6) is an NPhard combinatorial problem. To overcome this difficulty, a whole literature has emerged suggesting approximations or convex relaxations of (6). However, it was proposed to replace the problem l (6) by its convex relaxation l : s = argmin s subject to Θ s = y (8) The classical approach for solving inverse problems of this type is the mean square (minimization of l2 norm) which measures the signal energy: s = argmin s subject to Θ s = y (9) This optimization admits a simple analytical solution s = Θ T (ΘΘ T ) y, unfortunately, the minimization l 2 returns a non-sparse solution. s = argmin s 2 subject to Θ s = y () Practically, the Basis Pursuit algorithm obtains the decomposition that minimizes l norm of coefficients found in the sparse representation. To ensure an accurate reconstruction of each signal x k- sparse, the measurement matrix Θ must satisfy the restricted isometry condition (RIC), we define the restricted isometry constant as the smallest amount that satisfies: [6],[7] ( δ k ) s 2 2 Θs 2 2 2 ( + δ k ) s 2 The RIC of order k with a value of δ close to means that Θ approximately preserves the norm of any vector k- 59

sparse. An equivalent description of the restricted isometry condition is that all subsets of k columns taken from Θ are almost orthogonal. B. Greedy Methods The greedy algorithms are based on the iterative approximation of the coefficients of the signal until a convergence criterion is met, or alternatively by obtaining an improved evaluation of the sparse signal at each iteration. The first greedy algorithm that was developed for sparse reconstruction is the Matching pursuit [8] This algorithm was then modified by adding steps of directional optimization and orthogonalization. We focus in our paper on Orthogonal matching pursuit algorithm which can be explained as follows: Since x has only k nonzero components, the vector y = Фx is a linear combination of k columns of Ф. So OMP algorithm is used in this case to identify the correct signal x and therefore we need to determine which column of Ф participates in obtaining the measurement vector y. Algorithm [9] Input Measurement matrix Ф M N. The measurement vector y of dimension M. The level of sparsity k. Procedure: Initialize the residual r =y, the set of index, Г = and the approximation x =. Find the column of Ф most correlated with the residual: α i = Ф T i r n i max = arg maxi α i Г n = Г n U i max Solve the problem of least squares to obtain the new estimated signal..5 5 5 Fig. 2 Sparse signal We use a random measurement matrix of dimension 6 5..5 2 3 4 5 6 Fig.3 Measurement vector y. OMP algorithm is applied to reconstruct the sparse signal from the measurements vector y. x = arg min y Ф Г n x Г n 2 Calculate the residual r n = y Фx Output: An estimated x in R N A residual of dimension M...5 5 5 Fig. 4 Reconstructed signal using OMP. The algorithm stops after k iterations Г n : Set containing the indices of the selected elements (Indices of the columns of the matrix Ф which were used.) Ф Г n : Sub- matrix Ф which contains only the columns whose indices are in Г n. Here is an example of the D sparse signal reconstruction using the OMP algorithm; we generate a k- sparse signal of dimension N. Then the signal is projected on an M N measurement matrix, we obtains the measurement vector y of dimension M. The reconstructed signal (red) is identical to the original signal (blue) VI. EXPERIMENTAL RESULTS We will present in this section the results we obtained using the two reconstructions methods previously described in section V. The experimental environment is an Acer computer with Intel(R) Core i5-48m CPU (2.66GHz) and 3G memory, running Windows seven and Matlab R2a. Suppose we have an image X of size N N and we vectorize it into a column vector x of size N by concatenating the individual columns of X in order. 6

When designing a CS reconstruction, the identification of the sparsifying basis is the key for the success of the reconstruction. Here, we use the wavelet decomposition as sparsifying basis Ψ, and random basis as sampling scheme Φ. For the BP algorithm implementation, we use a part of the l- Magic package proposed by [].Then we use the Orthogonal Matching Pursuit greedy algorithm described in previous section to compare between the two methods in term of reconstruction quality. We vary the percentage of samples (number of measurements) taken to reconstruct the image, compare the computation time of each algorithm (BP and OMP) and calculate the PSNR for each reconstruction test. The PSNR computes the peak signal-to-noise ratio, in decibels, between two images. This ratio is often used as a quality measurement between the original and a compressed image. The higher the PSNR, the better the quality of the compressed or reconstructed image. [] Fig.3 Reconstructed image using Basis Pursuit from 4% randomly selected wavelet coefficient. PSNR(x, x )= log 255 2 MSE (x,x) () MSE(x, x ) = M (x MN i= N j= ij x ij ) 2 (2) TABLE I COMPARISON BETWEEN TWO APPROACH BP OMP Percentage of 2% 4% 6% 2% 4% 6% samples Computation 8.93 3.78 5.93.4 2.3 3.8 time (s) PSNR (db) 25.63 3.47 35.76 24.3 28.2 32.6 Table summarizes the different results obtained and compares the speed of reconstruction for each algorithm and the quality of CS reconstruction in term of PSNR. The first observation we can make is that OMP is much faster than BP. For the two reconstructions BP and OMP, we note that, as the number of random measurements (samples) taken to perform reconstruction increases, the PSNR is better, and the computation time increases. The results we obtained coincides with what has been described theoretically in Section V, in fact, BP reconstruction gives better results but is slower than OMP reconstruction. We can say that OMP provides a good compromise between quality and speed of reconstruction. Here is an example of image reconstruction using both algorithms described in the previous section: Fig.5 Original 256 256 cameraman image. Fig. 4. Reconstructed image using Orthogonal Matching Pursuit from 4% randomly selected wavelet coefficient VII. CONCLUSION Compressive sampling is used in all applications where the task is to reconstruct a signal or an image from linear measurements. Its use is advantageous whenever the acquisition of many measures would be expensive, time consuming, difficult, dangerous or impossible. Many sparse reconstruction algorithms have been developed and continue to be developed to date. In this paper, we focused on basis pursuit and orthogonal matching pursuit algorithms and we have shown that they are efficient algorithms for sparse image reconstruction from random measurements, we compared these two algorithms and we concluded that the OMP provides a good compromise between quality and speed of reconstruction Multiple perspectives can be considered, for example, reconstruct the image from deterministic measures instead of random measures,optimize existing reconstruction algorithms and apply it to more specific area of imaging (radar, astronomy and medical imaging). Relatively speaking, the emulation created by Compressive Sampling is comparable to that generated by wavelets there are two decades ago!! ACKNOWLEDGEMENTS This work was partially supported by the Franco-Algerian cooperation program PHC TASSILI grants. Entitled "Système Conjoint de Compression et d Indexation Basé-Objet pour la 6

Vidéo (SCCIBOV)", the project gathers members of Institut National des Sciences Appliquées (INSA) de Rennes, Université de Technologie de Compiègne (UTC), the University of Sidi Bel Abbès and the University of Mascara. REFERENCES [] M. Aharon, M. Elad, and A. Bruckstein. An algorithm for designing overcompletes dictionaries for sparse representation, IEEE Transactions on Signal Processing, 54():43 4322, November 26. [2] E. J. Candés, J. Romberg, Sparsity and Incoherence in compressive sampling, inverse Prob, vol. 23, 27. [3] R. G. Baranuik, Compressive sensing, lecture note, IEEE signal processing magazine, 27 [4] E. J. Candés, T. Tao, Decoding by linear programming, IEEE Trans. Inform. Theory, 5:423 425, 25 [5] E.J. Candés, Compressive sampling Proceedings of the International Congress of Mathematicians, Madrid, Spain, 26. [6] M. A. Davenport, M. F. Durate, Y. C. Eldar, G. Kutyniok, Introduction to compressive sampling Chapter in Compressed Sensing: Theory and Applications, Cambridge University Press, 2.. [7] E. J. Candès, The restricted isometry property and it implications for compressed sensing," Compte Rendus de l'academie des Sciences, vol. 346, pp. 589-592, 28 [8] J.A. Tropp, S.J. Wright, Computational methods for sparse solution of linear inverse problems Technical Report No. 29- March 29 [9] J. A. Tropp, A. C. Gilbert, Signal recovery from partial information via orthogonal matching pursuit, vol. 53, num. 2, pp. 4655-4666, Dec. 27 [] E. Candés. J.Romberg, l-magic, Recovery of sparse signal via convexe programming, Caltech, 25. [] A. Horé,D.Ziou Image quality metrics:psnr vs. SSIM International Conference on Pattern Recognition,2. 62