Cosine Transform Priors for Enhanced Decoding of Compressed Images.

Similar documents
Image Compression Techniques

Image Compression Algorithm and JPEG Standard

Professor Laurence S. Dooley. School of Computing and Communications Milton Keynes, UK

CMPT 365 Multimedia Systems. Media Compression - Image

Review and Implementation of DWT based Scalable Video Coding with Scalable Motion Coding.

IMAGE COMPRESSION. Image Compression. Why? Reducing transportation times Reducing file size. A two way event - compression and decompression

Features. Sequential encoding. Progressive encoding. Hierarchical encoding. Lossless encoding using a different strategy

Robert Matthew Buckley. Nova Southeastern University. Dr. Laszlo. MCIS625 On Line. Module 2 Graphics File Format Essay

Compression of Stereo Images using a Huffman-Zip Scheme

IMAGE COMPRESSION USING HYBRID QUANTIZATION METHOD IN JPEG

CHAPTER 6. 6 Huffman Coding Based Image Compression Using Complex Wavelet Transform. 6.3 Wavelet Transform based compression technique 106

Multimedia Systems Image III (Image Compression, JPEG) Mahdi Amiri April 2011 Sharif University of Technology

15 Data Compression 2014/9/21. Objectives After studying this chapter, the student should be able to: 15-1 LOSSLESS COMPRESSION

Video Compression An Introduction

JPEG IMAGE CODING WITH ADAPTIVE QUANTIZATION

Redundant Data Elimination for Image Compression and Internet Transmission using MATLAB

JPEG compression of monochrome 2D-barcode images using DCT coefficient distributions

7.5 Dictionary-based Coding

An introduction to JPEG compression using MATLAB

IMAGE COMPRESSION SYSTEMS A JPEG PERSPECTIVE

Lossless Image Compression having Compression Ratio Higher than JPEG

A Miniature-Based Image Retrieval System

JPEG Compression. What is JPEG?

AUDIOVISUAL COMMUNICATION

Lecture 8 JPEG Compression (Part 3)

Compression II: Images (JPEG)

Wireless Communication

REVIEW ON IMAGE COMPRESSION TECHNIQUES AND ADVANTAGES OF IMAGE COMPRESSION

Compression Part 2 Lossy Image Compression (JPEG) Norm Zeck

Video Compression Method for On-Board Systems of Construction Robots

MULTIMEDIA COMMUNICATION

Index. 1. Motivation 2. Background 3. JPEG Compression The Discrete Cosine Transformation Quantization Coding 4. MPEG 5.

BIG DATA-DRIVEN FAST REDUCING THE VISUAL BLOCK ARTIFACTS OF DCT COMPRESSED IMAGES FOR URBAN SURVEILLANCE SYSTEMS

BLIND IMAGE QUALITY ASSESSMENT WITH LOCAL CONTRAST FEATURES

Computer Vision 2. SS 18 Dr. Benjamin Guthier Professur für Bildverarbeitung. Computer Vision 2 Dr. Benjamin Guthier

Introduction ti to JPEG

JPEG. Wikipedia: Felis_silvestris_silvestris.jpg, Michael Gäbler CC BY 3.0

Towards copy-evident JPEG images

Video Codec Design Developing Image and Video Compression Systems

Color Imaging Seminar. Yair Moshe

JPEG: An Image Compression System. Nimrod Peleg update: Nov. 2003

Digital Image Representation Image Compression

Advanced Video Coding: The new H.264 video compression standard

Multimedia Signals and Systems Still Image Compression - JPEG

A Comparative Study of DCT, DWT & Hybrid (DCT-DWT) Transform

Lecture 5: Compression I. This Week s Schedule

Forensic analysis of JPEG image compression

Motion Estimation for Video Coding Standards

Lossless Image Compression with Lossy Image Using Adaptive Prediction and Arithmetic Coding

ICA mixture models for image processing

IMAGE COMPRESSION. Chapter - 5 : (Basic)

A NEW ENTROPY ENCODING ALGORITHM FOR IMAGE COMPRESSION USING DCT

DigiPoints Volume 1. Student Workbook. Module 8 Digital Compression

Digital Image Processing

Detecting Salient Contours Using Orientation Energy Distribution. Part I: Thresholding Based on. Response Distribution

AN ANALYTICAL STUDY OF LOSSY COMPRESSION TECHINIQUES ON CONTINUOUS TONE GRAPHICAL IMAGES

Reduction of Blocking artifacts in Compressed Medical Images

DIGITAL TELEVISION 1. DIGITAL VIDEO FUNDAMENTALS

Topic 5 Image Compression

Chapter 7: Competitive learning, clustering, and self-organizing maps

ISSN (ONLINE): , VOLUME-3, ISSUE-1,

Image Compression With Haar Discrete Wavelet Transform

Video Compression MPEG-4. Market s requirements for Video compression standard

Markov Random Fields and Gibbs Sampling for Image Denoising

Enhancing DubaiSat-1 Satellite Imagery Using a Single Image Super-Resolution

Operation of machine vision system

ROI Based Image Compression in Baseline JPEG

A new robust watermarking scheme based on PDE decomposition *

Medical Image Compression using DCT and DWT Techniques

Performance Evaluation of the TINA Medical Image Segmentation Algorithm on Brainweb Simulated Images

Low Cost Motion Capture

Overview. Videos are everywhere. But can take up large amounts of resources. Exploit redundancy to reduce file size

CHAPTER 4 REVERSIBLE IMAGE WATERMARKING USING BIT PLANE CODING AND LIFTING WAVELET TRANSFORM

Final Review. Image Processing CSE 166 Lecture 18

DCT Based, Lossy Still Image Compression

CSEP 521 Applied Algorithms Spring Lossy Image Compression

Video Quality Analysis for H.264 Based on Human Visual System

A Review of Image Compression Techniques

G64PMM - Lecture 3.2. Analogue vs Digital. Analogue Media. Graphics & Still Image Representation

Image coding and compression

Overcompressing JPEG images with Evolution Algorithms

Multimedia Communications. Transform Coding

A Novel Statistical Distortion Model Based on Mixed Laplacian and Uniform Distribution of Mpeg-4 FGS

SIGNAL COMPRESSION. 9. Lossy image compression: SPIHT and S+P

AN ALGORITHM FOR BLIND RESTORATION OF BLURRED AND NOISY IMAGES

Engineering Mathematics II Lecture 16 Compression

New Approach of Estimating PSNR-B For Deblocked

Source Coding Techniques

So, what is data compression, and why do we need it?

Image Compression for Mobile Devices using Prediction and Direct Coding Approach

Image, video and audio coding concepts. Roadmap. Rationale. Stefan Alfredsson. (based on material by Johan Garcia)

Image Coding. Image Coding

Part 3: Image Processing

Image Pyramids and Applications

VC 12/13 T16 Video Compression

JPEG: An Image Compression System

A COMPRESSION TECHNIQUES IN DIGITAL IMAGE PROCESSING - REVIEW

A Comparison of Still-Image Compression Standards Using Different Image Quality Metrics and Proposed Methods for Improving Lossy Image Quality

FRACTAL IMAGE COMPRESSION OF GRAYSCALE AND RGB IMAGES USING DCT WITH QUADTREE DECOMPOSITION AND HUFFMAN CODING. Moheb R. Girgis and Mohammed M.

A Novel Approach for Deblocking JPEG Images

Transcription:

Cosine Transform Priors for Enhanced Decoding of Compressed Images. Amos Storkey and Michael Allan School of Informatics University of Edinburgh Forrest Hill, Edinburgh, EH QL Abstract. Image compression methods such as JPEG use quantisation of discrete cosine transform (DCT) coefficients of image blocks to produce lossy compression. During decoding, an inverse DCT of the quantised values is used to obtain the lossy image. These methods suffer from blocky effects from the region boundaries, and can produce poor representations of regions containing sharp edges. Such problems can be obvious artefacts in compressed images but also cause significant problems for many super-resolution algorithms. Prior information about the DCT coefficients of an image and the continuity between image blocks can be used to improve the decoding using the same compressed image information. This paper analyses empirical priors for DCT coefficients, and shows how they can be combined with block edge contiguity information to produce decoding methods which reduce the blockiness of images. We show that the use of DCT priors is generic can be useful in many other circumstances. Introduction A number of image compression methods, most notably including baseline JPEG (joint photographic experts group), use quantisation of the discrete cosine transform (DCT) coefficients in order to obtain a lossy compressed representation of an image. Put simply, baseline JPEG splits each image into 8x8 blocks and then performs a DCT on each image block. These are then quantised according to a preset quantisation schema which depends on the compression rate required. The quantised coefficients are then losslessly compressed and encoded to a bitstream, usually using Huffman codes. To decode the jpeg, the quantised coefficients are the obtained from the bitstream using the relevant lossless decompression. The quantised coefficients are then used directly in the inverse DCT to recreate the image. The deficits of this scheme are that it can produce blocky artefacts [] as each 8x8 block is treated independently, and that it can produce poor representation of regions with significant high frequency information. In this paper we recognise the fact that the quantised DCT coefficients provide upper and lower bounds for the true coefficients. It is also possible to obtain empirical prior DCT coefficient distributions from the examination of many other 8x8 patches from a database of uncompressed images. Furthermore we can examine the pixel differences across block boundaries in the uncompressed image database and use that information as a prior measure for the blockiness effects of compressed images.

w w p ( w f ) p ( w f ) D Q w D Q w R e a l In n e r B o r d e r P r e d i c t e d B o r d e r (a) (b) (c) Fig.. Prior distribution of DCT coefficients given known quantised coefficients f. (a) Zeroquantised coefficient. (b) Non-zero quantised coefficient. (c) Predicted border-pixels are given by inner border-pixels. The Probabilistic Model Consider first a single 8x8 image patch, and the set of pixels bordering each edge of that 8x8 patch. Let w denote the vector of DCT coefficients, v denote the set of quantised coefficients (and implicitly the constraints they give to the real coefficients). Let y denote the border-pixel intensities. Suppose we have some prior distribution over DCT coefficients P (w) and are given the true border-pixel intensities y, and quantised DCT coefficients. Then we wish to calculate the posterior distribution P (w y, v) P (y w)p (w v) () Here P (w v) is the prior distribution P (w) constrained to the region implied by the quantised coefficients v, and renormalised. This is illustrated in Figure. P (y w) is given by the model P (y x) of the observed border pixels given the predicted border pixels x produced by the extrapolating the DCT basis functions pixel over the boundary of the 8x8 region. This is simple to calculate using the basis function symmetry, and amounts to using the inner border pixels as the predicted border pixels (see figure ). Prior distributions To use the above model two prior distributions are needed. First a prior distribution P (w) for the DCT coefficients is required. Second we need the distribution of the true border pixels P (y x) given the predicted values. Forms for both of these can be obtained empirically from a set of training images. The source images used in this report can be found at http://www.hlab.phys.rug.nl/ archive.html. The images were taken with a Kodak DCS4 digital camera. For details and a description of the calibration see the Methods section of [7]. The images contain many similar scenes of urban areas, landscapes and woodland areas. The linear intensity images have been used rather than the de-blurred set in order to reduce the possibility of artefact introduction. The camera used to capture the images does not deliver the the two outermost pixels along the edges and so it was necessary to remove them. To maintain an image that contains a whole number of 8x8 blocks, the eight outermost pixels along the edges were cropped. Also to reduce processing time the images were reduced

in size by a factor of two along both the width and height using pixel averaging. The -bit images were intensity-scaled linearly to between and to comply with the JPEG standard. The first twenty images in the set were used as training images. The. x (Bin Size =.7799, No. Samples = 97).7 (Bin Size =.6, No. Samples = 4788).6....4.... (a) 4 4 (b). (Bin Size =.68, No. Samples = 4788).6 (No. Samples = 6).4......8.6..4. (c) Predicted Border Pixel Error (all) (d) Fig.. Distribution of DCT coefficients: (a) The DC coefficient, (b) The (,) coefficient, (c) the (4,6) coefficient. (d) Distribution of differences between the predicted and true border pixels prior over discrete cosine transform coefficients was modelled as a factorised distribution P (w) = i P (w i) where w i is the ith DCT coefficient. Then the prior for each coefficient was set using empirical values the training images. Histograms of the priors are given in Figure. Note that the lowest frequency coefficient (commonly called the DC coefficient) has a different structure from the other (AC) coefficients. The AC coefficients appear to have the same form, but have different distribution widths, where the higher frequency components are more tightly distributed around zero. The prior over Objective Function = 6449.6 4 6 7 8 9 4 6 8 4 6 Log Likelihood = 4. Original door scene Original x block scene Fig.. Two original scenes

the border pixels was also factorised into independent distributions for each pixel. The distribution of the difference between the predicted pixel value and the true pixel value was used as the model for P (y w). The distribution obtained is illustrated in Figure d. More general forms for P (y w) were tried, but they had little effect on the final outcome. Gaussian mixture models were fit to these histograms to provide a working functional representation of the distribution. Objective Function = 4.78 Objective Function = 9.887 Optimised Obj. Fn. =.67 Initial Obj. Fn. = 47. 4 4 4 6 6 6 7 7 7 8 8 8 4 6 7 8 Log Likelihood = 6.67 4 6 7 8 Log Likelihood = 6.4476 4 6 7 8 Log Likelihood =.49 (a) (b) (c) Fig. 4. Results of optimisation. (a) Original patch (b) Patch reconstructed from quantised DCT, (c) reconstructed from optimised DCT 4 Intermediate results Although the form of model described is not the final model, as the true border pixel information is not available, it is instructive to see how it performs. Conjugate gradient optimisation was used to calculate the maximum a posteriori (MAP) DCT coefficients given the border pixel data and the quantised DCT coefficients. To test how well the original values can be recovered if a nearby local minima is found, the coefficients were initialised at the original values. For the image patch illustrated in figure b, and with a standard lossy decoding given by Figure 4a, we obtain Figure 4b using this approach. A representation of the true, quantised and optimised DCT coefficients are given in Figure. We can also initialise the values at the quantised coefficients. Figure 6 shows the use of this approach on a test image. MSE DQ Opt = 7.6 MSE Orig Opt = 44.96 MSE Orig DQ = 9.78 4 6 Coefficient Limits De quantised Coefficients Optimised Coefficients Original Coefficients 7 4 6 Number Fig.. Comparison of the different DCT coefficients.

Quantifiable assessment of image quality is notoriously hard and generally an unreliable measure of performance. However it is possible to use a perceptual error measure such as those of [4, 8, ]. For example, with the image illustrated the perceptual error measure of [8] improves from 8.46 to 8.9. In general we find that for regions with fewer patches containing clear edges or linear features, most error measures (mean square error, signal to noise ratio, peak signal to noise ratio, perceptual error measure) improve, whereas there is loss in systems where edges occur. This is due to the fact that the current factorised prior does not contain edges and features as part of its model. For fuller comparisons with other techniques see []. Objective Function = 989.6 Optimised Obj. Fn. = 48.8 4 4 6 6 7 7 8 8 9 9 4 6 8 4 6 Log Likelihood = 4878. Reconstructed using quantised coefficients 4 6 8 4 6 Log Likelihood = 689.6 Reconstructed using optimised coefficients Fig. 6. Results on door scene from Full model As the true pixel data is not available for the border pixels, the model just described is not directly useful. However it can be incorporated into a Markov Random Field model which can then be optimised using an iterative checkerboard approach. The simplest model is to use a model of the form P (X) P (X i ) P (Xi k Xj k ) () i (i,j) A,k where X i are the pixel values of the ith image patch and Xi s is the kth border pixel of the ith patch, and A is the set of indices (a, b) of adjacent image patches. Note that P (X i ) and P (Xi k Xk j ) are now clique potentials and not probability distributions. This model can be optimised using an iterative checkerboard scheme. First the coefficients of the black coloured patches are fixed and i P (X i) i W,j A(i),k P (Xk i Xk j ) is optimised, for W denoting the indices of the white patches and A(i) the adjacent patches to patch i. Then the white patches are fixed to their optimised values and the corresponding equation for the black patches is optimised. This process is repeated until a suitable convergence criterion is satisfied. Due to the symmetry of the P (Xi k Xk j ), each step is guaranteed to increase the global P (X). Note each P (X i ) is implicitly given by the prior over DCT coefficients and the quantised coefficients. Again the optimisation was initialised at the known quantised values. This approach was used on the patch illustrated in figure 7. The blocky artefacts of the quantised coefficients are significantly reduced.

Reconstructed using quantised coefficients. Reconstructed using optimised coefficients. 6 Continued work Fig. 7. Results of checkerboard iteration on x block scene Blockiness artefacts are only one of the problems of compressed images. More significant issues arise from within block edges. However the process of allowing freedom to choose the coefficients from within the range of values given by the quantised coefficients can be used in conjunction with any other information about the image patches. This might include more general patch priors than those given by factorised coefficients. We are currently working on a fuller quantitative analysis of these techniques. One area of interest is that the artefacts of jpeg are often serious problems in superresolution. [] provides a good example of this. As this work places jpeg modelling within a probabilistic framework, we are working on ways to combine it directly with a number of super-resolution methods such as [6]. Acknowledgements Amos Storkey thanks Microsoft Research, Cambridge for his fellowship. References. Albert J. Ahumada and Rensheng Horng. De-blocking dct compressed images. In Human Vision, Visual Proc. and Digital Disp. V, SPIE Proc., volume 79, pages 9 6, 994.. M. Allan. Probabilistic methods for improving image decoding. Master s thesis, Division of Informatics, University of Edinburgh, Forrest Hill, Edinburgh, EH QL, UK,.. W. T. Freeman, T. R. Jones, and E. C. Pasztor. Example-based super-resolution. IEEE Computer Graphics and Applications,. 4. John Lazzaro and John Wawrzynek. Jpeg quality transcoding using neural networks trained with a perceptual error measure. In Neural Computation, volume, pages 67 96, 999.. M. Miyahara, K. Kotani, and V.R. Algazi. Objective picture quality scale (pqs) for image coding. In International Conference on Image Processing, volume, 99. 6. A. J. Storkey. Dynamic structure super-resolution. In S. Thrun S. Becker and K. Obermayer, editors, Advances in Neural Information Processing Systems (NIPS), pages 9. MIT Press,. 7. J.H. van Hateren and A. van der Schaaf. Independent component filters of natural images compared with simple cells in primary visual cortex. In R.Soc.Lon.B, pages 9 66, 998. 8. Andrew B. Watson. Dctune: A technique for visual optimisation of dct quantisation matrices for individual images. In Society for Information Display Digest of Technical Papers, volume XXIV, pages 946 949, 99.