Image Compression with Competitive Networks and Pre-fixed Prototypes*

Similar documents
Compression of Image Using VHDL Simulation

CHAPTER 6 COUNTER PROPAGATION NEURAL NETWORK FOR IMAGE RESTORATION

AN EFFICIENT CODEBOOK INITIALIZATION APPROACH FOR LBG ALGORITHM

A Review on LBG Algorithm for Image Compression

Optimization of Bit Rate in Medical Image Compression

CHAPTER 6 INFORMATION HIDING USING VECTOR QUANTIZATION

A LOSSLESS INDEX CODING ALGORITHM AND VLSI DESIGN FOR VECTOR QUANTIZATION

Image Compression Using SOFM

A combined fractal and wavelet image compression approach

Feature-Guided K-Means Algorithm for Optimal Image Vector Quantizer Design

Comparative Study on VQ with Simple GA and Ordain GA

THREE DESCRIPTIONS OF SCALAR QUANTIZATION SYSTEM FOR EFFICIENT DATA TRANSMISSION

Available online at ScienceDirect. Procedia Computer Science 89 (2016 )

An Adaptive and Deterministic Method for Initializing the Lloyd-Max Algorithm

CHAPTER 6. 6 Huffman Coding Based Image Compression Using Complex Wavelet Transform. 6.3 Wavelet Transform based compression technique 106

SSIM based image quality assessment for vector quantization based lossy image compression using LZW coding

Multilevel Compression Scheme using Vector Quantization for Image Compression

CSEP 521 Applied Algorithms Spring Lossy Image Compression

Image Compression Using BPD with De Based Multi- Level Thresholding

Differential Compression and Optimal Caching Methods for Content-Based Image Search Systems

KeyWords: Image Compression, LBG, ENN, BPNN, FBP.

MRT based Adaptive Transform Coder with Classified Vector Quantization (MATC-CVQ)

K-Means Clustering. Sargur Srihari

Statistical Image Compression using Fast Fourier Coefficients

A new predictive image compression scheme using histogram analysis and pattern matching

Novel Lossy Compression Algorithms with Stacked Autoencoders

Compression of 3-Dimensional Medical Image Data Using Part 2 of JPEG 2000

4. Cluster Analysis. Francesc J. Ferri. Dept. d Informàtica. Universitat de València. Febrer F.J. Ferri (Univ. València) AIRF 2/ / 1

Searching similar images - Vector Quantization with S-tree

Chapter 7: Competitive learning, clustering, and self-organizing maps

A Very Low Bit Rate Image Compressor Using Transformed Classified Vector Quantization

AUDIO COMPRESSION USING WAVELET TRANSFORM

Module 8: Video Coding Basics Lecture 42: Sub-band coding, Second generation coding, 3D coding. The Lecture Contains: Performance Measures

International Journal of Modern Trends in Engineering and Research e-issn No.: , Date: 2-4 July, 2015

Figure (5) Kohonen Self-Organized Map

Highly Secure Invertible Data Embedding Scheme Using Histogram Shifting Method

Artificial Neural Networks Unsupervised learning: SOM

2. CNeT Architecture and Learning 2.1. Architecture The Competitive Neural Tree has a structured architecture. A hierarchy of identical nodes form an

Analysis of Image Compression Approaches Using Wavelet Transform and Kohonen's Network

Modified SPIHT Image Coder For Wireless Communication

Medical Image Compression using DCT and DWT Techniques

RGB Digital Image Forgery Detection Using Singular Value Decomposition and One Dimensional Cellular Automata

k-means Clustering David S. Rosenberg April 24, 2018 New York University

ANALYSIS OF SPIHT ALGORITHM FOR SATELLITE IMAGE COMPRESSION

Genetic algorithm for optimal imperceptibility in image communication through noisy channel

Segmentation using Codebook Index Statistics for Vector Quantized Images

1 Case study of SVM (Rob)

A Review on Digital Image Compression Techniques

Fingerprint Image Compression

A Color Image Digital Watermarking Scheme Based on SOFM

Hybrid image coding based on partial fractal mapping

Today. Lecture 4: Last time. The EM algorithm. We examine clustering in a little more detail; we went over it a somewhat quickly last time

Two Algorithms for Approximation in Highly Complicated Planar Domains

Hierarchical Ordering for Approximate Similarity Ranking

Fractal Compression. Related Topic Report. Henry Xiao. Queen s University. Kingston, Ontario, Canada. April 2004

FRACTAL IMAGE COMPRESSION OF GRAYSCALE AND RGB IMAGES USING DCT WITH QUADTREE DECOMPOSITION AND HUFFMAN CODING. Moheb R. Girgis and Mohammed M.

So, what is data compression, and why do we need it?

A Study on the Effect of Codebook and CodeVector Size on Image Retrieval Using Vector Quantization

Image compression. Stefano Ferrari. Università degli Studi di Milano Methods for Image Processing. academic year

Dictionary Based Compression for Images

Topic 5 Image Compression

MRT based Fixed Block size Transform Coding

Wavelet Based Image Compression Using ROI SPIHT Coding

5.7. Fractal compression Overview

Comparative Analysis of Image Compression Algorithms for Deep Space Communications Systems

SVM-based Filter Using Evidence Theory and Neural Network for Image Denosing

On the Selection of Image Compression Algorithms

IMAGE CLASSIFICATION USING COMPETITIVE NEURAL NETWORKS

Image Compression: An Artificial Neural Network Approach

SATELLITE IMAGE COMPRESSION TECHNIQUE BASED ON THE EVIDENCE THEORY

Codebook generation for Image Compression with Simple and Ordain GA

Function approximation using RBF network. 10 basis functions and 25 data points.

Two-Dimensional Fitting of Brightness Profiles in Galaxy Images with a Hybrid Algorithm

Image Compression Using K-Space Transformation Technique

A Study of Image Compression Based Transmission Algorithm Using SPIHT for Low Bit Rate Application

Video Compression An Introduction

Region Segmentation for Facial Image Compression

A Joint Histogram - 2D Correlation Measure for Incomplete Image Similarity

On the Selection of Image Compression Algorithms

Lecture 5: Compression I. This Week s Schedule

A Reversible Data Hiding Scheme for BTC- Compressed Images

DCT SVD Based Hybrid Transform Coding for Image Compression

A LOW-COMPLEXITY AND LOSSLESS REFERENCE FRAME ENCODER ALGORITHM FOR VIDEO CODING

Unsupervised Learning

Chapter 4. The Classification of Species and Colors of Finished Wooden Parts Using RBFNs

PRINCIPAL COMPONENT ANALYSIS IMAGE DENOISING USING LOCAL PIXEL GROUPING

Image Compression Algorithm for Different Wavelet Codes

3. MODIFIED BLOCK TRUNCATION CODING 3.1 MODIFIED BTC METHODS FOR IMPROVED CONTRAST

MEMORY EFFICIENT WDR (WAVELET DIFFERENCE REDUCTION) using INVERSE OF ECHELON FORM by EQUATION SOLVING

CHAPTER 6 MODIFIED FUZZY TECHNIQUES BASED IMAGE SEGMENTATION

Unsupervised Learning

TERM PAPER ON The Compressive Sensing Based on Biorthogonal Wavelet Basis

Adaptive Wavelet Image Denoising Based on the Entropy of Homogenus Regions

A novel approach for vector quantization using a neural network, mean shift, and principal component analysis-based seed re-initialization

CHAPTER 3 TUMOR DETECTION BASED ON NEURO-FUZZY TECHNIQUE

VLSI Implementation of Daubechies Wavelet Filter for Image Compression

WAVELET USE FOR IMAGE RESTORATION

A Neural Network Based Technique for Data Compression

A Image Comparative Study using DCT, Fast Fourier, Wavelet Transforms and Huffman Algorithm

A Comparative Study of Conventional and Neural Network Classification of Multispectral Data

Transcription:

Image Compression with Competitive Networks and Pre-fixed Prototypes* Enrique Merida-Casermeiro^, Domingo Lopez-Rodriguez^, and Juan M. Ortiz-de-Lazcano-Lobato^ ^ Department of Applied Mathematics, University of Malaga, Malaga, Spain; {merida,dlopez}@ctima.uma.es Department of Computer Science and Artificial Intelligence, University of Malaga, Malaga, Spain; jmortiz@lcc.unia.es Abstract. Image compression techniques have required much attention from the neural networks community for the last years. In this work we intend to develop a new algorithm to perform image compression based on adding some pre-fixed prototypes to those obtained by a competitive neural network. Prototypes are selected to get a better representation of the compressed image, improving the computational time needed to encode the image and decreasing the code-book storage necessities of the standard approach. This new method has been tested with some wellknown images and results proved that our proposal outperforms classical methods in terms of maximizing peak-signal-to-noise-ratio values. 1 Introduction The storage or transmission of images are tasks demanding either large capacity and/or bandwidth. Thus, image compression is key in the development of various multimedia computer services and telecommunication applications, such as medical imaging [?], satellite transmission, teleconferencing, pattern recognition [?], etc. The goal of image coding is to reduce both the distortion introduced in the coding process and the bit rate (or, equivalently, the compression rate) to an acceptable level. Recent publications show a substantial increase in the use of neural networks for image compression and coding. For a review in neural techniques for image compression, refer to [?]. Vector Quantization (VQ) is a lossy compression technique that can achieve high compression rates with good visual fidehty, see for example [?]. VQ is a coding method designed to represent a multidimensional space by means of a finite number of vectors, called representatives, prototypes or code-vectors. A vector quantizer statistically encodes data vectors in order to quantize and compress the data, by mapping each input vector in the p-dimensional Euclidean space R^ into one of the K prototypes. This work has been partially supported by Junta de Andalucia project number P06-TIC-01615. Please use the following format when citing this chapter: Merida-Casermeiro, E., Lopez-Rodriguez, D., Ortiz-de-Lazcano-Lobato, J. M., 2007, in IFIP International Federation for Information Processing, Volume 247, Artificial Intelligence and Innovations 2007: From Theory to Applications, eds. Boukis, C, Pnevmatikakis, L., Polymenakos, L., (Boston: Springer), pp. 339-346.

340 Enrique Merida-Casermeiro et al. According to Shannon's rate distortion theory, VQ can always achieve better compression performance than any conventional coding technique based on the encoding of scalar quantities [?j. Linde, Buzo and Gray [?] proposed the well-known LBG algorithm for VQ which made no use of differentiation, and it is the standard approach to compute the codebook (the set of prototypes). Competitive networks are designed to cluster the input data. Thus, by using VQ techniques in this type of networks, tasks such as data coding and compression can be performed. The basic structure of a network of this type is as follows: given an input vector from a j9-dimensional space, K neurons compute the VQ code-book in which each neuron relates to one code-vector via its coupling weights. The weight w^ = {wi^i^...,wi^p) associated with the i-th neuron is eventually trained to represent the code-vector c^ in the code-book. As the network is being trained, all the weights will be optimized to represent the best possible partition of all the input vectors. The aim of this paper is to present a new technique for image compression based on competitive neural networks. This technique allows to reduce the number of code-vectors needed to achieve a high quality code-book. This method is based on the use of a combination of multiple prototypes to store one input vector, achieving a better representation of the input dataset. 2 The Standard Approach Let us consider an image I{i,j)- In order to compress the image by means of VQ competitive networks, the image is subdivided into N x M square sub-images of size kx k, called windows. These windows are our input patterns with p = k"^ components (these patterns are obtained by arranging the pixel values row by row from top to bottom). The compression process consists in selecting a reduced set of K representative windows (corresponding to the solution prototypes) and replacing each window of the original image with the closest representative window among the prototypes. Thus, the compressed image f is built. In this work we have used the standard competitive learning (SCL) rule to build the code-book. Let X = {xi,..., x^} be the set of input patterns, and let {wi,..., WK} represent the code-book. The algorithm is as follows: L Choose an input pattern, Xi. 2. Compute the winning prototype (the closest to x^), Wc, verifying \\xi-wcf =. min \\xi-wjf j = l,...,k 3. Update the vector Wc according to Wc(t + 1) = Wc(t) + a(t) {xi - v^c(t))

Image Compression with Competitive Networks 341 where a{t) is the learning rate parameter, usually convergent to 0 as the number of iterations t tends to infinity. A typical value for a(t) is given by a linear decrease from 0.9 to 0 along the iterations. 4. Increase t, and repeat steps 1-4 until convergence is detected. The bottleneck of this algorithm is the computation of the closest prototype, since it involves the calculation of K distances. Thus, the time spent by this algorithm to build the code-book when the whole set of input patterns is presented to the net is proportional to n K. Though many other learning rules have been developed to obtain quasioptimal code-books [?,?,?,?,?], we have used the SCL algorithm since its performance is very well-known and allows us to compare our technique to the standard approach. 3 Pre-fixed Prototypes Our proposal diff"ers from the standard approach in several points: - Some of the prototypes used in the compression are a priori known by the encoder (transmitting the image) and decoder (receiving the image). Thus, with some prototypes pre-fixed in advance, the size of the code-book is reduced considerably. So, if, for example, we have K 32 prototypes, and we fix K' 16 of them, then the encoder only has to transmit K K' = 16 prototypes after the encoding phase. - In addition, since not all prototypes need to be computed, the encoding algorithm (the competitive neural network) will spend less time in computing the code-book. Thus, with our proposal, we achieve a high quality compression by reducing computational time as well as the code-book storage space. Our technique maintains the same compression rate as the standard approach for the given number of prototypes. This means that the transmission of the image from the encoder to the decoder (receiver) is improved, since the codebook storage space is halved. The encoding algorithm is based in applying the standard competitive learning rule (SCL) mentioned above, but with the restriction of the pre-fixed prototypes: 1. Choose an input pattern, Xi. 2. Compute the winning prototype (the closest to Xi), Wc, verifying lki-wc ^= jnin \\xi--wj\\'^ 3. If the winning unit Wc represents one of the fixed prototypes, it does not learn. Otherwise, update the vector Wc according to Wc(t + 1) = We{t) + a{t) {Xi - Wc(t)) where a(t) is the learning rate parameter.

342 Enrique Merida-Casermeiro et al. 4. Increase t, and repeat steps 1-4 until convergence is detected. It seems reasonable to use K' = ^^ that is, to fix half the prototypes, since it allows an easy codification of each input pattern: if b bits mo,..., rrib-i are used to encode a pattern, then mo will indicate whether the corresponding prototype is a priori known by both encoder and decoder (mo = 0) or not (mo =!) The other b 1 bits are used to encode the prototype number, from 0 to K\ So, the decoder, when a sequence of bits arrives, studies mo and looks for the prototype in the corresponding part of the code-book (the fixed or the transmitted, depending on mo). The intuitive idea behind the utilization of pre-fixed prototypes is that these prototypes are designed to fit in areas of the image in which there is little detail (sky, rivers, etc.), whilst the learnt prototypes focus on the areas of great detail. 4 Experimental Results In this section we compare the efficiency of our proposal in image compression with respect to the standard approach. We have considered a test set of images formed by 4 images (the well-known Lenna image plus 3 images from MatLab Image Processing Toolbox, see Fig. 1). Each image in our experiments has 256 x 256 pixels, and window size is 4x4. So, the number of input patterns for the standard algorithm is 4096. The number K of representative windows varies in the set /C = {32,64}. Two measures has been used in this work to compare the efficiency of our technique - The peak-signal-to-noise-ratio (PSNR), defined as: / 255 PSNR = 201og,o(^^^ and measured in decibels (db), where the image has 256 gray levels and RMSE is the root mean square error between two images (the original and the compressed). Ideally, a value of PSNR = cx) is the goal to attain, since it corresponds to a lossless compression. Thus, one tries to obtain the maximum value possible for PSNR. - The 1-norm ( H^) of the difference between the original image I and the compressed image /', defined as: 7-/' i=max^ /(j,j)-7'(j,i) i where I{i,j) represents the value of the pixel (i,j) in the image /. A lower value of this norm indicates a better approximation of the compressed image to the original one. Some of the test images are shown in Fig. 1. The compressed images are shown in Figs. 2 and 3. In these figures, we can observe that our compressed images obtain higher visual fidelity than the standard approach.

Image Compression with Competitive Networks 343 Fig. 1. Sample images from the test set: lenna (top left), cameraman (top right), rice (bottom left) and kids (bottom right). For our proposal, y fixed prototypes were built by considering vectors of the form w^ = (Q,..., Ci) G {0,..., 255}^, that is, all components are equal, each prototype representing a window with a constant grey-level in all its pixels. Ci values (i = 1,..., y) were equally spaced in {0,..., 255}. The same pre-fixed prototypes were used for every test image. In Table 1, there are represented the PSNR values of the compressed images using both the standard and our approach. It can be noted that if our method is used with K^ = K/2 fixed prototypes, where K is the number of the standard approach, the PSNR value obtained is at least comparable, and better in some cases, to that of the standard approach. This surprising fact may be explained as follows: since K' = ^ prototypes are pre-fixed, the dimensionality of the problem is reduced to half. As the dimension is reduced, the number of possible local minima of the distortion function is also reduced. So, the learning phase can avoid certain bad local minima present when all K prototypes are considered. This is also possible since prototypes with a constant gray-level in all its components are usually present in wide regions of most images (regions with little detail). I can also be observed that the 1-norm of our proposal is lower in most cases than the corresponding of the standard approach.

344 Enrique Merida-Casermeiro et al. Table 1. PSNR and 111 results for the considered compressed images. The quotient ^new At is defined by At = ^old Standard Approach Prop. Image cameraman rice lenna kids cameraman rice lenna kids K 32 32 32 32 64 64 64 64 PSNR 23.94 28.00 25.41 27.18 25.32 29.41 26.73 27.92 Mil 23.6 11.8 16.9 12.1 19.3 8.6 13.6 11.8 ^old 33.65 33.31 33.46 33.56 53.40 45.54 46.31 53.73 PSNR 24.13 28.47 25.39 26.79 25.16 30.33 26.48 27.87 II 111 20.1 10.5 16.0 13.8 18.9 7.6 15.7 11.24 ^new 31.38 32.01 31.12 31.17 45.64 43.50 43.00 42.18 At 0.93 0.96 0.93 0.93 0.85 0.95 0.93 0.78 The time spent by our algorithm to build the code-book is also lower than the standard, as shown in the last column of Table 1, where At indicates the quotient between the time spent by the standard approach and the time used by our proposal. 5 Conclusions In this work we have presented a new algorithm to perform image compression based on combining fixed prototypes with the solution prototypes obtained by a competitive learning rule. This new method is able to represent compressed images with higher visual fidelity than the standard approach, at the same time that achieves greater PSNR values in many cases. In addition, with this method, the computation of the code-book is less time-consuming, since fixed prototypes are never updated. We have tested our approach using the standard competitive learning rule. Better results are expected if learning rules as [?,?,?] are used instead. Our future work covers the study of new algorithms taking advantage of combining fixed and variable prototypes to form diflferent windows in the compressed image. The development of a competitive learning rule to optimize the code-book is also an issue of research. The application of this technique to specific scenarios or groups of images (biomedical images, satellite images, etc.) can improve their processing, segmentation, compression and transmission.

Image Compression with Competitive Networks 345 Fig. 2. Compressed images (from top to bottom): standard approach with 32 prototypes, our proposal with 32 prototypes, standard approach with 64 prototypes and our proposal with 64 prototypes.

346 Enrique Merida-Casermeiro et al. V» V :^v -. V Fig. 3. Compressed images (from top to bottom): standard approach with 32 prototypes, our proposal with 32 prototypes, standard approach with 64 prototypes and our proposal with 64 prototypes.