Scalable Coding of Image Collections with Embedded Descriptors

Similar documents
Embedded Indexing in Scalable Video Coding

Embedded Indexing in Scalable Video Coding

HIERARCHICAL VIDEO SUMMARIES BY DENDROGRAM CLUSTER ANALYSIS

Modified SPIHT Image Coder For Wireless Communication

Differential Compression and Optimal Caching Methods for Content-Based Image Search Systems

A Image Comparative Study using DCT, Fast Fourier, Wavelet Transforms and Huffman Algorithm

MRT based Adaptive Transform Coder with Classified Vector Quantization (MATC-CVQ)

Review and Implementation of DWT based Scalable Video Coding with Scalable Motion Coding.

Compression of RADARSAT Data with Block Adaptive Wavelets Abstract: 1. Introduction

CHAPTER 4 REVERSIBLE IMAGE WATERMARKING USING BIT PLANE CODING AND LIFTING WAVELET TRANSFORM

QUANTIZER DESIGN FOR EXPLOITING COMMON INFORMATION IN LAYERED CODING. Mehdi Salehifar, Tejaswi Nanjundaswamy, and Kenneth Rose

Blind Measurement of Blocking Artifact in Images

JPIP Proxy Server for remote browsing of JPEG2000 images

JPEG 2000 vs. JPEG in MPEG Encoding

CMPT 365 Multimedia Systems. Media Compression - Image

Automatic Categorization of Image Regions using Dominant Color based Vector Quantization

A new predictive image compression scheme using histogram analysis and pattern matching

An Iterative Joint Codebook and Classifier Improvement Algorithm for Finite- State Vector Quantization

SIGNAL COMPRESSION. 9. Lossy image compression: SPIHT and S+P

Pre- and Post-Processing for Video Compression

Module 6 STILL IMAGE COMPRESSION STANDARDS

Pattern based Residual Coding for H.264 Encoder *

CHAPTER 6. 6 Huffman Coding Based Image Compression Using Complex Wavelet Transform. 6.3 Wavelet Transform based compression technique 106

Optimization of Bit Rate in Medical Image Compression

Interactive Progressive Encoding System For Transmission of Complex Images

Mesh Based Interpolative Coding (MBIC)

Weighting and selection of features.

DIGITAL IMAGE PROCESSING WRITTEN REPORT ADAPTIVE IMAGE COMPRESSION TECHNIQUES FOR WIRELESS MULTIMEDIA APPLICATIONS

MULTIMEDIA COMMUNICATION

Reversible Wavelets for Embedded Image Compression. Sri Rama Prasanna Pavani Electrical and Computer Engineering, CU Boulder

A Miniature-Based Image Retrieval System

AN EFFICIENT CODEBOOK INITIALIZATION APPROACH FOR LBG ALGORITHM

EXPLORING ON STEGANOGRAPHY FOR LOW BIT RATE WAVELET BASED CODER IN IMAGE RETRIEVAL SYSTEM

Repeating Segment Detection in Songs using Audio Fingerprint Matching

Image Retrieval Based on its Contents Using Features Extraction

7.5 Dictionary-based Coding

FPGA IMPLEMENTATION OF BIT PLANE ENTROPY ENCODER FOR 3 D DWT BASED VIDEO COMPRESSION

Advances of MPEG Scalable Video Coding Standard

Implication of variable code block size in JPEG 2000 and its VLSI implementation

A QUAD-TREE DECOMPOSITION APPROACH TO CARTOON IMAGE COMPRESSION. Yi-Chen Tsai, Ming-Sui Lee, Meiyin Shen and C.-C. Jay Kuo

Lecture 5: Error Resilience & Scalability

International Journal of Advanced Research in Computer Science and Software Engineering

Overcompressing JPEG images with Evolution Algorithms

Content Based Image Retrieval Using Color Quantizes, EDBTC and LBP Features

CODING METHOD FOR EMBEDDING AUDIO IN VIDEO STREAM. Harri Sorokin, Jari Koivusaari, Moncef Gabbouj, and Jarmo Takala

A SCALABLE SPIHT-BASED MULTISPECTRAL IMAGE COMPRESSION TECHNIQUE. Fouad Khelifi, Ahmed Bouridane, and Fatih Kurugollu

JPEG 2000 compression

Rate-distortion Optimized Streaming of Compressed Light Fields with Multiple Representations

The Analysis and Detection of Double JPEG2000 Compression Based on Statistical Characterization of DWT Coefficients

Multi-View Image Coding in 3-D Space Based on 3-D Reconstruction

Digital Image Processing

Bit-Plane Decomposition Steganography Using Wavelet Compressed Video

Compression of Hyperspectral Images With LVQ-SPECK. CIPR Technical Report TR

Optimized Progressive Coding of Stereo Images Using Discrete Wavelet Transform

Optimal Estimation for Error Concealment in Scalable Video Coding

Complexity Reduced Mode Selection of H.264/AVC Intra Coding

Highly Secure Invertible Data Embedding Scheme Using Histogram Shifting Method

Image Compression for Mobile Devices using Prediction and Direct Coding Approach

A combined fractal and wavelet image compression approach

CS 335 Graphics and Multimedia. Image Compression

Xiaoqing Zhu, Sangeun Han and Bernd Girod Information Systems Laboratory Department of Electrical Engineering Stanford University

Title Adaptive Lagrange Multiplier for Low Bit Rates in H.264.

DIGITAL IMAGE WATERMARKING BASED ON A RELATION BETWEEN SPATIAL AND FREQUENCY DOMAINS

IMAGE PROCESSING (RRY025) LECTURE 13 IMAGE COMPRESSION - I

An Optimized Template Matching Approach to Intra Coding in Video/Image Compression

ANALYSIS OF SPIHT ALGORITHM FOR SATELLITE IMAGE COMPRESSION

Module 9 AUDIO CODING. Version 2 ECE IIT, Kharagpur

3D Mesh Sequence Compression Using Thin-plate Spline based Prediction

Image and Video Watermarking

DIGITAL TELEVISION 1. DIGITAL VIDEO FUNDAMENTALS

Image Compression Using BPD with De Based Multi- Level Thresholding

Enhanced Hybrid Compound Image Compression Algorithm Combining Block and Layer-based Segmentation

Context based optimal shape coding

International Journal of Emerging Technology and Advanced Engineering Website: (ISSN , Volume 2, Issue 4, April 2012)

Star Diamond-Diamond Search Block Matching Motion Estimation Algorithm for H.264/AVC Video Codec

Fundamentals of Video Compression. Video Compression

IMAGE COMPRESSION. Image Compression. Why? Reducing transportation times Reducing file size. A two way event - compression and decompression

LETTER Improvement of JPEG Compression Efficiency Using Information Hiding and Image Restoration

Audio-coding standards

06/12/2017. Image compression. Image compression. Image compression. Image compression. Coding redundancy: image 1 has four gray levels

EFFICIENT METHODS FOR ENCODING REGIONS OF INTEREST IN THE UPCOMING JPEG2000 STILL IMAGE CODING STANDARD

A Hybrid Temporal-SNR Fine-Granular Scalability for Internet Video

Neural Network based textural labeling of images in multimedia applications

Compression of Image Using VHDL Simulation

Video compression with 1-D directional transforms in H.264/AVC

A Comparison of Still-Image Compression Standards Using Different Image Quality Metrics and Proposed Methods for Improving Lossy Image Quality

Part 1 of 4. MARCH

Scalable Medical Data Compression and Transmission Using Wavelet Transform for Telemedicine Applications

Reconstruction PSNR [db]

AN EFFICIENT BATIK IMAGE RETRIEVAL SYSTEM BASED ON COLOR AND TEXTURE FEATURES

BLIND QUALITY ASSESSMENT OF JPEG2000 COMPRESSED IMAGES USING NATURAL SCENE STATISTICS. Hamid R. Sheikh, Alan C. Bovik and Lawrence Cormack

JPEG: An Image Compression System. Nimrod Peleg update: Nov. 2003

A COMPRESSION TECHNIQUES IN DIGITAL IMAGE PROCESSING - REVIEW

Authentication and Secret Message Transmission Technique Using Discrete Fourier Transformation

Performance Comparison between DWT-based and DCT-based Encoders

A Study on the Effect of Codebook and CodeVector Size on Image Retrieval Using Vector Quantization

( ) ; For N=1: g 1. g n

An Efficient Mode Selection Algorithm for H.264

Fingerprint Image Compression

Data Hiding in Video

Transcription:

Scalable Coding of Image Collections with Embedded Descriptors N. Adami, A. Boschetti, R. Leonardi, P. Migliorati Department of Electronic for Automation, University of Brescia Via Branze, 38, Brescia, Italy Abstract With the increasing popularity of repositories of personal images, the problem of effective encoding and retrieval of similar image collections has become very important. In this paper we propose an efficient method for the joint scalable encoding of image-data and visual-descriptors, applied to collections of similar images. From the generated compressed bit stream, it is possible to extract and decode the visual information at different granularity levels, enabling the so called Midstream Content Access. The proposed approach is based on the appropriate combination of Vector Quantization (VQ) and JPEG2000 image coding. Specifically, the images are encoded at a first draft level using an optimal visual-codebook, while the residual errors are encoded using a JPEG2000 approach. In this way, the codebook of the VQ is freely available as an efficient visual descriptor of the considered image collection. This scalable representation supports fast browsing and retrieval of image collections providing also a coding efficiency comparable with those of standard image coding methods. I. INTRODUCTION Recently the problem of effective encoding and retrieval of similar image collections has become very important. Traditional image coding approaches address the problem of Rate-Distortion (RD) optimization, trying to obtain the maximum quality (in terms of SNR) at a given bit-rate. An evolution of these coding schemes tries to obtain the scalability of the coded bit stream without affecting too much the coding performance in terms of RD. Scalable Coding (SC) methods, such as for example JPEG2000 (J2K) [1], generate a bit stream with a unique feature, the possibility of extracting decodable sub-streams corresponding to a scaled version, i.e., with a lower spatial resolution and/or a lower quality, of the original image. Moreover, this is achieved providing coding performances comparable with those of single point coding methods and requiring a very low sub-stream extraction complexity, actually comparable with read and write operations. Considering the problem of content based analysis and retrieval of multimedia documents, a specific importance is related to the low-level descriptors of the considered audiovisual information. Usually these descriptors are extracted ad hoc from the considered signal, independently from the adopted coding scheme, which requires some more computation at the expense of a more complex end-user system. The effectiveness of the low-level descriptors in terms of retrieval capabilities is usually measured considering the Recall and Precision (RP) performance. It could be an interesting approach to consider jointly the problem of data encoding and low-level descriptors representation in a unique data stream, optimized jointly with respect to the RD and the RP performance. In this respect, in [2] the problem of including in the overall coding objectives also the Content access work (in the cited paper, the so called fourth criterion ) added to the three classic criterion, Bit rate, Distortion, Computational cost, has been introduced as an innovative concept and an interesting new research field. This basic idea was then further developed in [3], where an image CODEC providing an easy access to the spatial image content was proposed and evaluated. Another example of application of this idea can be found in [4], [5], where an image coding method explicitly designed to ease the task of content access has been introduced. Specifically in [5] the proposed coding method uses quite sophisticated approaches such as Segmentation- Based Image Coding, Vector Quantization, Coloured Pattern Appearance Model, and leads the possibility to access image content descriptors with reduced computational complexity. Following this direction, in this paper we have considered the specific case of scalable joint data and descriptor encoding of image collections. The proposed approach is based on a suitable combination of Vector Quantization and JPEG image coding. The basic idea is to represent a collection of similar images at different levels of resolution, using a progressive encoding method based on low-level visual descriptors, namely the Visual Codebook (VC) of a Vector Quantizer (VQ), used to generate a prediction of the image that will be refined in the next resolution levels. Both the visual descriptors and the prediction error are then encoded using a standard JPEG encoder. In this way, the visual descriptors are naturally embedded in the coded bit stream, without additional costs for their extraction. Moreover, the availability of the Visual Codebook allows for an efficient characterization of the image collection as a whole, very useful for both contentbased retrieval and duplicate detection of image collections. The coding efficiency is also good, as demonstrated in the experimental evaluations.

Predictive image coding + - Residue encoding JPEG2K Images collection Image prediction indexes coding Visual Codebook generation Visual codebook generation VC ˆ VC JPEG2K Multiplexer R G B Quantization and Rate controller Collection Conllection Repository Fig. 1. The proposed hybrid encoder. Fig. 3. Transformation of an image block into a vector assuming l = 3. Image 1 Image 2 Image n VC Indexes residue Indexes residue.. Indexes residue Fig. 2. High level bit stream structure. II. THE PROPOSED HYBRID ENCODER FOR COLLECTIONS OF SIMILAR IMAGES The hybrid codec proposed in this work, is shown in Fig. 1. Two main blocks can be identified: the optimal visual codebook estimator and the predictive encoder. When a new group of similar images is added to the collection repository the coding process starts by finding the set of visual codebooks that provides the best RD tradeoff and, at the same time, allows to discriminate visually different groups of images. Once this set of visual codewords is found, all the images in the group are vector quantized generating a low quality version of the original content. In the second stage, the quantized images are used as a prediction for the corresponding input signal. The visual codebook and the reconstruction errors are then coded by using J2K, while the vector indexes are compressed by using a dedicated entropy coder. All this information is then packaged as shown in Fig. 2. This bit stream structure provides scalability at several levels of quality, depending on the way J2K is used. At least it is possible to extract and decode the images at two different qualities: the vector quantized version (base layer) of the original images, and the maximum coded quality obtained by adding all the residual information (enhancement layer). Moreover, if the enhancement layer is encoded by enabling quality scalability (in J2K) more working points can be obtained. By slightly modifying the proposed coding scheme it is also possible to get true spatial scalability. To achieve this goal, all the original images should be initially processed by applying the same wavelet decomposition used by the J2K encoder. Vector quantization and consequently prediction should then be applied to the low-pass spatial sub-band. A. Generation of the optimal codebook The algorithm used to train the vector quantizer is the accelerated k-means described in [6]. The input images are initially decomposed into a set of vectors, as shown in Fig. 3, which are used as training data. A certain number of Visual Codebooks V C(l, k) is generated varying the block size l and the number of codewords k in order to jointly find the best set of image blocks predictor and the best set of image block descriptors for the whole collection. This can be achieved by finding the pair ( l, k) which minimize the following cost function: J(l, k) = V Q cost (l, k) + P R cost (l, k) (1) where V Q cost (l, k) is the cost function associated to the vector quantization of all the images in the considered collection, and P R cost (l, k) expresses how the codebook V C(l, k) allows for a good discrimination of the considered images with respect to the collections already stored in the database. The function is given by the weighted sum V Q cost (l, k) = α(r V C(l,k) + βd V C(l,k)) where R V C(l,k) and D V C(l,k)) are the rate and the distortion associated to the vector quantization process. The second term of Equation (1) is given by P R cost (l, k) = γ(1 P rec(l, k)) where P rec(l, k) is the precision of inter-collection classification, i.e., the percentage of correctly classified images with respect to the collections in the database. B. Inter-collection classification To enable image retrieval and/or classification a robust mechanism to evaluate image similarities is needed. Concern-

ing the choice of the low-level features used to represent the visual-content and the metric adopted to estimate similarities, a variety of methods can be found in the literature. However, even if some sophisticated techniques have been proposed, in most cases image indexes have been generated by means of color information, often simply relying on the color histogram in one of the available color spaces or on color spatial distribution. In this work, we describe each collection of similar images with a dictionary of visual codewords, obtained from a vector quantization process. Traditionally employed for coding purposes [7], visual codebooks have been successfully proposed in [8] as an effective low-level feature for video indexing. Assuming to quantize the visual content by using a VC, the similarity between two images and/or collections can be estimated by evaluating the distortion introduced if the role of two VCs is exchanged. More formally, let C i, i = 1,.., n be an image (or a collection), N i the number of Visual Vectors (VV) obtained after its decomposition into blocks (see Fig. 3) and let V C j be a generic visual codebook generated by a vector quantizer. The reconstruction error can then be measured by evaluating the average distortion D V Cj (S i ), defined as: D V Cj (S i ) = 1 N i N i p=1 vv i (p) vc j (q) 2, (2) where vc j (q) is the codeword V C j with the smallest euclidean distance from the visual vector vv i (p), i.e.: q = arg min z vv i (p) vc j (z) 2. (3) Now, given two codebooks V C h and V C j, the value: D V Ch (C i ) D V Cj (C i ) (4) can be interpreted as the similarity between the two codebooks, when applied to the same visual content C i. A symmetric form, used in [8] to estimate the similarity measure between different images C i and C j can, thus, be defined as: φ(c i, C j ) = D V Cj (C i ) D V Ci (C i ) + D V Ci (C j ) D V Cj (C j ) (5) where V C i and V C j are in this case the optimal codebook for the image class C i and C j respectively. The smaller φ(.) is, the more similar the images are. Note that the similarity is based on the cross-effect of the two codebooks on the two considered image sets. In fact, it may be possible that the majority of blocks of one class (for example C i ), can be very well represented by a subset of codewords of codebook V C j. Therefore V C j can represent C i with a small average distortion, even if the visual-content of the two shots is only partly similar. On the other hand, it is possible that codebook V C i does not lead to a small distortion when applied to C j. So Classification Precision 1 0.95 0.9 0.85 0.8 0.75 0.7 Fig. 4. 22 20 l=4 l=8 l=16 8 16 32 64 128 256 512 Number of code-words (k) Examples of inter collection classification precision curves. l=4 l=8 l=16 0 0.1 0.2 0.3 0.4 Fig. 5. Rate-Distortion curve obtained by applying only VQ (average value for a whole image collection). the cross-effect of codebooks on the two classes can generally reduce the number of wrong classifications. C. Selection and coding of the best visual codebook Obviously the optimization of the cost function J(.) is a computationally intensive task, since it is needed to compute both the RD function cost associated to the vector quantization for several values (l, k). Concerning inter classification precision, it is required to previously estimate its average behaviors by using a ground truth. Examples of the V Q cost (l, k) and P R cost (l, k) function are reported in Figures 4 and 5 respectively. As it can be noticed, at some combinations of l and k there are possible saturation effects. If this is the case, at least one of the two member of equation 1 can be removed from the optimization procedure. Given the best VC, all its visual codewords are arranged in order to store the VC in an encoded image. Basically this task is the opposite process used to generate the visual vector

set from an input image (see Fig. 3). The VC image is then encoded, lossless or lossy, by using J2K. The compressed stream is sent to the multiplexer while the decoded version of this signal ( V ˆ C) is provided to the following coding stage. D. Predictive image coding Once the optimal visual codebook is generated and compressed, the scalable encoding of the whole collection is performed. Each image is initially quantized by applying the visual codebook V ˆ C (see Fig. 1). The generated visual codeword indexes are then entropy coded and sent to the stream multiplexer. The quantized image is then used as a prediction for the corresponding original input one. Finally the residual image, given by the difference between the input signal and its prediction, is encoded by using the J2K encoder. III. SYSTEM PERFORMANCE EVALUATION In this section the coding performance and the browsing/retrieval capabilities provided by the proposed method are evaluated. A. Coding effectiveness The coding effectiveness of the proposed hybrid codec has been evaluated comparing its rate distortion curves with those achieved by separately encoding all the images with J2K. This has been accomplished by encoding several collections and averaging the results by using different sizes of the image blocks (l) and different numbers of visual codewords (k). The best performances have been obtained, with l = 16 and few visual codewords k [8, 64] (see Fig. 6), where the proposed hybrid encoder for rates ranging in [0.5, 2.5] bit/pixel, or equivalently for quality values of [35 : 40]dB behaves as well as his competitor. As shown in Fig. 7, decreasing the block size to l = 8 lowers also the performance for k = 64. When a small block size, e.g., l = 4 and/or a large number of visual codewords are used, the compression effectiveness significantly decreases. This behaviour can be explained by analysing the rate distortion curves associated to the pure vector quantization process and the characteristics of the residual error. In fact, the compression ratio provided by vector quantization systems are quite far from those provided by traditional image coding methods, especially at high rates. Moreover, as it has been previously shown, depending on the vector quantization algorithm used, saturation effects can occur (see Fig. 5). This implies that the prediction stage should use the lowest possible rate which provides the desired inter collection discrimination capabilities. Moreover, increasing the rate allocated to the vector quantization process increases the relative amount of high frequencies in the residual signal. Since J2K has been designed and tuned mainly to compress natural images, the performance of the residual image coding 40 38 36 34 32 30 28 l=16, k=8 l=16, k=64 J2K 0 0.5 1 1.5 2 2.5 Fig. 6. Rate-Distortion curve comparison for l = 16 and k = 8, 64. 40 38 36 34 32 30 28 l=8, k=8 l=8, k=64 j2k 0 0.5 1 1.5 2 2.5 Fig. 7. Rate-Distortion curve comparison for l = 8 and k = 8, 64. are significantly lowered. With respect to the use of pure J2K system, the complexity of the proposed method is clearly higher. However, this penalty is well balanced by the provided embedding of visual descriptors in the final bit stream, which can be used in image browsing and retrieval applications. B. Scalable Browsing of Image Collection In Fig. 8 it is shown an example of the different levels of details at which the collection repository can be browsed. The lowest level corresponds to the representation of the content only by using the associated visual codebook. Anyway, also this rough representation gives a quick idea of the color patterns that characterize the images of the given collection. It is also possible to visualize the content through the quantized version of the most representative image in the collection (second line in Fig. 8) which is the image that has been quantized by using the highest number of different visual codewords. Data can also be browsed in more detailed views

Fig. 8. Examples of possible browsing resolution levels and the associated rates. TABLE I CLASSIFICATIONS RESULTS. Beach Mountain Boat House Beach 7 0 1 0 Mountain 0 7 0 1 Boat 0 1 7 0 House 0 0 0 8 according to the available quality layers. C. Retrieval and Classification Capabilities Beyond coding and fast browsing, the ability of automatically assigning a new image to a collection in the database has been also evaluated. In principle, this can be used to potentially avoid the proliferation of new collections but also to find collections visually similar to a query image. The image classification task has been realized by using the similarity estimation, described in Section II-B, which relies on the cross comparison of visual codebook. The considered test set is composed by four classes of images similar to those presented in Section III-B. Analyzing the confusion matrix reported in Tab. I the classification results are quite good, even if at the moment they are only indicative and a wider training and testing is in progress. IV. CONCLUSION This paper proposes an efficient method for scalable coding of collections of similar images, with embedded low-level visual descriptors. The key aspect that characterizes the proposed approach is the ability to embed in the compressed stream visual descriptors that possibly allow for a new scalability dimension. This paper also shows, how it is possible to enable browsing of image collections at different levels of detail: visual codebook, thumbnail, and images, by extracting and decoding the corresponding portion of the original bit stream. It is also possible to get a rough classification of the similarity of new images with respect to the collections in the repository. The coding efficiency is also good, as demonstrated in the experimental evaluations. Despite the promised potentials, several issues should be investigated in order to generalize and further enhance the presented results. The most difficult aspects will concern the generalization of the proposed approach to other application, and the determination of new low level descriptors useful for both data compression and data retrieval. In this respect, an analysis of the performance degradation produced by increasing the number of collections in the database should also be performed. REFERENCES [1] D. S. Taubman and M. W. Marcellin, JPEG 2000: Image Compression Fundamentals, Standards and Practice. Norwell, MA, USA: Kluwer Academic Publishers, 2001. [2] R. W. Picard, Content access for image/video coding: the fourth criterion,, MIT Media Laboratory Perceptual Computing Section, Cambridge, USA, Tech. Rep. 295, 1994. [3] A. Hanjalic, R. Lagendijk, and J. Biemond, Efficient image codec with reduced content access work, Image Processing, 1999. ICIP 99. Proceedings. 1999 International Conference on, vol. 3, pp. 807 811 vol.3, 1999. [4] G. Schaefer and G. Qiu, Midstream content access of visual pattern coded imagery, Computer Vision and Pattern Recognition Workshop, pp. 144 144, 27-02 June 2004. [5] G. Qiu, Embedded colour image coding for content-based retrieval, Journal of Visual Communication and Image Representation, vol. 15, no. 4, pp. 507 521, 2004. [6] C. Elkan, Using the triangle inequality to accelerate k-means, in ICML, 2003, pp. 147 153. [7] A. Gersho and R. M. Gray, Vector quantization and signal compression. Norwell, MA, USA: Kluwer Academic Publishers, 1991. [8] C. Saraceno and L. R., Indexing audio-visual databases through a joint audio and video processing, International Journal of Imaging Systems and Technology, vol. 9, no. 5, pp. 320 331, 1998.