Real-Time Fusion of Multi-Focus Images for Visual Sensor Networks

Similar documents
Domain. Faculty of. Abstract. is desirable to fuse. the for. algorithms based popular. The key. combination, the. A prominent. the

Performance Evaluation of Biorthogonal Wavelet Transform, DCT & PCA Based Image Fusion Techniques

University of Bristol - Explore Bristol Research. Peer reviewed version. Link to published version (if available): /ICIP.2005.

Image Fusion Using Double Density Discrete Wavelet Transform

A Study and Evaluation of Transform Domain based Image Fusion Techniques for Visual Sensor Networks

Compressive Sensing for Multimedia. Communications in Wireless Sensor Networks

Integrated PCA & DCT Based Fusion Using Consistency Verification & Non-Linear Enhancement

Reduction of Blocking artifacts in Compressed Medical Images

A Image Comparative Study using DCT, Fast Fourier, Wavelet Transforms and Huffman Algorithm

FRACTAL IMAGE COMPRESSION OF GRAYSCALE AND RGB IMAGES USING DCT WITH QUADTREE DECOMPOSITION AND HUFFMAN CODING. Moheb R. Girgis and Mohammed M.

A Comparative Analysis of Different Image Fusion Techniques

Visible and Long-Wave Infrared Image Fusion Schemes for Situational. Awareness

Image Compression. -The idea is to remove redundant data from the image (i.e., data which do not affect image quality significantly)

Compression of RADARSAT Data with Block Adaptive Wavelets Abstract: 1. Introduction

Image Compression Algorithm and JPEG Standard

Multi-focus Image Fusion Using Stationary Wavelet Transform (SWT) with Principal Component Analysis (PCA)

Video Compression An Introduction

Optimizing the Deblocking Algorithm for. H.264 Decoder Implementation

Video Compression Method for On-Board Systems of Construction Robots

15 Data Compression 2014/9/21. Objectives After studying this chapter, the student should be able to: 15-1 LOSSLESS COMPRESSION

Index. 1. Motivation 2. Background 3. JPEG Compression The Discrete Cosine Transformation Quantization Coding 4. MPEG 5.

Data Hiding in Video

BLIND MEASUREMENT OF BLOCKING ARTIFACTS IN IMAGES Zhou Wang, Alan C. Bovik, and Brian L. Evans. (

Optimized Progressive Coding of Stereo Images Using Discrete Wavelet Transform

HYBRID TRANSFORMATION TECHNIQUE FOR IMAGE COMPRESSION

Robust Image Watermarking based on DCT-DWT- SVD Method

Comparison of DCT, DWT Haar, DWT Daub and Blocking Algorithm for Image Fusion

JPEG compression of monochrome 2D-barcode images using DCT coefficient distributions

Reducing/eliminating visual artifacts in HEVC by the deblocking filter.

Video Codec Design Developing Image and Video Compression Systems

Survey on Multi-Focus Image Fusion Algorithms

Review and Implementation of DWT based Scalable Video Coding with Scalable Motion Coding.

A Miniature-Based Image Retrieval System

JPEG 2000 vs. JPEG in MPEG Encoding

Interactive Progressive Encoding System For Transmission of Complex Images

Efficient Image Compression of Medical Images Using the Wavelet Transform and Fuzzy c-means Clustering on Regions of Interest.

CS 260: Seminar in Computer Science: Multimedia Networking

Video Compression Standards (II) A/Prof. Jian Zhang

Enhancing the Image Compression Rate Using Steganography

Compression of Stereo Images using a Huffman-Zip Scheme

Image Interpolation using Collaborative Filtering

IMAGE COMPRESSION USING HYBRID TRANSFORM TECHNIQUE

A Comparison of Still-Image Compression Standards Using Different Image Quality Metrics and Proposed Methods for Improving Lossy Image Quality

The Analysis and Detection of Double JPEG2000 Compression Based on Statistical Characterization of DWT Coefficients

System Modeling and Implementation of MPEG-4. Encoder under Fine-Granular-Scalability Framework

DIGITAL IMAGE WATERMARKING BASED ON A RELATION BETWEEN SPATIAL AND FREQUENCY DOMAINS

Region-Based Image Fusion Using Complex Wavelets

Image Error Concealment Based on Watermarking

ISSN (ONLINE): , VOLUME-3, ISSUE-1,

REVIEW ON IMAGE COMPRESSION TECHNIQUES AND ADVANTAGES OF IMAGE COMPRESSION

Blur Space Iterative De-blurring

JPEG Compression Using MATLAB

Image Segmentation Techniques for Object-Based Coding

Comparative Evaluation of DWT and DT-CWT for Image Fusion and De-noising

Multi-View Image Coding in 3-D Space Based on 3-D Reconstruction

6. Multimodal Biometrics

Multi-focus image fusion using de-noising and sharpness criterion

DCT-BASED IMAGE QUALITY ASSESSMENT FOR MOBILE SYSTEM. Jeoong Sung Park and Tokunbo Ogunfunmi

CS 335 Graphics and Multimedia. Image Compression

A Novel Explicit Multi-focus Image Fusion Method

A Novel Image Super-resolution Reconstruction Algorithm based on Modified Sparse Representation

Performance Comparison between DWT-based and DCT-based Encoders

Wavelet Transform (WT) & JPEG-2000

ADVANCED IMAGE PROCESSING METHODS FOR ULTRASONIC NDE RESEARCH C. H. Chen, University of Massachusetts Dartmouth, N.

International Journal of Engineering Research-Online A Peer Reviewed International Journal Articles available online

Adaptive Wavelet Image Denoising Based on the Entropy of Homogenus Regions

A Novel Approach for Deblocking JPEG Images

CSE237A: Final Project Mid-Report Image Enhancement for portable platforms Rohit Sunkam Ramanujam Soha Dalal

A QUAD-TREE DECOMPOSITION APPROACH TO CARTOON IMAGE COMPRESSION. Yi-Chen Tsai, Ming-Sui Lee, Meiyin Shen and C.-C. Jay Kuo

Image denoising in the wavelet domain using Improved Neigh-shrink

IMAGE PROCESSING USING DISCRETE WAVELET TRANSFORM

Multimedia Systems Image III (Image Compression, JPEG) Mahdi Amiri April 2011 Sharif University of Technology

Using Shift Number Coding with Wavelet Transform for Image Compression

Digital Image Processing

Block-Matching based image compression

IMAGE COMPRESSION USING HYBRID QUANTIZATION METHOD IN JPEG

Module 8: Video Coding Basics Lecture 42: Sub-band coding, Second generation coding, 3D coding. The Lecture Contains: Performance Measures

International Journal of Research in Computer and Communication Technology, Vol 4, Issue 11, November- 2015

A COMPRESSION TECHNIQUES IN DIGITAL IMAGE PROCESSING - REVIEW

An introduction to JPEG compression using MATLAB

IMAGE COMPRESSION. Image Compression. Why? Reducing transportation times Reducing file size. A two way event - compression and decompression

Fast Implementation of VC-1 with Modified Motion Estimation and Adaptive Block Transform

Hybrid Fused Displays: Between Pixel- and Region-Based Image Fusion

Introduction ti to JPEG

BIG DATA-DRIVEN FAST REDUCING THE VISUAL BLOCK ARTIFACTS OF DCT COMPRESSED IMAGES FOR URBAN SURVEILLANCE SYSTEMS

Digital Image Representation Image Compression

Features. Sequential encoding. Progressive encoding. Hierarchical encoding. Lossless encoding using a different strategy

A Robust Digital Watermarking Scheme using BTC-PF in Wavelet Domain

ECE 533 Digital Image Processing- Fall Group Project Embedded Image coding using zero-trees of Wavelet Transform

INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY

Compression Part 2 Lossy Image Compression (JPEG) Norm Zeck

MULTI-FOCUS IMAGE FUSION USING GUIDED FILTERING

SSIM based image quality assessment for vector quantization based lossy image compression using LZW coding

Part 1 of 4. MARCH

A Novel Statistical Distortion Model Based on Mixed Laplacian and Uniform Distribution of Mpeg-4 FGS

Wavelet Based Image Compression Using ROI SPIHT Coding

WAVELET BASED NONLINEAR SEPARATION OF IMAGES. Mariana S. C. Almeida and Luís B. Almeida

A Laplacian Based Novel Approach to Efficient Text Localization in Grayscale Images

CMPT 365 Multimedia Systems. Media Compression - Image

Module 7 VIDEO CODING AND MOTION ESTIMATION

Transcription:

Real-Time Fusion of Multi-Focus Images for Visual Sensor Networks Mohammad Bagher Akbari Haghighat, Ali Aghagolzadeh, and Hadi Seyedarabi Faculty of Electrical and Computer Engineering, University of Tabriz, Tabriz, Iran haghighat@ieee.org, aghagol@tabrizu.ac.ir, seyedarabi@tabrizu.ac.ir Abstract The objective of image fusion is to combine information from multiple images of the same scene in order to deliver only the useful information. The discrete cosine transform (DCT) based methods of image fusion are more suitable and time-saving in real-time systems using DCT based standards of still image or video. In this paper an efficient approach for fusion of multi-focus images based on variance calculated in DCT domain is presented. The experimental results on several images show the efficiency improvement of our method both in quality and complexity reduction in comparison with several recent proposed techniques. Keywords Multi-focus; image fusion; visual sensor networks; DCT domain; JPEG; consistency verification I. INTRODUCTION In all sensor networks, every sensor can apperceive environment, produce and transfer data. Visual Sensor Networks (VSN) is the term used in the literature to refer to a system with a large number of cameras, geographically spread resources and many monitoring points [1]. In VSN, sensors are cameras which can record either still images or video sequences. Therefore, the processing of output information is related to image processing and machine vision subjects. A distinguished feature of visual sensors or cameras is the great amount of generated data. This characteristic makes necessary more local processing to deliver only the useful information represented in a conceptualized level []. Image fusion is generally defined as the process of combining multiple source images into a smaller set of images, usually a single one, which contains a more accurate description of the scene than any of the individual source images. The aim of image fusion, besides reducing the amount of data in network transmissions, is to create new images that are more suitable for the purposes of human or machine perception, and for further image-processing [3]. Due to the limited depth of focus in optical lenses, only objects at one particular depth in the scene are in focus and those in front of or behind the focus plane will be blurred. In VSN we have the opportunity of extending the depth of focuses using more than one camera. So far, several researches have focused on image fusion performed on the images in the spatial domain [4]-[8]. The algorithms based on multi-scale decompositions are more popular. The basic idea is to perform a multi-scale transform on each source image, and then integrate all these decompositions to produce a composite representation. The fused image is finally reconstructed by performing the inverse multi-scale transform. Examples of this approach include the Laplacian, gradient, or morphological pyramids, and the superior ones, discrete wavelet transform (DWT) [9] and shift invariant discrete wavelet transform (SIDWT) [1] which are treated in literature as standard methods. A good survey on these works may be found in [11] and [1]. In general, most of the spatial domain image fusion methods are complex and time-consuming which are hard to be performed on real-time applications. Moreover, when the source images are coded in Joint Photographic Experts Group (JPEG) standard or when the fused image will be saved or transmitted in JPEG format, the fusion approaches which are applied in DCT domain will be very efficient. We denote A and B as the output images of two cameras that have been compressed in JPEG coding standard in the sensor agent and transmitted to fusion agent of VSN. In the case of using the spatial domain fusion methods, these images must initially be decoded and transferred into the spatial domain. Then, after applying the fusion procedure, the fused image must be coded again in order to be stored or transmitted to an upper node. Recently, Tang [13] has considered the above mentioned issue of complexity reduction and proposed two image fusion techniques in DCT domain, namely, Average and DCT + Contrast. Tang has implemented his proposed methods on 8 8 DCT blocks defined in JPEG standard. Average obtains the DCT representation of the fused image by simply taking the average of all the DCT coefficients of all the input images. This simple method of averaging leads to undesirable side effects including blurring. For the second technique called Contrast, fusion criterion or activity level is based on a contrast measure which is calculated for every 63 AC coefficients of the blocks from source images. Then the contrast measures of each coefficient in the corresponding blocks in source images are compared. Then the coefficient with the highest activity level is selected. The DCT block of the output image is made up of AC coefficients with the highest contrast in comparing procedure, and DC coefficient of each block in the output image is the average of DC coefficients of the corresponding blocks in the input images. This algorithm is also complex in calculating the contrast measure for each coefficient. Furthermore, it suffers 978-1-444-978-9/1/$6. 1 IEEE

from side effects including blocking artifacts due to the manipulation in the diverse selection of DCT coefficients. In order to reduce the complication for the real-time applications and also enhance the quality of the output image, an image fusion technique in DCT domain is proposed in this paper. Here, the variance of 8 8 blocks calculated from DCT coefficients is used as a contrast criterion for the activity measure. Then, a consistency verification (CV) stage increases the quality of output image. Simulation results and comparisons show the considerable improvement in the quality of the output image and reduction of computation complexity. This paper is organized as follows: in section II, the basic concepts of our approach are discussed which is based on variance calculation in DCT domain. Section III describes our proposed method of image fusion. In section IV the simulation results are presented and analyzed. Section V concludes the paper. II. DCT BLOCKS ANALYSIS A. DCT in Compression Standards Several commercial standards widely used in image and video compression are based on DCT. Some examples include JPEG still image coding standard [14], Motion-JPEG, MPEG and the ITU H6X video coding standards [15]. In JPEG encoder, the given image is first partitioned into nonoverlapping 8 8 blocks. The two-dimensional DCT is then performed on each 8 8 block. Once the DCT coefficients are obtained, they are quantized using a specified quantization table. Quantization of the DCT coefficients is a lossy process, and in this step, many small coefficients (usually high frequency components) are quantized to zeros. The zigzag scan of the DCT coefficients matrix followed by entropy coding makes use of this property to lower the bit rate required to encode the coefficients. In the decoder side, each block of the compressed data is decoded and then dequantized by with the quantization table and then transformed into an image block by inverse DCT. B. Variance in DCT Domain as a Contrast Criterion Two-dimensional DCT transform of an N N block of an image x(m,n) is defined as: α( k) α( l) dkl (,) = N (m+ 1) πk (n+ 1) πk (1) xmn (, ) cos cos m= n= where k,l=,1,,n-1 and 1, if k = α( k) = () 1, otherwise The inverse DCT (IDCT) is also defined as: α( k) α( l) xmn (, ) = dkl (, ) k= l= N (m+ 1) πk (n+ 1) πk cos cos where m,n=,1,,n-1. In (1), d(,) is the DC coefficient and the other d(k,l)s are the AC coefficients of the block. The normalized transform coefficients dkl ˆ(, ) s are defined as below: dkl (, ) dkl ˆ(, ) = (4) N Mean value, μ, and variance,, of an N N block in the spatial domain are calculated as: m= n= (3) 1 μ = xmn (, ) (5) N σ 1 = x ( m, n) μ N m= n= (6) As we know the mean value, μ, of the N N block is given by d ˆ(,). With a little mathematical calculation, we can also provide the variance of the block from DCT coefficients. We have: x ( m, n) = x( m, n). x( m, n) m= n= m= n= = xmn (, ) m= n= α ( k) α ( l) (m+ 1) k (n 1) k. dkl (, ).cos π +.cos π k= l= N By interchanging the positions of the summations, we will have: α( k) α( l) = dkl (, ) k= l= N (m+ 1) πk (n+ 1) πk xmn (, ) cos cos m= n= α( k) α( l) N d( k, l) = dkl (, ) N α ( k ) α ( l ) k= l= = k= l= dkl (, ). dkl (, ) (7) x ( m, n) = d ( k, l) m= n= k= l=

σ Then, from (6) and (7), we have: 1 = (, ) x m n μ N m= n= d ( k, l) σ = ˆ (,) (8) d k= l= N In conclusion, the variance of an N N block of pixels can be exactly calculated from its DCT coefficients by sum of the squared normalized AC coefficient of the DCT block. III. PROPOSED METHOD VARIANCE BASED IMAGE FUSION IN DCT DOMAIN In multi-focus images, the focused area is more informative. This information is behind the details of that region which is corresponding to the high variance. Variance value is usually assumed as a contrast criterion in image processing applications. It is shown in Sec. II that the calculation of variance in DCT domain is very easy. Therefore, we can use the variance value as the activity level of the 8 8 blocks of the source images. Fig. 1 illustrates the general framework of a JPEG encoder combining with image fusion for VSN. Here, for simplicity, we consider the processing of just two source images A and B, but the algorithm can be extended straightforwardly for more than two source images. Moreover, it is assumed that the source images were registered. As Fig. 1 shows, after dividing the source images into 8 8 blocks and calculating the DCT coefficients for each block, the fusion algorithm is performed. Here the variances of the corresponding blocks from source images are calculated by (8) as the activity measures. Then, the block with high activity level is selected as the appropriate one for the fused image. Finally, the DCT presentation of the fused image is made up of blocks with the larger variances that are from source image either A or B. A. Consistency Verification Assume a region in the scene including several blocks that is only in the depth of focus of image A. Following the above mentioned process, all the blocks in this region must be chosen from image A. However, there may be some errors that can lead to erroneous selection of some blocks from image B. This defect can be solved with a consistency verification procedure. The defective decision map is demonstrated in Fig. 1. Li et al. [9] introduced a majority filter that can be used for consistency verification. If the center block comes from source image B while the majority of the surrounding blocks come from source image A, the center sample is switched to source image A. The fused image is finally obtained based on the modified decision map. Here, consistency verification is applied in a 3 3 neighborhood window. If the window size increases, the quality of the output image obviously will raise at the expense of complexity increase. But instead of using a 5 5 window and considering 4 items, using a 3 3 window for two consequential times and considering a total of 16 items, will be more efficient in both quality and complexity. IV. EXPERIMENTAL RESULTS AND ANALYSIS In this section, experimental results of the proposed image fusion method are given and evaluated by comparing with the results obtained by four other techniques. Two of them are DCT based image fusion techniques Average and DCT + Contrast proposed in [13] and the others are standard wavelet based fusion proposed in [9] and the shift invariant wavelet based technique (SIDWT) [1] that have been briefly described in section I. A. Performance Measure Although there have been many attempts, no universally accepted criterion has emerged for objectively evaluating image fusion performance as yet. This problem lies in the difficulty in defining an ideal fused image. Besides the subjective evaluation, in order to have a quantitative comparison of the fusion methods, we adopt the quality measure proposed by Li et al. [9]. Here the out-of-focus images are artificially created by low-pass filtering of some areas in an ideal image. The structural similarity measure (SSIM) [16], as a well-known quality criterion, is used for objectively evaluation in this section. Moreover, the run-time of the algorithms, for the same circumstances of simulations, can be considered as a complexity criterion. B. Experimental Circumstances The simulations of the fusion methods have been conducted with an Intel CoreDuo P84,.6 GHz, FSB 166 MHz processor. For the wavelet based methods, the DWT with DBSS(,) and the SIDWT for Haar basis with three levels of decomposition have been considered. For simulation of these methods, the Image Fusion Toolbox, kindly provided by Oliver Rockinger [17], is used. C. Fusion Results Evaluation In the first stage, the algorithms are performed on eighteen couples of artificially created source images by blurring six standard original images shown in Fig., with three disks of different radiuses 5, 7 and 9 pixels. The images are balancing in the sense that the blurring occurs at both the left and the right halves of the images. The average SSIM values of 18 experimental images are given in TABLE I. The first two columns are the average SSIM values of the DCT based algorithms proposed by Tang [13]. The third and fourth columns are corresponding to DWT and SIDWT techniques proposed in [9] and [1], respectively. The fifth column titled as Variance is related to the proposed algorithm without consistency verification and finally the last column demonstrates the average SSIM value of the proposed algorithm after applying consistency verification (CV) in a 3 3 neighborhood which has the most value of SSIM.

Fig. 1. A general framework of a JPEG encoder combining with image fusion and the proposed image fusion method. The average period of the run-time for reconstruction of every 8 8 block of images in the DCT based methods are given in TABLE II. As mentioned before, the DCT based algorithms are considerably faster and less computationally expensive than the other techniques when the images to be fused are saved in JPEG format or when the fused image is saved or transmitted in JPEG format. From the average SSIM values given in TABLE I, it can be clearly seen that the proposed algorithm even without CV is more efficient than the other DCT based algorithms. Also, after applying CV, there is a considerable improvement in quality even more than the wavelet based algorithms at the expense of a little complexity. Moreover, the averages runtime values given in TABLE II, demonstrate the simplicity of our proposed algorithm even after applying CV which had a noticeable quality improvement. The subjective test of the resultant images approves the objective measures. Due to the lack of space in this paper, the resultant images of only one test image is shown in Fig. 3. By carefully observing the fusion results, it is concluded that the method Average results in the side effect of reduction in contrast or blurring the fused image (Fig. 3.c). The method Contrast results in blocking artifacts (Fig. 3.d). In addition there are some artifacts for the wavelet based methods DWT and SIDWT, respectively, given in Fig. 3.e and Fig. 3.f. As it can be seen in the magnified images corresponding to DWT and SIDWT, respectively in Figs. 3.k and 3.l, the wavelet based algorithms suffer from a kind of ringing artifact. This ringing artifact is a common problem in wavelet based algorithms. This is due to its finite number of basis functions and using high-pass filter embedded in the details Fig.. Images used for simulation. TABLE I THE AVERAGE SSIM VALUE OF THE ALGORITHMS DWT SIDWT Variance Average Contrast Variance + CV.9787.9467.9811.98389.97654.9987 TABLE II THE AVERAGE RUN-TIME VALUES OF THE ALGORITHMS (IN MICROSECONDS PER 8 8 BLOCK) Variance Average Contrast Variance + CV 11.3933 64.354 16.1441 64.56163 branch of the wavelet tree. In the standard wavelet decomposition, the down-sampling process enlarges the region of ringing effect. Figs. 3.g and 3.h are the results of the proposed method without CV and after applying CV, respectively. It is clearly seen from the images that the proposed algorithm has a considerable improvement in fused image quality. Another experiment was conducted on Book database. The results are shown in Fig. 4. For further comparing, the

Fig. 3. The Cameraman source images and the fusion results: (a) The first source image with focus on the right. (b) The second source image with focus on the left. (c) Average result. (d) Contrast Result. (e) DWT result. (f) SIDWT result. (g) Result of the proposed Variance algorithm. (h) Result of the proposed algorithm with consistency verification Variance + CV. (i), (j), (k), (l), (m), (n) are the local magnified version of (c), (d), (e), (f), (g) and (h), respectively. same experiment has been conducted on two other wellknown images Pepsi and Clock from Rockinger s database [17] and similar results are obtained. Eventually, these experiments prove the objective results as the superiority of our proposed method. V. CONCLUSION In this paper, a new DCT based fusion technique for multifocus images is proposed. The method is based on variance definition in DCT domain. Many experiments on evaluating the fusion performance have been conducted and the results show that the proposed method outperforms the previous DCT based methods both in quality and complexity reduction. Moreover, it is more efficient than the other methods when the source images are in JPEG format or the fused image is saved or transmitted in JPEG format, especially in visual sensor networks. A shortcoming of our method without consistency verification is that it is a bit weak on the boundaries between focused and out-of-focus areas that is solved here by CV. REFERENCES [1] [] [3] [4] [5] [6] O. Perez, M. A. Patricio. J. Garcia, J. Carbo, and J. M. Molina, Fusion of Surveillance Information for Visual Sensor Networks, Proceedings of the IEEE Ninth International Conference on Information Fusion (ICIF), pp. 1 8, 6. F. Castanedo, J. Garcia, M. A. Patricio, and J. M. Molina, Analysis of Distributed Fusion Alternativesin Coordinated Vision Agents, Proceedings of the IEEE Eleventh International Conference on Information Fusion (ICIF), pp. 1 6, 8. D. Drajic, and N. Cvejic, Adaptive Fusion of Multimodal Surveillance Image Sequences in Visual Sensor Networks, IEEE Transactions on Consumer Electronics, vol. 53, no. 4, pp. 1456 146, November 7. J. J. Lewis, R. J. O Callaghan, S. G. Nikolov, D. R. Bull, and N. Canagarajah, Pixel- and Region-Based Image Fusion with Complex Wavelets, Information Fusion, vol. 8,no., pp. 119 13, 7. S. Li, and B. Yang, Multifocus Image Fusion Using Region Segmentation and Spatial Frequency, Image and Vision Computing, vol. 6, no. 7, pp. 971 979, July 8. L. Xu, M. Roux, H. Mingyi, and F. Schmitt, A new method of image fusion based on redundant wavelet transform, Proceedings of the IEEE Fifth International Conference on Visual Information Engineering, pp. 1 17, August 8.

Fig. 4. The Book source images and the fusion results. (a) The first source image with focus on the right. (b) The second source image with focus on the left. (c) Average result. (d) Contrast Result. (e) DWT result. (f) SIDWT result. (g) The result of the proposed Variance algorithm. (h) The result of the proposed algorithm with consistency verification Variance + CV. (i), (j), (k), (l), (m), (n) are the local magnified versions of (c), (d), (e), (f), (g) and (h), respectively. [7] T. Zaveri, M. Zaveri, V. Shah and N. Patel, A Novel Region Based Multifocus Image Fusion Method, Proceeding of IEEE International Conference on Digital Image Processing (ICDIP), pp. 5 54, March 9. [8] M. H. Arif and S. S. Shah, Block Level Multi-Focus Image Fusion using Wavelet Transform, Proceeding of IEEE International Conference on Signal Acquisition and Processing (ICSAP), pp. 13 16, April 9. [9] H. Li, B. Manjunath and S. Mitra, Multisensor Image Fusion Using the Wavelet Transform, Graphical Models Image Processing, vol. 57, no. 3, pp. 35 45, May 1995. [1] O. Rockinger, Image sequence fusion using a shift-invariant wavelet transform, Proceedings of IEEE International Conference on Image Processing, vol. 3, pp. 88 91, 1997. [11] R. S. Blum, and Z. Liu, Multi-sensor image fusion and its applications, CRC Press, Taylor & Francis Group, 6. [1] A. Goshtasby, and S. Nikolov, Image fusion: Advances in the state of the art, Information Fusion, vol. 8, no., pp. 114 118, April 7. [13] J. Tang, A contrast based image fusion technique in the DCT domain, Digital Signal Processing, vol. 14,no. 3, pp. 18 6, 4. [14] G. K. Wallace, The JPEG Still Picture Compression Standard, Communications of the ACM, vol. 34, no. 4, pp. 3 44, April 1991. [15] I. E. G. Richardson, Video Codec Design, John Wiley & Sons Ltd,. [16] Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, Image quality assessment: From error visibility to structural similarity, IEEE Transactions on Image Processing, vol. 13, no. 4, pp. 6 61, April 4. [17] http://www.metapix.de/toolbox.htm