3634 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 20, NO. 12, DECEMBER 2011

Size: px
Start display at page:

Download "3634 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 20, NO. 12, DECEMBER 2011"

Transcription

1 3634 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 20, NO. 12, DECEMBER 2011 Generalized Random Walks for Fusion of Multi-Exposure Images Rui Shen, Student Member, IEEE, Irene Cheng, Senior Member, IEEE, Jianbo Shi, and Anup Basu, Senior Member, IEEE Abstract A single captured image of a real-world scene is usually insufficient to reveal all the details due to under- or over-exposed regions. To solve this problem, images of the same scene can be first captured under different exposure settings and then combined into a single image using image fusion techniques. In this paper, we propose a novel probabilistic model-based fusion technique for multi-exposure images. Unlike previous multi-exposure fusion methods, our method aims to achieve an optimal balance between two quality measures, i.e., local contrast and color consistency, while combining the scene details revealed under different exposures. A generalized random walks framework is proposed to calculate a globally optimal solution subject to the two quality measures by formulating the fusion problem as probability estimation. Experiments demonstrate that our algorithm generates high-quality images at low computational cost. Comparisons with a number of other techniques show that our method generates better results in most cases. Index Terms Image enhancement, image fusion, multi-exposure fusion, random walks. I. INTRODUCTION ANATURAL scene often has a high dynamic range (HDR) that exceeds the capture range of common digital cameras. Therefore, a single digital photo is often insufficient to provide all the details in a scene due to under- or over-exposed regions. On the other hand, given an HDR image, current displays are only capable of handling a very limited dynamic range. In the last decade, researchers have explored various directions to resolve the contradiction between the HDR nature of real-world scenes and the low dynamic range (LDR) limitation of current image acquisition and display technologies. Although cameras with spatially varying pixel exposures [1], cameras that automatically adjust exposure for different parts of a scene [2], [3], Manuscript received June 27, 2010; revised November 07, 2010 and January 17, 2011; accepted April 10, Date of publication May 05, 2011; date of current version November 18, This work was supported in part by Killam Trusts, icore, and Alberta Advanced Education and Technology. The associate editor coordinating the review of this manuscript and approving it for publication was Dr. Jesus Malo. R. Shen, I. Cheng, and A. Basu are with the Department of Computing Science, University of Alberta, Edmonton, AB T6G 2E8, Canada ( rshen@ualberta.ca; locheng@ualberta.ca; basu@ualberta.ca). J. Shi is with the Department of Computer and Information Science, University of Pennsylvania, Philadelphia, PA USA ( jshi@cis.upenn. edu). Color versions of one or more of the figures in this paper are available online at Digital Object Identifier /TIP and displays that directly display HDR images [4] have been developed by previous researchers, their technologies are only at a prototyping stage and unavailable to ordinary users. Instead of employing specialized image sensors, an HDR image can be reconstructed digitally using HDR reconstruction (HDR-R) techniques from a set of images of the same scene taken by a conventional LDR camera [5] or a panoramic camera [6] under different exposure settings. These HDR images usually have higher fidelity than LDR images, which benefits many applications, such as physically-based rendering and remote sensing [7]. As for viewing on ordinary displays, an HDR image is compressed into an LDR image using tone mapping (TM) methods [8], [9]. This two-phase workflow, HDR-R+TM (HDR-R and TM), has several advantages: no specialized hardware is required; various operations can be performed on the HDR images, such as virtual exposure; and user interactions are allowed in the TM phase to generate a tone-mapped image with desired appearance. However, this workflow is usually not as efficient as image fusion (IF, e.g., [10], [11]), which directly combines the captured multi-exposure images into a single LDR image without involving HDR-R, as shown in Fig. 1. Another advantage of IF is that IF does not need the calibration of the camera response function (CRF), which is required in HDR-R if the CRF is not linear. IF is preferred for quickly generating a well-exposed image from an input set of multi-exposure images, especially when the number of input images is small and speed is crucial. Since its introduction in the 1980s, IF has been employed in various applications, such as multi-sensor fusion [12], [13] (combining information from multi-modality sensors), multifocus fusion [14], [15] (extending depth-of-field from multifocus images), and multi-exposure fusion [11], [16] (merging details of the same scene revealed under different exposures). Although some general fusion approaches [17], [18] have been proposed by previous researchers, they are not optimized for individual applications and have only been applied to gray-level images. In this paper, we only focus on multi-exposure fusion and propose a novel fusion algorithm that is both efficient and effective. This direct fusion of multi-exposure images removes the need for generating an intermediate HDR image. A fused image contains all the information present in different images and is ready for viewing on conventional displays. Unlike previous multi-exposure fusion methods [10], [11], our algorithm is based on a probabilistic model and global optimization taking neighborhood information into account. A generalized random walks (GRW) framework is proposed to calculate an optimal set of probabilities subject to two quality measures: local contrast /$ IEEE

2 SHEN et al.: GENERALIZED RANDOM WALKS FOR FUSION OF MULTI-EXPOSURE IMAGES 3635 Fig. 1. Comparison between multi-exposure fusion and the HDR reconstruction and tone mapping workflow. (The Garage image sequence courtesy of Shree Nayar.) and color consistency. The local contrast measure preserves details; the color consistency measure, which is not considered by previous methods, includes both consistency in a neighborhood and consistency with the natural scene. Although used for multiexposure fusion in this paper, this proposed GRW provides a general framework for solving problems that can be formulated as a labeling problem [19], i.e., estimating the probability of a site (e.g., a pixel) being assigned a label based on known information. The proposed fusion algorithm has low computational complexity and produces a final fused LDR image with fine details and an optimal balance between the two quality measures. By defining problem-specific quality measures, the proposed algorithm may also be applied to other fusion problems. The rest of the paper is organized as follows. Section II reviews previous methods. Differences between our algorithm, previous IF methods, and methods used in the HDR-R+TM workflow, are also discussed. Section III explains our algorithm in detail. Section IV discusses experimental results and performance, along with comparisons with other IF methods and the HDR-R+TM workflow. Finally, Section V gives the conclusions and future work. II. RELATED WORK A. Multi-Exposure Image Fusion Image fusion methods combine information in different images into a single composite image. Here we only discuss those IF methods that are most relevant to our algorithm. Please refer to [20] for an excellent survey on IF methods in different applications. For multi-exposure images, IF methods directly work on the input LDR images and focus on enhancing dynamic range while preserving details. Goshtasby [10] partitions the input images into uniform blocks and tries to maximize the information in each block based on an entropy measure. However, the approach may generate artifacts on object boundaries if the block is not sufficiently small. Instead of dividing images into blocks, Cheng and Basu [21] combine images in a column-bycolumn fashion. This algorithm maximizes the contrast within a column by selecting pixels from different images. However, since no neighborhood information is considered, it cannot preserve color consistency and artifacts may occur on object boundaries. Cho and Hong [22] focus on substituting the under- or over-exposed regions in one image, which are determined by a saturation mask, with the well-exposed regions in another image. Region boundaries are blended by minimizing an energy function defined in the gradient domain. Although this algorithm works better on object boundaries, it is only applicable to two images. Our algorithm works at the pixel level and applies a global optimization taking neighborhood information into account, which avoids the boundary artifacts. Image fusion can also be interpreted as an analogy to alpha blending. Raman and Chaudhuri [23] generate the fused image by solving an unconstrained optimization problem. The weight function for each pixel is modeled based on local contrast/variance in a way that the fused image tends to have uniform illumination or contrast. Raman and Chaudhuri [24] generate mattes for each pixel in an image using the difference between the original pixel value and the pixel value after bilateral filtering. This measure strengthens weak edges, but may not be sufficient to enhance the overall contrast. Our algorithm defines two quality measures and finds the optimal balance between them, i.e., enhancing local contrast while imposing color consistency. Multi-resolution analysis based fusion has demonstrated good performance in enhancing main image features. Bogoni and Hansen [16] propose a pyramid-based pattern-selective fusion technique. Laplacian and Gaussian pyramids are constructed for the luminance and chrominance components, respectively. However, the color scheme of the fused image may be very close to the average image because pixels with saturation closest to the average saturation are selected for blending. Mertens et al. [11] construct Laplacian pyramids for the input images and Gaussian pyramids for the weight maps. A weight for a pixel is determined by three quality measures: local contrast, saturation, and well-exposedness. The Laplacian and Gaussian pyramids are blended at each level, and then collapsed to form the final image. However, with only local calculation involved and no measure to preserve color consistency, the fusion results may be unnatural. Our algorithm also uses local contrast as one quality measure, but another quality measure that we consider is color consistency, which is not employed by [11]. Moreover, our algorithm does not use pyramid decomposition but solves a linear system defined at the pixel level. B. HDR Reconstruction and Tone Mapping Although the HDR-R+TM workflow is usually used in different scenarios than IF methods, we still give a brief discussion on those HDR-R and TM methods sharing some similar features to our IF algorithm, because the original input and the final output of this workflow is the same as IF. Given an input LDR sequence and exposure times associated with each image in the sequence, HDR-R techniques [5] first recover the CRF, which is a mapping from the exposure at each pixel location to

3 3636 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 20, NO. 12, DECEMBER 2011 Fig. 2. Processing procedure of the proposed fusion algorithm. (The window image sequence courtesy of Shree Nayar.) the pixel s digital LDR value and then use the CRF to reconstruct an HDR image via a weighting function. Debevec and Malik [5] recover the CRF by minimizing a quadratic objective function defined on exposures, pixels LDR values, and exposure times. Then, a hat-shaped weighting function is used to reconstruct the HDR image. Granados et al. [25] assume the CRF is linear and focus on the development of an optimal weighting function using a compound-gaussian noise model. Our IF algorithm also solves a quadratic objective function, but the function is defined on local features and the solution leads to probability maps. Given an HDR image, tone mapping methods [26] aim to reduce its dynamic range while preserving details. TM usually works solely in the luminance channel. Global TM methods [27], [28], which apply spatially invariant mapping of luminance values, usually have high speed, while local TM methods [8], [9], [29], [30], which apply spatially varying mapping, usually produce images with better qualities especially when strong local contrast is present [9], [30]. Reinhard et al. [9] use a multiscale local contrast measure to compress the luminance values. Li et al. [29] decompose the luminance channel of the input HDR image into multiscale subbands and apply local gain control to the subbands. Shan et al. [30] define a linear system for each overlapping window in the HDR image using local contrast, and the solution of each linear system are two coefficients that map luminance values from HDR to LDR. Although our IF algorithm also uses local contrast to define a linear system, the local contrast in our algorithm is calculated in a different manner and another quality measure (i.e., color consistency) is also considered. Furthermore, our IF algorithm defines a linear system on pixels from all the original LDR images, and the solution is a set of probabilities that determine the contributions from each pixel of each original LDR image to its corresponding pixel in the fused image. Krawczyk et al. [8] segment an HDR image into frameworks with consistent luminance and compute the belongingness of each pixel to each framework using the framework s centroid, which results in a set of probability maps. Our IF algorithm also generates probability maps, but directly from the original LDR sequence with no HDR or segmentation involved. One typical problem with some local TM methods is halo artifacts introduced due to contrast reversals [26]. Our IF algorithm balances between contrast and consistency in a neighborhood, which can prevent contrast reversals. Another problem with some TM methods is that color artifacts like oversaturation may be introduced into the results, because operations are usually performed in the luminance channel without involving chrominance [26]. This problem does not apply to our IF algorithm, because every color channel is treated equally. III. PROBABILISTIC FUSION Unlike most previous multi-exposure fusion methods, we consider image fusion as a probabilistic composition process, as illustrated in Fig The initial probability that a pixel in the fused image belongs to each input image is estimated based on local features. Taking neighborhood information into account, the final probabilities are obtained by global optimization using the proposed generalized random walks. These probability maps serve as weights in the linear fusion process to produce the fused image. In a probability map, the brighter a pixel is, the higher the probability. These processes are explained in detail below. A. Problem Formulation The fusion of a set of multi-exposure images can be formulated as a probabilistic composition process. Let denote the set of input images and the set of labels, where a label is associated with the th input image. s are assumed to be already registered and have the same size with pixels each. Normally,. Let us define a set of variables such that an is associated with the th pixel in the fused image and takes a value from. Then, each pixel in the fused image can be represented as where denotes the th pixel in and the probability of pixel being assigned label given with. This probabilistic formulation converts the fusion problem to the calculation of marginal probabilities given the input images subject to some quality measures and helps to achieve an optimal balance between different quality measures. If every pixel is given equal probability, i.e.,, is simply the average of s. Although it is also possible to only combine pixels with highest 1 The initial probability maps are normalized at each pixel. (1)

4 SHEN et al.: GENERALIZED RANDOM WALKS FOR FUSION OF MULTI-EXPOSURE IMAGES 3637 the weights between and, we can define a node compatibility function on with the following form: Fig. 3. Graph used in GRW. The yellow nodes are scene nodes and the orange nodes are label nodes.. (2) Because the graph is undirected, we have. Let denote the potential associated with. Based on the relationship between RW and electrical networks [33], the total energy of the system given in Fig. 3 is (3) probabilities instead of using (1), artifacts may appear at locations with significant brightness changes between different input images. If s are simply viewed and calculated as local weights followed by applying some relatively simple smoothing filters, such as Gaussian filter and bilateral filter, either artifacts (like halos) may appear at object boundaries or it is difficult to determine the termination criteria of the filtering; therefore, the results are usually unsatisfactory [11]. An alternative is to use multi-resolution fusion techniques [11], [16], where the weights are blended in each level to produce more satisfactory results. However, the weights in each level are still determined locally, which may not be optimal in a large neighborhood. In contrast to multi-resolution techniques, we propose an efficient single-resolution fusion technique by formulating s in (1) as probabilities in GRW. A set of optimal s that balances the influence of different quality measures is computed from global optimization in GRW. Experiments (Section IV) show that the results of the proposed probabilistic fusion technique are comparable to the results of multi-resolution techniques and the results of the HDR-R+TM workflow. B. Generalized Random Walks In this section, we propose a generalized random walks (GRW) framework based on the random walks (RW) algorithm [31], [32] and the relationship between RW and electrical networks [33]. 1) Image Representation: As shown in Fig. 3, the variable set and the label set are represented in a weighted undirected graph similar to [31]. Each variable is associated with a pixel location, and each label is associated with an input image in our case. The graph is constructed as and including edges both within and between and. The yellow nodes are scene nodes and the orange nodes are label nodes. For a scene node, edges are drawn between it and each of its immediate neighbors in (4-connectivity is assumed in our case). In addition, edges are drawn between a scene node and each label node. is a function defined on that models the compatibility/similarity between nodes and, and is a function defined on that models the compatibility between and. 2) Dirichlet Problem: Let be arranged in a way that the first nodes are label nodes, i.e.,, and the rest nodes are scene nodes, i.e.,. With two positive coefficients and introduced to balance Our goal is to find a function defined on that minimizes this quadratic energy with boundary values defined on.if satisfies, then it is called harmonic, and the harmonic function is guaranteed to minimize such quadratic energy [32]. The problem of finding this harmonic function is called the Dirichlet problem. The harmonic function can be computed efficiently using matrix operations. Similar to [32], a Laplacian matrix can be constructed following (4); however, unlike [32], here contains both the label nodes and the scene nodes and becomes a matrix: otherwise. where is the degree of the node defined on its immediate neighborhood. Then, (3) can be rewritten in matrix form as where and ; is the upper left submatrix of that encodes the interactions within ; is the lower right submatrix that encodes the interactions within ; and is the upper right submatrix that encodes the interactions between and. Hence, the minimum energy solution can be obtained by setting with respect to, i.e., solving the following equation: In some cases, part of may be already labeled. These prelabeled nodes can also be represented naturally in the current framework without altering the structure of the graph. Suppose is one of the pre-labeled nodes and is assigned label. Then, we can simply assign a sufficiently large value to and solve the same (6) for the unlabeled scene nodes. 3) Probability Calculation: The probability that a scene node is assigned the th label given all the observed data can be considered as the probability that a random walker starting at a scene node first reaches the label node on the graph. Thus, can be computed for each (4) (5) (6)

5 3638 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 20, NO. 12, DECEMBER 2011 pair of by solving Dirichlet problems in iterations. Note that the probabilities here are used in the context of random walks [31] [33], which is different from the log-probabilities used in Markov random field energy minimization [19]. can be considered as the initial probability that the scene node is assigned label given data associated with and data associated with, i.e., the probability that a random walker transits from to in a single move: can be considered as the probability that the scene nodes and are assigned the same label given and, i.e., the probability that a random walker transits from to in a single move: When constructing s or s, it is assumed that the probability that takes a different label from or is zero. This assumption ensures the smoothness of the labeling. s and s are initialized at the beginning of the algorithm and remain the same in each iteration. Let be the potential associated with node in the th iteration, which we define to be proportional to : where is a positive constant. Since,. is a binary function on :, when ;, otherwise. For any,. Once s and s are defined, the probabilities s can then be determined from (6) and (9). The RW algorithm [32] requires some variables to be instantiated, i.e., some scene nodes to be pre-labeled. This requirement is relaxed in GRW. The RW algorithm with prior models (RWPM) [31] is proposed for image segmentation and derives a similar linear system to (6) from a Bayesian point of view. A small inaccuracy in [31] is that a weighting parameter is missing from the right-hand side in their (11). When setting in GRW, we can get a linear system in the same format as derived in [31] with the missing parameter added. The nodewise priors in [31], which correspond to our compatibility function, are required to be defined following a probability expression. Although it is mentioned in [31] that the solution of RWPM is equivalent to that of the original RW [32] on an augmented graph considering the label nodes as the extra pre-labeled nodes, the requirement on the format of the nodewise priors limits the choice of. This requirement is relaxed in GRW, where we formulate the problem from the original RW point of view [33], where probabilities are considered as the transition probabilities of a random walker moving between nodes. The edge weighting function in [31], which corresponds to our compatibility function, serves as a regularization term. In GRW, and are not probability quantities; instead, they represent compatibility/similarity and are used to define the transition probabilities. In GRW, we have relaxed the extra requirements in [31], [32] and provided a more (7) (8) (9) flexible framework, where the compatibility functions (and potential function) may be defined in any form according to the need of a particular problem. For the fusion problem, the form of the compatibility functions is presented in Section III-C. Although in this paper GRW is proposed to solve the multi-exposure fusion problem, it can actually be applied to many different problems that can be formulated as estimating the probability of a site (e.g., a pixel) being assigned a label given known information. C. Compatibility Functions The compatibility functions and are defined to represent respectively the two quality measures used in the proposed fusion algorithm, i.e., local contrast and color consistency. Since image contrast is usually related to variations in image luminance [26], the local contrast measure should be biased towards pixels from the images that provide more local variations in luminance. Let denote the second-order partial derivative computed in the luminance channel at the th pixel in image, which is a indicator of local contrast. The higher the magnitude of (denoted by ) is, the more variations occur near the pixel, which may indicate more local contrast. In [11], a Laplacian filter is used to calculate the variations. Here both Laplacian filter and central difference in a 3 3 neighborhood were tested. With all other settings the same, central difference produces slightly better visual quality in the fused images. Therefore, central difference is used to approximate the second-order derivative. However, if the frequency (i.e., number of occurrences) of a value in is very low, the associated pixels may be noise or belong to unimportant features. Therefore, unlike previous methods [11], [23], our contrast measure is normalized by the frequencies of each. In addition, s are modified using a sigmoid-shaped function to suppress the difference in high contrast regions. Because of the nonlinear human perception of contrast [34], such a mapping scheme helps us avoid overemphasis on high local variations. Hence, taking into account both the magnitude and the frequency of the contrast indicator, the compatibility between a pixel and a label is computed as (10) where represents the frequency of the value in ; is the Guassian error function, which is monotonically increasing and sigmoid shaped; the exponent is equal to the number of input images and controls the shape of by giving less emphasis on the difference in high contrast regions as the number of input images increases; and is a weighting coefficient, which we take as the variance of all s. The other quality measure used in our algorithm is color consistency, which is not considered in previous methods [10], [11]. Although Bogoni and Hansen [16] also suggested a color consistency criterion by assuming that the hue component for all the input images is constant, this assumption is not true if the images are not taken with close exposure times. In addition, they

6 SHEN et al.: GENERALIZED RANDOM WALKS FOR FUSION OF MULTI-EXPOSURE IMAGES 3639 Fig. 4. Comparison of our method with EF, EntropyF, and VC using the Chairs image sequence. The result of EF is comparable to ours. The result of EntropyF suffers a little over-exposure. The result of VC shows serious halo artifacts on object boundaries. (a) Input sequence (top) and final probability maps (bottom). (b) Proposed. (c) EF [11]. (d) EntropyF [10]. (e) VC [23]. (Input sequence courtesy of Shree Nayar.) do not consider consistency in a neighborhood. Our color consistency measure imposes not only color consistency in a neighborhood but also consistency with the natural scene. If two adjacent pixels in most images have similar colors, then it is very likely that they have similar colors in the fused image. Also, if the colors at the same pixel location in different images under appropriate exposures (not under- or over-exposed) are similar, they indicate the true color of the scene and the fused color should not deviate from these colors. Therefore, the following equation is used to evaluate the similarity/compatibility between adjacent pixels in the input image set using all three channels in the RGB color space: (11) where and are adjacent pixels in image, is the exponential function; denotes Euclidean distance, denotes the average pixel, and and are free parameters. Although the two quality measures are defined locally, a global optimization using GRW is carried out to produce a fused image that maximizes contrast and details, as well as imposing color consistency. Once and are defined using (10) and (11), the probabilities s are calculated using (2) (6). Here, we fix and only use the ratio to determine the relative weight between and. The basic steps of our algorithm are summarized in Algorithm 1. Algorithm 1 Basic steps of the proposed fusion algorithm 1: Construct function from (10) 2: Construct function from (11) 3: Construct function from (2) 4: Construct from (4) 5: for to do 6: Calculate for all by solving (6) 7: end for 8: Compute the fused image from (1) D. Acceleration To accelerate the computation as well as reduce memory usage, the final probability maps are computed at a lower resolution of the Laplacian matrix and then interpolated back to the original resolution before being used in (1). The contrast indicator of each pixel is calculated at the original resolution. Then, the images are divided into uniform blocks of size. The average of s in a block is used to calculate the compatibility between that block and the label. The compatibility between two adjacent blocks is computed as of the average pixel in one block and the average pixel in the other block. IV. EXPERIMENTAL RESULTS AND DISCUSSION A. Comparison with Other Image Fusion Methods and Some Tone Mapping Methods Only three free parameters are used in our algorithm, and we take,, in all experiments unless otherwise mentioned. 12 LDR image sequences were used in the experiments. Fig. 4 shows the comparison of our IF algorithm with three other IF methods on the Chairs image sequence. The input image sequence and final probability maps are given in Fig. 4(a). Brighter pixels in a probability map stand for higher probabilities. The result of entropy fusion (EntropyF) [10] is taken from its project webpage. 2 The result of variational composition (VC) [23] is taken from its paper. The results of exposure fusion (EF) [11] in all experiments are generated by the Matlab implementation provided by its authors. Its default parameter setting is used in all experiments. The result of EF is comparable to ours. The result of EntropyF suffers a little over-exposure. The result of VC shows serious halo artifacts on object boundaries. Figs. 5 and 6 give comparison of our IF algorithm with EF and three TM methods on two image sequences. In all experiments, the intermediate HDR images for photographic tone reproduction (PTR) [9], subband compression (SC) [29] and linear windowed (LW) [30] are generated from the corresponding LDR sequences using HDRShop, 3 which employs the HDR-R algorithm in [5]. The results of PTR are generated by an HDRShop plugin. 4 The results of SC and LW are generated by the Matlab implementations provided by their respective authors. There is no constant set of parameters in their original papers, because TM methods usually depend on user-controlled parameters to generate desired tone-mapped images. However, in order to give a relatively fair image quality comparison with our and other IF methods where constant sets of parameters are used throughout the experiments, we use the default parameter settings in their programs in all experiments. Fig. 5 shows the experimental results on the Belgium House sequence. The last row gives a closeup view of the window regions. Although the result of SC preserves as much detail as ours, it looks a little pink due to color distortion. The results of EF and PTR suffer over-exposure for all the window regions. The result of LW shows some color artifacts, e.g., color reversal

7 3640 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 20, NO. 12, DECEMBER 2011 Fig. 5. Comparison of our algorithm with EF, PTR, SC, and LW using the Belgium House image sequence. The intermediate HDR image for PTR, SC, and LW is generated by HDRShop. The results of EF and PTR suffer over-exposure in the window regions. The results of SC and LW show some color distortion (looking pink). (a) Input image sequence (top) and final probability maps (bottom). (b) Proposed. (c) EF [11]. (d) PTR [9]. (e) SC [29]. (f) LW [30]. (Input sequence courtesy of Dani Lischinski.) Fig. 6. Comparison of our algorithm with EF, PTR, SC and LW using the House image sequence. The intermediate HDR image for PTR, SC, and LW is generated by HDRShop. EF introduces color artifacts that assign two chairs of the same type different colors. The result of PTR is a little dark for the indoor scene. The result of SC looks a little pink and dark for the entire image. The result of LW shows some color distortion. Our result reveals more detail than EF, especially in the door lock region. (a) Input image sequence (left four) and final probability maps (right four). (b) Proposed. (c) EF [11]. (d) PTR [9]. (e) SC [29]. (f) LW [30]. (Input sequence courtesy of Tom Mertens.) of the blackboard. Color artifacts in the results of TM methods are usually caused by operations carried out solely in the luminance channel without involving chrominance [26]. Our method treats each color channel equally and imposes color consistency, which helps to avoid color artifacts. Note that adjusting the parameters in SC and LW may reduce color distortion and generate more pleasing images, which is considered as a common user interaction in TM methods. However, we use constant sets of parameters for both TM and IF methods in our experiments in order to give a relatively fair comparison of these two types of methods. Fig. 6 shows the results on the House sequence. The last row gives a closeup view of the door lock region. The result of PTR is a little dark for the indoor scene. The result of SC reveals many details of the outdoor scene but looks a little pink and dark for the entire image. The result of LW also reveals many details but shows some color distortion of the outdoor scene. Although both our method and EF use local variations to indicate local contrasts, we use a nonlinear function to modify the contrast indicators. This nonlinear mapping is to reflect the nonlinear human perception of contrast [34]. In addition, the well-exposedness measure in EF is biased towards pixels with a specific luminance. Therefore, our result reveals more detail than EF, especially in the door lock region. The two chairs in the original sequence have the same color, but for EF the color difference between them is quite obvious. Our method keeps the colors consistent while EF fails, because our method imposes consistency in large neighborhoods and consistency with the natural scene via the color consistency measure. Our algorithm works well for various scenes and combinations of exposure settings. More results on different image sequences are given in Fig. 7. B. Computational Complexity In the initialization step, the input image sequence is converted to gray scale for calculating s, and an average image

8 SHEN et al.: GENERALIZED RANDOM WALKS FOR FUSION OF MULTI-EXPOSURE IMAGES 3641 Fig. 7. Results on different image sequences. Our algorithm works well for various scenes and combinations of exposure settings. Left most: National Cathedral; left: Chateau; right: Lizard; right most: Room. (Input sequences courtesy of Max Lyons, HDRsoft.com, Eric Reinhard, and Grzegorz Krawczyk, respectively.) TABLE I COMPUTATIONAL TIMES OF THE PROPOSED ALGORITHM AND EF ON THE TEST IMAGE SEQUENCES. TIMES ARE RECORDED IN SECONDS is also computed for calculating s. The complexity of this step is, where is the number of pixels in an input image and is the number of images in the input sequence. Since the number of operations is proportional to the total number of input pixels, the complexity for computing the compatibilities is also. We employ CHOLMOD 5 [35], which is based on supernodal Cholesky factorization, to solve the linear system in (6). The complexity of this direct solver is proportional to the number of the nonzero entries in, and in our case it is. Since there are altogether linear systems to be solved, the complexity of this step is. With only linear operations involved, the complexity of the fusion step is. Therefore, the total complexity of our algorithm is. Our algorithm is currently implemented in Matlab. The computational times for the 12 image sequences are reported in Table I, along with the comparison with EF [11]. Both our and EF s Matlab implementations were executed on the same computer with a 2.53-GHz CPU and 2-GB memory available for Matlab. Times for reading the input sequences and writing the output images are excluded. The times of EF on the National Cathedral and the Lizard sequences are not given, because its Matlab implementation requires more than 2-GB memory for the computation of those sequences. Our algorithm takes only 25% of the total time of EF on the average. C. Analysis of Free Parameters The effectiveness of acceleration is illustrated in Fig. 8(a). All the 12 image sequences in Table I were used in this and following analyses. For illustration, we plot three representative image sequences at different scales (i.e., Igloo, Memorial Church, and Belgium House) in the graph. We fix, 5 CHOLMOD has been included in Matlab since version 7.2. Fig. 8. Analysis of acceleration with different block width. Error is defined as the relative difference from the results generated with =1. Significant efficiency improvement is only observed when 5 and significant error increase only occurs when 5. (a) Effectiveness of acceleration. (b) Error introduced. in this analysis. The horizontal axis represents the block width and the vertical axis represents the computational time in seconds. Time used in initialization is excluded because the

9 3642 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 20, NO. 12, DECEMBER 2011 Fig. 9. Errors introduced by acceleration on the Belgium House image sequence. The result generated with = 1is used as the reference image. The error increases when increases, as shown in the color-coded error maps (warmer colors indicate larger errors). However, the total error is still below 9% even when =10. (a) =1. (b) =2. (c) =4. (d) =5. (e) =6. (f) =10. Fig. 10. Sensitivity analysis of the free parameter. The error increases dramatically as becomes too small. It increases slowly when 0:4 and converges after =1:0. Fig. 12. Sensitivity analysis of the free parameter. The error increases dramatically as becomes too small, but increases slowly when 5:0. Fig. 11. Sensitivity analysis of the free parameter on the Igloo image sequence. The result generated with = 0:1 is used as the reference image. When decreases, the image gets brighter, and therefore more information is revealed in the under-exposed regions. However, if becomes too small, adjacent pixels with large difference may be ignored, which leads to artifacts at object boundaries. (a) = 0:1. (b) = 0:01. (c) = 0:05. (d) =0:5. (e) =2:0. (Input sequence courtesy of Shree Nayar.) acceleration does not affect this step. Computational time decreases as increases. However, significant efficiency improvement is only observed when. Errors introduced by acceleration are shown against block width in Fig. 8(b). The error in a pixel is calculated using Euclidean distance between a pixel in a resulting image with and the corresponding pixel in the reference image with, i.e.,. The pixel values are normalized between [0,1]. The total error in a resulting image is measured using the root mean squared error (RMSE), i.e.,. The error increases as the block size increases. However, significant error increase only occurs when Fig. 13. Sensitivity analysis of the free parameter on the Memorial Church image sequence. The result generated with = 1:0 is used as the reference image. When decreases, more detail is revealed in the over-exposed regions. However, if is too small, artifacts may occur at object boundaries. (a) =1:0. (b) =0:01. (c) =0:2. (d) =5:0. (e) =10:0. (Input sequence courtesy of Paul Debevec.). Even when, the total error is still below 9%. Some results on the Belgium House image sequence with different s and their corresponding color-coded error maps are shown in Fig. 9. In an error map, warmer colors indicate larger errors. The error increases as increases. In order to balance the speed and error, we tested different values and found that using a block width generates reasonably good results. In the analysis of,wefix,. The error is defined similarly as in the analysis of and the results with are used as reference images. The analysis on three representative image sequences is plotted in Fig. 10. Some results on the Igloo sequence with different s and their corresponding color-coded error maps are shown in Fig. 11. The error increases dramatically as becomes too small. It increases

10 SHEN et al.: GENERALIZED RANDOM WALKS FOR FUSION OF MULTI-EXPOSURE IMAGES 3643 Fig. 14. Comparison of our algorithm with EF [11] using the Lamp image sequence. The results of EF are generated on a smaller scale (1/4) of the input sequence. Although the result of EF looks more colorful in some regions than our result, it preserves less detail. Especially for the lamp, the bulb and the tag beside it are clearly visible in our result but washed out in the result of EF. Although histogram equalization causes our result to lose some information in the lamp region, our result still preserves more detail than the result of EF, e.g., in the paper region. (a) Proposed before histogram equalization. (b) EF before histogram equalization. (c) Proposed after histogram equalization. (d) EF after histogram equalization. (Input sequence courtesy of Martin Cadík.) Fig. 15. Analysis of our algorithm s sensitivity to Gaussian white noise and comparison with other algorithms using the House image sequence. A close-up view of regions near the window is placed beside each image to give a clearer view of the effect of noise. As the Gaussian noise variance increases, errors introduced by the noise become more obvious. Our algorithm, EF, and PTR are all affected by the noise in the input image. (a) Original image (left) and its corresponding color-coded probability map (right) calculated by the proposed method (warmer colors represent higher probabilities). (b) Gaussian noise-corrupted image with 0 mean and 0.01 variance (left) and its corresponding color-coded probability map (right) calculated by the proposed method. (c) Fused image generated by the proposed method. (d) Fused image generated by EF [11]. (e) Tone-mapped image generated by PTR [9]. slowly when and converges after. When decreases, the image gets brighter, and therefore more information is revealed in the under-exposed regions. However, if becomes too small, adjacent pixels with large difference may be ignored, which leads to artifacts at object boundaries. Therefore, we suggest using. In the analysis of,wefix,. The error is defined similarly as in the analysis of and the results with are used as reference images. The analysis on three representative image sequences is plotted in Fig. 12. Some results on the Memorial Church sequence with different s and their corresponding color-coded error maps are shown in Fig. 13. The error increases dramatically as becomes too small, but increases slowly when. When decreases, more detail is revealed in the over-exposed regions. However, if is too small, like in the analysis of, artifacts may occur at object boundaries. Therefore, we suggest using. D. Post-Processing for Further Enhancement Although our algorithm preserves more detail than EF, the results of EF for some image sequences may have higher contrast (sharper) in certain regions. An example is given in Fig. 14(a) and (b). The tag near the bulb is clearly visible in our result but washed out in the result of EF. The region near the red book looks more colorful in the result of EF. Higher contrast can also be obtained from our method by applying histogram equalization as a post-processing step. In order to provide a fair comparison, we added histogram equalization to both methods. The images in Fig. 14(c) and (d) illustrate that the contrast in our result is improved while EF suffers a loss of detail during the process. Although histogram equalization causes our result to lose some information in the lamp region, our result still preserves more detail than the result of EF, e.g., in the paper region. E. Effect of Noise One limitation of our algorithm is that it is sensitive to noise in the input image sequence. One example is given in Fig. 15. White Gaussian noise with 0 mean and variance from to 0.01 with increments of is added to one of the four input images (pixel values are scaled to the range [0,1]). For brevity, only the corrupted image with variance 0.01 is shown in Fig. 15, along with the original image. In our algorithm, the initial probabilities are determined by local contrast, and this local mea-

11 3644 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 20, NO. 12, DECEMBER 2011 Fig. 16. Fusion result using our algorithm after adding a denoising step. The probability map calculated for this denoised image and the fused image are close to the ones computed from the clean input images. However, the object boundaries are a little blurred during the denoising process, which also affects the fusion result. (a) Denoised image. (b) Color-coded probability map. (c) Fused image. sure is sensitive to white Gaussian noise because it is calculated by local variance. Since a global optimization scheme is used afterwards, the error caused by the noise tends to propagate to a larger neighborhood. The color-coded probability maps are given in Fig. 15(a) and (b), where warmer colors represent higher probabilities. Compared to the probability map generated for the original image, higher probabilities are assigned to pixels in the corrupted image, especially in the textureless regions like the walls. Even if we use the correctly computed probability maps, pixels from the noisy images still contribute to the fused image through the use of (1). Therefore, the noisy images have significant influence on the fused images, as shown in Fig. 15(c). A close-up view of regions near the window is placed beside each image to give a clearer view of the effect of noise. As the Gaussian noise variance increases, errors introduced by the noise become more obvious. However, the results by EF are also affected by the noise as shown in Fig. 15(d). In addition, the noise in an input image also affects the HDR reconstruction process, which results in a noisy HDR image. The tone mapping process fails to correct the noise in this HDR image and the noise remains in the tone-mapped image. PTR [9], SC [29], and LW [30] generate similarly affected results. Therefore, only the result of PTR is given in Fig. 15(e). One solution to this problem is to add a pre-processing step to reduce the noise in the input images. One example is given in Fig. 16. The input sequence is the House image sequence, where one image is corrupted with white Gaussian noise with 0 mean and 0.01 variance as shown in Fig. 15(b). Fig. 16(a) gives the denoised image using the method proposed by Portilla et al. [36]. The noisy image is first decomposed into subbands using the steerable pyramid transform [37], and then Bayesian least square estimation and Gaussian scale mixture model are employed to denoise each subband. The probability map calculated for this denoised image and the fused image are given in Fig. 16(b) and (c), respectively. They are close to the ones computed from the clean input images [see Figs. 15(a) and 6(b)]. However, the object boundaries are a little blurred during the denoising process, which also affects the fusion result. In the future, we will explore the possibility of incorporating a noise term into our fusion model to make our algorithm more robust to noise. F. Analysis of Different Forms of Compatibility Function for Contrast Measure The compatibility function used for contrast measure can take different forms. One alternative is to use the contrast transducer function proposed by Wilson [34], which is a sigmoid-shaped function amplifying difference at low contrasts and suppressing difference at high contrasts. Wilson s transducer function takes the following form for contrasts below or near the contrast threshold where : (12) where represents local contrast; is a parameter obtained by setting the contrast detection threshold to 0.75; is an empirically determined parameter; is called the contrast sensitivity and in our experiment we set, where represents the maximum magnitude of the local contrast detected from the input sequence. Because of this setting of, all contrasts are below or near the threshold. Therefore, (12) is used in our experiment instead of the unified transducer function in [34] that combines (12) and a function for high suprathreshold contrasts. In our experiment, the local contrast at the th pixel in image is calculated based on Weber contrast (13) where we take as the luminance difference between the pixel and the average pixel of its 3 3 neighborhood and as the average luminance of the input sequence. can also be taken as the local average luminance, but this may make biased towards under-exposed pixels. Fig. 17 gives a comparison between the results obtained by our current compatibility function [see (10)] and by Wilson s transducer function coupled with Weber contrast [see (12) and (13)] using the Memorial Church image sequence. With all other parameter settings the same, i.e.,,, and, Wilson s transducer function with Weber contrast preserves more detail in the window regions, but produces contrast reversals at some places near the windows. These contrast reversals are caused by combining pixels from input images with large exposure gaps. When is increased to give more emphasis on color consistency, the contrast reversals disappear although there is some loss of detail, as shown in Fig. 17(c). However, the brightness of the entire fused image also decreases. The current form of the compatibility function gives a better balance between local contrast and color consistency. We will analyze compatibility functions of other forms (e.g., [38]) and other quality measures (e.g., [39]) that may enhance our algorithm s performance in the future.

12 SHEN et al.: GENERALIZED RANDOM WALKS FOR FUSION OF MULTI-EXPOSURE IMAGES 3645 Fig. 17. Comparison between the results obtained from our current compatibility function [Equation (10)] and from Wilson s transducer function with Weber contrast [Equations (12) and (13)] using the Memorial Church image sequence. With all other parameter settings the same, Wilson s transducer function with Weber contrast preserves more detail in the window regions, but produces contrast reversals at some places near the windows. When is increased to 100 to give more emphasis on color consistency, the contrast reversals disappear although with some loss of detail. However, the brightness of the entire fused image also decreases. (a) Equation (10) with =0:1, =1:0, and =4. (b) Equations (12) and (13) with =0:1, =1:0, and =4. (c) Equations (12) and (13) with = 0:1, = 100, and = 4. V. CONCLUSION AND FUTURE WORK In this paper, we proposed a new fusion algorithm for multi-exposure images considering fusion as a probabilistic composition process. A generalized random walks framework was proposed to compute the probabilities. Two quality measures were considered: local contrast and color consistency. Unlike previous fusion methods, our algorithm achieves an optimal balance between the two measures via a global optimization. Experimental results demonstrated that our probabilistic fusion produces good results, in which contrast is enhanced and details are preserved with high computational efficiency. Compared to other fusion methods and tone mapping methods, our algorithm produces images with comparable or even better qualities. In future work, we will explore more effective quality measures and the possibility of incorporating multi-resolution technique in the fusion process to further enhance our technique for different fusion problems. We will also explore the possibility of applying the generalized random walks framework to other image processing problems. ACKNOWLEDGMENT The authors would like to thank L. Grady for providing the source code of the original random walks algorithm. The authors would also like to thank the reviewers for their valuable comments. REFERENCES [1] S. K. Nayar and T. Mitsunaga, High dynamic range imaging: Spatially varying pixel exposures, in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2000, vol. 1, pp [2] H. Mannami, R. Sagawa, Y. Mukaigawa, T. Echigo, and Y. Yagi, High dynamic range camera using reflective liquid crystal, in Proc. Int. Conf. Comput. Vis., 2007, pp [3] J. Tumblin, A. Agrawal, and R. Raskar, Why i want a gradient camera, in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2005, vol. 1, pp [4] H. Seetzen, W. Heidrich, W. Stuerzlinger, G. Ward, L. Whitehead, M. Trentacoste, A. Ghosh, and A. Vorozcovs, High dynamic range display systems, in Proc. ACM SIGGRAPH, 2004, pp [5] P. E. Debevec and J. Malik, Recovering high dynamic range radiance maps from photographs, in Proc. ACM SIGGRAPH, 1997, pp [6] M. Aggarwal and N. Ahuja, High dynamic range panoramic imaging, in Proc. Int. Conf. Comput. Vis., 2001, pp [7] E. Reinhard, G. Ward, S. Pattanaik, P. Debevec, W. Heidrich, and K. Myszkowski, High Dynamic Range Imaging: Acquisition, Display, and Image-Based Lighting, 2nd ed. Waltham, MA: Morgan Kaufmann, [8] G. Krawczyk, K. Myszkowski, and H.-P. Seidel, Lightness perception in tone reproduction for high dynamic range images, Comput. Graph. Forum, vol. 24, no. 3, pp , [9] E. Reinhard, M. Stark, P. Shirley, and J. Ferwerda, Photographic tone reproduction for digital images, in Proc. ACM SIGGRAPH, 2002, pp [10] A. Goshtasby, Fusion of multi-exposure images, Image Vision Comput., vol. 23, no. 6, pp , [11] T. Mertens, J. Kautz, and F. Van Reeth, Exposure fusion, in Proc. Pacific Graphics, 2007, pp [12] M. Kumar and S. Dass, A total variation-based algorithm for pixellevel image fusion, IEEE Trans. Image Process., vol. 18, no. 9, pp , Sep [13] S. Zheng, W.-Z. Shi, J. Liu, G.-X. Zhu, and J.-W. Tian, Multisource image fusion method using support value transform, IEEE Trans. Image Process., vol. 16, no. 7, pp , Jul [14] S. Li, J. T.-Y. Kwok, I. W.-H. Tsang, and Y. Wang, Fusing images with different focuses using support vector machines, IEEE Trans. Neural Netw., vol. 15, no. 6, pp , Nov [15] H. Zhao, Q. Li, and H. Feng, Multi-focus color image fusion in the HSI space using the sum-modified-laplacian and a coarse edge map, Image Vis. Comput., vol. 26, no. 9, pp , [16] L. Bogoni and M. Hansen, Pattern-selective color image fusion, Pattern Recognit., vol. 34, no. 8, pp , [17] G. Piella, Image fusion for enhanced visualization: A variational approach, Int. J. Comput. Vis., vol. 83, no. 1, pp. 1 11, [18] V. S. Petrovic and C. S. Xydeas, Gradient-based multiresolution image fusion, IEEE Trans. Image Process., vol. 13, no. 2, pp , Feb [19] S. Z. Li, Markov random field models in computer vision, in Proc. Eur. Conf. Comput. Vis., 1994, pp [20] M. I. Smith and J. P. Heather, A review of image fusion technology in 2005, in Proc. SPIE, 2005, vol. 5782, pp [21] I. Cheng and A. Basu, Contrast enhancement from multiple panoramic images, in Proc. ICCV Workshop OMNIVIS, 2007, pp [22] W.-H. Cho and K.-S. Hong, Extending dynamic range of two color images under different exposures, in Proc. Int. Conf. Pattern Recognit., 2004, vol. 4, pp [23] S. Raman and S. Chaudhuri, A matte-less, variational approach to automatic scene compositing, in Proc. Int. Conf. Comput. Vis., 2007, pp [24] S. Raman and S. Chaudhuri, Bilateral filter based compositing for variable exposure photography, in Proc. Eurographics Short Papers, [25] M. Granados, B. Ajdin, M. Wand, C. Theobalt, H.-P. Seidel, and H. P. A. Lensch, Optimal HDR reconstruction with linear digital cameras, in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2010, pp [26] M. Cadík, M. Wimmer, L. Neumann, and A. Artusi, Evaluation of HDR tone mapping methods using essential perceptual attributes, Comput. Graph., vol. 32, no. 3, pp , [27] G. W. Larson, H. Rushmeier, and C. Piatko, A visibility matching tone reproduction operator for high dynamic range scenes, IEEE Trans. Vis. Comput. Graph., vol. 3, no. 4, pp , [28] F. Drago, K. Myszkowski, T. Annen, and N. Chiba, Adaptive logarithmic mapping for displaying high contrast scenes, Comput. Graph. Forum, vol. 22, no. 3, pp , [29] Y. Li, L. Sharan, and E. H. Adelson, Compressing and companding high dynamic range images with subband architectures, in Proc. ACM SIGGRAPH, 2005, pp

13 3646 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 20, NO. 12, DECEMBER 2011 [30] Q. Shan, J. Jia, and M. S. Brown, Globally optimized linear windowed tone mapping, IEEE Trans. Vis. Comput. Graph., vol. 16, no. 4, pp , Jul.-Aug [31] L. Grady, Multilabel random walker image segmentation using prior models, in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2005, vol. 1, pp [32] L. Grady, Random walks for image segmentation, IEEE Trans. Pattern Anal. Mach. Intell., vol. 28, no. 11, pp , Nov [33] P. G. Doyle and J. L. Snell, Random Walks and Electric Networks. Washington DC: MAA, [34] H. R. Wilson, A transducer function for threshold and suprathreshold human vision, Biolog. Cybern., vol. 38, no. 3, pp , [35] Y. Chen, T. A. Davis, W. W. Hager, and S. Rajamanickam, Algorithm 887: CHOLMOD, supernodal sparse Cholesky factorization and update/downdate, ACM Trans. Math. Softw., vol. 35, no. 3, pp. 1 14, [36] J. Portilla, V. Strela, M. J. Wainwright, and E. P. Simoncelli, Image denoising using scale mixtures of gaussians in the wavelet domain, IEEE Trans. Image Process., vol. 12, no. 11, pp , Nov [37] E. P. Simoncelli and W. T. Freeman, The steerable pyramid: A flexible architecture for multi-scale derivative computation, in Proc. Int. Conf. Image Process., 1995, vol. 3, pp [38] M. A. García-Pérez and R. Alcalá-Quintana, The transducer model for contrast detection and discrimination: Formal relations, implications, and an empirical test, Spatial Vis., vol. 20, no. 1 2, pp. 5 43, [39] S. Winkler, Vision models and quality metrics for image processing applications, Ph.D. dissertation, EPFL, Lausanne, Switzerland, Rui Shen (S 07) received the B.Eng. degree in computer science and technology from Beihang University, Beijing, China, and the M.S. degree in computing science from the University of Alberta, Edmonton, AB, Canada. He is currently working toward the Ph.D. degree in computing science at the University of Alberta. He is the author or coauthor of eight papers published in international conferences and journals. His current research interests include probability models, sensor fusion, stereo vision, and image processing. Mr. Shen is the recipient of the icore Graduate Student Scholarship in the Information and Communications Technology and the Izaak Walton Killam Memorial Scholarship. Irene Cheng (M 02 SM 09) received the Ph.D. degree in computing science from the University of Alberta, Edmonton, AB, Canada. She is now the Scientific Director of the Multimedia Research Group and an Adjunct faculty in the Department of Computing Science, University of Alberta. She also holds an Adjunct position in the Faculty of Medicine and Dentistry. She has two books and more than 100 papers published in international journals and peer-reviewed conferences. Her research interests include incorporating human perception Just-Noticeable-Difference following psychophysical methodology, to improve multimedia transmission techniques. She is also engaged in research on 3-D TV and perceptually motivated technologies in multimedia, including online education. Dr. Cheng is the Chair of the IEEE NCS EMBS Chapter, the Chair of the Communications Society MMTC 3DRPC Interest Group, and a board member of the SMC TC on Human Perception in Vision, Graphics and Multimedia. She is a General Chair in IEEE ICME 2011 and is awarded visiting professorship at INSA Lyon in Jianbo Shi received the B.A. degree in computer science and mathematics from Cornell University, Ithaca, NY, in 1994 and the Ph.D. degree in computer science from the University of California at Berkeley in 1998 with a thesis on Normalize Cuts image segmentation algorithm. He joined the Robotics Institute at Carnegie Mellon University, Pittsburgh, PA, in 1999 as a research faculty, where he led the Human Identification at Distance (HumanID) project, developing vision techniques for human identification and activity inference. In January 2003, he joined the Department of Computer and Information Science at the University of Pennsylvania, where he is currently an Associate Professor. His current research focus is on object recognition-segmentation, mid-level shape representation, and human behavior analysis in video. Prof. Shi received a U.S. National Science Foundation CAREER award on learning to see a unified segmentation and recognition approach, in Anup Basu (M 90 SM 02) received the Ph.D. degree in computer science from the University of Maryland, College Park. He was a Visiting Professor at the University of California, Riverside, a Guest Professor at the Technical University of Austria, Graz, and the Director at the Hewlett-Packard Imaging Systems Instructional Laboratory, University of Alberta, Edmonton, Canada, where, since 1999, he has been a Professor at the Department of Computing Science, and is currently an icore-nserc Industry Research Chair. He originated the use of foveation for image, video, stereo, and graphics communication in the early 1990s, an approach that is now widely used in industrial standards. He also developed the first robust (correspondence free) 3-D motion estimation algorithm, using multiple cameras, a robust (and the first correspondence free) active camera calibration method, a single camera panoramic stereo, and several new approaches merging foveation and stereo with application to 3-D TV visualization and better depth estimation. His current research interests include 3-D/4-D image processing and visualization especially for medical applications, multimedia in education and games, and wireless 3-D multimedia transmission.

Generalized Random Walks for Fusion of Multi-Exposure Images

Generalized Random Walks for Fusion of Multi-Exposure Images IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. X, NO. X, XXX 2011 1 Generalized Random Walks for Fusion of Multi-Exposure Images Rui Shen, Student Member, IEEE, Irene Cheng, Senior Member, IEEE, Jianbo Shi,

More information

Generalized Random Walks for Fusion of Multi-Exposure Images

Generalized Random Walks for Fusion of Multi-Exposure Images IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 0, NO. 1, DECEMBER 011 1 Generalized Random Walks for Fusion of Multi-Exposure Images Rui Shen, Student Member, IEEE, Irene Cheng, Senior Member, IEEE, Jianbo

More information

Nonlinear Multiresolution Image Blending

Nonlinear Multiresolution Image Blending Nonlinear Multiresolution Image Blending Mark Grundland, Rahul Vohra, Gareth P. Williams and Neil A. Dodgson Computer Laboratory, University of Cambridge, United Kingdom October, 26 Abstract. We study

More information

High Dynamic Range Imaging.

High Dynamic Range Imaging. High Dynamic Range Imaging High Dynamic Range [3] In photography, dynamic range (DR) is measured in exposure value (EV) differences or stops, between the brightest and darkest parts of the image that show

More information

Computer Vision 2. SS 18 Dr. Benjamin Guthier Professur für Bildverarbeitung. Computer Vision 2 Dr. Benjamin Guthier

Computer Vision 2. SS 18 Dr. Benjamin Guthier Professur für Bildverarbeitung. Computer Vision 2 Dr. Benjamin Guthier Computer Vision 2 SS 18 Dr. Benjamin Guthier Professur für Bildverarbeitung Computer Vision 2 Dr. Benjamin Guthier 3. HIGH DYNAMIC RANGE Computer Vision 2 Dr. Benjamin Guthier Pixel Value Content of this

More information

EE795: Computer Vision and Intelligent Systems

EE795: Computer Vision and Intelligent Systems EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 WRI C225 Lecture 04 130131 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Histogram Equalization Image Filtering Linear

More information

Literature Survey for Fusion of Multi-Exposure Images

Literature Survey for Fusion of Multi-Exposure Images Literature Survey for Fusion of Multi-Exposure Images R. Ganga PG Scholar, Department of ECE, PET Engineering College, Tirunelveli, India. T. Agnes Ramena Assistant Professor, Department of ECE, PET Engineering

More information

Advanced phase retrieval: maximum likelihood technique with sparse regularization of phase and amplitude

Advanced phase retrieval: maximum likelihood technique with sparse regularization of phase and amplitude Advanced phase retrieval: maximum likelihood technique with sparse regularization of phase and amplitude A. Migukin *, V. atkovnik and J. Astola Department of Signal Processing, Tampere University of Technology,

More information

Digital Image Processing. Prof. P. K. Biswas. Department of Electronic & Electrical Communication Engineering

Digital Image Processing. Prof. P. K. Biswas. Department of Electronic & Electrical Communication Engineering Digital Image Processing Prof. P. K. Biswas Department of Electronic & Electrical Communication Engineering Indian Institute of Technology, Kharagpur Lecture - 21 Image Enhancement Frequency Domain Processing

More information

Detail-Enhanced Exposure Fusion

Detail-Enhanced Exposure Fusion Detail-Enhanced Exposure Fusion IEEE Transactions on Consumer Electronics Vol. 21, No.11, November 2012 Zheng Guo Li, Jing Hong Zheng, Susanto Rahardja Presented by Ji-Heon Lee School of Electrical Engineering

More information

Filtering and Enhancing Images

Filtering and Enhancing Images KECE471 Computer Vision Filtering and Enhancing Images Chang-Su Kim Chapter 5, Computer Vision by Shapiro and Stockman Note: Some figures and contents in the lecture notes of Dr. Stockman are used partly.

More information

Low Contrast Image Enhancement Using Adaptive Filter and DWT: A Literature Review

Low Contrast Image Enhancement Using Adaptive Filter and DWT: A Literature Review Low Contrast Image Enhancement Using Adaptive Filter and DWT: A Literature Review AARTI PAREYANI Department of Electronics and Communication Engineering Jabalpur Engineering College, Jabalpur (M.P.), India

More information

Edges and Binary Images

Edges and Binary Images CS 699: Intro to Computer Vision Edges and Binary Images Prof. Adriana Kovashka University of Pittsburgh September 5, 205 Plan for today Edge detection Binary image analysis Homework Due on 9/22, :59pm

More information

IMAGE DENOISING TO ESTIMATE THE GRADIENT HISTOGRAM PRESERVATION USING VARIOUS ALGORITHMS

IMAGE DENOISING TO ESTIMATE THE GRADIENT HISTOGRAM PRESERVATION USING VARIOUS ALGORITHMS IMAGE DENOISING TO ESTIMATE THE GRADIENT HISTOGRAM PRESERVATION USING VARIOUS ALGORITHMS P.Mahalakshmi 1, J.Muthulakshmi 2, S.Kannadhasan 3 1,2 U.G Student, 3 Assistant Professor, Department of Electronics

More information

Digital Image Processing. Prof. P.K. Biswas. Department of Electronics & Electrical Communication Engineering

Digital Image Processing. Prof. P.K. Biswas. Department of Electronics & Electrical Communication Engineering Digital Image Processing Prof. P.K. Biswas Department of Electronics & Electrical Communication Engineering Indian Institute of Technology, Kharagpur Image Segmentation - III Lecture - 31 Hello, welcome

More information

Motion Estimation. There are three main types (or applications) of motion estimation:

Motion Estimation. There are three main types (or applications) of motion estimation: Members: D91922016 朱威達 R93922010 林聖凱 R93922044 謝俊瑋 Motion Estimation There are three main types (or applications) of motion estimation: Parametric motion (image alignment) The main idea of parametric motion

More information

Segmentation and Grouping

Segmentation and Grouping Segmentation and Grouping How and what do we see? Fundamental Problems ' Focus of attention, or grouping ' What subsets of pixels do we consider as possible objects? ' All connected subsets? ' Representation

More information

Overview. Related Work Tensor Voting in 2-D Tensor Voting in 3-D Tensor Voting in N-D Application to Vision Problems Stereo Visual Motion

Overview. Related Work Tensor Voting in 2-D Tensor Voting in 3-D Tensor Voting in N-D Application to Vision Problems Stereo Visual Motion Overview Related Work Tensor Voting in 2-D Tensor Voting in 3-D Tensor Voting in N-D Application to Vision Problems Stereo Visual Motion Binary-Space-Partitioned Images 3-D Surface Extraction from Medical

More information

Image Compression. -The idea is to remove redundant data from the image (i.e., data which do not affect image quality significantly)

Image Compression. -The idea is to remove redundant data from the image (i.e., data which do not affect image quality significantly) Introduction Image Compression -The goal of image compression is the reduction of the amount of data required to represent a digital image. -The idea is to remove redundant data from the image (i.e., data

More information

Motivation. Intensity Levels

Motivation. Intensity Levels Motivation Image Intensity and Point Operations Dr. Edmund Lam Department of Electrical and Electronic Engineering The University of Hong ong A digital image is a matrix of numbers, each corresponding

More information

Multi-Camera Calibration, Object Tracking and Query Generation

Multi-Camera Calibration, Object Tracking and Query Generation MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Multi-Camera Calibration, Object Tracking and Query Generation Porikli, F.; Divakaran, A. TR2003-100 August 2003 Abstract An automatic object

More information

Physics-based Fast Single Image Fog Removal

Physics-based Fast Single Image Fog Removal Physics-based Fast Single Image Fog Removal Jing Yu 1, Chuangbai Xiao 2, Dapeng Li 2 1 Department of Electronic Engineering, Tsinghua University, Beijing, 100084, China 2 College of Computer Science and

More information

Region-based Segmentation

Region-based Segmentation Region-based Segmentation Image Segmentation Group similar components (such as, pixels in an image, image frames in a video) to obtain a compact representation. Applications: Finding tumors, veins, etc.

More information

Logical Templates for Feature Extraction in Fingerprint Images

Logical Templates for Feature Extraction in Fingerprint Images Logical Templates for Feature Extraction in Fingerprint Images Bir Bhanu, Michael Boshra and Xuejun Tan Center for Research in Intelligent Systems University of Califomia, Riverside, CA 9252 1, USA Email:

More information

MRT based Adaptive Transform Coder with Classified Vector Quantization (MATC-CVQ)

MRT based Adaptive Transform Coder with Classified Vector Quantization (MATC-CVQ) 5 MRT based Adaptive Transform Coder with Classified Vector Quantization (MATC-CVQ) Contents 5.1 Introduction.128 5.2 Vector Quantization in MRT Domain Using Isometric Transformations and Scaling.130 5.2.1

More information

Visible and Long-Wave Infrared Image Fusion Schemes for Situational. Awareness

Visible and Long-Wave Infrared Image Fusion Schemes for Situational. Awareness Visible and Long-Wave Infrared Image Fusion Schemes for Situational Awareness Multi-Dimensional Digital Signal Processing Literature Survey Nathaniel Walker The University of Texas at Austin nathaniel.walker@baesystems.com

More information

RESTORATION OF DEGRADED DOCUMENTS USING IMAGE BINARIZATION TECHNIQUE

RESTORATION OF DEGRADED DOCUMENTS USING IMAGE BINARIZATION TECHNIQUE RESTORATION OF DEGRADED DOCUMENTS USING IMAGE BINARIZATION TECHNIQUE K. Kaviya Selvi 1 and R. S. Sabeenian 2 1 Department of Electronics and Communication Engineering, Communication Systems, Sona College

More information

x' = c 1 x + c 2 y + c 3 xy + c 4 y' = c 5 x + c 6 y + c 7 xy + c 8

x' = c 1 x + c 2 y + c 3 xy + c 4 y' = c 5 x + c 6 y + c 7 xy + c 8 1. Explain about gray level interpolation. The distortion correction equations yield non integer values for x' and y'. Because the distorted image g is digital, its pixel values are defined only at integer

More information

CHAPTER 6 PERCEPTUAL ORGANIZATION BASED ON TEMPORAL DYNAMICS

CHAPTER 6 PERCEPTUAL ORGANIZATION BASED ON TEMPORAL DYNAMICS CHAPTER 6 PERCEPTUAL ORGANIZATION BASED ON TEMPORAL DYNAMICS This chapter presents a computational model for perceptual organization. A figure-ground segregation network is proposed based on a novel boundary

More information

BLIND QUALITY ASSESSMENT OF JPEG2000 COMPRESSED IMAGES USING NATURAL SCENE STATISTICS. Hamid R. Sheikh, Alan C. Bovik and Lawrence Cormack

BLIND QUALITY ASSESSMENT OF JPEG2000 COMPRESSED IMAGES USING NATURAL SCENE STATISTICS. Hamid R. Sheikh, Alan C. Bovik and Lawrence Cormack BLIND QUALITY ASSESSMENT OF JPEG2 COMPRESSED IMAGES USING NATURAL SCENE STATISTICS Hamid R. Sheikh, Alan C. Bovik and Lawrence Cormack Laboratory for Image and Video Engineering, Department of Electrical

More information

HIGH DYNAMIC RANGE IMAGE TONE MAPPING BY OPTIMIZING TONE MAPPED IMAGE QUALITY INDEX

HIGH DYNAMIC RANGE IMAGE TONE MAPPING BY OPTIMIZING TONE MAPPED IMAGE QUALITY INDEX HIGH DYNAMIC RANGE IMAGE TONE MAPPING BY OPTIMIZING TONE MAPPED IMAGE QUALITY INDEX Kede Ma, Hojatollah Yeganeh, Kai Zeng and Zhou Wang Department of ECE, University of Waterloo July 17, 2014 2 / 32 Outline

More information

A New Technique of Extraction of Edge Detection Using Digital Image Processing

A New Technique of Extraction of Edge Detection Using Digital Image Processing International OPEN ACCESS Journal Of Modern Engineering Research (IJMER) A New Technique of Extraction of Edge Detection Using Digital Image Processing Balaji S.C.K 1 1, Asst Professor S.V.I.T Abstract:

More information

Analysis and extensions of the Frankle-McCann

Analysis and extensions of the Frankle-McCann Analysis and extensions of the Frankle-McCann Retinex algorithm Jounal of Electronic Image, vol.13(1), pp. 85-92, January. 2004 School of Electrical Engineering and Computer Science Kyungpook National

More information

MRT based Fixed Block size Transform Coding

MRT based Fixed Block size Transform Coding 3 MRT based Fixed Block size Transform Coding Contents 3.1 Transform Coding..64 3.1.1 Transform Selection...65 3.1.2 Sub-image size selection... 66 3.1.3 Bit Allocation.....67 3.2 Transform coding using

More information

Norbert Schuff VA Medical Center and UCSF

Norbert Schuff VA Medical Center and UCSF Norbert Schuff Medical Center and UCSF Norbert.schuff@ucsf.edu Medical Imaging Informatics N.Schuff Course # 170.03 Slide 1/67 Objective Learn the principle segmentation techniques Understand the role

More information

IMPLEMENTATION OF THE CONTRAST ENHANCEMENT AND WEIGHTED GUIDED IMAGE FILTERING ALGORITHM FOR EDGE PRESERVATION FOR BETTER PERCEPTION

IMPLEMENTATION OF THE CONTRAST ENHANCEMENT AND WEIGHTED GUIDED IMAGE FILTERING ALGORITHM FOR EDGE PRESERVATION FOR BETTER PERCEPTION IMPLEMENTATION OF THE CONTRAST ENHANCEMENT AND WEIGHTED GUIDED IMAGE FILTERING ALGORITHM FOR EDGE PRESERVATION FOR BETTER PERCEPTION Chiruvella Suresh Assistant professor, Department of Electronics & Communication

More information

Broad field that includes low-level operations as well as complex high-level algorithms

Broad field that includes low-level operations as well as complex high-level algorithms Image processing About Broad field that includes low-level operations as well as complex high-level algorithms Low-level image processing Computer vision Computational photography Several procedures and

More information

Motivation. Gray Levels

Motivation. Gray Levels Motivation Image Intensity and Point Operations Dr. Edmund Lam Department of Electrical and Electronic Engineering The University of Hong ong A digital image is a matrix of numbers, each corresponding

More information

An Algorithm for Blurred Thermal image edge enhancement for security by image processing technique

An Algorithm for Blurred Thermal image edge enhancement for security by image processing technique An Algorithm for Blurred Thermal image edge enhancement for security by image processing technique Vinay Negi 1, Dr.K.P.Mishra 2 1 ECE (PhD Research scholar), Monad University, India, Hapur 2 ECE, KIET,

More information

Image Restoration and Reconstruction

Image Restoration and Reconstruction Image Restoration and Reconstruction Image restoration Objective process to improve an image Recover an image by using a priori knowledge of degradation phenomenon Exemplified by removal of blur by deblurring

More information

Panoramic Image Stitching

Panoramic Image Stitching Mcgill University Panoramic Image Stitching by Kai Wang Pengbo Li A report submitted in fulfillment for the COMP 558 Final project in the Faculty of Computer Science April 2013 Mcgill University Abstract

More information

Perceptual Effects in Real-time Tone Mapping

Perceptual Effects in Real-time Tone Mapping Perceptual Effects in Real-time Tone Mapping G. Krawczyk K. Myszkowski H.-P. Seidel Max-Planck-Institute für Informatik Saarbrücken, Germany SCCG 2005 High Dynamic Range (HDR) HDR Imaging Display of HDR

More information

Anno accademico 2006/2007. Davide Migliore

Anno accademico 2006/2007. Davide Migliore Robotica Anno accademico 6/7 Davide Migliore migliore@elet.polimi.it Today What is a feature? Some useful information The world of features: Detectors Edges detection Corners/Points detection Descriptors?!?!?

More information

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing Visual servoing vision allows a robotic system to obtain geometrical and qualitative information on the surrounding environment high level control motion planning (look-and-move visual grasping) low level

More information

IMAGE ENHANCEMENT USING NONSUBSAMPLED CONTOURLET TRANSFORM

IMAGE ENHANCEMENT USING NONSUBSAMPLED CONTOURLET TRANSFORM IMAGE ENHANCEMENT USING NONSUBSAMPLED CONTOURLET TRANSFORM Rafia Mumtaz 1, Raja Iqbal 2 and Dr.Shoab A.Khan 3 1,2 MCS, National Unioversity of Sciences and Technology, Rawalpindi, Pakistan: 3 EME, National

More information

Lecture 7: Most Common Edge Detectors

Lecture 7: Most Common Edge Detectors #1 Lecture 7: Most Common Edge Detectors Saad Bedros sbedros@umn.edu Edge Detection Goal: Identify sudden changes (discontinuities) in an image Intuitively, most semantic and shape information from the

More information

Image Restoration and Reconstruction

Image Restoration and Reconstruction Image Restoration and Reconstruction Image restoration Objective process to improve an image, as opposed to the subjective process of image enhancement Enhancement uses heuristics to improve the image

More information

Analysis of Image and Video Using Color, Texture and Shape Features for Object Identification

Analysis of Image and Video Using Color, Texture and Shape Features for Object Identification IOSR Journal of Computer Engineering (IOSR-JCE) e-issn: 2278-0661,p-ISSN: 2278-8727, Volume 16, Issue 6, Ver. VI (Nov Dec. 2014), PP 29-33 Analysis of Image and Video Using Color, Texture and Shape Features

More information

ECG782: Multidimensional Digital Signal Processing

ECG782: Multidimensional Digital Signal Processing Professor Brendan Morris, SEB 3216, brendan.morris@unlv.edu ECG782: Multidimensional Digital Signal Processing Spring 2014 TTh 14:30-15:45 CBC C313 Lecture 03 Image Processing Basics 13/01/28 http://www.ee.unlv.edu/~b1morris/ecg782/

More information

Supervised texture detection in images

Supervised texture detection in images Supervised texture detection in images Branislav Mičušík and Allan Hanbury Pattern Recognition and Image Processing Group, Institute of Computer Aided Automation, Vienna University of Technology Favoritenstraße

More information

Satellite Image Processing Using Singular Value Decomposition and Discrete Wavelet Transform

Satellite Image Processing Using Singular Value Decomposition and Discrete Wavelet Transform Satellite Image Processing Using Singular Value Decomposition and Discrete Wavelet Transform Kodhinayaki E 1, vinothkumar S 2, Karthikeyan T 3 Department of ECE 1, 2, 3, p.g scholar 1, project coordinator

More information

ECG782: Multidimensional Digital Signal Processing

ECG782: Multidimensional Digital Signal Processing Professor Brendan Morris, SEB 3216, brendan.morris@unlv.edu ECG782: Multidimensional Digital Signal Processing Spatial Domain Filtering http://www.ee.unlv.edu/~b1morris/ecg782/ 2 Outline Background Intensity

More information

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS Cognitive Robotics Original: David G. Lowe, 004 Summary: Coen van Leeuwen, s1460919 Abstract: This article presents a method to extract

More information

FEATURE EXTRACTION TECHNIQUES FOR IMAGE RETRIEVAL USING HAAR AND GLCM

FEATURE EXTRACTION TECHNIQUES FOR IMAGE RETRIEVAL USING HAAR AND GLCM FEATURE EXTRACTION TECHNIQUES FOR IMAGE RETRIEVAL USING HAAR AND GLCM Neha 1, Tanvi Jain 2 1,2 Senior Research Fellow (SRF), SAM-C, Defence R & D Organization, (India) ABSTRACT Content Based Image Retrieval

More information

Filtering Images. Contents

Filtering Images. Contents Image Processing and Data Visualization with MATLAB Filtering Images Hansrudi Noser June 8-9, 010 UZH, Multimedia and Robotics Summer School Noise Smoothing Filters Sigmoid Filters Gradient Filters Contents

More information

Accurate 3D Face and Body Modeling from a Single Fixed Kinect

Accurate 3D Face and Body Modeling from a Single Fixed Kinect Accurate 3D Face and Body Modeling from a Single Fixed Kinect Ruizhe Wang*, Matthias Hernandez*, Jongmoo Choi, Gérard Medioni Computer Vision Lab, IRIS University of Southern California Abstract In this

More information

Topic 4 Image Segmentation

Topic 4 Image Segmentation Topic 4 Image Segmentation What is Segmentation? Why? Segmentation important contributing factor to the success of an automated image analysis process What is Image Analysis: Processing images to derive

More information

Image Segmentation Techniques for Object-Based Coding

Image Segmentation Techniques for Object-Based Coding Image Techniques for Object-Based Coding Junaid Ahmed, Joseph Bosworth, and Scott T. Acton The Oklahoma Imaging Laboratory School of Electrical and Computer Engineering Oklahoma State University {ajunaid,bosworj,sacton}@okstate.edu

More information

CHAPTER 3 IMAGE ENHANCEMENT IN THE SPATIAL DOMAIN

CHAPTER 3 IMAGE ENHANCEMENT IN THE SPATIAL DOMAIN CHAPTER 3 IMAGE ENHANCEMENT IN THE SPATIAL DOMAIN CHAPTER 3: IMAGE ENHANCEMENT IN THE SPATIAL DOMAIN Principal objective: to process an image so that the result is more suitable than the original image

More information

TERM PAPER ON The Compressive Sensing Based on Biorthogonal Wavelet Basis

TERM PAPER ON The Compressive Sensing Based on Biorthogonal Wavelet Basis TERM PAPER ON The Compressive Sensing Based on Biorthogonal Wavelet Basis Submitted By: Amrita Mishra 11104163 Manoj C 11104059 Under the Guidance of Dr. Sumana Gupta Professor Department of Electrical

More information

MULTI-FOCUS IMAGE FUSION USING GUIDED FILTERING

MULTI-FOCUS IMAGE FUSION USING GUIDED FILTERING INTERNATIONAL JOURNAL OF RESEARCH IN COMPUTER APPLICATIONS AND ROBOTICS ISSN 2320-7345 MULTI-FOCUS IMAGE FUSION USING GUIDED FILTERING 1 Johnson suthakar R, 2 Annapoorani D, 3 Richard Singh Samuel F, 4

More information

A Simple Automated Void Defect Detection for Poor Contrast X-ray Images of BGA

A Simple Automated Void Defect Detection for Poor Contrast X-ray Images of BGA Proceedings of the 3rd International Conference on Industrial Application Engineering 2015 A Simple Automated Void Defect Detection for Poor Contrast X-ray Images of BGA Somchai Nuanprasert a,*, Sueki

More information

Rendering and Modeling of Transparent Objects. Minglun Gong Dept. of CS, Memorial Univ.

Rendering and Modeling of Transparent Objects. Minglun Gong Dept. of CS, Memorial Univ. Rendering and Modeling of Transparent Objects Minglun Gong Dept. of CS, Memorial Univ. Capture transparent object appearance Using frequency based environmental matting Reduce number of input images needed

More information

Image Denoising Based on Hybrid Fourier and Neighborhood Wavelet Coefficients Jun Cheng, Songli Lei

Image Denoising Based on Hybrid Fourier and Neighborhood Wavelet Coefficients Jun Cheng, Songli Lei Image Denoising Based on Hybrid Fourier and Neighborhood Wavelet Coefficients Jun Cheng, Songli Lei College of Physical and Information Science, Hunan Normal University, Changsha, China Hunan Art Professional

More information

Compression of RADARSAT Data with Block Adaptive Wavelets Abstract: 1. Introduction

Compression of RADARSAT Data with Block Adaptive Wavelets Abstract: 1. Introduction Compression of RADARSAT Data with Block Adaptive Wavelets Ian Cumming and Jing Wang Department of Electrical and Computer Engineering The University of British Columbia 2356 Main Mall, Vancouver, BC, Canada

More information

Image restoration. Restoration: Enhancement:

Image restoration. Restoration: Enhancement: Image restoration Most images obtained by optical, electronic, or electro-optic means is likely to be degraded. The degradation can be due to camera misfocus, relative motion between camera and object,

More information

Schedule for Rest of Semester

Schedule for Rest of Semester Schedule for Rest of Semester Date Lecture Topic 11/20 24 Texture 11/27 25 Review of Statistics & Linear Algebra, Eigenvectors 11/29 26 Eigenvector expansions, Pattern Recognition 12/4 27 Cameras & calibration

More information

Feature Detectors - Canny Edge Detector

Feature Detectors - Canny Edge Detector Feature Detectors - Canny Edge Detector 04/12/2006 07:00 PM Canny Edge Detector Common Names: Canny edge detector Brief Description The Canny operator was designed to be an optimal edge detector (according

More information

An Edge-Based Approach to Motion Detection*

An Edge-Based Approach to Motion Detection* An Edge-Based Approach to Motion Detection* Angel D. Sappa and Fadi Dornaika Computer Vison Center Edifici O Campus UAB 08193 Barcelona, Spain {sappa, dornaika}@cvc.uab.es Abstract. This paper presents

More information

Supplemental Document for Deep Photo Style Transfer

Supplemental Document for Deep Photo Style Transfer Supplemental Document for Deep Photo Style Transfer Fujun Luan Cornell University Sylvain Paris Adobe Eli Shechtman Adobe Kavita Bala Cornell University fujun@cs.cornell.edu sparis@adobe.com elishe@adobe.com

More information

Computer Vision 2. SS 18 Dr. Benjamin Guthier Professur für Bildverarbeitung. Computer Vision 2 Dr. Benjamin Guthier

Computer Vision 2. SS 18 Dr. Benjamin Guthier Professur für Bildverarbeitung. Computer Vision 2 Dr. Benjamin Guthier Computer Vision 2 SS 18 Dr. Benjamin Guthier Professur für Bildverarbeitung Computer Vision 2 Dr. Benjamin Guthier 1. IMAGE PROCESSING Computer Vision 2 Dr. Benjamin Guthier Content of this Chapter Non-linear

More information

EE368 Project Report CD Cover Recognition Using Modified SIFT Algorithm

EE368 Project Report CD Cover Recognition Using Modified SIFT Algorithm EE368 Project Report CD Cover Recognition Using Modified SIFT Algorithm Group 1: Mina A. Makar Stanford University mamakar@stanford.edu Abstract In this report, we investigate the application of the Scale-Invariant

More information

Computational Photography and Capture: (Re)Coloring. Gabriel Brostow & Tim Weyrich TA: Frederic Besse

Computational Photography and Capture: (Re)Coloring. Gabriel Brostow & Tim Weyrich TA: Frederic Besse Computational Photography and Capture: (Re)Coloring Gabriel Brostow & Tim Weyrich TA: Frederic Besse Week Date Topic Hours 1 12-Jan Introduction to Computational Photography and Capture 1 1 14-Jan Intro

More information

Comparative Analysis of Edge Detection Algorithms Based on Content Based Image Retrieval With Heterogeneous Images

Comparative Analysis of Edge Detection Algorithms Based on Content Based Image Retrieval With Heterogeneous Images Comparative Analysis of Edge Detection Algorithms Based on Content Based Image Retrieval With Heterogeneous Images T. Dharani I. Laurence Aroquiaraj V. Mageshwari Department of Computer Science, Department

More information

A reversible data hiding based on adaptive prediction technique and histogram shifting

A reversible data hiding based on adaptive prediction technique and histogram shifting A reversible data hiding based on adaptive prediction technique and histogram shifting Rui Liu, Rongrong Ni, Yao Zhao Institute of Information Science Beijing Jiaotong University E-mail: rrni@bjtu.edu.cn

More information

Photometric Stereo with Auto-Radiometric Calibration

Photometric Stereo with Auto-Radiometric Calibration Photometric Stereo with Auto-Radiometric Calibration Wiennat Mongkulmann Takahiro Okabe Yoichi Sato Institute of Industrial Science, The University of Tokyo {wiennat,takahiro,ysato} @iis.u-tokyo.ac.jp

More information

Edge and Texture. CS 554 Computer Vision Pinar Duygulu Bilkent University

Edge and Texture. CS 554 Computer Vision Pinar Duygulu Bilkent University Edge and Texture CS 554 Computer Vision Pinar Duygulu Bilkent University Filters for features Previously, thinking of filtering as a way to remove or reduce noise Now, consider how filters will allow us

More information

Segmentation of Images

Segmentation of Images Segmentation of Images SEGMENTATION If an image has been preprocessed appropriately to remove noise and artifacts, segmentation is often the key step in interpreting the image. Image segmentation is a

More information

EECS 556 Image Processing W 09. Image enhancement. Smoothing and noise removal Sharpening filters

EECS 556 Image Processing W 09. Image enhancement. Smoothing and noise removal Sharpening filters EECS 556 Image Processing W 09 Image enhancement Smoothing and noise removal Sharpening filters What is image processing? Image processing is the application of 2D signal processing methods to images Image

More information

A Modified SVD-DCT Method for Enhancement of Low Contrast Satellite Images

A Modified SVD-DCT Method for Enhancement of Low Contrast Satellite Images A Modified SVD-DCT Method for Enhancement of Low Contrast Satellite Images G.Praveena 1, M.Venkatasrinu 2, 1 M.tech student, Department of Electronics and Communication Engineering, Madanapalle Institute

More information

OPTIMIZED QUALITY EVALUATION APPROACH OF TONED MAPPED IMAGES BASED ON OBJECTIVE QUALITY ASSESSMENT

OPTIMIZED QUALITY EVALUATION APPROACH OF TONED MAPPED IMAGES BASED ON OBJECTIVE QUALITY ASSESSMENT OPTIMIZED QUALITY EVALUATION APPROACH OF TONED MAPPED IMAGES BASED ON OBJECTIVE QUALITY ASSESSMENT ANJIBABU POLEBOINA 1, M.A. SHAHID 2 Digital Electronics and Communication Systems (DECS) 1, Associate

More information

Parametric Texture Model based on Joint Statistics

Parametric Texture Model based on Joint Statistics Parametric Texture Model based on Joint Statistics Gowtham Bellala, Kumar Sricharan, Jayanth Srinivasa Department of Electrical Engineering, University of Michigan, Ann Arbor 1. INTRODUCTION Texture images

More information

UNIT - 5 IMAGE ENHANCEMENT IN SPATIAL DOMAIN

UNIT - 5 IMAGE ENHANCEMENT IN SPATIAL DOMAIN UNIT - 5 IMAGE ENHANCEMENT IN SPATIAL DOMAIN Spatial domain methods Spatial domain refers to the image plane itself, and approaches in this category are based on direct manipulation of pixels in an image.

More information

Exposure Fusion Based on Shift-Invariant Discrete Wavelet Transform *

Exposure Fusion Based on Shift-Invariant Discrete Wavelet Transform * JOURNAL OF INFORMATION SCIENCE AND ENGINEERING 27, 197-211 (2011) Exposure Fusion Based on Shift-Invariant Discrete Wavelet Transform * JINHUA WANG 1,2, DE XU 1, CONGYAN LANG 1 AND BING LI 1 1 Institute

More information

Statistical image models

Statistical image models Chapter 4 Statistical image models 4. Introduction 4.. Visual worlds Figure 4. shows images that belong to different visual worlds. The first world (fig. 4..a) is the world of white noise. It is the world

More information

Light source separation from image sequences of oscillating lights

Light source separation from image sequences of oscillating lights 2014 IEEE 28-th Convention of Electrical and Electronics Engineers in Israel Light source separation from image sequences of oscillating lights Amir Kolaman, Rami Hagege and Hugo Guterman Electrical and

More information

CHAPTER 6 DETECTION OF MASS USING NOVEL SEGMENTATION, GLCM AND NEURAL NETWORKS

CHAPTER 6 DETECTION OF MASS USING NOVEL SEGMENTATION, GLCM AND NEURAL NETWORKS 130 CHAPTER 6 DETECTION OF MASS USING NOVEL SEGMENTATION, GLCM AND NEURAL NETWORKS A mass is defined as a space-occupying lesion seen in more than one projection and it is described by its shapes and margin

More information

EE795: Computer Vision and Intelligent Systems

EE795: Computer Vision and Intelligent Systems EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 14 130307 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Stereo Dense Motion Estimation Translational

More information

Digital Image Processing

Digital Image Processing Digital Image Processing Third Edition Rafael C. Gonzalez University of Tennessee Richard E. Woods MedData Interactive PEARSON Prentice Hall Pearson Education International Contents Preface xv Acknowledgments

More information

IMAGE DE-NOISING IN WAVELET DOMAIN

IMAGE DE-NOISING IN WAVELET DOMAIN IMAGE DE-NOISING IN WAVELET DOMAIN Aaditya Verma a, Shrey Agarwal a a Department of Civil Engineering, Indian Institute of Technology, Kanpur, India - (aaditya, ashrey)@iitk.ac.in KEY WORDS: Wavelets,

More information

Color Characterization and Calibration of an External Display

Color Characterization and Calibration of an External Display Color Characterization and Calibration of an External Display Andrew Crocker, Austin Martin, Jon Sandness Department of Math, Statistics, and Computer Science St. Olaf College 1500 St. Olaf Avenue, Northfield,

More information

Multi-exposure Fusion Features

Multi-exposure Fusion Features Multi-exposure Fusion Features Paper Number: 94 No Institute Given Abstract. This paper introduces a process where fusion features assist matching scale invariant feature transform (SIFT) image features

More information

Corner Detection. Harvey Rhody Chester F. Carlson Center for Imaging Science Rochester Institute of Technology

Corner Detection. Harvey Rhody Chester F. Carlson Center for Imaging Science Rochester Institute of Technology Corner Detection Harvey Rhody Chester F. Carlson Center for Imaging Science Rochester Institute of Technology rhody@cis.rit.edu April 11, 2006 Abstract Corners and edges are two of the most important geometrical

More information

Computer Vision & Digital Image Processing. Image segmentation: thresholding

Computer Vision & Digital Image Processing. Image segmentation: thresholding Computer Vision & Digital Image Processing Image Segmentation: Thresholding Dr. D. J. Jackson Lecture 18-1 Image segmentation: thresholding Suppose an image f(y) is composed of several light objects on

More information

Multimedia Systems Image III (Image Compression, JPEG) Mahdi Amiri April 2011 Sharif University of Technology

Multimedia Systems Image III (Image Compression, JPEG) Mahdi Amiri April 2011 Sharif University of Technology Course Presentation Multimedia Systems Image III (Image Compression, JPEG) Mahdi Amiri April 2011 Sharif University of Technology Image Compression Basics Large amount of data in digital images File size

More information

Non-Linear Masking based Contrast Enhancement via Illumination Estimation

Non-Linear Masking based Contrast Enhancement via Illumination Estimation https://doi.org/10.2352/issn.2470-1173.2018.13.ipas-389 2018, Society for Imaging Science and Technology Non-Linear Masking based Contrast Enhancement via Illumination Estimation Soonyoung Hong, Minsub

More information

Chapter 11 Representation & Description

Chapter 11 Representation & Description Chain Codes Chain codes are used to represent a boundary by a connected sequence of straight-line segments of specified length and direction. The direction of each segment is coded by using a numbering

More information

DESIGN OF A NOVEL IMAGE FUSION ALGORITHM FOR IMPULSE NOISE REMOVAL IN REMOTE SENSING IMAGES BY USING THE QUALITY ASSESSMENT

DESIGN OF A NOVEL IMAGE FUSION ALGORITHM FOR IMPULSE NOISE REMOVAL IN REMOTE SENSING IMAGES BY USING THE QUALITY ASSESSMENT DESIGN OF A NOVEL IMAGE FUSION ALGORITHM FOR IMPULSE NOISE REMOVAL IN REMOTE SENSING IMAGES BY USING THE QUALITY ASSESSMENT P.PAVANI, M.V.H.BHASKARA MURTHY Department of Electronics and Communication Engineering,Aditya

More information

Biometrics Technology: Image Processing & Pattern Recognition (by Dr. Dickson Tong)

Biometrics Technology: Image Processing & Pattern Recognition (by Dr. Dickson Tong) Biometrics Technology: Image Processing & Pattern Recognition (by Dr. Dickson Tong) References: [1] http://homepages.inf.ed.ac.uk/rbf/hipr2/index.htm [2] http://www.cs.wisc.edu/~dyer/cs540/notes/vision.html

More information

convolution shift invariant linear system Fourier Transform Aliasing and sampling scale representation edge detection corner detection

convolution shift invariant linear system Fourier Transform Aliasing and sampling scale representation edge detection corner detection COS 429: COMPUTER VISON Linear Filters and Edge Detection convolution shift invariant linear system Fourier Transform Aliasing and sampling scale representation edge detection corner detection Reading:

More information