Single Image Super-Resolution via Sparse Representation over Directionally Structured Dictionaries Based on the Patch Gradient Phase Angle

Similar documents
IMAGE DENOISING TO ESTIMATE THE GRADIENT HISTOGRAM PRESERVATION USING VARIOUS ALGORITHMS

Learning based face hallucination techniques: A survey

Image Interpolation using Collaborative Filtering

IMAGE DENOISING USING NL-MEANS VIA SMOOTH PATCH ORDERING

Single-Image Super-Resolution Using Multihypothesis Prediction

Image Deblurring Using Adaptive Sparse Domain Selection and Adaptive Regularization

Single Image Interpolation via Adaptive Non-Local Sparsity-Based Modeling

Augmented Coupled Dictionary Learning for Image Super-Resolution

Single image super-resolution by directionally structured coupled dictionary learning

Patch-Based Color Image Denoising using efficient Pixel-Wise Weighting Techniques

Image Restoration and Background Separation Using Sparse Representation Framework

IMAGE RESTORATION VIA EFFICIENT GAUSSIAN MIXTURE MODEL LEARNING

IMAGE SUPER-RESOLUTION BASED ON DICTIONARY LEARNING AND ANCHORED NEIGHBORHOOD REGRESSION WITH MUTUAL INCOHERENCE

IMPROVED MOTION-BASED LOCALIZED SUPER RESOLUTION TECHNIQUE USING DISCRETE WAVELET TRANSFORM FOR LOW RESOLUTION VIDEO ENHANCEMENT

Edge-directed Image Interpolation Using Color Gradient Information

NTHU Rain Removal Project

Edge-Directed Image Interpolation Using Color Gradient Information

Structured Face Hallucination

Department of Electronics and Communication KMP College of Engineering, Perumbavoor, Kerala, India 1 2

Sparse Representation Based Super-Resolution Algorithm using Wavelet Domain Interpolation and Nonlocal Means

Image Super-Resolution Reconstruction Based On L 1/2 Sparsity

Low Light Image Enhancement via Sparse Representations

A Single Image Compression Framework Combined with Sparse Representation-Based Super- Resolution

Denoising an Image by Denoising its Components in a Moving Frame

MULTI-POSE FACE HALLUCINATION VIA NEIGHBOR EMBEDDING FOR FACIAL COMPONENTS. Yanghao Li, Jiaying Liu, Wenhan Yang, Zongming Guo

One Network to Solve Them All Solving Linear Inverse Problems using Deep Projection Models

Self-Learning of Edge-Preserving Single Image Super-Resolution via Contourlet Transform

An Optimized Pixel-Wise Weighting Approach For Patch-Based Image Denoising

An Improved Approach For Mixed Noise Removal In Color Images

Image Interpolation Using Multiscale Geometric Representations

Image Resizing Based on Gradient Vector Flow Analysis

Robust Single Image Super-resolution based on Gradient Enhancement

A A A. Fig.1 image patch. Then the edge gradient magnitude is . (1)

Exploiting Self-Similarities for Single Frame Super-Resolution

A Self-Learning Optimization Approach to Single Image Super-Resolution using Kernel ridge regression model

Image Quality Assessment Techniques: An Overview

arxiv: v1 [cs.cv] 3 Jan 2017

Boosting face recognition via neural Super-Resolution

Single Image Super-Resolution via Iterative Collaborative Representation

Image Inpainting Using Sparsity of the Transform Domain

Centralized Sparse Representation for Image Restoration

IMPLEMENTATION OF THE CONTRAST ENHANCEMENT AND WEIGHTED GUIDED IMAGE FILTERING ALGORITHM FOR EDGE PRESERVATION FOR BETTER PERCEPTION

SUPPLEMENTARY MATERIAL

Single Image Super-resolution using Deformable Patches

COMPRESSED FACE HALLUCINATION. Electrical Engineering and Computer Science University of California, Merced, CA 95344, USA

Ms.DHARANI SAMPATH Computer Science And Engineering, Sri Krishna College Of Engineering & Technology Coimbatore, India

arxiv: v1 [cs.cv] 8 Feb 2018

Image Restoration Using DNN

Deblurring Text Images via L 0 -Regularized Intensity and Gradient Prior

Anchored Neighborhood Regression for Fast Example-Based Super-Resolution

Structure-adaptive Image Denoising with 3D Collaborative Filtering

An improved non-blind image deblurring methods based on FoEs

A DEEP DICTIONARY MODEL FOR IMAGE SUPER-RESOLUTION. Jun-Jie Huang and Pier Luigi Dragotti

An Approach for Reduction of Rain Streaks from a Single Image

Optimizing the Deblocking Algorithm for. H.264 Decoder Implementation

Single-image super-resolution in RGB space via group sparse representation

Interpolation Based Image Super Resolution by Support-Vector-Regression

Example-Based Image Super-Resolution Techniques

Fast single image super-resolution based on sigmoid transformation

IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 20, NO. 6, JUNE

ICA mixture models for image processing

Comparative Analysis of Edge Based Single Image Superresolution

Learning how to combine internal and external denoising methods

Non-local Means for Stereo Image Denoising Using Structural Similarity

Optimal Denoising of Natural Images and their Multiscale Geometry and Density

SSIM Image Quality Metric for Denoised Images

Image Super-Resolution via Sparse Representation

Virtual Training Samples and CRC based Test Sample Reconstruction and Face Recognition Experiments Wei HUANG and Li-ming MIAO

A Novel Image Super-resolution Reconstruction Algorithm based on Modified Sparse Representation

A Novel Multi-Frame Color Images Super-Resolution Framework based on Deep Convolutional Neural Network. Zhe Li, Shu Li, Jianmin Wang and Hongyang Wang

Blind Deblurring using Internal Patch Recurrence. Tomer Michaeli & Michal Irani Weizmann Institute

Influence of Training Set and Iterative Back Projection on Example-based Super-resolution

Deep Back-Projection Networks For Super-Resolution Supplementary Material

IMAGE SUPER RESOLUTION USING NON SUB-SAMPLE CONTOURLET TRANSFORM WITH LOCAL TERNARY PATTERN

Bayesian region selection for adaptive dictionary-based Super-Resolution

Recovering Realistic Texture in Image Super-resolution by Deep Spatial Feature Transform. Xintao Wang Ke Yu Chao Dong Chen Change Loy

A Sparse Coding based Approach for the Resolution Enhancement and Restoration of Printed and Handwritten Textual Images

Blind Image Deblurring Using Dark Channel Prior

3194 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 25, NO. 7, JULY 2016

Inverse Problems and Machine Learning

Image Processing Via Pixel Permutations

GRID WARPING IN TOTAL VARIATION IMAGE ENHANCEMENT METHODS. Andrey Nasonov, and Andrey Krylov

On Single Image Scale-Up using Sparse-Representation

A Comparison of Still-Image Compression Standards Using Different Image Quality Metrics and Proposed Methods for Improving Lossy Image Quality

Fast and Effective Interpolation Using Median Filter

A Comparative Analysis of Noise Reduction Filters in Images Mandeep kaur 1, Deepinder kaur 2

A New Orthogonalization of Locality Preserving Projection and Applications

Bilevel Sparse Coding

FAST SINGLE-IMAGE SUPER-RESOLUTION WITH FILTER SELECTION. Image Processing Lab Technicolor R&I Hannover

Bidirectional Recurrent Convolutional Networks for Video Super-Resolution

Image Super-Resolution Using Compressed Sensing Based on Learning Sub Dictionary

Enhancing DubaiSat-1 Satellite Imagery Using a Single Image Super-Resolution

Texture Sensitive Image Inpainting after Object Morphing

Face Hallucination Based on Eigentransformation Learning

Generalized Tree-Based Wavelet Transform and Applications to Patch-Based Image Processing

PRINCIPAL COMPONENT ANALYSIS IMAGE DENOISING USING LOCAL PIXEL GROUPING

BLIND IMAGE QUALITY ASSESSMENT WITH LOCAL CONTRAST FEATURES

REJECTION-BASED CLASSIFICATION FOR ACTION RECOGNITION USING A SPATIO-TEMPORAL DICTIONARY. Stefen Chan Wai Tim, Michele Rombaut, Denis Pellerin

A Novel Approach for Deblocking JPEG Images

Locally Adaptive Learning for Translation-Variant MRF Image Priors

Transcription:

2014 UKSim-AMSS 8th European Modelling Symposium Single Image Super-Resolution via Sparse Representation over Directionally Structured Dictionaries Based on the Patch Gradient Phase Angle Mahmoud Nazzal, Faezeh Yeganli and Huseyin Ozkaramanli Electrical and Electronic Engineering Department Eastern Mediterranean University Gazimagusa, via Mersin 10, Turkey mahmoud.nazzal@cc.emu.edu.tr, faezeh.yeganli@cc.emu.edu.tr, huseyin.ozkaramanli@emu.edu.tr Abstract We propose a single-image super-resolution algorithm based on sparse representation over a set of cluster dictionary pairs. For each cluster, a directionally structured dictionary pair is designed. The dominant angle in the patch gradient phase matrix is employed as an approximately scaleinvariant measure. This measure serves for patch clustering and sparse model selection. The dominant phase angle of each low resolution patch is found and used to identify its corresponding cluster. Then, the sparse coding coefficients of this patch with respect to the low resolution cluster dictionary are calculated. These coefficients are imposed on the high resolution dictionary of the same cluster to obtain a high resolution patch estimate. In experiments conducted on several images, the proposed algorithm is shown to outperform the algorithm that uses a single universal dictionary pair, and to be competitive to the state-of-the art algorithm. This is validated in terms of PSNR, SSIM and visual comparison. Keywords single-image super-resolution, sparse representation, cluster dictionary pairs, gradient operator. I. INTRODUCTION Singe-image super-resolution (SR) is the process of reconstructing a high resolution (HR) image (I H ) from a low resolution (LR) image (I L ) corresponding to the same scene. The relationship between I L and I H is typically characterized by a blurring and downsampling operation, as described in (1). I L ΨI H, (1) where Ψ is the blurring and downsampling operator. It is well-known that this process is an ill-posed inverse problem. Approaches to SR can be classified into three main categories. First, interpolation methods which estimate the unknown pixel values based on the known ones. Several advanced interpolation techniques have been proposed to over-come the low-pass nature of the interpolation process. Second is the reconstruction-based methods, where constrains are applied and exploited in reconstructing the HR image. The third category is the learning-based approaches where a training process is carried out and then used during a testing process. The fundamental idea in these approaches is to utilize the correspondence between the training and testing sets. One of the most successful learning-based approaches is the sparse representationbased approach. In this framework, dictionaries are trained over example signals and then used in the testing process. In the context of sparse representation-based SR, sparsity of the representation of LR and HR image patches is used as a natural image prior to regularize the SR problem. This usage allows for assuming that the sparse coding coefficients of an image patch are scale-invariant as proposed by Yang et al. [1], [2]. Another prior that has been successfully employed in regard is the gradient operator. This operator has been employed as a natural image prior to solve ill-posed image processing problems such as SR [3], [4] and denoising [5]. The early usage of the gradient as a prior was proposed by Rudin et al. [6] in the total variation denoising framework. Another line of research considered approximating the gradient distribution such as [7] and [8]. More recent works have considered the invariance or preservation of the gradient histogram and applied it as a prior. An example is the work conducted by Zuo et al. [5] in image denoising, where they enforced the histogram of the denoised image to be similar to that of the noisy one. Another example is the work by Sun et al. [3], [4] which defines a gradient profile prior that is invariant to scale, and uses it as a generic image prior for SR. In view of the success of sparse representation in various signal and image processing applications, it is well-known that its major added-benefit is the signalfitting capability due to the use of learned dictionaries. Compared to fixed basis functions, such dictionaries are proven to better fit natural signals as they are trained over examples signals [9]. Since the representation is intended to be sparse, learned dictionaries have to be overcomplete (redundant) to allow for a good representation quality. However, high redundancy has its own drawbacks. It has been shown that having high dictionary redundancy tends to cause instability of the representation and therefore to degrade the representation quality [10], [11]. In other words, a good dictionary needs to possess a good capability in representing signals at a reasonable 978-1-4799-7412-2/14 $31.00 2014 IEEE DOI 10.1109/EMS.2014.70 209

degree of redundancy. Motivated by the above findings, several attempts have recently been made towards designing compact dictionaries without sacrificing the representation power. Intuitively, the variability of signals within a class is less than the general signal variability. A recent research trend considered learning class-dependent dictionaries. Examples of this trend include clustering the training and testing data as proposed by Dong et al. [12], where they applied K-means clustering for separating signals into several clusters, and learned cluster sub-dictionaries. The same clustering criterion is used to classify a signal into a cluster and use its sub-dictionary for the purpose of sparse coding. Along this line, researchers aimed at designing directionally structured dictionaries that correspond to directionally sensitive signal classes. In [13], Yang et al. employed multiple geometric cluster dictionaries. Each cluster is concerned with a certain directional structure. Sparse coding of a signal is carried out over the cluster dictionary that best fits this signal based on its structure. In [14], Yu et al. designed a structural dictionary composed of several orthogonal bases that correspond to different structures. Again, sparse coding of a signal is done by first selecting the best fitting basis (model) according to the signal s structure, and then calculating the sparse coding coefficient with respect to this basis. In this paper, we classify the training and testing signals into a set of directionally structured clusters. We employ the dominant phase angle (DPA) defined via the gradient operator as a classification criterion. We empirically show an acceptable degree of scale-invariance of the DPA as a natural image quantity. For each cluster, a directionally structured coupled dictionary pair is learned and then used for the purpose of sparse coding. We show that the designed dictionaries inherit the intended directional structure of their respective clusters. The proposed SR algorithm is shown to be superior to that algorithm of Yang et al. [2] that uses a universal dictionary pair. Besides, it is shown to be competitive to the state-of-the-art SR algorithms. This result is validated qualitatively in terms of the peak signal-to-noise ratio (PSNR) and structural similarity index (SSIM) measures, and qualitatively in terms of visual comparisons. II. SINGLE IMAGE SUPER-RESOLUTION VIA SPARSE REPRESENTATION The sparse-representation SR algorithm of Yang et al. [2] is based on using patches of a LR image to reconstruct the underlying patches in the unknown HR image. Sparsity of the representation is applied as a generic natural image property that is invariant with respect to scale. Their work is based on applying two constraints on the reconstructed HR image. First is a reconstruction constraint enforcing the blurred and downsampled version of the HR image estimate to be consistent with the given LR image. Second, is a sparsity constraint which assumes that the sparse coding coefficients of a LR patch are similar to those of the underlying HR one, or scale-invariant in other words. In [1], the authors used a set of sampled patch pairs as a dictionary for sparse coding. However, this paradigm is very slow in practice. Therefore, in [2], Yang et al. employed a pair of coupled dictionaries learned from such patch pairs. They first calculated the sparse coding coefficients of a LR patch with respect to the LR dictionary. Then, they imposed these coefficients on the HR dictionary to find a HR patch estimate. For the sake of local consistency of reconstructed HR patches, it is advantageous to divide the LR image into overlapping patches [1], [2]. Correspondingly, the reconstructed HR patches are also overlapped. They are then merged at the overlap locations to generate a HR image estimate. The LR and HR dictionaries are simultaneously learned in a coupled manner. Patches of HR training images are extracted and column-stacked to form an array of HR training patches. The mean value of each patch is subtracted to allow for a better training. At the same time, LR images are obtained by applying a blurring and downsampling operation of the HR ones. Then, LR images are upsampled to the so called middle resolution (MR) to allow for better feature extraction.. Afterwards, first and second order gradient filters are operated over the MR images to extract the features. The next step is to combine the extracted features of each MR patch into a single column, and then to combine feature vectors column-wise to obtain the corresponding LR training patch array. Eventually, LR and HR training patches are used to train for a pair of coupled LR and HR dictionaries, respectively. This coupling is vital for the validity of the sparse coding coefficient invariance assumption. Given a HR vector patch x H, one may find the sparse coding coefficient vector α H of this patch over a dictionary in the same resolution level D H. Vector selection techniques such the least absolute shrinkage and selection operator (LASSO) [15] can be applied for this purpose. A sparse approximation of x H can be written as x H D H α H, (2) Analogously, one may obtain a sparse approximation for the corresponding LR patch x L. This requires a sparse coding coefficient vector α L over a LR dictionary D L. This approximation can be written as x L D L α L. (3) The same blurring and downsampling operator Ψ shown in (1) can be used to relate x L and x H. With the assumption that Ψ relates also the atoms of D L and D H, one 210

may write x L Ψx H ΨD H α H D L α H, (4) Based on the above analysis, the following result in concluded α H α L. Consequently, a reconstruction for x H can be obtained using D H and α L, as follows x H D H α L. (5) III. THE PROPOSED SUPER-RESOLUTION ALGORITHM In this section, we introduce the DPA as an approximately scale-invariant patch measure and empirically study its behavior with respect to scale. Then, we detail the training and testing stages of the proposed algorithm. Intuitively, phase values can characterize the directional structure of an image patch. One may define a dominant phase angle (DPA) in the phase matrix that best describes the directionality. This requires quantizing the the phase matrix and establishing a histogram for the quantized angles. It is empirically found that if a certain angle value is repeated more frequently than half of the number of elements in the phase matrix, then the patch can be considered a directional patch. The directionality of this patch is characterized by that dominant angle. Otherwise, it is considered as a non-directional patch. In this work, we opt to quantize angles into values of 0,45,90 and 135. TABLE I THE NUMBER OF HR PATCHES CLASSIFIED BASED ON DPA IN EACH CLUSTER (TOP) AND DPA SCALE-INVARIANCE RATIO IN THE SAME CLUSTER (BOTTOM). THE LISTINGS OF THE CLUSTER CONTAINING THE LARGEST NUMBER OF IMAGE PATCHES ARE IN BOLD FACE. Image C 1 C 2 C 3 C 4 C 5 Barbara 1819 580 466 508 3852 45.5 43.3 87.3 52.4 72.0 BSDS 198054 555 126 130 163 3266 71.9 58.7 61.5 55.2 82.7 Buildposter 1082 151 1761 135 11325 83.9 80.8 93.9 80.7 88.4 Butterfly 126 224 269 47 1098 81.0 74.6 92.9 72.3 66.0 Fence 624 49 252 12 827 68.9 65.3 59.5 66.7 76.5 Flowers 453 249 412 185 3681 72.4 64.3 71.1 68.6 79.6 Lena 894 171 157 474 5529 85.9 69.6 73.2 82.1 78.2 ppt3 636 346 814 299 7497 78.8 62.1 87.0 61.9 86.3 Text Image 1 288 1 8 3 11036 12.8 0.0 12.5 0.0 76.1 Average 66.4 60.1 85.7 66.2 81.0 A. Approximate scale-invariance of the dominant gradient phase angle measure The image gradient operator is believed to be crucial to the perception and analysis of natural images [5]. In [3], [4], Sun et al. defined the gradient profile prior as a 1-D profile of gradient magnitudes perpendicular to image structures, and studied its behavior with respect to scale. They reported that this profile follows a certain distribution which is independent of image resolution. Then they applied this prior to the problem of single-image SR. It is well-acknowledged that the phase is generally more informative than the magnitude. It seems thus advantageous to think about exploiting information in the phase of the gradient operator. One can define the phase matrix of the gradient operator based on the horizontal (G h ) and vertical (G v ) gradients as follows Φ = arctan( Gv ). (6) Gh Figure 1. Test images from left to right and top to bottom: Barbara, BSDS 198054, Buildposter, Butterfly, Fence, Flowers, Lena, ppt3 and Text Image 1. To investigate the impact of scale on DPA, the following experiment is conducted on the images shown in Figure 1. Each image is divided into non-overlapping 8 8 patches. A LR version of each image is obtained by filtering it with a bicubic filter and downsampling it by 2 in both dimensions. Clusters C 1 through C 4 are defined corresponding to DPA values of 0,45,90 and 135 respectively, while C 5 corresponds to patches that do not have a specific directional nature. Then the DPA value of each HR and LR patch is calculated as explained earlier. The DPA value of each patch is used to cluster it into one of the clusters. The DPA invariance is defined as the ratio between the number of LR patches correctly classified into a certain cluster and the total number of HR patches in that cluster. DPA invariance ratios for the images shown in Figure 1 are listed in Table I. One can see in Table I that the DPA invariance ratios are generally greater than 50 %. DPA invariance values are high for C 5, they are still high in C 1 and C 3 as compared to the cases of C 2 211

C 1 C 2 C 3 C 4 C 5 Figure 2. Reshaped example atoms of HR dictionaries in C 1 through C 5. and C 4. B. Clustering and sparse model selection with the dominant gradient phase angle measure In this work, we classify the training image patch pairs into the aforementioned five clusters. Then, we use the dictionary learning method of Yang et al. [2] to train for coupled LR and HR cluster dictionary pairs. Features extracted from the LR image patches at the MR level are used to train for the LR dictionary. The DPA of each LR patch at the MR level is used to classify it into a certain cluster. Then, the corresponding HR patch and the MR extracted features are inserted to the HR and LR training sets of that cluster, respectively. Algorithm 1 outlines the main steps of the training stage. In this setting, the purpose is to design cluster dictionaries that possess the intended directional structures of their respective clusters. Algorithm 1 The Proposed Dictionary Learning Stage. 1: INPUT: HR Training Image Set. 2: OUTPUT: A Set of Directionally Structured Cluster Dictionary Pairs. 3: Obtain a LR image for each HR one by blurring and downsampling. 4: Upsample each LR image to obtain a MR image. 5: Divide the HR and MR images into patches. 6: Extract features from the MR image by filtering. 7: Divide feature images into column patches. 8: Combine HR patches to form a HR training set and MR features to form a LR training set. 9: for Each patch in the LR training set, do 10: Calculate the DPA of the corresponding patch in the MR image, and identify the cluster number. 11: Set the MR features and the HR patch to the LR and HR training sets of this cluster. 12: end for 13: Learn a pair of coupled dictionaries for each cluster. In the reconstruction stage, a LR image is first upsampled with bicubic interpolation to the MR level. The MR image is then divided into overlapping patches. Next, the DPA of each MR patch is calculated and used to identify the cluster this patch belongs to. The next step is to calculate the sparse coding coefficients of the features extracted from the MR patch over the cluster LR dictionary. The corresponding HR patch is then reconstructed by imposing these coefficients on the HR dictionary of the same cluster. Algorithm 2 The Proposed Reconstruction Stage. 1: INPUT: A LR Image, Cluster Dictionary Pairs. 2: OUTPUT: A HR Image Estimate 3: Upsample the LR image to the required resolution level (MR). 4: Extract feature images from the MR image 5: Extract patches from the feature images and group them columnwise 6: Divide the MR image into overlapping patches. 7: for Each MR patch do 8: Calculate DPA of the MR patch. 9: Determine the cluster this patch belongs to. 10: Calculate the sparse coding coefficients of the corresponding feature vector over the cluster LR dictionary. 11: Reconstruct a HR patch as the product of the cluster HR dictionary and the calculated coefficients. 12: end for 13: Obtain a HR image estimate by merging overlapping HR patches. The final step is to reshape the overplaying HR patches and merge them to form a HR image estimate. A summary of the proposed reconstruction algorithm is outlined in Algorithm 2. Since the proposed algorithm relies on designing cluster dictionaries, they need not to be highly redundant. In his work, we choose to design 600-atom dictionaries as a good compromise between computational complexity and representation quality. This value is empirically determined to meet these requirements. The sparse coding stage is the most computationally expensive stage in the sparse reconstruction framework as it relies on vector selection [6]. Therefore, employing five compact dictionaries is expected to substantially reduce the sparse coding computational complexity and thus reduce the overall SR computational complexity, as compared to the case of employing a single highly redundant dictionary. As an example, Yang et al. [2] use a 1000-atom dictionary pair. However, the proposed algorithm requires calculating the DPA value of each MR patch. In view of these observations, it can be seen that the proposed algorithm s computation complexity is comparable to that of the algorithm of Yang et al. [2]. IV. EXPERIMENTAL RESULTS The performance of the proposed algorithm is compared to bicubic interpolation, the algorithm of Yang et al. [2] which uses a single dictionary pair and the algorithms 212

(a) (b) (c) (d) (e) (f) (g) (h) (i) Figure 3. Super-resolution comparison of a scene in the Butterfly image. (a) Original scene, (b) Bicubic interpolation result, (c) Peleg et al. s [16] result, (d) Yang et al. s [2] result, (e) He et al. s [17] result and (d) The proposed algorithm s result. The last row shows the difference between the original scene and (g) Yang et al. s result, (h) He et al. s result and (i) The proposed algorithm s result. of Peleg et al. [16] and He et al. [17] as state-of-theart algorithms. Comparisons are conducted in terms of the PSNR and SSIM [18] measures. In accordance with the common practice, we transform a color image into the luminance and chrominance color space and apply the SR algorithm only to the luminance component. The two chrominance components are reconstructed by bicubic interpolation. Accordingly, PSNR is calculated with the luminance components, where SSIM is calculated as the average of its values between the three color components. Results are obtained by running source codes obtained from the authors of [2], [16] and [17]. It is noted that we have paid attention to make equality of the simulation parameters of the four algorithms. Besides, each algorithm is set to work with the optimal parameters suggested by its authors. In this experiment, the proposed algorithm uses a 6 6 patch size, and five cluster dictionary pairs. We use the images in the Flickr image dataset [19] as a training set. Image patches are classified based on DPA into the five clusters. Then, 40,000 patch pairs are randomly selected and used to train for a cluster dictionary pair. The algorithm of Yang et al. [2] is set to the same parameters with the same training data set to train for a 1000-atom dictionary pair. The algorithms of Peleg et al. [16] and He et al. [17] are set to work according to the default values suggested by the authors. Figure 2 shows example reshaped atoms in the HR cluster dictionaries of the proposed algorithm. The atoms of these dictionaries clearly possess the directional structure of their respective clusters. From left to right, atoms of the first four dictionaries are generally perpendicular to the directions of 0,45,90 and 135 respectively. However, atoms in the fifth dictionary have a chaotic directional in accordance with the non-directional nature of C 5. Table II lists PSNR and SSIM values of the test images shown in Figure 1 with the aforementioned approaches. The performance of the proposed algorithm is shown with dictionaries of 600, 800 and 1000 atoms, as denoted by P 600, P 800 and, P 1000 respectively. With 600-atom dictio- 213

TABLE II QUANTITATIVE ASSESSMENT OF BICUBIC INTERPOLATION, THE ALGORITHMS OF PELEG ET AL. [16], YANGETAL. [2], HE ETAL. [17] AND THE PROPOSED ALGORITHM IN TERMS OF PSNR (TOP) AND SSIM (BOTTOM). Image Bicubic Yang [2] Peleg [16] He [17] P 1000 P 800 P 600 Barbara 25.35 25.86 25.76 25.84 25.89 25.87 25.87 0.7930 0.8357 0.8359 0.8372 0.8359 0.8351 0.8350 BSDS 198054 24.75 26.85 26.76 26.98 27.16 27.15 27.09 0.8267 0.8816 0.8760 0.8839 0.8847 0.8843 0.8837 Buildposter 29.95 32.40 32.07 32.77 32.76 32.80 32.60 0.9239 0.9526 0.9478 0.9572 0.9564 0.9563 0.9536 Butterfly 27.46 31.26 30.96 31.44 31.76 31.80 31.73 0.8985 0.9457 0.9227 0.9463 0.9485 0.9487 0.9480 Fence 25.05 26.34 26.17 26.22 26.41 26.42 26.49 0.7449 0.8037 0.7967 0.8045 0.8046 0.8053 0.8066 Flowers 30.42 32.76 32.57 32.98 32.94 32.92 32.88 0.8828 0.9005 0.8522 0.9018 0.9010 0.9009 0.9007 Lena 34.71 36.36 36.59 36.58 36.34 36.44 36.31 0.8507 0.8631 0.8387 0.8647 0.8630 0.8636 0.8628 ppt3 26.85 29.68 29.71 29.79 30.13 30.03 30.04 0.9372 0.9604 0.9494 0.9621 0.9640 0.9639 0.9634 Text Image 1 17.52 18.58 18.73 18.53 18.85 18.85 18.79 0.7246 0.7974 0.8118 0.7950 0.8119 0.8121 0.8080 Average 26.90 28.90 28.81 29.02 29.14 29.14 29.09 0.8425 0.8823 0.8701 0.8836 0.8855 0.8856 0.8846 naries, the proposed algorithm outperforms the algorithm of Yang et al. [2] and Peleg et al. [16] with average PSNR improvements of 0.19 db and 0.28 db, respectively. However, it is competitive to the algorithm of He et al. [17]. The same result is concluded in terms of SSIM. It is conclusive that using more redundancy in the dictionaries of the proposed algorithm will not significantly improve the performance. Besides, the proposed algorithm is particularly superior in handling images rich of edges in different orientations. In Figure 3 we compare our approach to bicubic interpolation and the algorithms of Peleg et al. [16], Yang et al. [2] and He et al. [17]. Figure 3 (g), (h) and (i) show the differences between the ground-truth scene and its reconstructions from Yang et al., He et al. and the proposed algorithm, respectively. Bicubic interpolation result is over-smooth. The reconstructions of Peleg et al. and Yang et al. have sharper edges. The algorithm of He et al. has further sharper edges. However, the proposed algorithm s reconstruction is the best to approximate the ground-truth scene with less artifacts. This is particularly observable in the edges and details of the butterfly s wing. This is assured by observing that the difference image of the proposed algorithm is the darkest one. V. CONCLUSION A new single-image super-resolution algorithm is proposed based on sparse coding over directionally-structured cluster dictionary pairs. Clustering is done based on the dominant angle in the gradient operator phase matrix. This angle is shown as an approximately scale-invariant patch measure. The same measure is used as a sparse model selection criterion in the reconstruction phase. With compact cluster dictionaries, the proposed algorithm is shown to be superior to the algorithm of Yang et al. [2] with an average 0.19 db PSNR improvement that uses a single dictionary pair. It is also shown to be competitive with the state-of-the-art algorithms. This result is validated in terms of PSNR and SSIM as quantities metrics and visual comparison. REFERENCES [1] J. Yang, J. Wright, Y. Ma and T. Huang, Image super-resolution as sparserepresentation of raw image patches, Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), 2008, pp. 1-8. [2] J. Yang, J. Wright, T.S. Huang and Y. Ma, Image super-resolution via sparse representation, IEEE Trans. Image Process., vol. 19, no. 11, 2010, pp. 2861-2873. [3] J. Sun, Z. Xu and H.Y. Shum, Imagesuper-resolution using gradient profile prior, Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), 2008, pp. 1-8. [4] J. Sun, Z. Xu and H.Y. Shum, Gradient Profile Prior and Its Applications in Image Super-Resolution and Enhancement, IEEE Trans. Image Process., vol. 20, no. 6, 2011, pp. 1529-1542. [5] W. Zuo, L. Zhang, C. Song, D. Zhang and H. Gao, Gradient Histogram Estimation and Preservation for Texture Enhanced Image Denoising, IEEE Trans. on Image Process., vol. 23, no. 6, 2014, pp. 2459-2472. [6] L.I. Rudin, S. Osher, and E. Fatemi, Nonlinear total variation based noise removal algorithms, Physica D, vol. 60, 1992, pp. 259-268. [7] R. Fergus, B. Singh, A. Hertzmann, S.T. Roweis and W.T. Freeman, Removing camera shake from a single photograph, In ACM Transactions on Graphics (TOG). vol. 25, no. 3, 2006, pp. 787-794. [8] A. Levin, Y. Weiss, F. Durand and W. T. Freeman, Efficient marginal likelihood optimization in blind deconvolution, Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), 2011, pp. 2657-2664. [9] J. Mairal, F. Bach, J. Ponce and G. Sapiro, Online learning for matrix factorization and sparse coding, J. Mach. Learn. Res., vol. 11, 2010, pp. 19-60. [10] D. Thanou, D. I Shuman and P. Frossard, Parametric dictionary learning for graph signals, IEEE Glob. Conf. Signal Inf. Process. (GlobalSIP), 2013, pp. 487-490. [11] S. Mallat and Y. Guoshen, Super-resolution with sparse mixing estimators, IEEE Trans. Image Process., vol. 19, no. 11, 2010, pp. 2889-2900. [12] W. Dong, L. Zhang, G. Shi and X. Wu, Image deblurring and super-resolution by adaptive sparsedomain selection and adaptive regularization, IEEE Trans. Image Process., vol. 20, no. 7, 2011, pp. 1838-1857. [13] S. Yang, M. Wang, Y. Chen and Y. Sun, Single-image superresolution reconstruction via learned geometric dictionaries and clustered sparse coding, IEEE Trans. Image Process., vol. 21, no. 9, 2012, pp. 4016-4028. [14] G. Yu, G. Sapiro and S. Mallat, Image modeling and enhancement via structured sparse model selection, Proc. IEEE Int. Conf. Image Process. (ICIP), 2010, pp. 1641-1644. [15] R. Tibshirani, Regression shrinkage and selection via the lasso, J. Royal. Statist. Soc. B, vol. 58, 1996, pp. 267-288. [16] T. Peleg and M. Elad, A statistical prediction model based on sparse representations for single image super-resolution, IEEE Trans. Image Process., vol. 23, no. 6, 2014, pp. 2569-2582. [17] H. Li, H. Qi and R. Zaretzki, Beta process joint dictionary learning for coupled feature spaces with application to single image super-resolution, Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), 2013, pp. 345-352. [18] Z. Wang, A.C. Bovik, H.R. Sheikh and E.P. Simoncelli, Image quality assessment: From error visibility to structural similarity, IEEE Trans. Image Process., vol. 13, no. 4, 2004, pp. 600-612. [19] http://see.xidian.edu.cn/faculty/wsdong/wsdong_downl oads.htm 214