Salient Object Segmentation based on Linearly Combined Affinity Graphs

Size: px
Start display at page:

Download "Salient Object Segmentation based on Linearly Combined Affinity Graphs"

Transcription

1 Salient Object Segmentation based on Linearly Combined Affinity Graphs Çağlar Aytekin, Alexandros Iosifidis, Serkan Kiranyaz * and Moncef Gabbouj Department of Signal Processing, Tampere University of Technology, Tampere, Finland * Department of Electrical Engineering, Qatar University, Doha, Qatar {caglar.aytekin, alexandros.iosifidis, moncef.gabbouj}@tut.fi, mkiranyaz@qu.edu.qa Abstract In this paper, we propose a graph affinity learning method for a recently proposed graph-based salient object detection method, namely Extended Quantum Cuts (EQCut). We exploit the fact that the output of EQCut is differentiable with respect to graph affinities, in order to optimize linear combination coefficients for several differentiable affinity functions by applying error backpropagation. We show that the learnt linear combination of affinities improves the performance over the baseline method and achieves comparable (or even better) performance when compared to the state-of-the-art salient object segmentation methods Keywords graph affinity learning, salient object segmentation, spectral graph theor I. INTRODUCTION Visual saliency estimation is an important research topic in computer vision where the task is to highlight the objects that visually stand out from their surroundings. Earlier research endeavors have focused mostly on models that are inspired by the human visual system [2] due to the fact that visually important region selection is related to human attention mechanisms [1]. Accordingly, some works have based their evaluation on the performance of accurately predicting human eye fixations [3]-[7]. However, the output of such methods corresponds to a (sparse) set of points (image pixels) and, thus, it is not very useful in terms providing an initial region of interest for higher level computer vision tasks, such as object recognition. Instead, a closed object segment is a more preferred input for these tasks. Hence, recently, the focus on visual saliency estimation is mostly concentrated on salient object segmentation [8]-[18], a task which has been proved to be useful for higher level computer vision and pattern recognition problems, such as tracking [26], object region proposals [28] and object recognition [27]. Graph-based salient object detection methods rank among the top performing methods in the literature. According to a recent benchmark [42], 4 out of the 6 top performing methods are based on graphs. All these methods exploit different graphbased approaches. For instance, in [12], the shortest path problem in graph theory [47] was exploited in order to find a robust backgroundness measure for the image regions. Then, this measure was combined with a region contrast measure in order to evaluate saliency. In [13], a Markov Chain was constructed on an image graph model. Virtual boundary nodes on this graph are selected as absorbing nodes and the saliency is calculated as the absorption time of the transient nodes. The method in [14] exploits the minimum barrier distance on graphs, in order to efficiently evaluate the saliency of image regions based on their deviation from the appearance of image boundary regions. The method in [17] and [18] solves a spectral approximation of a graph based cut problem, in order to find salient regions that both differ from image boundaries in appearance and are in high contrast with their surrounding, and have a large area. In [38], the authors give a general formulation for several salient object detection methods [11], [13] following a graph-based diffusion interpretation and propose a new diffusion based model by applying in-depth spectral analysis of the diffusion matrix. Aside from the fact that the aforementioned methods exploit graphs, they differ in terms of their graph-construction in many ways. For example [11], [12], [13], [18] and [38] exploit superpixels as nodes in the given graph representation whereas [14] and [17] define pixels as graph nodes. A 4- connected neighborhood rule for nodes was adopted in [14] and [17], in [12] every node was connected to adjacent nodes only, whereas in [11], [13] and [38] the nodes are also connected to the ones sharing common boundaries with their neighboring nodes. In [18] a very dense graph was constructed connecting nodes up to a 16 th degree of neighborhood. In [11], [12], [13] and [38] all nodes corresponding to the regions on the image boundary are connected to each other, whereas in [17] and [18] an artificial background node was introduced and all image boundary regions were connected to this artificial node with high weights. While constructing the graph weights between a pair of nodes, [12], [13] and [38] use an exponentially decaying function with increasing Euclidian distance between appearance features corresponding to each superpixel, whereas [17] and [18] use the inverse of this distance. The above-described differences in graph construction have a high impact on the performance of salient object detection methods. For instance, the difference between the methods in [18] and [17] lies on graph construction approach. Even if all remaining steps are the same, different graphs lead into a significant difference in performance. As will be shown later in this paper, graph construction can be even more effective than

2 the method itself. In addition, there might be cases where a method can only work well with a specific graph type. Based on the observations above, in this paper, we will focus on the graph construction problem and in particular we aim to improve the graph weights (affinity) computation for a better salient object segmentation. Conventionally, graph-based methods use a pre-defined function with fixed parameters mapping Euclidian distance of appearance features between a pair of nodes to an affinity value. In this paper, we propose a method that learns the parameters of three popular affinity functions along with the weights of their linear combination. We employ Extended Quantum Cuts (EQCut) [18] as salient object segmentation method. We exploit the fact that EQCut produces saliency maps that are differentiable with respect to graph affinities and this in turn allows us to backpropagate the salient object detection error in order to optimize the aforementioned parameters of the overall system. The rest of the paper is organized as follows: In Section II, EQCut is briefly introduced and the proposed graph affinity learning strategy is described. In Section III we show that the learnt affinities lead to a significant performance improvement over the baseline EQCut. Finally, Section IV concludes the paper and suggests future topics for research. II. METHOD In this section we first briefly describe the conventional EQCut algorithm that will be exploited. Next, we describe the proposed affinity learning approach. A. Extended Quantum Cuts Extended Quantum Cuts (EQCut) [18] is based on a set of improvements over Quantum Cuts (QCut) [17], which is a spectral foreground detection method exploiting the link between graph theory and quantum mechanics. In a graph representation of an image, QCut approximates the following optimization problem: A cut(a, A ) = argmin A area(a), (1) where A is the foreground segment, A is the background segment (rest of the image), area(a) is the number of nodes in segment A and the cut-cost is defined as cut(a, A ) = i A,j A w i,j. Here, w i,j is the affinity (similarity) between nodes i and j. Let y b be a a binary indicator vector representing A, that takes a value of 1 in foreground nodes and 0 elsewhere. It has been proven in [17] that, if y b is relaxed to have real values, the relaxed solution to (1) is given as follows: y = z z, Hz = λz, (2) where λ is the minimum eigenvalue of the Hamiltonian matrix H, which is formulated as H = D W + V. Here, W is constructed from the affinities w i,j, D is a diagonal degree matrix whose elements are obtained by row-wise summation of W and V is a non-negative diagonal matrix of background priors for each node. The above representation exhibits a great similarity with the solutions for quantum states of a sub-atomic particle which were previously investigated for salient object segmentation in [19]. EQCut investigates the design of the graph that represents the image and its affinity matrix W, both of which highly affect the performance in salient object detection. The extensions proposed in [18] include a superpixel representation of the image, a novel graph affinity normalization strategy on a graph with higher connectivity and a multiresolution approach. The proposed extensions lead to a significant improvement over QCut [17]. It should be noted here that the graph affinity normalization in EQCut is based on the neighbourhood degrees between pairs of superpixels and they result into an asymmetric affinity matrix. In this paper we will exploit a symmetric affinity matrix by simply averaging the asymmetric one with its transposed. B. Proposed Method In this study, we use the same graph construction approach with the symmetric version of graph affinity normalization in EQCut and we will investigate the effect of optimally combining affinity functions in salient object detection performance based on EQCut. In [17] and [18], the following affinity function was exploited: w (1) 1 i,j = ε + f i f j 2, (3) where f i is the related feature vector for the superpixel i and ε is a small constant to prevent division by zero. A property of the above-given affinity function is that it greatly penalizes the Euclidian distance between two feature vectors. An exponential function which is relatively smoother was used in some salient object segmentation methods [12], [13], [38] expressed as: w (2) i,j = e f i f j σ 2, where σ is a smoothing factor controlling the affinity decrease rate with the increasing feature distance. Finally, a linear affinity function can also be used as follows: (4) w (3) i,j = 1 f i f j. (5) Note that in all affinity functions, the Euclidian distances are unit-normalized in such a way that among all possible feature distances in a given image, the maximum is set to 1. Hence, it is guaranteed that (5) does not produce negative affinities. At this point we would like to note that, for a given image, affinities obtained using (3) are normalized such that the maximum affinity will have a value of 1. The affinities obtained by (4) and (5) are not needed to be normalized since they already produce outputs between 0 and 1. The proposed method, instead of using a single affinity function with fixed parameters, makes use of a linear combination of them, as follows: w i,j = c 1 w (1) i,j + c 2 w (2) i,j + c 3 w (3) i,j. (6)

3 Furthermore, we consider ε in (3) and σ in (4) as parameters to be learned. Hence, we aim to learn the parameter set θ = {c 1, c 2, c 3, ε, σ} in order to achieve a high performance in salient object detection. In order to learn affinities that will perform well on salient object detection, we exploit the observation that the saliency maps obtained via (2) are differentiable with respect to affinities of the graph that is used. This property gives us the opportunity to backpropagate the error related to the output of EQCut with respect to the parameters we wish to optimize. Exploiting the eigenvector derivatives for symmetric matrices as presented in [48], one can write the following: z h = [z (λi H) ]Dpl, where λ is the minimum eigenvector of H, z is the corresponding eigenvector, h is the vectorized version of H by row-wise raster scan, (. ) denotes the Moore-Penrose pseudoinverse, Dpl is the duplication matrix [48] which converts the half-vectorizations of symmetric matrices to full vectorizations, I is the identity matrix and is the Kronecker product. In order to backpropagate the error e from EQCut output y to affinities, we can write the partial derivatives of the error corresponding to i th superpixel with respect to the entire H matrix as follows: (7) e i H = ( e i y i ) ( y i z i ) ( z i H ). (8) We can extract ( z i H ) component of (8) from (7) and ( y i z i ) = 2z i as observed from (2). Only remaining component ( e i y i ) depends on the error function that is chosen for performance evaluation of the EQcut output. We have used the following error function and the corresponding partial derivatives of it with respect to y: e i = y i g i, ( e i y i ) = sgn(y i g i ), where g i is the ground-truth label for superpixel i. The groundtruth map takes the value 1 for superpixels corresponding to salient regions and 0 otherwise. At this point it should be noted that since y is l 1 normalized due to (2), elements of y never gets the value 1. Furthermore, due to numerical precision they do not become exactly equal to 0 either. Hence, (9) always produces an error derivative of 1 for background regions and 1 for salient object regions. We have observed that this kind of enforcement yields much better performance, mainly because of the fact that it always forces the network to give higher values to salient regions and suppress background regions. Other error functions such as Euclidian distance between y and l 1 normalized ground truth map, produce small gradients once y is close to l 1 normalized ground truth map, hence converges to a suboptimal solution. (9) Based on the above discussion, (8) can be calculated. Accordingly, partial derivatives with respect to W can be calculated as follows: e i W = 1T d ei,h d T ei,h e i H, (10) where d ei,h is a vector obtained by the diagonal elements of e i H. Finally, the derivative for the average error e per node can be calculated as follows: e W = 1 N e i. W i (11) Next, partial derivatives of affinities with respect to parameter set θ are needed to be calculated. The partial derivatives of affinities with respect to linear combination coefficient c k, k=1,2,3, can be calculated as follows: w i,j c k = w (k) i,j, (12) The partial derivatives of affinities with respect to parameters ε and σ can be given as follows: w i,j ε w i,j σ c 1 = (ε + f i f j 2 2, (13) ) 2c (2) 2w i,j fi f j = σ 3, (14) Combining (7)-(14), one can obtain the partial derivatives of the error e from EQCut s output y with respect to the affinity calculation parameter set θ. Hence, given a training dataset, parameter set θ can be learned by iteratively backpropagating the error to update θ and calculating the new output until convergence based on a stopping rule that will be explained in the next section. III. EXPERIMENTAL RESULTS In this section we provide extensive experiments conducted in order to evaluate the performance of the proposed method exploiting learned affinities for EQCut-based saliency object detection. Comparisons are made with EQCut as well as the state-of-the-art salient object detection methods. A. Datasets SOD [33] dataset contains 300 images that includes many salient objects and a high background clutter. JUDD [41] dataset is the most challenging dataset containing 900 images in total with multiple salient objects and high background clutter. MSRA10k [9] contains 10k images with relatively easier examples with large objects having high contrast. Dut- OMRON [11] contains 5168 images with one or more salient objects per image and high background clutter. Pascal1500 [49] contains 1500 images with more than one salient object at a variety of locations and scales and challenging background.

4 B. Evaluation Metrics In our experiments, we use precision-recall curves and maximum and mean F 1 scores, as performance metrics. Precision is the ratio of the total number of true positives to the total number of detected samples. Recall is the ratio of the number of true positives to the number of true samples. For salient object detection, a true sample is a pixel in the groundtruth segmentation mask G. A detected sample at a threshold τ is the number of nonzero elements in a saliency map thresholded with τ, denoted by S τ. pre(τ) = G S τ, rec(τ) = G S τ S τ G (15) F-measure is defined according to precision and recall values as follows: in the original EQCut affinity calculation as in (3). As the distances increase, the affinities decrease more smoothly when compared to (3) due to contribution of (4) and (5). F 1 (τ) = 2pre(τ)rec(τ) pre(τ) + rec(τ) (16) Maximum F 1 measure is then simply obtained by taking the maximum of the F 1 measure over all thresholds applied. C. Training In all our experiments, Dut-OMRON was used during the training phase in order to learn the parameters θ, as it is the one of the largest datasets with the highest complexity. A portion of Dut-OMRON could also have been used as a validation set, however the images within each dataset have some correlation, for example Dut-OMRON dataset contains several images that exhibit great similarity from some classes. Therefore, in order to have a system that generalizes well, we have used an entirely different dataset (SOD) for validation purposes. We have initialized the parameters with θ 0 = {1,0,0, 10 6, 0.1}. This initialization corresponds to the affinity calculation in EQCut and σ = 0.1 was set as in [12], [13] and [38]. At each iteration, the images in Dut-OMRON dataset are randomly shuffled. Next, in an epoch, 50 training samples were chosen from the shuffled indices and the affinity function parameters were updated via error backpropagation with a minibatch size of 1. The step size is scaled by 0.99 in each consecutive iteration. At each epoch, the maximum F 1 measure for the validation dataset is measured. Training is stopped when the error converges, or when a pre-defined maximum iteration number is achieved. The parameter values corresponding to the epoch that produced the best performance on the validation set were selected as the final parameters. D. Comparison to EQCut The Here we compare the performance of the proposed method (EQCut-A) with EQCut. First, three separate parameter sets were learned for three superpixel granularities of EQCut (150, 300 and 600 superpixels) according to the details explained in the previous section. The log affinities of the learnt functions and the ones calculated by w (1) in (3), w (2) in (4) and w (3) in (5) are illustrated in Fig. 1. The learnt affinity functions lead to very high affinities in very small distances as Fig. 1 Log affinities for w (1), w (2), w (3) and learnt affinities for 150, 300 and 600 superpixel (SP) granularities. In Fig. 2, maximum and mean F1 measures for the test datasets (i.e., Pascal1500, MSRA10k and Judd) are illustrated. For each superpixel granularity and for each test set, we observe a clear improvement over the maximum and mean F1 measures. The improvement is much clearer as the dataset gets more difficult. For instance, MSRA10k is the simplest dataset and although there is a performance improvement with the learnt affinities, it is rather incremental. However, for challenging datasets Pascal1500 and Judd, the performance improvement is significant, reaching up to 7% relative improvement over the EQCut in maximum F1 measures, respectively. Fig. 2 Maximum F1 measure of baseline EQCut and EQCut with learned affinities (EQCut-A). Performances are given for Pascal (P), Judd (J) and MSRA10k(M) datasets and 150, 300 and 600 superpixel granularities.

5 Fig. 3 Performance comparisons of EQCut-A, EQCut, GP, RBD, ST, DSR, MC and DRFI in precision-recall curves (Column 1 for JUDD and 2 for MSRA10k datasets) and in maximum F1measures (Column 3 for JUDD and 4 for MSRA10k datasets) E. Comparison with the State of the Art In this section we compare the performance of EQCut-A with the state of the art salient object detection methods according to a recent benchmark study [42], namely EQCut [18], MC [13], RBD [12], ST [15], DSR [8] and DRFI [16]. Furthermore, we include one more method that became the state-of-the-art and published after the benchmark, namely GP [38]. For the comparisons in this section, we have employed the multi-resolution approach in [18] and combined all saliency maps at each superpixel granularity by simply averaging them. This was made for both EQCut and EQCut-A for superpixel granularities 150, 300 and 600. The performances are given in precision-recall curves and maximum F1 measures for Judd and MSRA10k datasets. Pascal-1500 dataset was excluded from comparisons since we couldn t reach all saliency map results/codes of the compared methods for this dataset. Nevertheless, Judd and MSRA10k datasets cover a large variety of cases where the former datasets include salient objects with highest difficulty, whereas the latter includes simplest cases in salient object detection task. In both datasets we observe consistent improvement over EQCut, moreover, in Judd database, EQCut-A performs better than all state-of-theart methods and in MSRA10k dataset, EQCut-A is better than all the methods except for DRFI, while it is comparable with DRFI in this dataset F. Computational Complexity In terms of computational complexity, the proposed affinity calculations do not bring any significant burden over EQCut. For the saliency calculations on SOD dataset, the average runtimes for EQCut and EQCUT-A are and seconds, respectively. Therefore, the total increase in the computational complexity is only around 11%. IV. CONCLUSION In this paper, we have proposed a graph affinity and parameter learning method for salient object detection. Specifically, linear combination coefficients combining several affinity functions and their parameters are learned altogether such that the resulting affinity matrix will lead to better saliency results via the exploited salient object detection method EQCut. The proposed affinities lead to a consistent improvement over EQCut in precision-recall curves and maximum F1 measures, yet they do not bring any significant computational complexity rise. Moreover, EQCut with the learnt affinities performs the best or comparable to the state-ofthe-art in salient object segmentation. REFERENCES [1] S. K. Ungerleider, and G. Leslie, "Mechanisms of visual attention in the human cortex," Annual Review of Neuroscience, vol. 23, no. 1, pp , [2] L. Itti, C. Koch, and E. Niebur, A model of saliency-based visual attention for rapid scene analysis, IEEE Trans. Pattern Anal. Mach. Intell., vol. 20, no. 11, pp , [3] J. Harel, C. Koch, and P. Perona, Graph-based visual saliency, Conference on Neural Information Processing Systems, pp , [4] I. Rigas, G. Economou and S. Fotopoulos, Efficient Modelling of Visual Saliency Based on Local Sparse Representation and the Use of Hamming Distance, Computer Vision and Image Understanding, vol. 134, pp 33-45, [5] D. Gao, V. Mahadevan, and N.Vasconcelos, The discriminant centersurround hypothesis for bottom-up saliency, Conference on Neural Information Processing Systems, pp , [6] T. Judd, K. Ehinger, F. Durand, and A. Torralba, Learning to predict where humans look, in Proc. IEEE International Conference on Computer Vision, pp , [7] J. Han, L. Sun, X. Hu, J. Han and L. Shao, Spatial and temporal visual attention prediction in videos using eye movement data, Neurocomputing, vol. 145, pp , [8] X. Li, H. Lu, L. Zhang, X. Ruan and M.H. Yang, "Saliency detection via dense and sparse reconstruction," in Proc. IEEE International Conference on Computer Vision, pp , [9] M.M. Cheng, G. Zhang, N. Mitra, X. Huang, and S.Hu, Global contrast based salient region detection, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 37, no.3, pp , [10] Q. Yan, L. Xu, J. Shi, and J. Jia, Hierarchical saliency detection, in Proc. IEEE Conference on Computer Vision and Pattern Recognition, pp , [11] C. Yang, L. Zhang, H. Lu, X. Ruan, and M.-H. Yang, Saliency detection via graph-based manifold ranking, IEEE Conference on Computer Vision and Pattern Recognition, pp , [12] W. Zhu, S. Liang, Y. Wei and J. Sun, Saliency Optimization from Robust Background Detection, IEEE Conference on Computer Vision and Pattern Recognition, pp , [13] B. Jiang, L. Zhang, H. Lu, C. Yang and M-H. Yang, Saliency Detection via Absorbing Markov Chain, IEEE International Conference on Computer Vision, pp , [14] J. Zhang, S. Sclaroff, Z. Lin, X. Shen, B. Price, and R. Mechi, Minimum Barrier Salient Object Detection at 80 FPS, IEEE International Conference on Computer Vision, pp , [15] Z. Liu, W. Zou, and O. Le Meur. "Saliency tree: A novel saliency detection framework." IEEE Transactions on Image Processing, vol. 23 no. 5, pp , [16] H. Jiang, J. Wang, Z. Yuan, Y. Wu, N. Zheng, and S. Li. "Salient object detection: A discriminative regional feature integration approach." IEEE

6 Conference on Computer Vision and Pattern Recognition, pp , [17] C. Aytekin, S. Kiranyaz and M. Gabbouj, Automatic Object Segmentation by Quantum Cuts, International Conference on Pattern Recognition, pp , [18] C. Aytekin, E. C. Ozan, S. Kiranyaz, and M. Gabbouj, Visual saliency by extended quantum cuts, IEEE International Conference on Image Processing, pp , [19] C. Aytekin, S. Kiranyaz and M. Gabbouj, Quantum mechanics in Computer Vision: Automatic Object Extraction, IEEE International Conference on Image Processing, pp , [20] D. Rudoy, D.B. Goldman, E. Shechtman and L. Z. Manor, Learning Video Saliency from Human Gaze Using Candidate Selection, IEEE Conference on Computer Vision and Pattern Recognition, pp , [21] H. Hadizadeh and I.V. Bajic, Saliency-Aware Video Compression, IEEE Transactions on Image Processing, vol. 23, no.1, pp , [22] R. Margolin, L. Z. Manor and A. Tal, Saliency for Image Manipulation, The Visual Computer, vol. 29, no. 5, pp , [23] X. Hou and L. Zhang, Thumbnail Generation based on Global Saliency, Advances in Cognitive Neurodynamics, pp , [24] M. Casares, S. Velipasalar and A. Pinto, Light-weight Salient Foreground Detection for Embedded Smart Cameras, Computer Vision and Image Understanding, vol. 114, no. 11, pp , [25] Z. Li, S. Qin and L. Itti, Visual attention guided bit allocation in video compression, Image and Vision Computing, vol. 29, no. 1, pp. 1-14, [26] C. Aytekin, E. Tunali, and S. Oz, Fast Semi-automatic Target Initialization based on Visual Saliency for Airborne Thermal Imagery, In International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications, pp , [27] E. Horbert, G. M. García, S. Frintrop and B. Leibe, Sequence-level object candidates based on saliency for generic object recognition on mobile systems, International Conference on Robotics and Automation, pp , [28] C. Aytekin, S. Kiranyaz, and M. Gabbouj, "Learning to rank salient segments extracted by multispectral quantum cuts." Pattern Recognition Letters, vol. 72, pp , [29] [30] R. Achanta, S. Hemami, F. Estrada and S. Susstrunk, "Frequency-tuned salient region detection." IEEE Conference on Computer Vision and Pattern Recognition, pp , [31] R. Achanta and S. Susstrunk, Saliency detection using maximum symmetric surround, International Conference on Image Processing, pp , [32] W. Zou, K. Kpalma, Z. Liu, and J. Ronsin, Segmentation driven lowrank matrix recovery for saliency detection, in Proc. British Machine Vision Conference, pp. 1-13, [33] V. Mohavedi and J. H. Elder, Design and perceptual validation of performance measures for salient object segmentation, IEEE Computer Vision and Pattern Recognition Workshops, pp , [34] S. Alpert, M. Galun, R. Basri, and A. Brandt, Image Segmentation by Probabilistic bottom-up aggregiation and cue integration, IEEE Conference on Computer Vision and Pattern Recognition, pp. 1-8, [35] G. Li, Y. Yizhou "Visual Saliency Based on Multiscale Deep Features," arxiv preprint arxiv: , [36] A. Borji, M. M. Cheng, H. Jiang and J. Li, "Salient object detection: A survey," arxiv preprint arxiv: , [37] R.R. Coifman, and S. Lafon. "Diffusion maps," Applied and computational harmonic analysis, vol. 21, no. 1, pp. 5-30, [38] P. Jiang, N. Vasconcelos, and J. Peng. "Generic Promotion of Diffusion- Based Salient Object Detection," IEEE International Conference on Computer Vision, pp , [39] D. Zhou, J. Weston, A. Gretton, O. Bousquet, and B. Schölkopf, "Ranking on data manifolds," Advances in Neural Information Processing Systems, pp , [40] D. Zhou, O. Bousquet, T.N. Lal, J. Weston, and B. Schölkopf, "Learning with local and global consistency," Advances in Neural Information processing systems, pp , [41] A. Borji. "What is a salient object? a dataset and a baseline model for salient object detection," IEEE Transactions on Image Processing, vol 24, no. 2, pp , [42] A. Borji, M. M. Cheng, H. Jiang and J. Li. "Salient object detection: A benchmark," IEEE Transactions on Image Processing, vol. 24, no. 12, pp , [43] L. M. Manevitz, and M. Yousef. "One-class SVMs for document classification," The Journal of Machine Learning Research, vol. 2 pp , [44] B. Schölkopf, R. Herbrich, and A. J. Smola, A generalized representer theorem, Computational Learning Theory, pp , [45] G.D. Poole, "Generalized M-matrices and applications," Mathematics of Computation, vol. 29, no.131, pp , [46] B. Scholkopf, and A.J. Smola, Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond, Cambridge MA: MIT press, [47] D.B. West, Introduction to graph theory (2 nd ed.), Upper Saddle River: Prentice Hall, [48] J.R. Magnus, On differentiating eigenvalues and eigenvectors, Econometric Theory, vol. 1, no. 2, pp , [49] W. Zou, K. Kpalma, Z. Liu and J. Rosnin, Segmentation driven lowrank matrix recovery for saliency detection, British Machine Vision Conference, 2013.

Supplementary Materials for Salient Object Detection: A

Supplementary Materials for Salient Object Detection: A Supplementary Materials for Salient Object Detection: A Discriminative Regional Feature Integration Approach Huaizu Jiang, Zejian Yuan, Ming-Ming Cheng, Yihong Gong Nanning Zheng, and Jingdong Wang Abstract

More information

Supplementary Material for submission 2147: Traditional Saliency Reloaded: A Good Old Model in New Shape

Supplementary Material for submission 2147: Traditional Saliency Reloaded: A Good Old Model in New Shape Supplementary Material for submission 247: Traditional Saliency Reloaded: A Good Old Model in New Shape Simone Frintrop, Thomas Werner, and Germán M. García Institute of Computer Science III Rheinische

More information

Data-driven Saliency Region Detection Based on Undirected Graph Ranking

Data-driven Saliency Region Detection Based on Undirected Graph Ranking Data-driven Saliency Region Detection Based on Undirected Graph Ranking Wenjie Zhang ; Qingyu Xiong 2 ; Shunhan Chen 3 College of Automation, 2 the School of Software Engineering, 3 College of Information

More information

Unsupervised Saliency Estimation based on Robust Hypotheses

Unsupervised Saliency Estimation based on Robust Hypotheses Utah State University DigitalCommons@USU Computer Science Faculty and Staff Publications Computer Science 3-2016 Unsupervised Saliency Estimation based on Robust Hypotheses Fei Xu Utah State University,

More information

Salient Region Detection and Segmentation in Images using Dynamic Mode Decomposition

Salient Region Detection and Segmentation in Images using Dynamic Mode Decomposition Salient Region Detection and Segmentation in Images using Dynamic Mode Decomposition Sikha O K 1, Sachin Kumar S 2, K P Soman 2 1 Department of Computer Science 2 Centre for Computational Engineering and

More information

Saliency Detection for Videos Using 3D FFT Local Spectra

Saliency Detection for Videos Using 3D FFT Local Spectra Saliency Detection for Videos Using 3D FFT Local Spectra Zhiling Long and Ghassan AlRegib School of Electrical and Computer Engineering, Georgia Institute of Technology, Atlanta, GA 30332, USA ABSTRACT

More information

SALIENT OBJECT DETECTION FOR RGB-D IMAGE VIA SALIENCY EVOLUTION.

SALIENT OBJECT DETECTION FOR RGB-D IMAGE VIA SALIENCY EVOLUTION. SALIENT OBJECT DETECTION FOR RGB-D IMAGE VIA SALIENCY EVOLUTION Jingfan Guo,2, Tongwei Ren,2,, Jia Bei,2 State Key Laboratory for Novel Software Technology, Nanjing University, China 2 Software Institute,

More information

Saliency Detection via Graph-Based Manifold Ranking

Saliency Detection via Graph-Based Manifold Ranking Saliency Detection via Graph-Based Manifold Ranking Chuan Yang, Lihe Zhang, Huchuan Lu, Xiang Ruan 2, and Ming-Hsuan Yang 3 Dalian University of Technology 2 OMRON Corporation 3 University of California

More information

IMAGE SALIENCY DETECTION VIA MULTI-SCALE STATISTICAL NON-REDUNDANCY MODELING. Christian Scharfenberger, Aanchal Jain, Alexander Wong, and Paul Fieguth

IMAGE SALIENCY DETECTION VIA MULTI-SCALE STATISTICAL NON-REDUNDANCY MODELING. Christian Scharfenberger, Aanchal Jain, Alexander Wong, and Paul Fieguth IMAGE SALIENCY DETECTION VIA MULTI-SCALE STATISTICAL NON-REDUNDANCY MODELING Christian Scharfenberger, Aanchal Jain, Alexander Wong, and Paul Fieguth Department of Systems Design Engineering, University

More information

Video saliency detection by spatio-temporal sampling and sparse matrix decomposition

Video saliency detection by spatio-temporal sampling and sparse matrix decomposition Video saliency detection by spatio-temporal sampling and sparse matrix decomposition * Faculty of Information Science and Engineering Ningbo University Ningbo 315211 CHINA shaofeng@nbu.edu.cn Abstract:

More information

AS a subfield of computer vision and pattern recognition,

AS a subfield of computer vision and pattern recognition, 1 Saliency Detection with Spaces of Background-based Distribution Tong Zhao, Lin Li, Xinghao Ding, Yue Huang and Delu Zeng* arxiv:1603.05335v1 [cs.cv] 17 Mar 2016 Abstract In this letter, an effective

More information

Image Compression and Resizing Using Improved Seam Carving for Retinal Images

Image Compression and Resizing Using Improved Seam Carving for Retinal Images Image Compression and Resizing Using Improved Seam Carving for Retinal Images Prabhu Nayak 1, Rajendra Chincholi 2, Dr.Kalpana Vanjerkhede 3 1 PG Student, Department of Electronics and Instrumentation

More information

Supplementary Material: Unconstrained Salient Object Detection via Proposal Subset Optimization

Supplementary Material: Unconstrained Salient Object Detection via Proposal Subset Optimization Supplementary Material: Unconstrained Salient Object via Proposal Subset Optimization 1. Proof of the Submodularity According to Eqns. 10-12 in our paper, the objective function of the proposed optimization

More information

Co-Saliency Detection Based on Hierarchical Segmentation

Co-Saliency Detection Based on Hierarchical Segmentation 88 IEEE SIGNAL PROCESSING LETTERS, VOL. 21, NO. 1, JANUARY 2014 Co-Saliency Detection Based on Hierarchical Segmentation Zhi Liu, Member, IEEE, Wenbin Zou, Lina Li, Liquan Shen, and Olivier Le Meur Abstract

More information

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference Minh Dao 1, Xiang Xiang 1, Bulent Ayhan 2, Chiman Kwan 2, Trac D. Tran 1 Johns Hopkins Univeristy, 3400

More information

ACCORDING to the principle of human visual perception,

ACCORDING to the principle of human visual perception, IEEE SIGNAL PROCESSING LETTERS, VOL. XX, NO. XX, XXXX 2016 1 Saliency Detection for Stereoscopic Images Based on Depth Confidence Analysis Multiple Cues Fusion Runmin Cong, Student Member, IEEE, Jianjun

More information

A Bi-directional Message Passing Model for Salient Object Detection

A Bi-directional Message Passing Model for Salient Object Detection A Bi-directional Message Passing Model for Salient Object Detection Lu Zhang, Ju Dai, Huchuan Lu, You He 2, ang Wang 3 Dalian University of Technology, China 2 Naval Aviation University, China 3 Alibaba

More information

Image Resizing Based on Gradient Vector Flow Analysis

Image Resizing Based on Gradient Vector Flow Analysis Image Resizing Based on Gradient Vector Flow Analysis Sebastiano Battiato battiato@dmi.unict.it Giovanni Puglisi puglisi@dmi.unict.it Giovanni Maria Farinella gfarinellao@dmi.unict.it Daniele Ravì rav@dmi.unict.it

More information

Hierarchical Co-salient Object Detection via Color Names

Hierarchical Co-salient Object Detection via Color Names Hierarchical Co-salient Object Detection via Color Names Jing Lou Fenglei Xu Qingyuan Xia Mingwu Ren School of Computer Science and Engineering Nanjing University of Science and Technology Nanjing 210094

More information

2.1 Optimized Importance Map

2.1 Optimized Importance Map 3rd International Conference on Multimedia Technology(ICMT 2013) Improved Image Resizing using Seam Carving and scaling Yan Zhang 1, Jonathan Z. Sun, Jingliang Peng Abstract. Seam Carving, the popular

More information

Small Object Segmentation Based on Visual Saliency in Natural Images

Small Object Segmentation Based on Visual Saliency in Natural Images J Inf Process Syst, Vol.9, No.4, pp.592-601, December 2013 http://dx.doi.org/10.3745/jips.2013.9.4.592 pissn 1976-913X eissn 2092-805X Small Object Segmentation Based on Visual Saliency in Natural Images

More information

Semi-supervised Data Representation via Affinity Graph Learning

Semi-supervised Data Representation via Affinity Graph Learning 1 Semi-supervised Data Representation via Affinity Graph Learning Weiya Ren 1 1 College of Information System and Management, National University of Defense Technology, Changsha, Hunan, P.R China, 410073

More information

A New Framework for Multiscale Saliency Detection Based on Image Patches

A New Framework for Multiscale Saliency Detection Based on Image Patches Neural Process Lett (2013) 38:361 374 DOI 10.1007/s11063-012-9276-3 A New Framework for Multiscale Saliency Detection Based on Image Patches Jingbo Zhou Zhong Jin Published online: 8 January 2013 Springer

More information

Human Detection and Tracking for Video Surveillance: A Cognitive Science Approach

Human Detection and Tracking for Video Surveillance: A Cognitive Science Approach Human Detection and Tracking for Video Surveillance: A Cognitive Science Approach Vandit Gajjar gajjar.vandit.381@ldce.ac.in Ayesha Gurnani gurnani.ayesha.52@ldce.ac.in Yash Khandhediya khandhediya.yash.364@ldce.ac.in

More information

Saliency Detection via Dense and Sparse Reconstruction

Saliency Detection via Dense and Sparse Reconstruction 23 IEEE International Conference on Computer Vision Saliency Detection via Dense and Sparse Reconstruction Xiaohui Li, Huchuan Lu, Lihe Zhang, Xiang Ruan 2, and Ming-Hsuan Yang 3 Dalian University of Technology

More information

Journal of Engineering Science and Technology Review 10 (3) (2017) Research Article

Journal of Engineering Science and Technology Review 10 (3) (2017) Research Article Jestr Journal of Engineering Science and Technology Review 10 (3) (017) 188-194 Research Article Salient Region Detection by Integrating Intrinsic and Extrinsic Cues without Prior Information Ji Ma 1,*,

More information

Hierarchical Saliency Detection Supplementary Material

Hierarchical Saliency Detection Supplementary Material Hierarchical Saliency Detection Supplementary Material Qiong Yan Li Xu Jianping Shi Jiaya Jia The Chinese University of Hong Kong {qyan,xuli,pshi,leoia}@cse.cuhk.edu.hk http://www.cse.cuhk.edu.hk/leoia/proects/hsaliency/

More information

An Improved Image Resizing Approach with Protection of Main Objects

An Improved Image Resizing Approach with Protection of Main Objects An Improved Image Resizing Approach with Protection of Main Objects Chin-Chen Chang National United University, Miaoli 360, Taiwan. *Corresponding Author: Chun-Ju Chen National United University, Miaoli

More information

DETECTION OF IMAGE PAIRS USING CO-SALIENCY MODEL

DETECTION OF IMAGE PAIRS USING CO-SALIENCY MODEL DETECTION OF IMAGE PAIRS USING CO-SALIENCY MODEL N S Sandhya Rani 1, Dr. S. Bhargavi 2 4th sem MTech, Signal Processing, S. J. C. Institute of Technology, Chickballapur, Karnataka, India 1 Professor, Dept

More information

A Survey on Detecting Image Visual Saliency

A Survey on Detecting Image Visual Saliency 1/29 A Survey on Detecting Image Visual Saliency Hsin-Ho Yeh Institute of Information Science, Acamedic Sinica, Taiwan {hhyeh}@iis.sinica.edu.tw 2010/12/09 2/29 Outline 1 Conclusions 3/29 What is visual

More information

An efficient face recognition algorithm based on multi-kernel regularization learning

An efficient face recognition algorithm based on multi-kernel regularization learning Acta Technica 61, No. 4A/2016, 75 84 c 2017 Institute of Thermomechanics CAS, v.v.i. An efficient face recognition algorithm based on multi-kernel regularization learning Bi Rongrong 1 Abstract. A novel

More information

Improving Image Segmentation Quality Via Graph Theory

Improving Image Segmentation Quality Via Graph Theory International Symposium on Computers & Informatics (ISCI 05) Improving Image Segmentation Quality Via Graph Theory Xiangxiang Li, Songhao Zhu School of Automatic, Nanjing University of Post and Telecommunications,

More information

Dynamic visual attention: competitive versus motion priority scheme

Dynamic visual attention: competitive versus motion priority scheme Dynamic visual attention: competitive versus motion priority scheme Bur A. 1, Wurtz P. 2, Müri R.M. 2 and Hügli H. 1 1 Institute of Microtechnology, University of Neuchâtel, Neuchâtel, Switzerland 2 Perception

More information

Saliency Detection using Region-Based Incremental Center-Surround Distance

Saliency Detection using Region-Based Incremental Center-Surround Distance Saliency Detection using Region-Based Incremental Center-Surround Distance Minwoo Park Corporate Research and Engineering Eastman Kodak Company Rochester, NY, USA minwoo.park@kodak.com Mrityunjay Kumar

More information

How to Evaluate Foreground Maps?

How to Evaluate Foreground Maps? How to Evaluate Foreground Maps? Ran Margolin Technion Haifa, Israel margolin@tx.technion.ac.il Lihi Zelnik-Manor Technion Haifa, Israel lihi@ee.technion.ac.il Ayellet Tal Technion Haifa, Israel ayellet@ee.technion.ac.il

More information

A Novel Image Super-resolution Reconstruction Algorithm based on Modified Sparse Representation

A Novel Image Super-resolution Reconstruction Algorithm based on Modified Sparse Representation , pp.162-167 http://dx.doi.org/10.14257/astl.2016.138.33 A Novel Image Super-resolution Reconstruction Algorithm based on Modified Sparse Representation Liqiang Hu, Chaofeng He Shijiazhuang Tiedao University,

More information

THE human visual system can spot salient objects in a cluttered

THE human visual system can spot salient objects in a cluttered 1892 IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 39, NO. 9, SEPTEMBER 2017 Ranking Saliency Lihe Zhang, Chuan Yang, Huchuan Lu, Senior Member, IEEE, Xiang Ruan, and Ming-Hsuan

More information

The Un-normalized Graph p-laplacian based Semi-supervised Learning Method and Speech Recognition Problem

The Un-normalized Graph p-laplacian based Semi-supervised Learning Method and Speech Recognition Problem Int. J. Advance Soft Compu. Appl, Vol. 9, No. 1, March 2017 ISSN 2074-8523 The Un-normalized Graph p-laplacian based Semi-supervised Learning Method and Speech Recognition Problem Loc Tran 1 and Linh Tran

More information

A Novel Approach to Saliency Detection Model and Its Applications in Image Compression

A Novel Approach to Saliency Detection Model and Its Applications in Image Compression RESEARCH ARTICLE OPEN ACCESS A Novel Approach to Saliency Detection Model and Its Applications in Image Compression Miss. Radhika P. Fuke 1, Mr. N. V. Raut 2 1 Assistant Professor, Sipna s College of Engineering

More information

SALIENT OBJECT DETECTION VIA BACKGROUND CONTRAST

SALIENT OBJECT DETECTION VIA BACKGROUND CONTRAST SALIENT OBJECT DETECTION VIA BACKGROUND CONTRAST Quan Zhou,, Nianyi Li 2,, Jianxin Chen, Liang Zhou, and Longin Jan Latecki 3 Key Lab of BB WL Comm & Sen Net Tech, Nanjing Univ of Posts & Telecom, Nanjing,

More information

A Survey of Light Source Detection Methods

A Survey of Light Source Detection Methods A Survey of Light Source Detection Methods Nathan Funk University of Alberta Mini-Project for CMPUT 603 November 30, 2003 Abstract This paper provides an overview of the most prominent techniques for light

More information

Learning to Match. Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li

Learning to Match. Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li Learning to Match Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li 1. Introduction The main tasks in many applications can be formalized as matching between heterogeneous objects, including search, recommendation,

More information

A FRAMEWORK OF EXTRACTING MULTI-SCALE FEATURES USING MULTIPLE CONVOLUTIONAL NEURAL NETWORKS. Kuan-Chuan Peng and Tsuhan Chen

A FRAMEWORK OF EXTRACTING MULTI-SCALE FEATURES USING MULTIPLE CONVOLUTIONAL NEURAL NETWORKS. Kuan-Chuan Peng and Tsuhan Chen A FRAMEWORK OF EXTRACTING MULTI-SCALE FEATURES USING MULTIPLE CONVOLUTIONAL NEURAL NETWORKS Kuan-Chuan Peng and Tsuhan Chen School of Electrical and Computer Engineering, Cornell University, Ithaca, NY

More information

IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 14, NO. 4, AUGUST

IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 14, NO. 4, AUGUST IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 14, NO. 4, AUGUST 2012 1275 Unsupervised Salient Object Segmentation Based on Kernel Density Estimation and Two-Phase Graph Cut Zhi Liu, Member, IEEE, Ran Shi, Liquan

More information

arxiv: v3 [cs.cv] 2 Oct 2018

arxiv: v3 [cs.cv] 2 Oct 2018 Coarse-to-Fine Salient Object Detection with Low-Rank Matrix Recovery Qi Zheng a, Shujian Yu b, Xinge You a, a Huazhong University of Science and Technology, Wuhan, China b University of Florida, Gainesville,

More information

Robust and Efficient Saliency Modeling from Image Co-Occurrence Histograms

Robust and Efficient Saliency Modeling from Image Co-Occurrence Histograms IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 36, NO. 1, JANUARY 2014 195 Robust and Efficient Saliency Modeling from Image Co-Occurrence Histograms Shijian Lu, Cheston Tan, and

More information

Robust Frequency-tuned Salient Region Detection

Robust Frequency-tuned Salient Region Detection Robust Frequency-tuned Salient Region Detection 1 Yong Zhang, 2 Yihua Lan, 3 Haozheng Ren, 4 Ming Li 1 School of Computer Engineering, Huaihai Institute of Technology, Lianyungang, China, zhyhglyg@126.com

More information

arxiv: v2 [cs.cv] 31 Mar 2015

arxiv: v2 [cs.cv] 31 Mar 2015 Visual Saliency Based on Multiscale Deep Features Guanbin Li Yizhou Yu Department of Computer Science, The University of Hong Kong https://sites.google.com/site/ligb86/mdfsaliency/ arxiv:1503.08663v2 [cs.cv]

More information

Machine Learning for Media Data Analysis

Machine Learning for Media Data Analysis 27 OKTOBER 2010 Machine Learning for Media Data Analysis Assistant Machine Learning and Computer Vision Aarhus University, alexandrosiosifidis@engaudk Overview Media Data Analysis applications Image analysis/recognition/segmentation

More information

Channel Locality Block: A Variant of Squeeze-and-Excitation

Channel Locality Block: A Variant of Squeeze-and-Excitation Channel Locality Block: A Variant of Squeeze-and-Excitation 1 st Huayu Li Northern Arizona University Flagstaff, United State Northern Arizona University hl459@nau.edu arxiv:1901.01493v1 [cs.lg] 6 Jan

More information

Random Walks on Graphs to Model Saliency in Images

Random Walks on Graphs to Model Saliency in Images Random Walks on Graphs to Model Saliency in Images Viswanath Gopalakrishnan Yiqun Hu Deepu Rajan School of Computer Engineering Nanyang Technological University, Singapore 639798 visw0005@ntu.edu.sg yqhu@ntu.edu.sg

More information

PERFORMANCE ANALYSIS OF COMPUTING TECHNIQUES FOR IMAGE DISPARITY IN STEREO IMAGE

PERFORMANCE ANALYSIS OF COMPUTING TECHNIQUES FOR IMAGE DISPARITY IN STEREO IMAGE PERFORMANCE ANALYSIS OF COMPUTING TECHNIQUES FOR IMAGE DISPARITY IN STEREO IMAGE Rakesh Y. Department of Electronics and Communication Engineering, SRKIT, Vijayawada, India E-Mail: rakesh.yemineni@gmail.com

More information

What Makes a Patch Distinct?

What Makes a Patch Distinct? What Makes a Patch Distinct? Ran Margolin Technion Haifa, Israel margolin@tx.technion.ac.il Ayellet Tal Technion Haifa, Israel ayellet@ee.technion.ac.il Lihi Zelnik-Manor Technion Haifa, Israel lihi@ee.technion.ac.il

More information

arxiv: v1 [cond-mat.dis-nn] 30 Dec 2018

arxiv: v1 [cond-mat.dis-nn] 30 Dec 2018 A General Deep Learning Framework for Structure and Dynamics Reconstruction from Time Series Data arxiv:1812.11482v1 [cond-mat.dis-nn] 30 Dec 2018 Zhang Zhang, Jing Liu, Shuo Wang, Ruyue Xin, Jiang Zhang

More information

arxiv: v1 [cs.cv] 10 May 2017

arxiv: v1 [cs.cv] 10 May 2017 Learning RGB-D Salient Object Detection using background enclosure, depth contrast, and top-down features arxiv:1705.03607v1 [cs.cv] 10 May 2017 Riku Shigematsu David Feng Australian National University

More information

Open Access Research on the Prediction Model of Material Cost Based on Data Mining

Open Access Research on the Prediction Model of Material Cost Based on Data Mining Send Orders for Reprints to reprints@benthamscience.ae 1062 The Open Mechanical Engineering Journal, 2015, 9, 1062-1066 Open Access Research on the Prediction Model of Material Cost Based on Data Mining

More information

International Conference on Information Sciences, Machinery, Materials and Energy (ICISMME 2015)

International Conference on Information Sciences, Machinery, Materials and Energy (ICISMME 2015) International Conference on Information Sciences, Machinery, Materials and Energy (ICISMME 2015) Brief Analysis on Typical Image Saliency Detection Methods Wenwen Pan, Xiaofei Sun, Xia Wang, Wei Zhang

More information

Evaluation of regions-of-interest based attention algorithms using a probabilistic measure

Evaluation of regions-of-interest based attention algorithms using a probabilistic measure Evaluation of regions-of-interest based attention algorithms using a probabilistic measure Martin Clauss, Pierre Bayerl and Heiko Neumann University of Ulm, Dept. of Neural Information Processing, 89081

More information

Object Extraction Using Image Segmentation and Adaptive Constraint Propagation

Object Extraction Using Image Segmentation and Adaptive Constraint Propagation Object Extraction Using Image Segmentation and Adaptive Constraint Propagation 1 Rajeshwary Patel, 2 Swarndeep Saket 1 Student, 2 Assistant Professor 1 2 Department of Computer Engineering, 1 2 L. J. Institutes

More information

Salient Region Detection via High-Dimensional Color Transform

Salient Region Detection via High-Dimensional Color Transform Salient Region Detection via High-Dimensional Color Transform Jiwhan Kim Dongyoon Han Yu-Wing Tai Junmo Kim Department of Electrical Engineering Korea Advanced Institute of Science and Technology (KAIST)

More information

Image Error Concealment Based on Watermarking

Image Error Concealment Based on Watermarking Image Error Concealment Based on Watermarking Shinfeng D. Lin, Shih-Chieh Shie and Jie-Wei Chen Department of Computer Science and Information Engineering,National Dong Hwa Universuty, Hualien, Taiwan,

More information

FASA: Fast, Accurate, and Size-Aware Salient Object Detection

FASA: Fast, Accurate, and Size-Aware Salient Object Detection FASA: Fast, Accurate, and Size-Aware Salient Object Detection Gökhan Yildirim, Sabine Süsstrunk School of Computer and Communication Sciences École Polytechnique Fédérale de Lausanne Abstract. Fast and

More information

Fast and Efficient Saliency Detection Using Sparse Sampling and Kernel Density Estimation

Fast and Efficient Saliency Detection Using Sparse Sampling and Kernel Density Estimation Fast and Efficient Saliency Detection Using Sparse Sampling and Kernel Density Estimation Hamed Rezazadegan Tavakoli, Esa Rahtu, and Janne Heikkilä Machine Vision Group, Department of Electrical and Information

More information

An ICA based Approach for Complex Color Scene Text Binarization

An ICA based Approach for Complex Color Scene Text Binarization An ICA based Approach for Complex Color Scene Text Binarization Siddharth Kherada IIIT-Hyderabad, India siddharth.kherada@research.iiit.ac.in Anoop M. Namboodiri IIIT-Hyderabad, India anoop@iiit.ac.in

More information

Multi-Scale Saliency Detection using Dictionary Learning

Multi-Scale Saliency Detection using Dictionary Learning Multi-Scale Saliency Detection using Dictionary Learning Shubham Pachori Electrical Engineering Indian Institute of Technology Gandhinagar, India 382355 Email: shubham pachori@iitgn.ac.in arxiv:1611.06307v3

More information

Video Inter-frame Forgery Identification Based on Optical Flow Consistency

Video Inter-frame Forgery Identification Based on Optical Flow Consistency Sensors & Transducers 24 by IFSA Publishing, S. L. http://www.sensorsportal.com Video Inter-frame Forgery Identification Based on Optical Flow Consistency Qi Wang, Zhaohong Li, Zhenzhen Zhang, Qinglong

More information

Main Subject Detection via Adaptive Feature Selection

Main Subject Detection via Adaptive Feature Selection Main Subject Detection via Adaptive Feature Selection Cuong Vu and Damon Chandler Image Coding and Analysis Lab Oklahoma State University Main Subject Detection is easy for human 2 Outline Introduction

More information

RSRN: Rich Side-output Residual Network for Medial Axis Detection

RSRN: Rich Side-output Residual Network for Medial Axis Detection RSRN: Rich Side-output Residual Network for Medial Axis Detection Chang Liu, Wei Ke, Jianbin Jiao, and Qixiang Ye University of Chinese Academy of Sciences, Beijing, China {liuchang615, kewei11}@mails.ucas.ac.cn,

More information

SEMI-BLIND IMAGE RESTORATION USING A LOCAL NEURAL APPROACH

SEMI-BLIND IMAGE RESTORATION USING A LOCAL NEURAL APPROACH SEMI-BLIND IMAGE RESTORATION USING A LOCAL NEURAL APPROACH Ignazio Gallo, Elisabetta Binaghi and Mario Raspanti Universitá degli Studi dell Insubria Varese, Italy email: ignazio.gallo@uninsubria.it ABSTRACT

More information

Salient Object Detection: A Discriminative Regional Feature Integration Approach

Salient Object Detection: A Discriminative Regional Feature Integration Approach 23 IEEE Conference on Computer Vision and Pattern Recognition Salient Object Detection: A Discriminative Regional Feature Integration Approach Huaizu Jiang Jingdong Wang Zejian Yuan Yang Wu Nanning Zheng

More information

Image retrieval based on bag of images

Image retrieval based on bag of images University of Wollongong Research Online Faculty of Informatics - Papers (Archive) Faculty of Engineering and Information Sciences 2009 Image retrieval based on bag of images Jun Zhang University of Wollongong

More information

Cost-alleviative Learning for Deep Convolutional Neural Network-based Facial Part Labeling

Cost-alleviative Learning for Deep Convolutional Neural Network-based Facial Part Labeling [DOI: 10.2197/ipsjtcva.7.99] Express Paper Cost-alleviative Learning for Deep Convolutional Neural Network-based Facial Part Labeling Takayoshi Yamashita 1,a) Takaya Nakamura 1 Hiroshi Fukui 1,b) Yuji

More information

Learning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009

Learning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009 Learning and Inferring Depth from Monocular Images Jiyan Pan April 1, 2009 Traditional ways of inferring depth Binocular disparity Structure from motion Defocus Given a single monocular image, how to infer

More information

Robust Ring Detection In Phase Correlation Surfaces

Robust Ring Detection In Phase Correlation Surfaces Griffith Research Online https://research-repository.griffith.edu.au Robust Ring Detection In Phase Correlation Surfaces Author Gonzalez, Ruben Published 2013 Conference Title 2013 International Conference

More information

Graph Autoencoder-Based Unsupervised Feature Selection

Graph Autoencoder-Based Unsupervised Feature Selection Graph Autoencoder-Based Unsupervised Feature Selection Siwei Feng Department of Electrical and Computer Engineering University of Massachusetts Amherst Amherst, MA, 01003 siwei@umass.edu Marco F. Duarte

More information

Image Classification based on Saliency Driven Nonlinear Diffusion and Multi-scale Information Fusion Ms. Swapna R. Kharche 1, Prof.B.K.

Image Classification based on Saliency Driven Nonlinear Diffusion and Multi-scale Information Fusion Ms. Swapna R. Kharche 1, Prof.B.K. Image Classification based on Saliency Driven Nonlinear Diffusion and Multi-scale Information Fusion Ms. Swapna R. Kharche 1, Prof.B.K.Chaudhari 2 1M.E. student, Department of Computer Engg, VBKCOE, Malkapur

More information

Tensor Sparse PCA and Face Recognition: A Novel Approach

Tensor Sparse PCA and Face Recognition: A Novel Approach Tensor Sparse PCA and Face Recognition: A Novel Approach Loc Tran Laboratoire CHArt EA4004 EPHE-PSL University, France tran0398@umn.edu Linh Tran Ho Chi Minh University of Technology, Vietnam linhtran.ut@gmail.com

More information

Using temporal seeding to constrain the disparity search range in stereo matching

Using temporal seeding to constrain the disparity search range in stereo matching Using temporal seeding to constrain the disparity search range in stereo matching Thulani Ndhlovu Mobile Intelligent Autonomous Systems CSIR South Africa Email: tndhlovu@csir.co.za Fred Nicolls Department

More information

SEMI-SUPERVISED LEARNING (SSL) for classification

SEMI-SUPERVISED LEARNING (SSL) for classification IEEE SIGNAL PROCESSING LETTERS, VOL. 22, NO. 12, DECEMBER 2015 2411 Bilinear Embedding Label Propagation: Towards Scalable Prediction of Image Labels Yuchen Liang, Zhao Zhang, Member, IEEE, Weiming Jiang,

More information

Multidirectional 2DPCA Based Face Recognition System

Multidirectional 2DPCA Based Face Recognition System Multidirectional 2DPCA Based Face Recognition System Shilpi Soni 1, Raj Kumar Sahu 2 1 M.E. Scholar, Department of E&Tc Engg, CSIT, Durg 2 Associate Professor, Department of E&Tc Engg, CSIT, Durg Email:

More information

Short Survey on Static Hand Gesture Recognition

Short Survey on Static Hand Gesture Recognition Short Survey on Static Hand Gesture Recognition Huu-Hung Huynh University of Science and Technology The University of Danang, Vietnam Duc-Hoang Vo University of Science and Technology The University of

More information

arxiv: v1 [cs.cv] 19 Apr 2016

arxiv: v1 [cs.cv] 19 Apr 2016 Deep Saliency with Encoded Low level Distance Map and High Level Features Gayoung Lee KAIST gylee1103@gmail.com Yu-Wing Tai SenseTime Group Limited yuwing@gmail.com Junmo Kim KAIST junmo.kim@kaist.ac.kr

More information

Focusing Attention on Visual Features that Matter

Focusing Attention on Visual Features that Matter TSAI, KUIPERS: FOCUSING ATTENTION ON VISUAL FEATURES THAT MATTER 1 Focusing Attention on Visual Features that Matter Grace Tsai gstsai@umich.edu Benjamin Kuipers kuipers@umich.edu Electrical Engineering

More information

Synthetic Saliency. Ben Weems Stanford University. Anthony Perez Stanford University. Karan Rai Stanford University. Abstract. 1.

Synthetic Saliency. Ben Weems Stanford University. Anthony Perez Stanford University. Karan Rai Stanford University. Abstract. 1. Synthetic Saliency Anthony Perez Stanford University aperez8@stanford.edu Karan Rai Stanford University karanrai@stanford.edu Ben Weems Stanford University bweems@stanford.edu Abstract There are many existing

More information

MATLAB User Guide for Depth Reconstruction from Sparse Samples

MATLAB User Guide for Depth Reconstruction from Sparse Samples MATLAB User Guide for Depth Reconstruction from Sparse Samples Lee-Kang Liu, Stanley H. Chan and Truong Q. Nguyen This MATLAB user guide presents the instructions of how to use the MATLAB functions accompanied

More information

Learning RGB-D Salient Object Detection using background enclosure, depth contrast, and top-down features

Learning RGB-D Salient Object Detection using background enclosure, depth contrast, and top-down features Learning RGB-D Salient Object Detection using background enclosure, depth contrast, and top-down features Riku Shigematsu David Feng Australian National University riku.research@gmail.com david.feng@data61.csiro.au

More information

Bilevel Sparse Coding

Bilevel Sparse Coding Adobe Research 345 Park Ave, San Jose, CA Mar 15, 2013 Outline 1 2 The learning model The learning algorithm 3 4 Sparse Modeling Many types of sensory data, e.g., images and audio, are in high-dimensional

More information

Image Quality Assessment Techniques: An Overview

Image Quality Assessment Techniques: An Overview Image Quality Assessment Techniques: An Overview Shruti Sonawane A. M. Deshpande Department of E&TC Department of E&TC TSSM s BSCOER, Pune, TSSM s BSCOER, Pune, Pune University, Maharashtra, India Pune

More information

THE HUMAN visual system interprets the world in a

THE HUMAN visual system interprets the world in a 1150 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 23, NO. 7, JULY 2013 Visual Saliency by Selective Contrast Qi Wang, Yuan Yuan, Senior Member, IEEE, and Pingkun Yan, Senior Member,

More information

STUDYING THE FEASIBILITY AND IMPORTANCE OF GRAPH-BASED IMAGE SEGMENTATION TECHNIQUES

STUDYING THE FEASIBILITY AND IMPORTANCE OF GRAPH-BASED IMAGE SEGMENTATION TECHNIQUES 25-29 JATIT. All rights reserved. STUDYING THE FEASIBILITY AND IMPORTANCE OF GRAPH-BASED IMAGE SEGMENTATION TECHNIQUES DR.S.V.KASMIR RAJA, 2 A.SHAIK ABDUL KHADIR, 3 DR.S.S.RIAZ AHAMED. Dean (Research),

More information

Supplementary material: Strengthening the Effectiveness of Pedestrian Detection with Spatially Pooled Features

Supplementary material: Strengthening the Effectiveness of Pedestrian Detection with Spatially Pooled Features Supplementary material: Strengthening the Effectiveness of Pedestrian Detection with Spatially Pooled Features Sakrapee Paisitkriangkrai, Chunhua Shen, Anton van den Hengel The University of Adelaide,

More information

An Efficient Saliency Based Lossless Video Compression Based On Block-By-Block Basis Method

An Efficient Saliency Based Lossless Video Compression Based On Block-By-Block Basis Method An Efficient Saliency Based Lossless Video Compression Based On Block-By-Block Basis Method Ms. P.MUTHUSELVI, M.E(CSE), V.P.M.M Engineering College for Women, Krishnankoil, Virudhungar(dt),Tamil Nadu Sukirthanagarajan@gmail.com

More information

Detect Globally, Refine Locally: A Novel Approach to Saliency Detection

Detect Globally, Refine Locally: A Novel Approach to Saliency Detection Detect Globally, Refine Locally: A Novel Approach to Saliency Detection Tiantian Wang, Lihe Zhang, Shuo Wang, Huchuan Lu, Gang Yang 2, Xiang Ruan 3, Ali Borji 4 Dalian University of Technology, 2 Northeastern

More information

An Efficient Salient Feature Extraction by Using Saliency Map Detection with Modified K-Means Clustering Technique

An Efficient Salient Feature Extraction by Using Saliency Map Detection with Modified K-Means Clustering Technique International Journal of Computational Engineering & Management, Vol. 15 Issue 5, September 2012 www..org 63 An Efficient Salient Feature Extraction by Using Saliency Map Detection with Modified K-Means

More information

CS 534: Computer Vision Segmentation and Perceptual Grouping

CS 534: Computer Vision Segmentation and Perceptual Grouping CS 534: Computer Vision Segmentation and Perceptual Grouping Ahmed Elgammal Dept of Computer Science CS 534 Segmentation - 1 Outlines Mid-level vision What is segmentation Perceptual Grouping Segmentation

More information

2 Proposed Methodology

2 Proposed Methodology 3rd International Conference on Multimedia Technology(ICMT 2013) Object Detection in Image with Complex Background Dong Li, Yali Li, Fei He, Shengjin Wang 1 State Key Laboratory of Intelligent Technology

More information

Limitations of Matrix Completion via Trace Norm Minimization

Limitations of Matrix Completion via Trace Norm Minimization Limitations of Matrix Completion via Trace Norm Minimization ABSTRACT Xiaoxiao Shi Computer Science Department University of Illinois at Chicago xiaoxiao@cs.uic.edu In recent years, compressive sensing

More information

Diffusion Wavelets for Natural Image Analysis

Diffusion Wavelets for Natural Image Analysis Diffusion Wavelets for Natural Image Analysis Tyrus Berry December 16, 2011 Contents 1 Project Description 2 2 Introduction to Diffusion Wavelets 2 2.1 Diffusion Multiresolution............................

More information

Design of Orthogonal Graph Wavelet Filter Banks

Design of Orthogonal Graph Wavelet Filter Banks Design of Orthogonal Graph Wavelet Filter Banks Xi ZHANG Department of Communication Engineering and Informatics The University of Electro-Communications Chofu-shi, Tokyo, 182-8585 JAPAN E-mail: zhangxi@uec.ac.jp

More information

Image Enhancement Techniques for Fingerprint Identification

Image Enhancement Techniques for Fingerprint Identification March 2013 1 Image Enhancement Techniques for Fingerprint Identification Pankaj Deshmukh, Siraj Pathan, Riyaz Pathan Abstract The aim of this paper is to propose a new method in fingerprint enhancement

More information