Salient Object Segmentation based on Linearly Combined Affinity Graphs
|
|
- Sara Anderson
- 6 years ago
- Views:
Transcription
1 Salient Object Segmentation based on Linearly Combined Affinity Graphs Çağlar Aytekin, Alexandros Iosifidis, Serkan Kiranyaz * and Moncef Gabbouj Department of Signal Processing, Tampere University of Technology, Tampere, Finland * Department of Electrical Engineering, Qatar University, Doha, Qatar {caglar.aytekin, alexandros.iosifidis, moncef.gabbouj}@tut.fi, mkiranyaz@qu.edu.qa Abstract In this paper, we propose a graph affinity learning method for a recently proposed graph-based salient object detection method, namely Extended Quantum Cuts (EQCut). We exploit the fact that the output of EQCut is differentiable with respect to graph affinities, in order to optimize linear combination coefficients for several differentiable affinity functions by applying error backpropagation. We show that the learnt linear combination of affinities improves the performance over the baseline method and achieves comparable (or even better) performance when compared to the state-of-the-art salient object segmentation methods Keywords graph affinity learning, salient object segmentation, spectral graph theor I. INTRODUCTION Visual saliency estimation is an important research topic in computer vision where the task is to highlight the objects that visually stand out from their surroundings. Earlier research endeavors have focused mostly on models that are inspired by the human visual system [2] due to the fact that visually important region selection is related to human attention mechanisms [1]. Accordingly, some works have based their evaluation on the performance of accurately predicting human eye fixations [3]-[7]. However, the output of such methods corresponds to a (sparse) set of points (image pixels) and, thus, it is not very useful in terms providing an initial region of interest for higher level computer vision tasks, such as object recognition. Instead, a closed object segment is a more preferred input for these tasks. Hence, recently, the focus on visual saliency estimation is mostly concentrated on salient object segmentation [8]-[18], a task which has been proved to be useful for higher level computer vision and pattern recognition problems, such as tracking [26], object region proposals [28] and object recognition [27]. Graph-based salient object detection methods rank among the top performing methods in the literature. According to a recent benchmark [42], 4 out of the 6 top performing methods are based on graphs. All these methods exploit different graphbased approaches. For instance, in [12], the shortest path problem in graph theory [47] was exploited in order to find a robust backgroundness measure for the image regions. Then, this measure was combined with a region contrast measure in order to evaluate saliency. In [13], a Markov Chain was constructed on an image graph model. Virtual boundary nodes on this graph are selected as absorbing nodes and the saliency is calculated as the absorption time of the transient nodes. The method in [14] exploits the minimum barrier distance on graphs, in order to efficiently evaluate the saliency of image regions based on their deviation from the appearance of image boundary regions. The method in [17] and [18] solves a spectral approximation of a graph based cut problem, in order to find salient regions that both differ from image boundaries in appearance and are in high contrast with their surrounding, and have a large area. In [38], the authors give a general formulation for several salient object detection methods [11], [13] following a graph-based diffusion interpretation and propose a new diffusion based model by applying in-depth spectral analysis of the diffusion matrix. Aside from the fact that the aforementioned methods exploit graphs, they differ in terms of their graph-construction in many ways. For example [11], [12], [13], [18] and [38] exploit superpixels as nodes in the given graph representation whereas [14] and [17] define pixels as graph nodes. A 4- connected neighborhood rule for nodes was adopted in [14] and [17], in [12] every node was connected to adjacent nodes only, whereas in [11], [13] and [38] the nodes are also connected to the ones sharing common boundaries with their neighboring nodes. In [18] a very dense graph was constructed connecting nodes up to a 16 th degree of neighborhood. In [11], [12], [13] and [38] all nodes corresponding to the regions on the image boundary are connected to each other, whereas in [17] and [18] an artificial background node was introduced and all image boundary regions were connected to this artificial node with high weights. While constructing the graph weights between a pair of nodes, [12], [13] and [38] use an exponentially decaying function with increasing Euclidian distance between appearance features corresponding to each superpixel, whereas [17] and [18] use the inverse of this distance. The above-described differences in graph construction have a high impact on the performance of salient object detection methods. For instance, the difference between the methods in [18] and [17] lies on graph construction approach. Even if all remaining steps are the same, different graphs lead into a significant difference in performance. As will be shown later in this paper, graph construction can be even more effective than
2 the method itself. In addition, there might be cases where a method can only work well with a specific graph type. Based on the observations above, in this paper, we will focus on the graph construction problem and in particular we aim to improve the graph weights (affinity) computation for a better salient object segmentation. Conventionally, graph-based methods use a pre-defined function with fixed parameters mapping Euclidian distance of appearance features between a pair of nodes to an affinity value. In this paper, we propose a method that learns the parameters of three popular affinity functions along with the weights of their linear combination. We employ Extended Quantum Cuts (EQCut) [18] as salient object segmentation method. We exploit the fact that EQCut produces saliency maps that are differentiable with respect to graph affinities and this in turn allows us to backpropagate the salient object detection error in order to optimize the aforementioned parameters of the overall system. The rest of the paper is organized as follows: In Section II, EQCut is briefly introduced and the proposed graph affinity learning strategy is described. In Section III we show that the learnt affinities lead to a significant performance improvement over the baseline EQCut. Finally, Section IV concludes the paper and suggests future topics for research. II. METHOD In this section we first briefly describe the conventional EQCut algorithm that will be exploited. Next, we describe the proposed affinity learning approach. A. Extended Quantum Cuts Extended Quantum Cuts (EQCut) [18] is based on a set of improvements over Quantum Cuts (QCut) [17], which is a spectral foreground detection method exploiting the link between graph theory and quantum mechanics. In a graph representation of an image, QCut approximates the following optimization problem: A cut(a, A ) = argmin A area(a), (1) where A is the foreground segment, A is the background segment (rest of the image), area(a) is the number of nodes in segment A and the cut-cost is defined as cut(a, A ) = i A,j A w i,j. Here, w i,j is the affinity (similarity) between nodes i and j. Let y b be a a binary indicator vector representing A, that takes a value of 1 in foreground nodes and 0 elsewhere. It has been proven in [17] that, if y b is relaxed to have real values, the relaxed solution to (1) is given as follows: y = z z, Hz = λz, (2) where λ is the minimum eigenvalue of the Hamiltonian matrix H, which is formulated as H = D W + V. Here, W is constructed from the affinities w i,j, D is a diagonal degree matrix whose elements are obtained by row-wise summation of W and V is a non-negative diagonal matrix of background priors for each node. The above representation exhibits a great similarity with the solutions for quantum states of a sub-atomic particle which were previously investigated for salient object segmentation in [19]. EQCut investigates the design of the graph that represents the image and its affinity matrix W, both of which highly affect the performance in salient object detection. The extensions proposed in [18] include a superpixel representation of the image, a novel graph affinity normalization strategy on a graph with higher connectivity and a multiresolution approach. The proposed extensions lead to a significant improvement over QCut [17]. It should be noted here that the graph affinity normalization in EQCut is based on the neighbourhood degrees between pairs of superpixels and they result into an asymmetric affinity matrix. In this paper we will exploit a symmetric affinity matrix by simply averaging the asymmetric one with its transposed. B. Proposed Method In this study, we use the same graph construction approach with the symmetric version of graph affinity normalization in EQCut and we will investigate the effect of optimally combining affinity functions in salient object detection performance based on EQCut. In [17] and [18], the following affinity function was exploited: w (1) 1 i,j = ε + f i f j 2, (3) where f i is the related feature vector for the superpixel i and ε is a small constant to prevent division by zero. A property of the above-given affinity function is that it greatly penalizes the Euclidian distance between two feature vectors. An exponential function which is relatively smoother was used in some salient object segmentation methods [12], [13], [38] expressed as: w (2) i,j = e f i f j σ 2, where σ is a smoothing factor controlling the affinity decrease rate with the increasing feature distance. Finally, a linear affinity function can also be used as follows: (4) w (3) i,j = 1 f i f j. (5) Note that in all affinity functions, the Euclidian distances are unit-normalized in such a way that among all possible feature distances in a given image, the maximum is set to 1. Hence, it is guaranteed that (5) does not produce negative affinities. At this point we would like to note that, for a given image, affinities obtained using (3) are normalized such that the maximum affinity will have a value of 1. The affinities obtained by (4) and (5) are not needed to be normalized since they already produce outputs between 0 and 1. The proposed method, instead of using a single affinity function with fixed parameters, makes use of a linear combination of them, as follows: w i,j = c 1 w (1) i,j + c 2 w (2) i,j + c 3 w (3) i,j. (6)
3 Furthermore, we consider ε in (3) and σ in (4) as parameters to be learned. Hence, we aim to learn the parameter set θ = {c 1, c 2, c 3, ε, σ} in order to achieve a high performance in salient object detection. In order to learn affinities that will perform well on salient object detection, we exploit the observation that the saliency maps obtained via (2) are differentiable with respect to affinities of the graph that is used. This property gives us the opportunity to backpropagate the error related to the output of EQCut with respect to the parameters we wish to optimize. Exploiting the eigenvector derivatives for symmetric matrices as presented in [48], one can write the following: z h = [z (λi H) ]Dpl, where λ is the minimum eigenvector of H, z is the corresponding eigenvector, h is the vectorized version of H by row-wise raster scan, (. ) denotes the Moore-Penrose pseudoinverse, Dpl is the duplication matrix [48] which converts the half-vectorizations of symmetric matrices to full vectorizations, I is the identity matrix and is the Kronecker product. In order to backpropagate the error e from EQCut output y to affinities, we can write the partial derivatives of the error corresponding to i th superpixel with respect to the entire H matrix as follows: (7) e i H = ( e i y i ) ( y i z i ) ( z i H ). (8) We can extract ( z i H ) component of (8) from (7) and ( y i z i ) = 2z i as observed from (2). Only remaining component ( e i y i ) depends on the error function that is chosen for performance evaluation of the EQcut output. We have used the following error function and the corresponding partial derivatives of it with respect to y: e i = y i g i, ( e i y i ) = sgn(y i g i ), where g i is the ground-truth label for superpixel i. The groundtruth map takes the value 1 for superpixels corresponding to salient regions and 0 otherwise. At this point it should be noted that since y is l 1 normalized due to (2), elements of y never gets the value 1. Furthermore, due to numerical precision they do not become exactly equal to 0 either. Hence, (9) always produces an error derivative of 1 for background regions and 1 for salient object regions. We have observed that this kind of enforcement yields much better performance, mainly because of the fact that it always forces the network to give higher values to salient regions and suppress background regions. Other error functions such as Euclidian distance between y and l 1 normalized ground truth map, produce small gradients once y is close to l 1 normalized ground truth map, hence converges to a suboptimal solution. (9) Based on the above discussion, (8) can be calculated. Accordingly, partial derivatives with respect to W can be calculated as follows: e i W = 1T d ei,h d T ei,h e i H, (10) where d ei,h is a vector obtained by the diagonal elements of e i H. Finally, the derivative for the average error e per node can be calculated as follows: e W = 1 N e i. W i (11) Next, partial derivatives of affinities with respect to parameter set θ are needed to be calculated. The partial derivatives of affinities with respect to linear combination coefficient c k, k=1,2,3, can be calculated as follows: w i,j c k = w (k) i,j, (12) The partial derivatives of affinities with respect to parameters ε and σ can be given as follows: w i,j ε w i,j σ c 1 = (ε + f i f j 2 2, (13) ) 2c (2) 2w i,j fi f j = σ 3, (14) Combining (7)-(14), one can obtain the partial derivatives of the error e from EQCut s output y with respect to the affinity calculation parameter set θ. Hence, given a training dataset, parameter set θ can be learned by iteratively backpropagating the error to update θ and calculating the new output until convergence based on a stopping rule that will be explained in the next section. III. EXPERIMENTAL RESULTS In this section we provide extensive experiments conducted in order to evaluate the performance of the proposed method exploiting learned affinities for EQCut-based saliency object detection. Comparisons are made with EQCut as well as the state-of-the-art salient object detection methods. A. Datasets SOD [33] dataset contains 300 images that includes many salient objects and a high background clutter. JUDD [41] dataset is the most challenging dataset containing 900 images in total with multiple salient objects and high background clutter. MSRA10k [9] contains 10k images with relatively easier examples with large objects having high contrast. Dut- OMRON [11] contains 5168 images with one or more salient objects per image and high background clutter. Pascal1500 [49] contains 1500 images with more than one salient object at a variety of locations and scales and challenging background.
4 B. Evaluation Metrics In our experiments, we use precision-recall curves and maximum and mean F 1 scores, as performance metrics. Precision is the ratio of the total number of true positives to the total number of detected samples. Recall is the ratio of the number of true positives to the number of true samples. For salient object detection, a true sample is a pixel in the groundtruth segmentation mask G. A detected sample at a threshold τ is the number of nonzero elements in a saliency map thresholded with τ, denoted by S τ. pre(τ) = G S τ, rec(τ) = G S τ S τ G (15) F-measure is defined according to precision and recall values as follows: in the original EQCut affinity calculation as in (3). As the distances increase, the affinities decrease more smoothly when compared to (3) due to contribution of (4) and (5). F 1 (τ) = 2pre(τ)rec(τ) pre(τ) + rec(τ) (16) Maximum F 1 measure is then simply obtained by taking the maximum of the F 1 measure over all thresholds applied. C. Training In all our experiments, Dut-OMRON was used during the training phase in order to learn the parameters θ, as it is the one of the largest datasets with the highest complexity. A portion of Dut-OMRON could also have been used as a validation set, however the images within each dataset have some correlation, for example Dut-OMRON dataset contains several images that exhibit great similarity from some classes. Therefore, in order to have a system that generalizes well, we have used an entirely different dataset (SOD) for validation purposes. We have initialized the parameters with θ 0 = {1,0,0, 10 6, 0.1}. This initialization corresponds to the affinity calculation in EQCut and σ = 0.1 was set as in [12], [13] and [38]. At each iteration, the images in Dut-OMRON dataset are randomly shuffled. Next, in an epoch, 50 training samples were chosen from the shuffled indices and the affinity function parameters were updated via error backpropagation with a minibatch size of 1. The step size is scaled by 0.99 in each consecutive iteration. At each epoch, the maximum F 1 measure for the validation dataset is measured. Training is stopped when the error converges, or when a pre-defined maximum iteration number is achieved. The parameter values corresponding to the epoch that produced the best performance on the validation set were selected as the final parameters. D. Comparison to EQCut The Here we compare the performance of the proposed method (EQCut-A) with EQCut. First, three separate parameter sets were learned for three superpixel granularities of EQCut (150, 300 and 600 superpixels) according to the details explained in the previous section. The log affinities of the learnt functions and the ones calculated by w (1) in (3), w (2) in (4) and w (3) in (5) are illustrated in Fig. 1. The learnt affinity functions lead to very high affinities in very small distances as Fig. 1 Log affinities for w (1), w (2), w (3) and learnt affinities for 150, 300 and 600 superpixel (SP) granularities. In Fig. 2, maximum and mean F1 measures for the test datasets (i.e., Pascal1500, MSRA10k and Judd) are illustrated. For each superpixel granularity and for each test set, we observe a clear improvement over the maximum and mean F1 measures. The improvement is much clearer as the dataset gets more difficult. For instance, MSRA10k is the simplest dataset and although there is a performance improvement with the learnt affinities, it is rather incremental. However, for challenging datasets Pascal1500 and Judd, the performance improvement is significant, reaching up to 7% relative improvement over the EQCut in maximum F1 measures, respectively. Fig. 2 Maximum F1 measure of baseline EQCut and EQCut with learned affinities (EQCut-A). Performances are given for Pascal (P), Judd (J) and MSRA10k(M) datasets and 150, 300 and 600 superpixel granularities.
5 Fig. 3 Performance comparisons of EQCut-A, EQCut, GP, RBD, ST, DSR, MC and DRFI in precision-recall curves (Column 1 for JUDD and 2 for MSRA10k datasets) and in maximum F1measures (Column 3 for JUDD and 4 for MSRA10k datasets) E. Comparison with the State of the Art In this section we compare the performance of EQCut-A with the state of the art salient object detection methods according to a recent benchmark study [42], namely EQCut [18], MC [13], RBD [12], ST [15], DSR [8] and DRFI [16]. Furthermore, we include one more method that became the state-of-the-art and published after the benchmark, namely GP [38]. For the comparisons in this section, we have employed the multi-resolution approach in [18] and combined all saliency maps at each superpixel granularity by simply averaging them. This was made for both EQCut and EQCut-A for superpixel granularities 150, 300 and 600. The performances are given in precision-recall curves and maximum F1 measures for Judd and MSRA10k datasets. Pascal-1500 dataset was excluded from comparisons since we couldn t reach all saliency map results/codes of the compared methods for this dataset. Nevertheless, Judd and MSRA10k datasets cover a large variety of cases where the former datasets include salient objects with highest difficulty, whereas the latter includes simplest cases in salient object detection task. In both datasets we observe consistent improvement over EQCut, moreover, in Judd database, EQCut-A performs better than all state-of-theart methods and in MSRA10k dataset, EQCut-A is better than all the methods except for DRFI, while it is comparable with DRFI in this dataset F. Computational Complexity In terms of computational complexity, the proposed affinity calculations do not bring any significant burden over EQCut. For the saliency calculations on SOD dataset, the average runtimes for EQCut and EQCUT-A are and seconds, respectively. Therefore, the total increase in the computational complexity is only around 11%. IV. CONCLUSION In this paper, we have proposed a graph affinity and parameter learning method for salient object detection. Specifically, linear combination coefficients combining several affinity functions and their parameters are learned altogether such that the resulting affinity matrix will lead to better saliency results via the exploited salient object detection method EQCut. The proposed affinities lead to a consistent improvement over EQCut in precision-recall curves and maximum F1 measures, yet they do not bring any significant computational complexity rise. Moreover, EQCut with the learnt affinities performs the best or comparable to the state-ofthe-art in salient object segmentation. REFERENCES [1] S. K. Ungerleider, and G. Leslie, "Mechanisms of visual attention in the human cortex," Annual Review of Neuroscience, vol. 23, no. 1, pp , [2] L. Itti, C. Koch, and E. Niebur, A model of saliency-based visual attention for rapid scene analysis, IEEE Trans. Pattern Anal. Mach. Intell., vol. 20, no. 11, pp , [3] J. Harel, C. Koch, and P. Perona, Graph-based visual saliency, Conference on Neural Information Processing Systems, pp , [4] I. Rigas, G. Economou and S. Fotopoulos, Efficient Modelling of Visual Saliency Based on Local Sparse Representation and the Use of Hamming Distance, Computer Vision and Image Understanding, vol. 134, pp 33-45, [5] D. Gao, V. Mahadevan, and N.Vasconcelos, The discriminant centersurround hypothesis for bottom-up saliency, Conference on Neural Information Processing Systems, pp , [6] T. Judd, K. Ehinger, F. Durand, and A. Torralba, Learning to predict where humans look, in Proc. IEEE International Conference on Computer Vision, pp , [7] J. Han, L. Sun, X. Hu, J. Han and L. Shao, Spatial and temporal visual attention prediction in videos using eye movement data, Neurocomputing, vol. 145, pp , [8] X. Li, H. Lu, L. Zhang, X. Ruan and M.H. Yang, "Saliency detection via dense and sparse reconstruction," in Proc. IEEE International Conference on Computer Vision, pp , [9] M.M. Cheng, G. Zhang, N. Mitra, X. Huang, and S.Hu, Global contrast based salient region detection, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 37, no.3, pp , [10] Q. Yan, L. Xu, J. Shi, and J. Jia, Hierarchical saliency detection, in Proc. IEEE Conference on Computer Vision and Pattern Recognition, pp , [11] C. Yang, L. Zhang, H. Lu, X. Ruan, and M.-H. Yang, Saliency detection via graph-based manifold ranking, IEEE Conference on Computer Vision and Pattern Recognition, pp , [12] W. Zhu, S. Liang, Y. Wei and J. Sun, Saliency Optimization from Robust Background Detection, IEEE Conference on Computer Vision and Pattern Recognition, pp , [13] B. Jiang, L. Zhang, H. Lu, C. Yang and M-H. Yang, Saliency Detection via Absorbing Markov Chain, IEEE International Conference on Computer Vision, pp , [14] J. Zhang, S. Sclaroff, Z. Lin, X. Shen, B. Price, and R. Mechi, Minimum Barrier Salient Object Detection at 80 FPS, IEEE International Conference on Computer Vision, pp , [15] Z. Liu, W. Zou, and O. Le Meur. "Saliency tree: A novel saliency detection framework." IEEE Transactions on Image Processing, vol. 23 no. 5, pp , [16] H. Jiang, J. Wang, Z. Yuan, Y. Wu, N. Zheng, and S. Li. "Salient object detection: A discriminative regional feature integration approach." IEEE
6 Conference on Computer Vision and Pattern Recognition, pp , [17] C. Aytekin, S. Kiranyaz and M. Gabbouj, Automatic Object Segmentation by Quantum Cuts, International Conference on Pattern Recognition, pp , [18] C. Aytekin, E. C. Ozan, S. Kiranyaz, and M. Gabbouj, Visual saliency by extended quantum cuts, IEEE International Conference on Image Processing, pp , [19] C. Aytekin, S. Kiranyaz and M. Gabbouj, Quantum mechanics in Computer Vision: Automatic Object Extraction, IEEE International Conference on Image Processing, pp , [20] D. Rudoy, D.B. Goldman, E. Shechtman and L. Z. Manor, Learning Video Saliency from Human Gaze Using Candidate Selection, IEEE Conference on Computer Vision and Pattern Recognition, pp , [21] H. Hadizadeh and I.V. Bajic, Saliency-Aware Video Compression, IEEE Transactions on Image Processing, vol. 23, no.1, pp , [22] R. Margolin, L. Z. Manor and A. Tal, Saliency for Image Manipulation, The Visual Computer, vol. 29, no. 5, pp , [23] X. Hou and L. Zhang, Thumbnail Generation based on Global Saliency, Advances in Cognitive Neurodynamics, pp , [24] M. Casares, S. Velipasalar and A. Pinto, Light-weight Salient Foreground Detection for Embedded Smart Cameras, Computer Vision and Image Understanding, vol. 114, no. 11, pp , [25] Z. Li, S. Qin and L. Itti, Visual attention guided bit allocation in video compression, Image and Vision Computing, vol. 29, no. 1, pp. 1-14, [26] C. Aytekin, E. Tunali, and S. Oz, Fast Semi-automatic Target Initialization based on Visual Saliency for Airborne Thermal Imagery, In International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications, pp , [27] E. Horbert, G. M. García, S. Frintrop and B. Leibe, Sequence-level object candidates based on saliency for generic object recognition on mobile systems, International Conference on Robotics and Automation, pp , [28] C. Aytekin, S. Kiranyaz, and M. Gabbouj, "Learning to rank salient segments extracted by multispectral quantum cuts." Pattern Recognition Letters, vol. 72, pp , [29] [30] R. Achanta, S. Hemami, F. Estrada and S. Susstrunk, "Frequency-tuned salient region detection." IEEE Conference on Computer Vision and Pattern Recognition, pp , [31] R. Achanta and S. Susstrunk, Saliency detection using maximum symmetric surround, International Conference on Image Processing, pp , [32] W. Zou, K. Kpalma, Z. Liu, and J. Ronsin, Segmentation driven lowrank matrix recovery for saliency detection, in Proc. British Machine Vision Conference, pp. 1-13, [33] V. Mohavedi and J. H. Elder, Design and perceptual validation of performance measures for salient object segmentation, IEEE Computer Vision and Pattern Recognition Workshops, pp , [34] S. Alpert, M. Galun, R. Basri, and A. Brandt, Image Segmentation by Probabilistic bottom-up aggregiation and cue integration, IEEE Conference on Computer Vision and Pattern Recognition, pp. 1-8, [35] G. Li, Y. Yizhou "Visual Saliency Based on Multiscale Deep Features," arxiv preprint arxiv: , [36] A. Borji, M. M. Cheng, H. Jiang and J. Li, "Salient object detection: A survey," arxiv preprint arxiv: , [37] R.R. Coifman, and S. Lafon. "Diffusion maps," Applied and computational harmonic analysis, vol. 21, no. 1, pp. 5-30, [38] P. Jiang, N. Vasconcelos, and J. Peng. "Generic Promotion of Diffusion- Based Salient Object Detection," IEEE International Conference on Computer Vision, pp , [39] D. Zhou, J. Weston, A. Gretton, O. Bousquet, and B. Schölkopf, "Ranking on data manifolds," Advances in Neural Information Processing Systems, pp , [40] D. Zhou, O. Bousquet, T.N. Lal, J. Weston, and B. Schölkopf, "Learning with local and global consistency," Advances in Neural Information processing systems, pp , [41] A. Borji. "What is a salient object? a dataset and a baseline model for salient object detection," IEEE Transactions on Image Processing, vol 24, no. 2, pp , [42] A. Borji, M. M. Cheng, H. Jiang and J. Li. "Salient object detection: A benchmark," IEEE Transactions on Image Processing, vol. 24, no. 12, pp , [43] L. M. Manevitz, and M. Yousef. "One-class SVMs for document classification," The Journal of Machine Learning Research, vol. 2 pp , [44] B. Schölkopf, R. Herbrich, and A. J. Smola, A generalized representer theorem, Computational Learning Theory, pp , [45] G.D. Poole, "Generalized M-matrices and applications," Mathematics of Computation, vol. 29, no.131, pp , [46] B. Scholkopf, and A.J. Smola, Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond, Cambridge MA: MIT press, [47] D.B. West, Introduction to graph theory (2 nd ed.), Upper Saddle River: Prentice Hall, [48] J.R. Magnus, On differentiating eigenvalues and eigenvectors, Econometric Theory, vol. 1, no. 2, pp , [49] W. Zou, K. Kpalma, Z. Liu and J. Rosnin, Segmentation driven lowrank matrix recovery for saliency detection, British Machine Vision Conference, 2013.
Supplementary Materials for Salient Object Detection: A
Supplementary Materials for Salient Object Detection: A Discriminative Regional Feature Integration Approach Huaizu Jiang, Zejian Yuan, Ming-Ming Cheng, Yihong Gong Nanning Zheng, and Jingdong Wang Abstract
More informationSupplementary Material for submission 2147: Traditional Saliency Reloaded: A Good Old Model in New Shape
Supplementary Material for submission 247: Traditional Saliency Reloaded: A Good Old Model in New Shape Simone Frintrop, Thomas Werner, and Germán M. García Institute of Computer Science III Rheinische
More informationData-driven Saliency Region Detection Based on Undirected Graph Ranking
Data-driven Saliency Region Detection Based on Undirected Graph Ranking Wenjie Zhang ; Qingyu Xiong 2 ; Shunhan Chen 3 College of Automation, 2 the School of Software Engineering, 3 College of Information
More informationUnsupervised Saliency Estimation based on Robust Hypotheses
Utah State University DigitalCommons@USU Computer Science Faculty and Staff Publications Computer Science 3-2016 Unsupervised Saliency Estimation based on Robust Hypotheses Fei Xu Utah State University,
More informationSalient Region Detection and Segmentation in Images using Dynamic Mode Decomposition
Salient Region Detection and Segmentation in Images using Dynamic Mode Decomposition Sikha O K 1, Sachin Kumar S 2, K P Soman 2 1 Department of Computer Science 2 Centre for Computational Engineering and
More informationSaliency Detection for Videos Using 3D FFT Local Spectra
Saliency Detection for Videos Using 3D FFT Local Spectra Zhiling Long and Ghassan AlRegib School of Electrical and Computer Engineering, Georgia Institute of Technology, Atlanta, GA 30332, USA ABSTRACT
More informationSALIENT OBJECT DETECTION FOR RGB-D IMAGE VIA SALIENCY EVOLUTION.
SALIENT OBJECT DETECTION FOR RGB-D IMAGE VIA SALIENCY EVOLUTION Jingfan Guo,2, Tongwei Ren,2,, Jia Bei,2 State Key Laboratory for Novel Software Technology, Nanjing University, China 2 Software Institute,
More informationSaliency Detection via Graph-Based Manifold Ranking
Saliency Detection via Graph-Based Manifold Ranking Chuan Yang, Lihe Zhang, Huchuan Lu, Xiang Ruan 2, and Ming-Hsuan Yang 3 Dalian University of Technology 2 OMRON Corporation 3 University of California
More informationIMAGE SALIENCY DETECTION VIA MULTI-SCALE STATISTICAL NON-REDUNDANCY MODELING. Christian Scharfenberger, Aanchal Jain, Alexander Wong, and Paul Fieguth
IMAGE SALIENCY DETECTION VIA MULTI-SCALE STATISTICAL NON-REDUNDANCY MODELING Christian Scharfenberger, Aanchal Jain, Alexander Wong, and Paul Fieguth Department of Systems Design Engineering, University
More informationVideo saliency detection by spatio-temporal sampling and sparse matrix decomposition
Video saliency detection by spatio-temporal sampling and sparse matrix decomposition * Faculty of Information Science and Engineering Ningbo University Ningbo 315211 CHINA shaofeng@nbu.edu.cn Abstract:
More informationAS a subfield of computer vision and pattern recognition,
1 Saliency Detection with Spaces of Background-based Distribution Tong Zhao, Lin Li, Xinghao Ding, Yue Huang and Delu Zeng* arxiv:1603.05335v1 [cs.cv] 17 Mar 2016 Abstract In this letter, an effective
More informationImage Compression and Resizing Using Improved Seam Carving for Retinal Images
Image Compression and Resizing Using Improved Seam Carving for Retinal Images Prabhu Nayak 1, Rajendra Chincholi 2, Dr.Kalpana Vanjerkhede 3 1 PG Student, Department of Electronics and Instrumentation
More informationSupplementary Material: Unconstrained Salient Object Detection via Proposal Subset Optimization
Supplementary Material: Unconstrained Salient Object via Proposal Subset Optimization 1. Proof of the Submodularity According to Eqns. 10-12 in our paper, the objective function of the proposed optimization
More informationCo-Saliency Detection Based on Hierarchical Segmentation
88 IEEE SIGNAL PROCESSING LETTERS, VOL. 21, NO. 1, JANUARY 2014 Co-Saliency Detection Based on Hierarchical Segmentation Zhi Liu, Member, IEEE, Wenbin Zou, Lina Li, Liquan Shen, and Olivier Le Meur Abstract
More informationDetecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference
Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference Minh Dao 1, Xiang Xiang 1, Bulent Ayhan 2, Chiman Kwan 2, Trac D. Tran 1 Johns Hopkins Univeristy, 3400
More informationACCORDING to the principle of human visual perception,
IEEE SIGNAL PROCESSING LETTERS, VOL. XX, NO. XX, XXXX 2016 1 Saliency Detection for Stereoscopic Images Based on Depth Confidence Analysis Multiple Cues Fusion Runmin Cong, Student Member, IEEE, Jianjun
More informationA Bi-directional Message Passing Model for Salient Object Detection
A Bi-directional Message Passing Model for Salient Object Detection Lu Zhang, Ju Dai, Huchuan Lu, You He 2, ang Wang 3 Dalian University of Technology, China 2 Naval Aviation University, China 3 Alibaba
More informationImage Resizing Based on Gradient Vector Flow Analysis
Image Resizing Based on Gradient Vector Flow Analysis Sebastiano Battiato battiato@dmi.unict.it Giovanni Puglisi puglisi@dmi.unict.it Giovanni Maria Farinella gfarinellao@dmi.unict.it Daniele Ravì rav@dmi.unict.it
More informationHierarchical Co-salient Object Detection via Color Names
Hierarchical Co-salient Object Detection via Color Names Jing Lou Fenglei Xu Qingyuan Xia Mingwu Ren School of Computer Science and Engineering Nanjing University of Science and Technology Nanjing 210094
More information2.1 Optimized Importance Map
3rd International Conference on Multimedia Technology(ICMT 2013) Improved Image Resizing using Seam Carving and scaling Yan Zhang 1, Jonathan Z. Sun, Jingliang Peng Abstract. Seam Carving, the popular
More informationSmall Object Segmentation Based on Visual Saliency in Natural Images
J Inf Process Syst, Vol.9, No.4, pp.592-601, December 2013 http://dx.doi.org/10.3745/jips.2013.9.4.592 pissn 1976-913X eissn 2092-805X Small Object Segmentation Based on Visual Saliency in Natural Images
More informationSemi-supervised Data Representation via Affinity Graph Learning
1 Semi-supervised Data Representation via Affinity Graph Learning Weiya Ren 1 1 College of Information System and Management, National University of Defense Technology, Changsha, Hunan, P.R China, 410073
More informationA New Framework for Multiscale Saliency Detection Based on Image Patches
Neural Process Lett (2013) 38:361 374 DOI 10.1007/s11063-012-9276-3 A New Framework for Multiscale Saliency Detection Based on Image Patches Jingbo Zhou Zhong Jin Published online: 8 January 2013 Springer
More informationHuman Detection and Tracking for Video Surveillance: A Cognitive Science Approach
Human Detection and Tracking for Video Surveillance: A Cognitive Science Approach Vandit Gajjar gajjar.vandit.381@ldce.ac.in Ayesha Gurnani gurnani.ayesha.52@ldce.ac.in Yash Khandhediya khandhediya.yash.364@ldce.ac.in
More informationSaliency Detection via Dense and Sparse Reconstruction
23 IEEE International Conference on Computer Vision Saliency Detection via Dense and Sparse Reconstruction Xiaohui Li, Huchuan Lu, Lihe Zhang, Xiang Ruan 2, and Ming-Hsuan Yang 3 Dalian University of Technology
More informationJournal of Engineering Science and Technology Review 10 (3) (2017) Research Article
Jestr Journal of Engineering Science and Technology Review 10 (3) (017) 188-194 Research Article Salient Region Detection by Integrating Intrinsic and Extrinsic Cues without Prior Information Ji Ma 1,*,
More informationHierarchical Saliency Detection Supplementary Material
Hierarchical Saliency Detection Supplementary Material Qiong Yan Li Xu Jianping Shi Jiaya Jia The Chinese University of Hong Kong {qyan,xuli,pshi,leoia}@cse.cuhk.edu.hk http://www.cse.cuhk.edu.hk/leoia/proects/hsaliency/
More informationAn Improved Image Resizing Approach with Protection of Main Objects
An Improved Image Resizing Approach with Protection of Main Objects Chin-Chen Chang National United University, Miaoli 360, Taiwan. *Corresponding Author: Chun-Ju Chen National United University, Miaoli
More informationDETECTION OF IMAGE PAIRS USING CO-SALIENCY MODEL
DETECTION OF IMAGE PAIRS USING CO-SALIENCY MODEL N S Sandhya Rani 1, Dr. S. Bhargavi 2 4th sem MTech, Signal Processing, S. J. C. Institute of Technology, Chickballapur, Karnataka, India 1 Professor, Dept
More informationA Survey on Detecting Image Visual Saliency
1/29 A Survey on Detecting Image Visual Saliency Hsin-Ho Yeh Institute of Information Science, Acamedic Sinica, Taiwan {hhyeh}@iis.sinica.edu.tw 2010/12/09 2/29 Outline 1 Conclusions 3/29 What is visual
More informationAn efficient face recognition algorithm based on multi-kernel regularization learning
Acta Technica 61, No. 4A/2016, 75 84 c 2017 Institute of Thermomechanics CAS, v.v.i. An efficient face recognition algorithm based on multi-kernel regularization learning Bi Rongrong 1 Abstract. A novel
More informationImproving Image Segmentation Quality Via Graph Theory
International Symposium on Computers & Informatics (ISCI 05) Improving Image Segmentation Quality Via Graph Theory Xiangxiang Li, Songhao Zhu School of Automatic, Nanjing University of Post and Telecommunications,
More informationDynamic visual attention: competitive versus motion priority scheme
Dynamic visual attention: competitive versus motion priority scheme Bur A. 1, Wurtz P. 2, Müri R.M. 2 and Hügli H. 1 1 Institute of Microtechnology, University of Neuchâtel, Neuchâtel, Switzerland 2 Perception
More informationSaliency Detection using Region-Based Incremental Center-Surround Distance
Saliency Detection using Region-Based Incremental Center-Surround Distance Minwoo Park Corporate Research and Engineering Eastman Kodak Company Rochester, NY, USA minwoo.park@kodak.com Mrityunjay Kumar
More informationHow to Evaluate Foreground Maps?
How to Evaluate Foreground Maps? Ran Margolin Technion Haifa, Israel margolin@tx.technion.ac.il Lihi Zelnik-Manor Technion Haifa, Israel lihi@ee.technion.ac.il Ayellet Tal Technion Haifa, Israel ayellet@ee.technion.ac.il
More informationA Novel Image Super-resolution Reconstruction Algorithm based on Modified Sparse Representation
, pp.162-167 http://dx.doi.org/10.14257/astl.2016.138.33 A Novel Image Super-resolution Reconstruction Algorithm based on Modified Sparse Representation Liqiang Hu, Chaofeng He Shijiazhuang Tiedao University,
More informationTHE human visual system can spot salient objects in a cluttered
1892 IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 39, NO. 9, SEPTEMBER 2017 Ranking Saliency Lihe Zhang, Chuan Yang, Huchuan Lu, Senior Member, IEEE, Xiang Ruan, and Ming-Hsuan
More informationThe Un-normalized Graph p-laplacian based Semi-supervised Learning Method and Speech Recognition Problem
Int. J. Advance Soft Compu. Appl, Vol. 9, No. 1, March 2017 ISSN 2074-8523 The Un-normalized Graph p-laplacian based Semi-supervised Learning Method and Speech Recognition Problem Loc Tran 1 and Linh Tran
More informationA Novel Approach to Saliency Detection Model and Its Applications in Image Compression
RESEARCH ARTICLE OPEN ACCESS A Novel Approach to Saliency Detection Model and Its Applications in Image Compression Miss. Radhika P. Fuke 1, Mr. N. V. Raut 2 1 Assistant Professor, Sipna s College of Engineering
More informationSALIENT OBJECT DETECTION VIA BACKGROUND CONTRAST
SALIENT OBJECT DETECTION VIA BACKGROUND CONTRAST Quan Zhou,, Nianyi Li 2,, Jianxin Chen, Liang Zhou, and Longin Jan Latecki 3 Key Lab of BB WL Comm & Sen Net Tech, Nanjing Univ of Posts & Telecom, Nanjing,
More informationA Survey of Light Source Detection Methods
A Survey of Light Source Detection Methods Nathan Funk University of Alberta Mini-Project for CMPUT 603 November 30, 2003 Abstract This paper provides an overview of the most prominent techniques for light
More informationLearning to Match. Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li
Learning to Match Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li 1. Introduction The main tasks in many applications can be formalized as matching between heterogeneous objects, including search, recommendation,
More informationA FRAMEWORK OF EXTRACTING MULTI-SCALE FEATURES USING MULTIPLE CONVOLUTIONAL NEURAL NETWORKS. Kuan-Chuan Peng and Tsuhan Chen
A FRAMEWORK OF EXTRACTING MULTI-SCALE FEATURES USING MULTIPLE CONVOLUTIONAL NEURAL NETWORKS Kuan-Chuan Peng and Tsuhan Chen School of Electrical and Computer Engineering, Cornell University, Ithaca, NY
More informationIEEE TRANSACTIONS ON MULTIMEDIA, VOL. 14, NO. 4, AUGUST
IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 14, NO. 4, AUGUST 2012 1275 Unsupervised Salient Object Segmentation Based on Kernel Density Estimation and Two-Phase Graph Cut Zhi Liu, Member, IEEE, Ran Shi, Liquan
More informationarxiv: v3 [cs.cv] 2 Oct 2018
Coarse-to-Fine Salient Object Detection with Low-Rank Matrix Recovery Qi Zheng a, Shujian Yu b, Xinge You a, a Huazhong University of Science and Technology, Wuhan, China b University of Florida, Gainesville,
More informationRobust and Efficient Saliency Modeling from Image Co-Occurrence Histograms
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 36, NO. 1, JANUARY 2014 195 Robust and Efficient Saliency Modeling from Image Co-Occurrence Histograms Shijian Lu, Cheston Tan, and
More informationRobust Frequency-tuned Salient Region Detection
Robust Frequency-tuned Salient Region Detection 1 Yong Zhang, 2 Yihua Lan, 3 Haozheng Ren, 4 Ming Li 1 School of Computer Engineering, Huaihai Institute of Technology, Lianyungang, China, zhyhglyg@126.com
More informationarxiv: v2 [cs.cv] 31 Mar 2015
Visual Saliency Based on Multiscale Deep Features Guanbin Li Yizhou Yu Department of Computer Science, The University of Hong Kong https://sites.google.com/site/ligb86/mdfsaliency/ arxiv:1503.08663v2 [cs.cv]
More informationMachine Learning for Media Data Analysis
27 OKTOBER 2010 Machine Learning for Media Data Analysis Assistant Machine Learning and Computer Vision Aarhus University, alexandrosiosifidis@engaudk Overview Media Data Analysis applications Image analysis/recognition/segmentation
More informationChannel Locality Block: A Variant of Squeeze-and-Excitation
Channel Locality Block: A Variant of Squeeze-and-Excitation 1 st Huayu Li Northern Arizona University Flagstaff, United State Northern Arizona University hl459@nau.edu arxiv:1901.01493v1 [cs.lg] 6 Jan
More informationRandom Walks on Graphs to Model Saliency in Images
Random Walks on Graphs to Model Saliency in Images Viswanath Gopalakrishnan Yiqun Hu Deepu Rajan School of Computer Engineering Nanyang Technological University, Singapore 639798 visw0005@ntu.edu.sg yqhu@ntu.edu.sg
More informationPERFORMANCE ANALYSIS OF COMPUTING TECHNIQUES FOR IMAGE DISPARITY IN STEREO IMAGE
PERFORMANCE ANALYSIS OF COMPUTING TECHNIQUES FOR IMAGE DISPARITY IN STEREO IMAGE Rakesh Y. Department of Electronics and Communication Engineering, SRKIT, Vijayawada, India E-Mail: rakesh.yemineni@gmail.com
More informationWhat Makes a Patch Distinct?
What Makes a Patch Distinct? Ran Margolin Technion Haifa, Israel margolin@tx.technion.ac.il Ayellet Tal Technion Haifa, Israel ayellet@ee.technion.ac.il Lihi Zelnik-Manor Technion Haifa, Israel lihi@ee.technion.ac.il
More informationarxiv: v1 [cond-mat.dis-nn] 30 Dec 2018
A General Deep Learning Framework for Structure and Dynamics Reconstruction from Time Series Data arxiv:1812.11482v1 [cond-mat.dis-nn] 30 Dec 2018 Zhang Zhang, Jing Liu, Shuo Wang, Ruyue Xin, Jiang Zhang
More informationarxiv: v1 [cs.cv] 10 May 2017
Learning RGB-D Salient Object Detection using background enclosure, depth contrast, and top-down features arxiv:1705.03607v1 [cs.cv] 10 May 2017 Riku Shigematsu David Feng Australian National University
More informationOpen Access Research on the Prediction Model of Material Cost Based on Data Mining
Send Orders for Reprints to reprints@benthamscience.ae 1062 The Open Mechanical Engineering Journal, 2015, 9, 1062-1066 Open Access Research on the Prediction Model of Material Cost Based on Data Mining
More informationInternational Conference on Information Sciences, Machinery, Materials and Energy (ICISMME 2015)
International Conference on Information Sciences, Machinery, Materials and Energy (ICISMME 2015) Brief Analysis on Typical Image Saliency Detection Methods Wenwen Pan, Xiaofei Sun, Xia Wang, Wei Zhang
More informationEvaluation of regions-of-interest based attention algorithms using a probabilistic measure
Evaluation of regions-of-interest based attention algorithms using a probabilistic measure Martin Clauss, Pierre Bayerl and Heiko Neumann University of Ulm, Dept. of Neural Information Processing, 89081
More informationObject Extraction Using Image Segmentation and Adaptive Constraint Propagation
Object Extraction Using Image Segmentation and Adaptive Constraint Propagation 1 Rajeshwary Patel, 2 Swarndeep Saket 1 Student, 2 Assistant Professor 1 2 Department of Computer Engineering, 1 2 L. J. Institutes
More informationSalient Region Detection via High-Dimensional Color Transform
Salient Region Detection via High-Dimensional Color Transform Jiwhan Kim Dongyoon Han Yu-Wing Tai Junmo Kim Department of Electrical Engineering Korea Advanced Institute of Science and Technology (KAIST)
More informationImage Error Concealment Based on Watermarking
Image Error Concealment Based on Watermarking Shinfeng D. Lin, Shih-Chieh Shie and Jie-Wei Chen Department of Computer Science and Information Engineering,National Dong Hwa Universuty, Hualien, Taiwan,
More informationFASA: Fast, Accurate, and Size-Aware Salient Object Detection
FASA: Fast, Accurate, and Size-Aware Salient Object Detection Gökhan Yildirim, Sabine Süsstrunk School of Computer and Communication Sciences École Polytechnique Fédérale de Lausanne Abstract. Fast and
More informationFast and Efficient Saliency Detection Using Sparse Sampling and Kernel Density Estimation
Fast and Efficient Saliency Detection Using Sparse Sampling and Kernel Density Estimation Hamed Rezazadegan Tavakoli, Esa Rahtu, and Janne Heikkilä Machine Vision Group, Department of Electrical and Information
More informationAn ICA based Approach for Complex Color Scene Text Binarization
An ICA based Approach for Complex Color Scene Text Binarization Siddharth Kherada IIIT-Hyderabad, India siddharth.kherada@research.iiit.ac.in Anoop M. Namboodiri IIIT-Hyderabad, India anoop@iiit.ac.in
More informationMulti-Scale Saliency Detection using Dictionary Learning
Multi-Scale Saliency Detection using Dictionary Learning Shubham Pachori Electrical Engineering Indian Institute of Technology Gandhinagar, India 382355 Email: shubham pachori@iitgn.ac.in arxiv:1611.06307v3
More informationVideo Inter-frame Forgery Identification Based on Optical Flow Consistency
Sensors & Transducers 24 by IFSA Publishing, S. L. http://www.sensorsportal.com Video Inter-frame Forgery Identification Based on Optical Flow Consistency Qi Wang, Zhaohong Li, Zhenzhen Zhang, Qinglong
More informationMain Subject Detection via Adaptive Feature Selection
Main Subject Detection via Adaptive Feature Selection Cuong Vu and Damon Chandler Image Coding and Analysis Lab Oklahoma State University Main Subject Detection is easy for human 2 Outline Introduction
More informationRSRN: Rich Side-output Residual Network for Medial Axis Detection
RSRN: Rich Side-output Residual Network for Medial Axis Detection Chang Liu, Wei Ke, Jianbin Jiao, and Qixiang Ye University of Chinese Academy of Sciences, Beijing, China {liuchang615, kewei11}@mails.ucas.ac.cn,
More informationSEMI-BLIND IMAGE RESTORATION USING A LOCAL NEURAL APPROACH
SEMI-BLIND IMAGE RESTORATION USING A LOCAL NEURAL APPROACH Ignazio Gallo, Elisabetta Binaghi and Mario Raspanti Universitá degli Studi dell Insubria Varese, Italy email: ignazio.gallo@uninsubria.it ABSTRACT
More informationSalient Object Detection: A Discriminative Regional Feature Integration Approach
23 IEEE Conference on Computer Vision and Pattern Recognition Salient Object Detection: A Discriminative Regional Feature Integration Approach Huaizu Jiang Jingdong Wang Zejian Yuan Yang Wu Nanning Zheng
More informationImage retrieval based on bag of images
University of Wollongong Research Online Faculty of Informatics - Papers (Archive) Faculty of Engineering and Information Sciences 2009 Image retrieval based on bag of images Jun Zhang University of Wollongong
More informationCost-alleviative Learning for Deep Convolutional Neural Network-based Facial Part Labeling
[DOI: 10.2197/ipsjtcva.7.99] Express Paper Cost-alleviative Learning for Deep Convolutional Neural Network-based Facial Part Labeling Takayoshi Yamashita 1,a) Takaya Nakamura 1 Hiroshi Fukui 1,b) Yuji
More informationLearning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009
Learning and Inferring Depth from Monocular Images Jiyan Pan April 1, 2009 Traditional ways of inferring depth Binocular disparity Structure from motion Defocus Given a single monocular image, how to infer
More informationRobust Ring Detection In Phase Correlation Surfaces
Griffith Research Online https://research-repository.griffith.edu.au Robust Ring Detection In Phase Correlation Surfaces Author Gonzalez, Ruben Published 2013 Conference Title 2013 International Conference
More informationGraph Autoencoder-Based Unsupervised Feature Selection
Graph Autoencoder-Based Unsupervised Feature Selection Siwei Feng Department of Electrical and Computer Engineering University of Massachusetts Amherst Amherst, MA, 01003 siwei@umass.edu Marco F. Duarte
More informationImage Classification based on Saliency Driven Nonlinear Diffusion and Multi-scale Information Fusion Ms. Swapna R. Kharche 1, Prof.B.K.
Image Classification based on Saliency Driven Nonlinear Diffusion and Multi-scale Information Fusion Ms. Swapna R. Kharche 1, Prof.B.K.Chaudhari 2 1M.E. student, Department of Computer Engg, VBKCOE, Malkapur
More informationTensor Sparse PCA and Face Recognition: A Novel Approach
Tensor Sparse PCA and Face Recognition: A Novel Approach Loc Tran Laboratoire CHArt EA4004 EPHE-PSL University, France tran0398@umn.edu Linh Tran Ho Chi Minh University of Technology, Vietnam linhtran.ut@gmail.com
More informationUsing temporal seeding to constrain the disparity search range in stereo matching
Using temporal seeding to constrain the disparity search range in stereo matching Thulani Ndhlovu Mobile Intelligent Autonomous Systems CSIR South Africa Email: tndhlovu@csir.co.za Fred Nicolls Department
More informationSEMI-SUPERVISED LEARNING (SSL) for classification
IEEE SIGNAL PROCESSING LETTERS, VOL. 22, NO. 12, DECEMBER 2015 2411 Bilinear Embedding Label Propagation: Towards Scalable Prediction of Image Labels Yuchen Liang, Zhao Zhang, Member, IEEE, Weiming Jiang,
More informationMultidirectional 2DPCA Based Face Recognition System
Multidirectional 2DPCA Based Face Recognition System Shilpi Soni 1, Raj Kumar Sahu 2 1 M.E. Scholar, Department of E&Tc Engg, CSIT, Durg 2 Associate Professor, Department of E&Tc Engg, CSIT, Durg Email:
More informationShort Survey on Static Hand Gesture Recognition
Short Survey on Static Hand Gesture Recognition Huu-Hung Huynh University of Science and Technology The University of Danang, Vietnam Duc-Hoang Vo University of Science and Technology The University of
More informationarxiv: v1 [cs.cv] 19 Apr 2016
Deep Saliency with Encoded Low level Distance Map and High Level Features Gayoung Lee KAIST gylee1103@gmail.com Yu-Wing Tai SenseTime Group Limited yuwing@gmail.com Junmo Kim KAIST junmo.kim@kaist.ac.kr
More informationFocusing Attention on Visual Features that Matter
TSAI, KUIPERS: FOCUSING ATTENTION ON VISUAL FEATURES THAT MATTER 1 Focusing Attention on Visual Features that Matter Grace Tsai gstsai@umich.edu Benjamin Kuipers kuipers@umich.edu Electrical Engineering
More informationSynthetic Saliency. Ben Weems Stanford University. Anthony Perez Stanford University. Karan Rai Stanford University. Abstract. 1.
Synthetic Saliency Anthony Perez Stanford University aperez8@stanford.edu Karan Rai Stanford University karanrai@stanford.edu Ben Weems Stanford University bweems@stanford.edu Abstract There are many existing
More informationMATLAB User Guide for Depth Reconstruction from Sparse Samples
MATLAB User Guide for Depth Reconstruction from Sparse Samples Lee-Kang Liu, Stanley H. Chan and Truong Q. Nguyen This MATLAB user guide presents the instructions of how to use the MATLAB functions accompanied
More informationLearning RGB-D Salient Object Detection using background enclosure, depth contrast, and top-down features
Learning RGB-D Salient Object Detection using background enclosure, depth contrast, and top-down features Riku Shigematsu David Feng Australian National University riku.research@gmail.com david.feng@data61.csiro.au
More informationBilevel Sparse Coding
Adobe Research 345 Park Ave, San Jose, CA Mar 15, 2013 Outline 1 2 The learning model The learning algorithm 3 4 Sparse Modeling Many types of sensory data, e.g., images and audio, are in high-dimensional
More informationImage Quality Assessment Techniques: An Overview
Image Quality Assessment Techniques: An Overview Shruti Sonawane A. M. Deshpande Department of E&TC Department of E&TC TSSM s BSCOER, Pune, TSSM s BSCOER, Pune, Pune University, Maharashtra, India Pune
More informationTHE HUMAN visual system interprets the world in a
1150 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 23, NO. 7, JULY 2013 Visual Saliency by Selective Contrast Qi Wang, Yuan Yuan, Senior Member, IEEE, and Pingkun Yan, Senior Member,
More informationSTUDYING THE FEASIBILITY AND IMPORTANCE OF GRAPH-BASED IMAGE SEGMENTATION TECHNIQUES
25-29 JATIT. All rights reserved. STUDYING THE FEASIBILITY AND IMPORTANCE OF GRAPH-BASED IMAGE SEGMENTATION TECHNIQUES DR.S.V.KASMIR RAJA, 2 A.SHAIK ABDUL KHADIR, 3 DR.S.S.RIAZ AHAMED. Dean (Research),
More informationSupplementary material: Strengthening the Effectiveness of Pedestrian Detection with Spatially Pooled Features
Supplementary material: Strengthening the Effectiveness of Pedestrian Detection with Spatially Pooled Features Sakrapee Paisitkriangkrai, Chunhua Shen, Anton van den Hengel The University of Adelaide,
More informationAn Efficient Saliency Based Lossless Video Compression Based On Block-By-Block Basis Method
An Efficient Saliency Based Lossless Video Compression Based On Block-By-Block Basis Method Ms. P.MUTHUSELVI, M.E(CSE), V.P.M.M Engineering College for Women, Krishnankoil, Virudhungar(dt),Tamil Nadu Sukirthanagarajan@gmail.com
More informationDetect Globally, Refine Locally: A Novel Approach to Saliency Detection
Detect Globally, Refine Locally: A Novel Approach to Saliency Detection Tiantian Wang, Lihe Zhang, Shuo Wang, Huchuan Lu, Gang Yang 2, Xiang Ruan 3, Ali Borji 4 Dalian University of Technology, 2 Northeastern
More informationAn Efficient Salient Feature Extraction by Using Saliency Map Detection with Modified K-Means Clustering Technique
International Journal of Computational Engineering & Management, Vol. 15 Issue 5, September 2012 www..org 63 An Efficient Salient Feature Extraction by Using Saliency Map Detection with Modified K-Means
More informationCS 534: Computer Vision Segmentation and Perceptual Grouping
CS 534: Computer Vision Segmentation and Perceptual Grouping Ahmed Elgammal Dept of Computer Science CS 534 Segmentation - 1 Outlines Mid-level vision What is segmentation Perceptual Grouping Segmentation
More information2 Proposed Methodology
3rd International Conference on Multimedia Technology(ICMT 2013) Object Detection in Image with Complex Background Dong Li, Yali Li, Fei He, Shengjin Wang 1 State Key Laboratory of Intelligent Technology
More informationLimitations of Matrix Completion via Trace Norm Minimization
Limitations of Matrix Completion via Trace Norm Minimization ABSTRACT Xiaoxiao Shi Computer Science Department University of Illinois at Chicago xiaoxiao@cs.uic.edu In recent years, compressive sensing
More informationDiffusion Wavelets for Natural Image Analysis
Diffusion Wavelets for Natural Image Analysis Tyrus Berry December 16, 2011 Contents 1 Project Description 2 2 Introduction to Diffusion Wavelets 2 2.1 Diffusion Multiresolution............................
More informationDesign of Orthogonal Graph Wavelet Filter Banks
Design of Orthogonal Graph Wavelet Filter Banks Xi ZHANG Department of Communication Engineering and Informatics The University of Electro-Communications Chofu-shi, Tokyo, 182-8585 JAPAN E-mail: zhangxi@uec.ac.jp
More informationImage Enhancement Techniques for Fingerprint Identification
March 2013 1 Image Enhancement Techniques for Fingerprint Identification Pankaj Deshmukh, Siraj Pathan, Riyaz Pathan Abstract The aim of this paper is to propose a new method in fingerprint enhancement
More information