Least Soft-thresold Squares Tracking
|
|
- Candace Bridges
- 6 years ago
- Views:
Transcription
1 Least Soft-thresold Squares Tracking Dong Wang Dalian University of Technology Huchuan Lu Dalian University of Technology Ming-Hsuan Yang University of California at Merced Abstract In this paper, we propose a generative tracking method based on a novel robust linear regression algorithm. In contrast to eisting methods, the proposed Least Soft-thresold Squares LSS) algorithm models the error term with the Gaussian-Laplacian distribution, which can be solved efficiently. Based on maimum joint likelihood of parameters, we derive a LSS distance to measure the difference between an observation sample and the dictionary. Compared with the distance derived from ordinary least squares methods, the proposed metric is more effective in dealing with outliers. In addition, we present an update scheme to capture the appearance change of the tracked target and ensure that the model is properly updated. Eperimental results on several challenging image sequences demonstrate that the proposed tracker achieves more favorable performance than the state-of-the-art methods.. Introduction Visual tracking plays a critical role in computer vision that finds many practical applications e.g., motion analysis, video surveillance, vehicle navigation and human-computer interaction). Although significant progress has been made in the past decades, developing a robust tracking algorithm is still a challenging problem due to numerous factors such as partial occlusion, illumination variation, pose change, comple motion, and background clutter. Tracking algorithms can be classified as either generative e.g., [7,, 20, 9, 2, 5]) or discriminative e.g., [2, 0, 4,, 27, 8, 30, 4]) methods. Generative methods focus on searching for the regions which are the most similar to the tracked targets, while discriminative methods cast tracking as a classification problem that distinguishes the tracked targets from the surrounding backgrounds. In this work, we propose a robust generative tracker which is able to handle partial occlusion and other challenging factors effectively. Among the generative methods, the trackers based on linear representation maintain holistic appearance information and therefore provide a compact notion of the thing being tracked, which may facilitate some advanced vision tasks [7, 20]. These methods often adopt a dictionary e.g., a set of basis vectors from a subspace or a series of templates) to describe the tracked target. A given candidate sample is linearly represented by the dictionary, and the representation coefficient and reconstruction error are computed, from which the corresponding likelihood belonging to the object class) is determined. Ross et al. [20] propose an incremental visual tracking IVT) method which represents the tracked target by a low dimensional PCA subspace a set of PCA basis vectors) and assumes that the error is Gaussian distributed with small variances i.e., small dense noise). Therefore, the representation coefficient can be obtained by a simple projection operator, which is equivalent to the ordinary least squares solution under the assumption that the dictionary atoms are orthogonal. The reconstruction error is computed by the objective function of the ordinary least squares methods. While the IVT method is effective to handle appearance change caused by illumination variation and pose variation, it is not robust to some challenging factors e.g., partial occlusion and background clutter) due to the following two reasons. First, ordinary least squares methods have been shown to be sensitive to outliers due to the formulation based on reconstruction error with Gaussian noise assumption. Second, the IVT method uses new observations to update the observation model without detecting outliers and processing them accordingly. Other recent tracking algorithms [6, 2] based on the Gaussian noise assumption or the ordinary least squares methods have similar problems as the IVT method. Motivated by the success of sparse representation-based face recognition [28], Mei et al. [9] develop a novel l tracker that uses a series of target templates and trivial templates to model the tracked target, where the target templates are used to describe the object class to be tracked and trivial templates are used to deal with outliers e.g., partial occlusion) with the sparsity constraints. For tracking, a candidate sample can be sparsely represented by both target and trivial templates, and its corresponding likelihood is determined by the reconstruction error with respect to target templates. We note that this formulation is a linear regres-
2 sion problem with sparsity constraints on the representation coefficients. Recently, several methods have been proposed to improve the l tracker in terms of both speed and accuracy by using accelerated proimal gradient algorithm [5], replacing raw piel templates with orthogonal basis vectors [24, 26, 25], modeling the similarity between different candidates [3], to name a few. Although these algorithms consider outliers by using additional trivial templates, this formulation can be generalized with better understanding. In this work, we show that the linear regression with the Gaussian-Laplacian noise assumption is more effective in dealing with outliers for object tracking. In addition, from the viewpoint of linear regression, it is not suitable to estimate the likelihood based on the reconstruction error with respect to target templates. We present a novel distance function to compute the distance between a candidate and the object class. In this paper, we present a generative tracking algorithm based on linear regression. The contributions of this work are as follows. First, we introduce a novel linear regression method, Least Soft-thresold Squares LSS), which assumes that the error vectors follow the i.i.d Gaussian- Laplacian distribution. Second, we present an efficient iteration method to solve the LSS problem and propose a LSS distance to measure the dissimilarity between the observation vector and dictionary. We note that the LSS method is related to the robust regression with the Huber loss function and is effective in detecting outliers. Compared with the least squares distance, the LSS distance is more effective in measuring the distance between the observation vector and dictionary when outliers occur. Third, we design a generative tracker by using the LSS method, where the dictionary consists of PCA basis vectors. The likelihood of each candidate is computed based on the LSS distance. Furthermore, we update the tracker by using an effective update scheme. Numerous eperiments on challenging image sequences with comparisons to state-of-the-art tracking methods demonstrate the effectiveness of the proposed model and algorithm. 2. Least Soft-thresold Squares 2.. Regression with Gaussian-Laplacian Noise The objective of linear regression is to fit a linear model i.e, estimate model parameters) to a series of noisy observations: y = A + e, ) where y R d is a d-dimensional observation vector, R k denotes the k-dimensional parameter or coefficient) vector to be estimated and A = [r ; r 2 ;...; r d ] R d k represents the input data matri r i is the i-th row of A). We denote e = y A = [e ; e 2 ;...; e d ] i.e, e i = y i r i, i =,..., d) as the error or residual term. The coefficient can be obtained by maimizing the posteriori probability p y), which is also equivalent to maimizing the joint likelihood probability p, y). Assume there is a uniform prior, the coefficient is estimated by = arg ma py ) = arg ma p e), which is the maimum likelihood estimation MLE). The errors e, e 2,..., e d are usually assumed to be independently and identically distributed i.i.d) according to some probability density function PDF) f θ e i ), where θ is the parameter set that characterizes the probability distribution. Thus, the likelihood of the estimator the joint probability of the error term e) is p e) = d i= f θ e i ). To maimize the likelihood function is equivalent to minimizing the objective function L θ e, e 2,..., e d ) = d i= ρ θ e i ), where ρ θ e i ) = log f θ e i ). When the error e = y A follows the Gaussian distribution e i N 0, σ 2 N ) ), the MLE solution is equivalent to the ordinary least squares OLS) solution = arg min 2 y A 2 2, 2) and the closed form solution is = A A ) A y. Although the OLS method is easy to solve, it is sensitive to outliers due to the Gaussian noise assumption. If the error e follows the Laplacian distribution e i L 0, σ L ) 2 ), the MLE solution is equivalent to least absolute deviations LAD) solution, = arg min y A. 3) Compared with the OLS method, the LAD method is robust to outliers. However, it is difficult to be solved by using either the simple-based methods [6] or the iteratively reweighted least squares methods [22]. In this paper, we model error vector e as an additive combination of two independent components: an i.i.d Gaussian noise vector n n i N 0, σ 2 N ) ) and an i.i.d Laplacian noise vector s s i L 0, σ L )), y = A + n + s, 4) where the Gaussian component models small dense noise and the Laplacian one aims to handle outliers 3. e i is a zero-mean Gaussian random variable with variance σn 2 and its ) PDF is f N e i ) = ep e2 i 2πσN 2σ N 2. 2 e i is a zero-mean Laplacian random variable with variance σl 2 and its PDF is f L e i ) = 2σL ep 2 ei σ L ). 3 The similar idea of decomposing the noise into two components e.g., decomposing the noise into sparse and non-sparse ones [3, 8] for achieving robust motion estimation) have appeared in computer vision community. While the goals are similar, the formulations and derivations are different. In this work, the proposed algorithm focuses on not only handling occlusion but also deriving a novel distance metric to compare the target candidate and the target template under the outlier condition, which will presented later.
3 The additive combination of i.i.d Gaussian and i.i.d Laplacian noise variables is also called Gaussian-Laplacian distribution [23]. Its joint PDF is p e) = d i= f N L e i ), where the PDF f N L e i ) is given by the convolution, f N L e i ) = f N n i ) f L s i ) = f L s i ) f N e i s i ) ds i = 2 2σ N ep e2 i 2σ 2 N ) erfc σn σl ei 2σN ) +erfc σn σ L + ei 2σN ) 5) where erfc ) = ep 2) erfc ) and erfc ) = 2 π, e t2 dt. Compared with the Gaussian and Laplacian distributions, the PDF of the Gaussian-Laplacian distribution is comple. Thus, it is difficult to obtain a simple objective function such as Eqs. 2 and 3) directly. Because of this, we treat the Laplacian noise term s as missing values with the same Laplacian prior, and therefore the joint likelihood p y, ) is converted to p y,, s). p y,, s) = p y, s) p, s) = p y A s) p s) = K ep { σ 2 N 2 y A s λ s )}, ) d ) d where K = 2σ 2 2σL 2πσN and λ = N. Thus, to maimize the joint likelihood p y,, s) is equivalent to minimizing the function 2 y A s λ s with respect to both and s Least Soft-thresold Squares Regression To maimize the joint likelihood of Eq. 6, we consider the objective function: σ L 6) L, s) = 2 y A s λ s, 7) and the optimal solution is [, ŝ] = arg min L, s). The,s above objective function is based on the standard least squares criterion and an l regularization term on s, and thus is conve but not differentiable everywhere. To the best of our knowledge, there is no closed-form solution for this optimization problem, so we present an iterative algorithm. Proposition : Given ŝ, the optimal can be computed by the ordinary least squares solution = A A ) A y ŝ). Proposition 2: Given, the optimal ŝ can be obtained by a soft-thresholding or shrinkage) operation S λ y A ), where S τ ) = ma τ, 0)sgn), where sgn ) is the sign function. Let P denote A A ) A, which can be precomputed before the iterative process. By Propositions and 2, the optimization can be solved efficiently. The iterative operation is terminated when a stopping criterion is met e.g., the difference of objective values between two iterations or the number of iterations). As our algorithm consists of two main components: ordinary least squares and soft-thresholding operation, we denote it as the Least Soft-threshold Squares Regression method. Table. Least Soft-threshold Squares Regression Input: An observation vector y, matri A, percomputed matri P = A A ) A, and a small constant λ. : Initialize s 0 = 0 and i = 0 2: Iterate 3: Obtain i+ via i+ = P y s i ) 4: Obtain s i+ via s i+ = S λ y A i+ ) 5: i i + 6: Until convergence or termination Output:, ŝ It is clear that L i+, s i+ ) L i+, s i ) L i, s i ) 4. Hence, the proposed iterative algorithm converges to a local minimal value. As the objective function is conve, it also obtains the global minimal solution. Remark : The proposed least soft-threshold squares regression is equivalent to robust regression with the Huber loss function: where = arg min d i= f e i), e i = y i r i, 8) f e) = { e 2 / 2, e λ λ e λ 2/ 2, e > λ is the Huber loss function and r i denotes the i-th row of A. A detailed eplanation can be found in the supplementary material. We note the approach to compute robust regression with the Huber loss function is less efficient than the proposed method as it is generally solved by using the iteratively reweighted least squares scheme which requires solving a weighted least squares problem i.e., pseudo-inverse) at each iteration. As mentioned before, the pseudo-inverse matri can be percalculated before the iterative process in the proposed LSS method. Remark 2: The non-zero components in s can be used to identify outliers. 4 L i+, s i ) = min L, s i ) L i, s i ), and L i+, s i+ ) = min s L i+, s) L i+, s i ) 9)
4 y 5 0 y 5 0 s z 5 0 st Iteration 5th Iteration 0th Iteration 20th Iteration z a) Observation samples. b) Iterative process in the LSS method. c) The Laplacian noise term. Figure. Robust line fitting by using the least soft-threshold squares LSS) regression. z Figure shows an eample of fitting a straight line when small Gaussian noise and outliers occur simultaneously. It requires to estimate an accurate parameter for a straight line function y = az + b, where z is the input variable, y is the output variable, and a and b are parameters to be determined. Figure a) shows ten observation sample pairs {z i, y i }, i =,..., 0, where z i = i in this case z 5 and z 9 are two outliers). Denote that y = [y ; y 2 ;...; y 0 ], z = [z ; z 2 ;...; z 0 ], A = [z, ] and = [a; b], we use the proposed least soft-threshold squares LSS) method to estimate the parameter. Figure b) illustrates the iterative process of the LSS method λ is set to ), from which it is clear that the proposed LSS is not sensitive to outliers. We note that the result after the first iteration is the same as the result obtained using the OLS method, which also shows that the proposed LSS method is more robust than the OLS method. In addition, as shown in Figure c), the non-zero Laplacian noise components correspond to the outliers Least Soft-thresold Squares Distance We represent the matri A = [a, a 2,..., a k ], where a i is the i-th column of A. The vector, A, can be viewed as a linear combination of the columns of A A = a + 2 a k a k ). The matri A is known as dictionary or basis matri, and the vector a i is called an atom or basis vector. For some vision applications such as tracking), it requires not only to estimate the coefficient accurately but also to define a distance between a noisy observation and the dictionary or the subspace. This generative perspective has commonly been eploited in subspace-based methods [7, 20]. The distance is usually defined to be inversely proportional to the maimum joint likelihood with respect to the coefficient, d y; A) log ma p y, ) 0) = log ma p y ) p ). Take the ordinary least squares method for eample i.e., uniform prior), we have log ma p y ) p ) log ma p y ) ) log ma ep 2 y A 2 2 min 2 y A 2 2. ) Recall that the OLS method assumes the observation vector with i.i.d Gaussian noise, the distance y and A can therefore be defined as, d OLS y; A) = min 2 y A 2 2 y A A ) A y. = ) Similarly, under the i.i.d Laplacian noise assumption, the distance between y and A can be defined as, d LAD y; A) = min y A, 3) which is difficult to be calculated as mentioned in Section 2.). In this work, we adopt the Gaussian-Laplacian distribution to model observation noise. Therefore, we define the distance between y and A under the i.i.d Gaussian- Laplacian noise assumption as, d LSS y; A) = min,s 2 y A s λ s. 4) As it is related to the Least Soft-thresold Squares LSS) regression, we denote it as the LSS distance. Figure 2 illustrates a toy eample of good and bad candidates for template matching with partial occlusion. For simplification, we merely use a single template shown in Figure 2a). Figure 2b) and c) show good and bad candidates shown in red and blue boes respectively). In Table 2, we report the OLS distance and the LSS distance between the template and different candidates. We denote the
5 Table 2. The OLS and LSS distances between the template and different candidates of Figure 2. d OLS d LSS d LSS λ = 0.05) λ = 0.) Good Candidate Bad Candidate template as t, the good candidate as y G and the bad candidate as y B respectively. In this eample, d OLS y B ; t) is smaller than d OLS y G ; t), which means the bad candidate is picked if the OLS distance is used. On the other hand, the good candidate is selected d LSS y G ; t) < d LSS y B ; t)) when the proposed LSS distance is used. Thus, we note that the proposed LSS distance is better than the OLS distance for handling outliers e.g., partial occlusion). where p t t ) is the motion model that describes the state transition between consecutive frames, and p y t t ) is the observation model that estimates the likelihood of an observed image patch belonging to the object class. The affine motion model is used in this work and the state transition is formulated by random walk, i.e., p t t ) = N t ; t, Σ), where Σ is a diagonal covariance matri that indicates the variances of affine parameters. Observation model: In this paper, we assume that the tracked target object is generated by a PCA subspace spanned by U and centered at µ) with i.i.d Gaussian- Laplacian noise y = µ + Uz + n + s, 7) where y denotes an observation vector, U represents a matri of column basis vectors, z indicates the coefficients of basis vectors, n is the Gaussian noise component and s is the Laplacian noise component. Based on the discussion in Section 2, under the i.i.d Gaussian-Laplacian noise assumption, the distance between the vector y and the subspace U, µ) is the least softthreshold squares distance, d y; U, µ) = min z,s 2 y Uz s λ s, 8) a) Template b) Good Candidate c) Bad Candidate Figure 2. A toy eample of good and bad candidates for template matching. 3. Least Soft-thresold Squares Tracking In this paper, visual tracking is treated as a dynamic Bayesian inference task with a hidden Markov model. Given a set of observed image vectors y :t = {y, y 2,..., y t } up to the t-th frame, the aim is to estimate the target state variable t by using the maimum a posteriori estimation, t = arg ma p i ) t y :t, 5) i t where i t indicates the i-th sample of the state t. Based on the Bayes theorem, the posterior distribution p t y :t ) can be estimated recursively by, p t y :t ) p y t t ) p t t ) p t y :t ) t, 6) where y = y µ. Thus, for each observation y i corresponding to a predicted state i, we firstly solve the following optimization problem, ] [ẑi, ŝi = arg min y z i,s i 2 i Uz i s i 2 + λ s i 2, 9) where i denotes the i-th sample of the state without loss of generality, we drop the frame inde t). As the PCA basis vectors U is orthogonal, the per-computed matri P can be simply set to U. After the optimal ẑi and ŝi are obtained, the least soft-threshold squares distance can be calculated by d y i ; U, µ ) = 2 y i 2 Uẑi ŝi + λ ŝi. Then 2 the observation likelihood can be measured by p y i i) = ep γd y i ; U, µ )), 20) where γ is a constant controlling the shape of the Gaussian kernel. Model Update: We note that the non-zero components in the Laplacian noise term can be used to identify outliers. Thus, we present a simple yet effective update scheme. After obtaining the best candidate state of each frame, [ we etract ] its corresponding observation vector y o = y [ o ; yo; 2...; y ] o d and infer the Laplacian noise term so = s o ; s 2 o...; s d o. Then we reconstruct the observation vector
6 by replacing the outliers with its corresponding parts of the mean vector µ, { yr i y i = o, s i o = 0 µ i, s i o 0, 2) where y r = [ y r; y 2 r;...; y d r ] denotes the reconstructed vector and µ = [ µ r; µ 2 r;...; µ d r] is the mean vector. The reconstructed sample is cumulated and then used to update the tracker i.e., PCA basis vectors U and the mean vector µ) by using an incremental principal component analysis PCA) method [20]. 4. Eperiments The proposed tracker is implemented in MATLAB and runs at 3.5 frames per second on a PC with Intel i CPU 3.4 GHz) with 32 GB memory. The regularization constant λ is set to 0. in all eperiments. For each sequence, the location of the tracked target is manually labeled in the first frame. We resize each image observation to piels and use 6 eigenvectors for PCA representation. As a trade-off between effectiveness and speed, 600 particles are adopted and our tracker is incrementally updated every 5 frames. The MATLAB source codes, datasets and supplementary materials are available on our websites ice.dlut.edu.cn/lu/publications.html, faculty.ucmerced.edu/mhyang/pubs.html). In this work, we use fifteen challenging image sequences from prior work [20, 9, 4, 5, 29] and the CAVIAR data set CAVIAR/CAVIARDATA/). The challenging factors of these sequences include partial occlusion, illumination variation, pose change, background clutter and motion blur. We evaluate the proposed tracker against ten state-of-theart algorithms, including the FragT [], IVT [20], MIL [4], VTD [5], TLD [4], APGL [5], MTT [3], LSAT [7], SCM [32], ASLSA [3] and OSPT [26] trackers. For fair evaluation, we use the source codes provided by the authors and run them with adjusted parameters. 4.. Quantitative Evaluation We evaluate the above-mentioned algorithms using two criteria: the center location error and the overlap rate. Table 3 reports the average center location errors in piels, where a smaller average error means a more accurate result. In addition, we use the segmentation criterion in the PAS- CAL VOC challenge [9] to evaluate the overlap rate. Given the tracking result bounding bo) of each frame R T and the corresponding ground truth bounding bo R G, the overlap. Table 4 reports the average overlap rates, where larger average scores mean more accurate results. score is defined as score = arear T R G ) arear T R G ) 4.2. Qualitative Evaluation Severe Occlusion: We test several sequences Occlusion, Occlusion2, Caviar, Caviar2, Caviar3, DavidOutdoor) with heavy or long-time partial occlusion, scale change and rotation. Figure 3 a-c) demonstrate that the proposed method performs well in terms of position, rotation and scale when the target undergoes severe occlusion. This can be attributes to two reasons: ) the proposed LSS distance takes outliers e.g., occlusion) into account eplicitly; and 2) the update scheme is able to avoid degrading the observation model by removing the outliers from new observed samples. In addition, the SCM and ASLAS methods also achieve good performance in most cases as both of them include part-based representations with overlapping patches. The IVT method is sensitive to partial occlusion Occlusion2, Caviar, Caviar3, DavidOutdoor) since the OLS distance is not effective to handle outliers. The MIL and TLD methods do not perform well when the target object is occluded by a similar object Caviar, Caviar2, Caviar3). This can be eplained by that the rectangle features they adopted generalized Haar-liker features or binary patterns) are less effective when similar objects occlude each other. Illumination Change: Figure 3 d) shows the tracking results in the sequences DavidIndoor, Car4, Singer) with significant illumination variation and both scale and pose change. We can see that the MIL and TLD methods are less effective in these cases e.g., DavidIndoor #0350 and Singer #0090). Due to the use of incremental PCA algorithm, the proposed tracker achieves good performance in dealing with the appearance change caused by light change. For the same reason, the IVT and ASLSA methods also perform well. Background Clutter: Figure 3 e) demonstrates the tracking results in the Car, Deer and Football sequences with background clutter. These videos also pose other challenging factors including illumination variation Car), fast motion Deer) and partial occlusion Football). As the proposed LSS distance encourages good matching results when outliers occur, our tracker performs better than other methods in these videos e.g., Deer #0052 and Football #035). Fast Motion: Figure 3 f) illustrates the tracking results on the Jumping, Owl and Face sequences. It is difficult to predict the locations of the tracked objects when they undergo abrupt motion. Furthermore, the appearance change caused by motion blur poses great challenges for capturing the tracked targets accurately and updating the observation models properly. We can see that the TLD and proposed methods perform better than other algorithms e.g., Owl #0248 and Face #0254). We note that the TLD method is equipped with a re-initialization mechanism which facilitates object tracking.
7 Table 3. Average center location error in piels). The best three results are shown in red, blue, and green fonts. Sequence FragT IVT MIL VTD TLD APGL MTT LSAT SCM ASLAS OSPT LSSTOurs) Occlusion Occlusion Caviar Caviar Caviar DavidOutdoor DavidIndoor Singer Car Car Deer Football Jumping Owl Face Average Table 4. Average overlap rate. The best three results are shown in red, blue, and green fonts. Sequence FragT IVT MIL VTD TLD APGL MTT LSAT SCM ASLAS OSPT LSSTOurs) Occlusion Occlusion Caviar Caviar Caviar DavidOutdoor DavidIndoor Singer Car Car Deer Football Jumping Owl Face Average Conclusion In this paper, we propose a Least Soft-thresold Squares LSS) regression method that assumes the noise is Gaussian-Laplacian distributed, and apply it to object tracking. We present an efficient iteration algorithm to solve the LSS problem, which achieves the global minimal solution. We derive a LSS distance to measure the difference between an observation sample and the dictionary. The LSS distance is effective in handling outliers and therefore provides an accurate match, which facilitates object tracking e.g., in dealing with partial occlusion). In addition, we develop a robust generative tracker based on the proposed LSS method and a simple update scheme. Both quantitative and qualitative evaluations on challenging image sequences show that the proposed tracker performs favorably against several state-of-the-art algorithms. In the future, we will etend the LSS method to solve other vision problems e.g., face recognition). Acknowledgements: D. Wang and H. Lu are supported by the National Natural Science Foundation of China # and # M.-H. Yang is supported by the US National Science Foundation CAREER Grant #49783 and IIS Grant # References [] A. Adam, E. Rivlin, and I. Shimshoni. Robust fragments-based tracking using the integral histogram. In CVPR, pages , [2] S. Avidan. Ensemble tracking. TPAMI, 292):26, [3] A. Ayvaci, M. Raptis, and S. Soatto. Occlusion detection and motion estimation with conve optimization. In NIPS, pages 00 08, 200. [4] B. Babenko, M.-H. Yang, and S. Belongie. Visual tracking with online multiple instance learning. In CVPR, pages , [5] C. Bao, Y. Wu, H. Ling, and H. Ji. Real time robust l tracker using accelerated proimal gradient approach. In CVPR, pages , 202. [6] I. Barrodale and F. D. K. Roberts. An improved algorithm for discrete l linear approimation. SIAM Journal on Numerical Analysis, 05): , 973. [7] M. J. Black. EigenTracking : Robust Matching and Tracking of Articulated Objects Using a View-Based Representation. IJCV, 26):63 84, 998. [8] Z. Chen, J. Wang, and Y. Wu. Decomposing and regularizing sparse/non-sparse components for motion field estimation. In CVPR, pages , 202. [9] M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn, and A. Zisserman. The pascal visual object classes voc) challenge. IJCV, 882): , 200. [0] H. Grabner and H. Bischof. On-line boosting and vision. In CVPR, pages , [] S. Hare, A. Saffari, and P. H. S. Torr. Struck: Structured output tracking with kernels. In ICCV, pages , 20. [2] W. Hu, X. Li, X. Zhang, X. Shi, S. J. Maybank, and Z. Zhang. Incremental tensor subspace learning and its applications to foreground segmentation and tracking. IJCV, 93): , 20. [3] X. Jia, H. Lu, and M.-H. Yang. Visual tracking via adaptive structural local sparse appearance model. In CVPR, pages , 202. [4] Z. Kalal, K. Mikolajczyk, and J. Matas. Tracking-learning-detection. TPAMI, 347): , 202.
8 a) Tracking results on sequence Occlusion and Occlusion2 with heavy occlusion and in-plane rotation. b) Tracking results on sequence Caviar and Caviar2 with partial occlusion and scale change. c) Tracking results on sequence Caviar3 and DavidOutdoor with severe occlusion, scale change and pose variation. d) Tracking results on sequence DavidIndoor, Car4 and Singer with illumination variation. e) Tracking results on sequence Car, Deer and Football with background clutter. f) Tracking results on sequence Jumping, Owl and Face with abrupt motion. IVT MIL TLD APGL SCM ASLAS Our Figure 3. Sample tracking results on fifteen challenging image sequences. This figure demonstrates the results of the IVT [20], MIL [4], TLD [4], APGL [5], SCM [32], ASLAS [3] and the proposed methods. More results can be found in the supplementary material. [5] J. Kwon and K. M. Lee. Visual tracking decomposition. In CVPR, pages , 200. [24] D. Wang and H. Lu. Object tracking via 2DPCA and ` -regularization. SPL, 9):7 74, 202. [6] X. Li, W. Hu, Z. Zhang, X. Zhang, M. Zhu, and J. Cheng. Visual tracking via incremental log-euclidean Riemannian subspace learning. In CVPR, pages 8, [25] D. Wang and H. Lu. On-line learning parts-based representation via incremental orthogonal projective non-negative matri factorization. SP, 93: , 203. [7] B. Liu, J. Huang, L. Yang, and C. A. Kulikowski. Robust tracking using local sparse appearance model and k-selection. In CVPR, pages , 20. [26] D. Wang, H. Lu, and M.-H. Yang. Online object tracking with sparse prototypes. TIP, 22):34 325, 203. [8] H. Lu, S. Lu, D. Wang, S. Wang, and H. Leung. Piel-wise spatial pyramidbased hybrid tracking. TCSVT, 229): , 202. [27] S. Wang, H. Lu, F. Yang, and M.-H. Yang. Superpiel tracking. In ICCV, pages , 20. [9] X. Mei and H. Ling. Robust visual tracking using ` minimization. In ICCV, pages , [28] J. Wright, A. Y. Yang, A. Ganesh, S. S. Sastry, and Y. Ma. Robust face recognition via sparse representation. TPAMI, 32):20 227, [20] D. Ross, J. Lim, R.-S. Lin, and M.-H. Yang. Incremental learning for robust visual tracking. IJCV, 77-3):25 4, [29] Y. Wu, H. Ling, J. Yu, F. Li, X. Mei, and E. Cheng. Blurred target tracking by blur-driven tracker. In ICCV, pages 00 07, 20. [2] J. Santner, C. Leistner, A. Saffari, T. Pock, and H. Bischof. PROST: Parallel robust online simple tracking. In CVPR, pages , 200. [30] K. Zhang, L. Zhang, and M.-H. Yang. Real-time compressive tracking. In ECCV, pages , 202. [22] E. J. Schlossmacher. An iterative technique for absolute deviations curve fitting. JOSA, 68344): , 973. [3] T. Zhang, B. Ghanem, S. Liu, and N. Ahuja. Robust visual tracking via multitask sparse learning. In CVPR, pages , 202. [23] I. W. Selesnick. The estimation of laplace random vectors in additive white gaussian noise. TSP, 568-): , [32] W. Zhong, H. Lu, and M.-H. Yang. Robust object tracking via sparsity-based collaborative model. In CVPR, pages , 202.
Structural Sparse Tracking
Structural Sparse Tracking Tianzhu Zhang 1,2 Si Liu 3 Changsheng Xu 2,4 Shuicheng Yan 5 Bernard Ghanem 1,6 Narendra Ahuja 1,7 Ming-Hsuan Yang 8 1 Advanced Digital Sciences Center 2 Institute of Automation,
More informationEnhanced Laplacian Group Sparse Learning with Lifespan Outlier Rejection for Visual Tracking
Enhanced Laplacian Group Sparse Learning with Lifespan Outlier Rejection for Visual Tracking Behzad Bozorgtabar 1 and Roland Goecke 1,2 1 Vision & Sensing, HCC Lab, ESTeM University of Canberra 2 IHCC,
More informationarxiv: v2 [cs.cv] 9 Sep 2016
arxiv:1608.08171v2 [cs.cv] 9 Sep 2016 Tracking Completion Yao Sui 1, Guanghui Wang 1,4, Yafei Tang 2, Li Zhang 3 1 Dept. of EECS, University of Kansas, Lawrence, KS 66045, USA 2 China Unicom Research Institute,
More informationOnline visual tracking by integrating spatio-temporal cues
Published in IET Computer Vision Received on 12th September 2013 Revised on 27th March 2014 Accepted on 14th May 2014 ISSN 1751-9632 Online visual tracking by integrating spatio-temporal cues Yang He,
More informationTHE goal of object tracking is to estimate the states of
2356 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, NO. 5, MAY 2014 Robust Object Tracking via Sparse Collaborative Appearance Model Wei Zhong, Huchuan Lu, Senior Member, IEEE, and Ming-Hsuan Yang, Senior
More informationAS ONE of the fundamental problems in computer vision,
34 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL., NO., JANUARY 3 Online Object Tracking With Sparse Prototypes Dong Wang, Huchuan Lu, Member, IEEE, and Ming-Hsuan Yang, Senior Member, IEEE Abstract Online
More informationSuperpixel Tracking. The detail of our motion model: The motion (or dynamical) model of our tracker is assumed to be Gaussian distributed:
Superpixel Tracking Shu Wang 1, Huchuan Lu 1, Fan Yang 1 abnd Ming-Hsuan Yang 2 1 School of Information and Communication Engineering, University of Technology, China 2 Electrical Engineering and Computer
More informationarxiv: v1 [cs.cv] 24 Feb 2014
EXEMPLAR-BASED LINEAR DISCRIMINANT ANALYSIS FOR ROBUST OBJECT TRACKING Changxin Gao, Feifei Chen, Jin-Gang Yu, Rui Huang, Nong Sang arxiv:1402.5697v1 [cs.cv] 24 Feb 2014 Science and Technology on Multi-spectral
More informationStruck: Structured Output Tracking with Kernels. Presented by Mike Liu, Yuhang Ming, and Jing Wang May 24, 2017
Struck: Structured Output Tracking with Kernels Presented by Mike Liu, Yuhang Ming, and Jing Wang May 24, 2017 Motivations Problem: Tracking Input: Target Output: Locations over time http://vision.ucsd.edu/~bbabenko/images/fast.gif
More informationVisual Tracking Using Pertinent Patch Selection and Masking
Visual Tracking Using Pertinent Patch Selection and Masking Dae-Youn Lee, Jae-Young Sim, and Chang-Su Kim School of Electrical Engineering, Korea University, Seoul, Korea School of Electrical and Computer
More informationTracking Completion. Institute of Automation, CAS, Beijing, China
Tracking Completion Yao Sui 1(B), Guanghui Wang 1,4, Yafei Tang 2, and Li Zhang 3 1 Department of EECS, University of Kansas, Lawrence, KS 66045, USA suiyao@gmail.com, ghwang@ku.edu 2 China Unicom Research
More informationTHE recent years have witnessed significant advances
IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, NO. 4, APRIL 2014 1639 Robust Superpixel Tracking Fan Yang, Student Member, IEEE, Huchuan Lu, Senior Member, IEEE, and Ming-Hsuan Yang, Senior Member, IEEE
More informationTracking Using Online Feature Selection and a Local Generative Model
Tracking Using Online Feature Selection and a Local Generative Model Thomas Woodley Bjorn Stenger Roberto Cipolla Dept. of Engineering University of Cambridge {tew32 cipolla}@eng.cam.ac.uk Computer Vision
More informationKeywords:- Object tracking, multiple instance learning, supervised learning, online boosting, ODFS tracker, classifier. IJSER
International Journal of Scientific & Engineering Research, Volume 5, Issue 2, February-2014 37 Object Tracking via a Robust Feature Selection approach Prof. Mali M.D. manishamali2008@gmail.com Guide NBNSCOE
More informationTracking. Hao Guan( 管皓 ) School of Computer Science Fudan University
Tracking Hao Guan( 管皓 ) School of Computer Science Fudan University 2014-09-29 Multimedia Video Audio Use your eyes Video Tracking Use your ears Audio Tracking Tracking Video Tracking Definition Given
More informationReal-time Object Tracking via Online Discriminative Feature Selection
IEEE TRANSACTION ON IMAGE PROCESSING 1 Real-time Object Tracking via Online Discriminative Feature Selection Kaihua Zhang, Lei Zhang, and Ming-Hsuan Yang Abstract Most tracking-by-detection algorithms
More informationMulti-Cue Visual Tracking Using Robust Feature-Level Fusion Based on Joint Sparse Representation
Multi-Cue Visual Tracking Using Robust Feature-Level Fusion Based on Joint Sparse Representation Xiangyuan Lan Andy J Ma Pong C Yuen Department of Computer Science, Hong Kong Baptist University {xylan,
More information2 Cascade detection and tracking
3rd International Conference on Multimedia Technology(ICMT 213) A fast on-line boosting tracking algorithm based on cascade filter of multi-features HU Song, SUN Shui-Fa* 1, MA Xian-Bing, QIN Yin-Shi,
More informationOnline Graph-Based Tracking
Online Graph-Based Tracking Hyeonseob Nam, Seunghoon Hong, and Bohyung Han Department of Computer Science and Engineering, POSTECH, Korea {namhs09,maga33,bhhan}@postech.ac.kr Abstract. Tracking by sequential
More informationOnline Multiple Support Instance Tracking
Online Multiple Support Instance Tracking Qiu-Hong Zhou Huchuan Lu Ming-Hsuan Yang School of Electronic and Information Engineering Electrical Engineering and Computer Science Dalian University of Technology
More informationOutline. Advanced Digital Image Processing and Others. Importance of Segmentation (Cont.) Importance of Segmentation
Advanced Digital Image Processing and Others Xiaojun Qi -- REU Site Program in CVIP (7 Summer) Outline Segmentation Strategies and Data Structures Algorithms Overview K-Means Algorithm Hidden Markov Model
More informationTracking Algorithms. Lecture16: Visual Tracking I. Probabilistic Tracking. Joint Probability and Graphical Model. Deterministic methods
Tracking Algorithms CSED441:Introduction to Computer Vision (2017F) Lecture16: Visual Tracking I Bohyung Han CSE, POSTECH bhhan@postech.ac.kr Deterministic methods Given input video and current state,
More informationReal Time Face Tracking By Using Naïve Bayes Classifier
Real Time Face Tracking By Using Naïve Bayes Classifier Krishna Priya Arni 1, Satish Ammuloju 2, Jagan Mohan Vithalachari Gollapalli 3, Dinesh Bhogapurapu 4) Shaik Azeez 5 Student, Department of ECE, Lendi
More informationEfficient Visual Tracking via Low-Complexity Sparse Representation
Efficient Visual Tracking via Low-Complexity Sparse Representation Weizhi Lu, Jinglin Zhang, Kidiyo Kpalma, Joseph Ronsin To cite this version: Weizhi Lu, Jinglin Zhang, Kidiyo Kpalma, Joseph Ronsin. Efficient
More informationBlind Image Deblurring Using Dark Channel Prior
Blind Image Deblurring Using Dark Channel Prior Jinshan Pan 1,2,3, Deqing Sun 2,4, Hanspeter Pfister 2, and Ming-Hsuan Yang 3 1 Dalian University of Technology 2 Harvard University 3 UC Merced 4 NVIDIA
More informationObject Tracking using HOG and SVM
Object Tracking using HOG and SVM Siji Joseph #1, Arun Pradeep #2 Electronics and Communication Engineering Axis College of Engineering and Technology, Ambanoly, Thrissur, India Abstract Object detection
More informationTransferring Visual Prior for Online Object Tracking
IEEE TRANSACTIONS ON IMAGE PROCESSING 1 Transferring Visual Prior for Online Object Tracking Qing Wang Feng Chen Jimei Yang Wenli Xu Ming-Hsuan Yang Abstract Visual prior from generic real-world images
More informationImplementation of Robust Visual Tracker by using Weighted Spatio-Temporal Context Learning Algorithm
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology ISSN 2320 088X IMPACT FACTOR: 5.258 IJCSMC,
More informationStructural Local Sparse Tracking Method Based on Multi-feature Fusion and Fractional Differential
Journal of Information Hiding and Multimedia Signal Processing c 28 ISSN 273-422 Ubiquitous International Volume 9, Number, January 28 Structural Local Sparse Tracking Method Based on Multi-feature Fusion
More informationEstimating Human Pose in Images. Navraj Singh December 11, 2009
Estimating Human Pose in Images Navraj Singh December 11, 2009 Introduction This project attempts to improve the performance of an existing method of estimating the pose of humans in still images. Tasks
More informationOBJECT tracking has been extensively studied in computer
IEEE TRANSAION ON IMAGE PROCESSING 1 Real-time Object Tracking via Online Discriminative Feature Selection Kaihua Zhang, Lei Zhang, and Ming-Hsuan Yang Abstract Most tracking-by-detection algorithms train
More informationAn efficient face recognition algorithm based on multi-kernel regularization learning
Acta Technica 61, No. 4A/2016, 75 84 c 2017 Institute of Thermomechanics CAS, v.v.i. An efficient face recognition algorithm based on multi-kernel regularization learning Bi Rongrong 1 Abstract. A novel
More informationIncremental Structured Dictionary Learning for Video Sensor-Based Object Tracking
Sensors 2014, 14, 3130-3155; doi:10.3390/s140203130 OPEN ACCESS sensors ISSN 1424-8220 www.mdpi.com/journal/sensors Article Incremental Structured Dictionary Learning for Video Sensor-Based Object Tracking
More informationEfficient visual tracking via low-complexity sparse representation
Lu et al. EURASIP Journal on Advances in Signal Processing (2015) 2015:21 DOI 10.1186/s13634-015-0200-7 RESEARCH Efficient visual tracking via low-complexity sparse representation Weizhi Lu 1*, Jinglin
More informationOnline Object Tracking: A Benchmark
Online Object Tracking: A Benchmark Yi Wu, Ming-Hsuan Yang University of California at Merced {ywu29,mhyang}@ucmerced.edu Jongwoo Lim Hanyang University jlim@hanyang.ac.kr Abstract Object tracking is one
More informationA Comparative Study of Object Trackers for Infrared Flying Bird Tracking
A Comparative Study of Object Trackers for Infrared Flying Bird Tracking Ying Huang, Hong Zheng, Haibin Ling 2, Erik Blasch 3, Hao Yang 4 School of Automation Science and Electrical Engineering, Beihang
More informationDeep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks
Deep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks Si Chen The George Washington University sichen@gwmail.gwu.edu Meera Hahn Emory University mhahn7@emory.edu Mentor: Afshin
More informationOnline Robust Non-negative Dictionary Learning for Visual Tracking
213 IEEE International Conference on Computer Vision Online Robust Non-negative Dictionary Learning for Visual Tracking Naiyan Wang Jingdong Wang Dit-Yan Yeung Hong Kong University of Science and Technology
More informationFace detection and recognition. Detection Recognition Sally
Face detection and recognition Detection Recognition Sally Face detection & recognition Viola & Jones detector Available in open CV Face recognition Eigenfaces for face recognition Metric learning identification
More informationRobust Face Recognition via Sparse Representation Authors: John Wright, Allen Y. Yang, Arvind Ganesh, S. Shankar Sastry, and Yi Ma
Robust Face Recognition via Sparse Representation Authors: John Wright, Allen Y. Yang, Arvind Ganesh, S. Shankar Sastry, and Yi Ma Presented by Hu Han Jan. 30 2014 For CSE 902 by Prof. Anil K. Jain: Selected
More informationUnsupervised Learning
Unsupervised Learning Pierre Gaillard ENS Paris September 28, 2018 1 Supervised vs unsupervised learning Two main categories of machine learning algorithms: - Supervised learning: predict output Y from
More informationSegmentation and Tracking of Partial Planar Templates
Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract
More informationSUPPLEMENTARY MATERIAL
SUPPLEMENTARY MATERIAL Zhiyuan Zha 1,3, Xin Liu 2, Ziheng Zhou 2, Xiaohua Huang 2, Jingang Shi 2, Zhenhong Shang 3, Lan Tang 1, Yechao Bai 1, Qiong Wang 1, Xinggan Zhang 1 1 School of Electronic Science
More informationEvaluation and comparison of interest points/regions
Introduction Evaluation and comparison of interest points/regions Quantitative evaluation of interest point/region detectors points / regions at the same relative location and area Repeatability rate :
More informationDeblurring Text Images via L 0 -Regularized Intensity and Gradient Prior
Deblurring Text Images via L -Regularized Intensity and Gradient Prior Jinshan Pan, Zhe Hu, Zhixun Su, Ming-Hsuan Yang School of Mathematical Sciences, Dalian University of Technology Electrical Engineering
More informationHow and what do we see? Segmentation and Grouping. Fundamental Problems. Polyhedral objects. Reducing the combinatorics of pose estimation
Segmentation and Grouping Fundamental Problems ' Focus of attention, or grouping ' What subsets of piels do we consider as possible objects? ' All connected subsets? ' Representation ' How do we model
More informationFace Detection and Recognition in an Image Sequence using Eigenedginess
Face Detection and Recognition in an Image Sequence using Eigenedginess B S Venkatesh, S Palanivel and B Yegnanarayana Department of Computer Science and Engineering. Indian Institute of Technology, Madras
More informationOBJECT tracking is an important problem in image
4454 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 21, NO. 1, OCTOBER 212 Object Tracking via Partial Least Squares Analysis Qing Wang, Student Member, IEEE, Feng Chen, Member, IEEE, Wenli Xu, and Ming-Hsuan
More informationPairwise Threshold for Gaussian Mixture Classification and its Application on Human Tracking Enhancement
Pairwise Threshold for Gaussian Mixture Classification and its Application on Human Tracking Enhancement Daegeon Kim Sung Chun Lee Institute for Robotics and Intelligent Systems University of Southern
More informationOBJECT tracking has been an important and active research
3296 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 21, NO. 7, JULY 2012 Transferring Visual Prior for Online Object Tracking Qing Wang, Feng Chen, Jimei Yang, Wenli Xu, and Ming-Hsuan Yang Abstract Visual
More informationSimultaneous Appearance Modeling and Segmentation for Matching People under Occlusion
Simultaneous Appearance Modeling and Segmentation for Matching People under Occlusion Zhe Lin, Larry S. Davis, David Doermann, and Daniel DeMenthon Institute for Advanced Computer Studies University of
More informationIn Defense of Sparse Tracking: Circulant Sparse Tracker
In Defense of Sparse Tracking: Circulant Sparse Tracker Tianzhu Zhang,2, Adel Bibi, and Bernard Ghanem King Abdullah University of Science and Technology (KAUST), Saudi Arabia 2 Institute of Automation,
More informationReal-Time Compressive Tracking
Real-Time Compressive Tracking Kaihua Zhang 1,LeiZhang 1, and Ming-Hsuan Yang 2 1 Depart. of Computing, Hong Kong Polytechnic University 2 Electrical Engineering and Computer Science, University of California
More informationLearning Data Terms for Non-blind Deblurring Supplemental Material
Learning Data Terms for Non-blind Deblurring Supplemental Material Jiangxin Dong 1, Jinshan Pan 2, Deqing Sun 3, Zhixun Su 1,4, and Ming-Hsuan Yang 5 1 Dalian University of Technology dongjxjx@gmail.com,
More informationVisual Motion Analysis and Tracking Part II
Visual Motion Analysis and Tracking Part II David J Fleet and Allan D Jepson CIAR NCAP Summer School July 12-16, 16, 2005 Outline Optical Flow and Tracking: Optical flow estimation (robust, iterative refinement,
More informationKey properties of local features
Key properties of local features Locality, robust against occlusions Must be highly distinctive, a good feature should allow for correct object identification with low probability of mismatch Easy to etract
More informationSparse Shape Registration for Occluded Facial Feature Localization
Shape Registration for Occluded Facial Feature Localization Fei Yang, Junzhou Huang and Dimitris Metaxas Abstract This paper proposes a sparsity driven shape registration method for occluded facial feature
More informationEfficient Visual Object Tracking with Online Nearest Neighbor Classifier
Efficient Visual Object Tracking with Online Nearest Neighbor Classifier Steve Gu and Ying Zheng and Carlo Tomasi Department of Computer Science, Duke University Abstract. A tracking-by-detection framework
More informationFast Visual Tracking via Dense Spatio-Temporal Context Learning
Fast Visual Tracking via Dense Spatio-Temporal Context Learning Kaihua Zhang 1, Lei Zhang 2, Qingshan Liu 1, David Zhang 2, and Ming-Hsuan Yang 3 1 S-mart Group, Nanjing University of Information Science
More informationLinearization to Nonlinear Learning for Visual Tracking
Linearization to Nonlinear Learning for Visual Tracking Bo Ma 1, Hongwei Hu 1, Jianbing Shen 1, Yuping Zhang 1, Fatih Porikli 2 1 Beijing Lab of Intelligent Information Technology, School of Computer Science,
More informationOnline Discriminative Tracking with Active Example Selection
1 Online Discriminative Tracking with Active Example Selection Min Yang, Yuwei Wu, Mingtao Pei, Bo Ma and Yunde Jia, Member, IEEE Abstract Most existing discriminative tracking algorithms use a sampling-and-labeling
More informationAS a fundamental component in computer vision system,
18 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 7, NO. 3, MARCH 018 Exploiting Spatial-Temporal Locality of Tracking via Structured Dictionary Learning Yao Sui, Guanghui Wang, Senior Member, IEEE, Li Zhang,
More informationReal-Time Visual Tracking: Promoting the Robustness of Correlation Filter Learning
Real-Time Visual Tracking: Promoting the Robustness of Correlation Filter Learning Yao Sui 1, Ziming Zhang 2, Guanghui Wang 1, Yafei Tang 3, Li Zhang 4 1 Dept. of EECS, University of Kansas, Lawrence,
More informationRobust Visual Tracking Via Consistent Low-Rank Sparse Learning
DOI 10.1007/s11263-014-0738-0 Robust Visual Tracking Via Consistent Low-Rank Sparse Learning Tianzhu Zhang Si Liu Narendra Ahuja Ming-Hsuan Yang Bernard Ghanem Received: 15 November 2013 / Accepted: 3
More informationCo-training Framework of Generative and Discriminative Trackers with Partial Occlusion Handling
Co-training Framework of Generative and Discriminative Trackers with Partial Occlusion Handling Thang Ba Dinh Gérard Medioni Institute of Robotics and Intelligent Systems University of Southern, Los Angeles,
More informationEfficient Detector Adaptation for Object Detection in a Video
2013 IEEE Conference on Computer Vision and Pattern Recognition Efficient Detector Adaptation for Object Detection in a Video Pramod Sharma and Ram Nevatia Institute for Robotics and Intelligent Systems,
More informationRobust Visual Tracking with Deep Convolutional Neural Network based Object Proposals on PETS
Robust Visual Tracking with Deep Convolutional Neural Network based Object Proposals on PETS Gao Zhu Fatih Porikli,2,3 Hongdong Li,3 Australian National University, NICTA 2 ARC Centre of Excellence for
More informationComputer Vision I - Filtering and Feature detection
Computer Vision I - Filtering and Feature detection Carsten Rother 30/10/2015 Computer Vision I: Basics of Image Processing Roadmap: Basics of Digital Image Processing Computer Vision I: Basics of Image
More informationRobust Incremental Subspace Learning for Object Tracking
Robust Incremental Subspace Learning for Object Tracking Gang Yu, Zhiwei Hu, and Hongtao Lu MOE-Microsoft Laboratory for Intelligent Computing and Intelligent Systems, Department of Computer Science and
More informationDynamic Shape Tracking via Region Matching
Dynamic Shape Tracking via Region Matching Ganesh Sundaramoorthi Asst. Professor of EE and AMCS KAUST (Joint work with Yanchao Yang) The Problem: Shape Tracking Given: exact object segmentation in frame1
More informationMultiple-Person Tracking by Detection
http://excel.fit.vutbr.cz Multiple-Person Tracking by Detection Jakub Vojvoda* Abstract Detection and tracking of multiple person is challenging problem mainly due to complexity of scene and large intra-class
More informationDiscrete Optimization of Ray Potentials for Semantic 3D Reconstruction
Discrete Optimization of Ray Potentials for Semantic 3D Reconstruction Marc Pollefeys Joined work with Nikolay Savinov, Christian Haene, Lubor Ladicky 2 Comparison to Volumetric Fusion Higher-order ray
More informationFast trajectory matching using small binary images
Title Fast trajectory matching using small binary images Author(s) Zhuo, W; Schnieders, D; Wong, KKY Citation The 3rd International Conference on Multimedia Technology (ICMT 2013), Guangzhou, China, 29
More informationBiometrics Technology: Image Processing & Pattern Recognition (by Dr. Dickson Tong)
Biometrics Technology: Image Processing & Pattern Recognition (by Dr. Dickson Tong) References: [1] http://homepages.inf.ed.ac.uk/rbf/hipr2/index.htm [2] http://www.cs.wisc.edu/~dyer/cs540/notes/vision.html
More informationA Generalized Search Method for Multiple Competing Hypotheses in Visual Tracking
A Generalized Search Method for Multiple Competing Hypotheses in Visual Tracking Muhammad H. Khan, Michel F. Valstar, and Tony P. Pridmore Computer Vision Laboratory, School of Computer Science, University
More informationRobust Visual Tracking via Convolutional Networks
SUBMITTED 1 Robust Visual Tracking via Convolutional Networks Kaihua Zhang, Qingshan Liu, Yi Wu, and Ming-Hsuan Yang arxiv:1501.04505v2 [cs.cv] 24 Aug 2015 Abstract Deep networks have been successfully
More informationMidterm Wed. Local features: detection and description. Today. Last time. Local features: main components. Goal: interest operator repeatability
Midterm Wed. Local features: detection and description Monday March 7 Prof. UT Austin Covers material up until 3/1 Solutions to practice eam handed out today Bring a 8.5 11 sheet of notes if you want Review
More informationObject Recognition Using Pictorial Structures. Daniel Huttenlocher Computer Science Department. In This Talk. Object recognition in computer vision
Object Recognition Using Pictorial Structures Daniel Huttenlocher Computer Science Department Joint work with Pedro Felzenszwalb, MIT AI Lab In This Talk Object recognition in computer vision Brief definition
More informationVISUAL TRACKING has been one of the fundamental
242 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 24, NO. 2, FEBRUARY 2014 Robust Visual Tracking via Multiple Kernel Boosting With Affinity Constraints Fan Yang, Student Member,
More informationInverse Sparse Group Lasso Model for Robust Object Tracking
1 Inverse Sparse Group Lasso Model for Robust Object Tracking Yun Zhou, Jianghong Han, Xiaohui Yuan, Senior Member, IEEE, Zhenchun Wei, and Richang Hong Abstract Sparse representation has been applied
More informationObject Tracking using Superpixel Confidence Map in Centroid Shifting Method
Indian Journal of Science and Technology, Vol 9(35), DOI: 10.17485/ijst/2016/v9i35/101783, September 2016 ISSN (Print) : 0974-6846 ISSN (Online) : 0974-5645 Object Tracking using Superpixel Confidence
More informationFMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu
FMA901F: Machine Learning Lecture 3: Linear Models for Regression Cristian Sminchisescu Machine Learning: Frequentist vs. Bayesian In the frequentist setting, we seek a fixed parameter (vector), with value(s)
More informationSelf Lane Assignment Using Smart Mobile Camera For Intelligent GPS Navigation and Traffic Interpretation
For Intelligent GPS Navigation and Traffic Interpretation Tianshi Gao Stanford University tianshig@stanford.edu 1. Introduction Imagine that you are driving on the highway at 70 mph and trying to figure
More informationMotion Tracking and Event Understanding in Video Sequences
Motion Tracking and Event Understanding in Video Sequences Isaac Cohen Elaine Kang, Jinman Kang Institute for Robotics and Intelligent Systems University of Southern California Los Angeles, CA Objectives!
More informationOnline Multi-Face Detection and Tracking using Detector Confidence and Structured SVMs
Online Multi-Face Detection and Tracking using Detector Confidence and Structured SVMs Francesco Comaschi 1, Sander Stuijk 1, Twan Basten 1,2 and Henk Corporaal 1 1 Eindhoven University of Technology,
More informationRobust and Fast Collaborative Tracking with Two Stage Sparse Optimization
Robust and Fast Collaborative Tracking with Two Stage Sparse Optimization Baiyang Liu 1,2, Lin Yang 2, Junzhou Huang 1, Peter Meer 3, Leiguang Gong 4, and Casimir Kulikowski 1 1 Department of Computer
More informationA novel template matching method for human detection
University of Wollongong Research Online Faculty of Informatics - Papers (Archive) Faculty of Engineering and Information Sciences 2009 A novel template matching method for human detection Duc Thanh Nguyen
More informationImage Processing. Image Features
Image Processing Image Features Preliminaries 2 What are Image Features? Anything. What they are used for? Some statements about image fragments (patches) recognition Search for similar patches matching
More informationGaussian Processes for Robotics. McGill COMP 765 Oct 24 th, 2017
Gaussian Processes for Robotics McGill COMP 765 Oct 24 th, 2017 A robot must learn Modeling the environment is sometimes an end goal: Space exploration Disaster recovery Environmental monitoring Other
More informationLearning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009
Learning and Inferring Depth from Monocular Images Jiyan Pan April 1, 2009 Traditional ways of inferring depth Binocular disparity Structure from motion Defocus Given a single monocular image, how to infer
More informationColored Point Cloud Registration Revisited Supplementary Material
Colored Point Cloud Registration Revisited Supplementary Material Jaesik Park Qian-Yi Zhou Vladlen Koltun Intel Labs A. RGB-D Image Alignment Section introduced a joint photometric and geometric objective
More informationBinary Principal Component Analysis
1 Binary Principal Component Analysis Feng Tang and Hai Tao University of California, Santa Cruz tang tao@soe.ucsc.edu Abstract Efficient and compact representation of images is a fundamental problem in
More informationVisual Tracking with Online Multiple Instance Learning
Visual Tracking with Online Multiple Instance Learning Boris Babenko University of California, San Diego bbabenko@cs.ucsd.edu Ming-Hsuan Yang University of California, Merced mhyang@ucmerced.edu Serge
More informationAction Recognition in Video by Sparse Representation on Covariance Manifolds of Silhouette Tunnels
Action Recognition in Video by Sparse Representation on Covariance Manifolds of Silhouette Tunnels Kai Guo, Prakash Ishwar, and Janusz Konrad Department of Electrical & Computer Engineering Motivation
More informationRobust Kernel Methods in Clustering and Dimensionality Reduction Problems
Robust Kernel Methods in Clustering and Dimensionality Reduction Problems Jian Guo, Debadyuti Roy, Jing Wang University of Michigan, Department of Statistics Introduction In this report we propose robust
More informationMixture Models and EM
Mixture Models and EM Goal: Introduction to probabilistic mixture models and the expectationmaximization (EM) algorithm. Motivation: simultaneous fitting of multiple model instances unsupervised clustering
More informationRobotics Programming Laboratory
Chair of Software Engineering Robotics Programming Laboratory Bertrand Meyer Jiwon Shin Lecture 8: Robot Perception Perception http://pascallin.ecs.soton.ac.uk/challenges/voc/databases.html#caltech car
More informationImage Denoising based on Adaptive BM3D and Singular Value
Image Denoising based on Adaptive BM3D and Singular Value Decomposition YouSai hang, ShuJin hu, YuanJiang Li Institute of Electronic and Information, Jiangsu University of Science and Technology, henjiang,
More informationObject tracking using a convolutional network and a structured output SVM
Computational Visual Media DOI 10.1007/s41095-017-0087-3 Vol. 3, No. 4, December 2017, 325 335 Research Article Object tracking using a convolutional network and a structured output SVM Junwei Li 1, Xiaolong
More informationEnsemble Tracking. Abstract. 1 Introduction. 2 Background
Ensemble Tracking Shai Avidan Mitsubishi Electric Research Labs 201 Broadway Cambridge, MA 02139 avidan@merl.com Abstract We consider tracking as a binary classification problem, where an ensemble of weak
More information