Least Soft-thresold Squares Tracking

Size: px
Start display at page:

Download "Least Soft-thresold Squares Tracking"

Transcription

1 Least Soft-thresold Squares Tracking Dong Wang Dalian University of Technology Huchuan Lu Dalian University of Technology Ming-Hsuan Yang University of California at Merced Abstract In this paper, we propose a generative tracking method based on a novel robust linear regression algorithm. In contrast to eisting methods, the proposed Least Soft-thresold Squares LSS) algorithm models the error term with the Gaussian-Laplacian distribution, which can be solved efficiently. Based on maimum joint likelihood of parameters, we derive a LSS distance to measure the difference between an observation sample and the dictionary. Compared with the distance derived from ordinary least squares methods, the proposed metric is more effective in dealing with outliers. In addition, we present an update scheme to capture the appearance change of the tracked target and ensure that the model is properly updated. Eperimental results on several challenging image sequences demonstrate that the proposed tracker achieves more favorable performance than the state-of-the-art methods.. Introduction Visual tracking plays a critical role in computer vision that finds many practical applications e.g., motion analysis, video surveillance, vehicle navigation and human-computer interaction). Although significant progress has been made in the past decades, developing a robust tracking algorithm is still a challenging problem due to numerous factors such as partial occlusion, illumination variation, pose change, comple motion, and background clutter. Tracking algorithms can be classified as either generative e.g., [7,, 20, 9, 2, 5]) or discriminative e.g., [2, 0, 4,, 27, 8, 30, 4]) methods. Generative methods focus on searching for the regions which are the most similar to the tracked targets, while discriminative methods cast tracking as a classification problem that distinguishes the tracked targets from the surrounding backgrounds. In this work, we propose a robust generative tracker which is able to handle partial occlusion and other challenging factors effectively. Among the generative methods, the trackers based on linear representation maintain holistic appearance information and therefore provide a compact notion of the thing being tracked, which may facilitate some advanced vision tasks [7, 20]. These methods often adopt a dictionary e.g., a set of basis vectors from a subspace or a series of templates) to describe the tracked target. A given candidate sample is linearly represented by the dictionary, and the representation coefficient and reconstruction error are computed, from which the corresponding likelihood belonging to the object class) is determined. Ross et al. [20] propose an incremental visual tracking IVT) method which represents the tracked target by a low dimensional PCA subspace a set of PCA basis vectors) and assumes that the error is Gaussian distributed with small variances i.e., small dense noise). Therefore, the representation coefficient can be obtained by a simple projection operator, which is equivalent to the ordinary least squares solution under the assumption that the dictionary atoms are orthogonal. The reconstruction error is computed by the objective function of the ordinary least squares methods. While the IVT method is effective to handle appearance change caused by illumination variation and pose variation, it is not robust to some challenging factors e.g., partial occlusion and background clutter) due to the following two reasons. First, ordinary least squares methods have been shown to be sensitive to outliers due to the formulation based on reconstruction error with Gaussian noise assumption. Second, the IVT method uses new observations to update the observation model without detecting outliers and processing them accordingly. Other recent tracking algorithms [6, 2] based on the Gaussian noise assumption or the ordinary least squares methods have similar problems as the IVT method. Motivated by the success of sparse representation-based face recognition [28], Mei et al. [9] develop a novel l tracker that uses a series of target templates and trivial templates to model the tracked target, where the target templates are used to describe the object class to be tracked and trivial templates are used to deal with outliers e.g., partial occlusion) with the sparsity constraints. For tracking, a candidate sample can be sparsely represented by both target and trivial templates, and its corresponding likelihood is determined by the reconstruction error with respect to target templates. We note that this formulation is a linear regres-

2 sion problem with sparsity constraints on the representation coefficients. Recently, several methods have been proposed to improve the l tracker in terms of both speed and accuracy by using accelerated proimal gradient algorithm [5], replacing raw piel templates with orthogonal basis vectors [24, 26, 25], modeling the similarity between different candidates [3], to name a few. Although these algorithms consider outliers by using additional trivial templates, this formulation can be generalized with better understanding. In this work, we show that the linear regression with the Gaussian-Laplacian noise assumption is more effective in dealing with outliers for object tracking. In addition, from the viewpoint of linear regression, it is not suitable to estimate the likelihood based on the reconstruction error with respect to target templates. We present a novel distance function to compute the distance between a candidate and the object class. In this paper, we present a generative tracking algorithm based on linear regression. The contributions of this work are as follows. First, we introduce a novel linear regression method, Least Soft-thresold Squares LSS), which assumes that the error vectors follow the i.i.d Gaussian- Laplacian distribution. Second, we present an efficient iteration method to solve the LSS problem and propose a LSS distance to measure the dissimilarity between the observation vector and dictionary. We note that the LSS method is related to the robust regression with the Huber loss function and is effective in detecting outliers. Compared with the least squares distance, the LSS distance is more effective in measuring the distance between the observation vector and dictionary when outliers occur. Third, we design a generative tracker by using the LSS method, where the dictionary consists of PCA basis vectors. The likelihood of each candidate is computed based on the LSS distance. Furthermore, we update the tracker by using an effective update scheme. Numerous eperiments on challenging image sequences with comparisons to state-of-the-art tracking methods demonstrate the effectiveness of the proposed model and algorithm. 2. Least Soft-thresold Squares 2.. Regression with Gaussian-Laplacian Noise The objective of linear regression is to fit a linear model i.e, estimate model parameters) to a series of noisy observations: y = A + e, ) where y R d is a d-dimensional observation vector, R k denotes the k-dimensional parameter or coefficient) vector to be estimated and A = [r ; r 2 ;...; r d ] R d k represents the input data matri r i is the i-th row of A). We denote e = y A = [e ; e 2 ;...; e d ] i.e, e i = y i r i, i =,..., d) as the error or residual term. The coefficient can be obtained by maimizing the posteriori probability p y), which is also equivalent to maimizing the joint likelihood probability p, y). Assume there is a uniform prior, the coefficient is estimated by = arg ma py ) = arg ma p e), which is the maimum likelihood estimation MLE). The errors e, e 2,..., e d are usually assumed to be independently and identically distributed i.i.d) according to some probability density function PDF) f θ e i ), where θ is the parameter set that characterizes the probability distribution. Thus, the likelihood of the estimator the joint probability of the error term e) is p e) = d i= f θ e i ). To maimize the likelihood function is equivalent to minimizing the objective function L θ e, e 2,..., e d ) = d i= ρ θ e i ), where ρ θ e i ) = log f θ e i ). When the error e = y A follows the Gaussian distribution e i N 0, σ 2 N ) ), the MLE solution is equivalent to the ordinary least squares OLS) solution = arg min 2 y A 2 2, 2) and the closed form solution is = A A ) A y. Although the OLS method is easy to solve, it is sensitive to outliers due to the Gaussian noise assumption. If the error e follows the Laplacian distribution e i L 0, σ L ) 2 ), the MLE solution is equivalent to least absolute deviations LAD) solution, = arg min y A. 3) Compared with the OLS method, the LAD method is robust to outliers. However, it is difficult to be solved by using either the simple-based methods [6] or the iteratively reweighted least squares methods [22]. In this paper, we model error vector e as an additive combination of two independent components: an i.i.d Gaussian noise vector n n i N 0, σ 2 N ) ) and an i.i.d Laplacian noise vector s s i L 0, σ L )), y = A + n + s, 4) where the Gaussian component models small dense noise and the Laplacian one aims to handle outliers 3. e i is a zero-mean Gaussian random variable with variance σn 2 and its ) PDF is f N e i ) = ep e2 i 2πσN 2σ N 2. 2 e i is a zero-mean Laplacian random variable with variance σl 2 and its PDF is f L e i ) = 2σL ep 2 ei σ L ). 3 The similar idea of decomposing the noise into two components e.g., decomposing the noise into sparse and non-sparse ones [3, 8] for achieving robust motion estimation) have appeared in computer vision community. While the goals are similar, the formulations and derivations are different. In this work, the proposed algorithm focuses on not only handling occlusion but also deriving a novel distance metric to compare the target candidate and the target template under the outlier condition, which will presented later.

3 The additive combination of i.i.d Gaussian and i.i.d Laplacian noise variables is also called Gaussian-Laplacian distribution [23]. Its joint PDF is p e) = d i= f N L e i ), where the PDF f N L e i ) is given by the convolution, f N L e i ) = f N n i ) f L s i ) = f L s i ) f N e i s i ) ds i = 2 2σ N ep e2 i 2σ 2 N ) erfc σn σl ei 2σN ) +erfc σn σ L + ei 2σN ) 5) where erfc ) = ep 2) erfc ) and erfc ) = 2 π, e t2 dt. Compared with the Gaussian and Laplacian distributions, the PDF of the Gaussian-Laplacian distribution is comple. Thus, it is difficult to obtain a simple objective function such as Eqs. 2 and 3) directly. Because of this, we treat the Laplacian noise term s as missing values with the same Laplacian prior, and therefore the joint likelihood p y, ) is converted to p y,, s). p y,, s) = p y, s) p, s) = p y A s) p s) = K ep { σ 2 N 2 y A s λ s )}, ) d ) d where K = 2σ 2 2σL 2πσN and λ = N. Thus, to maimize the joint likelihood p y,, s) is equivalent to minimizing the function 2 y A s λ s with respect to both and s Least Soft-thresold Squares Regression To maimize the joint likelihood of Eq. 6, we consider the objective function: σ L 6) L, s) = 2 y A s λ s, 7) and the optimal solution is [, ŝ] = arg min L, s). The,s above objective function is based on the standard least squares criterion and an l regularization term on s, and thus is conve but not differentiable everywhere. To the best of our knowledge, there is no closed-form solution for this optimization problem, so we present an iterative algorithm. Proposition : Given ŝ, the optimal can be computed by the ordinary least squares solution = A A ) A y ŝ). Proposition 2: Given, the optimal ŝ can be obtained by a soft-thresholding or shrinkage) operation S λ y A ), where S τ ) = ma τ, 0)sgn), where sgn ) is the sign function. Let P denote A A ) A, which can be precomputed before the iterative process. By Propositions and 2, the optimization can be solved efficiently. The iterative operation is terminated when a stopping criterion is met e.g., the difference of objective values between two iterations or the number of iterations). As our algorithm consists of two main components: ordinary least squares and soft-thresholding operation, we denote it as the Least Soft-threshold Squares Regression method. Table. Least Soft-threshold Squares Regression Input: An observation vector y, matri A, percomputed matri P = A A ) A, and a small constant λ. : Initialize s 0 = 0 and i = 0 2: Iterate 3: Obtain i+ via i+ = P y s i ) 4: Obtain s i+ via s i+ = S λ y A i+ ) 5: i i + 6: Until convergence or termination Output:, ŝ It is clear that L i+, s i+ ) L i+, s i ) L i, s i ) 4. Hence, the proposed iterative algorithm converges to a local minimal value. As the objective function is conve, it also obtains the global minimal solution. Remark : The proposed least soft-threshold squares regression is equivalent to robust regression with the Huber loss function: where = arg min d i= f e i), e i = y i r i, 8) f e) = { e 2 / 2, e λ λ e λ 2/ 2, e > λ is the Huber loss function and r i denotes the i-th row of A. A detailed eplanation can be found in the supplementary material. We note the approach to compute robust regression with the Huber loss function is less efficient than the proposed method as it is generally solved by using the iteratively reweighted least squares scheme which requires solving a weighted least squares problem i.e., pseudo-inverse) at each iteration. As mentioned before, the pseudo-inverse matri can be percalculated before the iterative process in the proposed LSS method. Remark 2: The non-zero components in s can be used to identify outliers. 4 L i+, s i ) = min L, s i ) L i, s i ), and L i+, s i+ ) = min s L i+, s) L i+, s i ) 9)

4 y 5 0 y 5 0 s z 5 0 st Iteration 5th Iteration 0th Iteration 20th Iteration z a) Observation samples. b) Iterative process in the LSS method. c) The Laplacian noise term. Figure. Robust line fitting by using the least soft-threshold squares LSS) regression. z Figure shows an eample of fitting a straight line when small Gaussian noise and outliers occur simultaneously. It requires to estimate an accurate parameter for a straight line function y = az + b, where z is the input variable, y is the output variable, and a and b are parameters to be determined. Figure a) shows ten observation sample pairs {z i, y i }, i =,..., 0, where z i = i in this case z 5 and z 9 are two outliers). Denote that y = [y ; y 2 ;...; y 0 ], z = [z ; z 2 ;...; z 0 ], A = [z, ] and = [a; b], we use the proposed least soft-threshold squares LSS) method to estimate the parameter. Figure b) illustrates the iterative process of the LSS method λ is set to ), from which it is clear that the proposed LSS is not sensitive to outliers. We note that the result after the first iteration is the same as the result obtained using the OLS method, which also shows that the proposed LSS method is more robust than the OLS method. In addition, as shown in Figure c), the non-zero Laplacian noise components correspond to the outliers Least Soft-thresold Squares Distance We represent the matri A = [a, a 2,..., a k ], where a i is the i-th column of A. The vector, A, can be viewed as a linear combination of the columns of A A = a + 2 a k a k ). The matri A is known as dictionary or basis matri, and the vector a i is called an atom or basis vector. For some vision applications such as tracking), it requires not only to estimate the coefficient accurately but also to define a distance between a noisy observation and the dictionary or the subspace. This generative perspective has commonly been eploited in subspace-based methods [7, 20]. The distance is usually defined to be inversely proportional to the maimum joint likelihood with respect to the coefficient, d y; A) log ma p y, ) 0) = log ma p y ) p ). Take the ordinary least squares method for eample i.e., uniform prior), we have log ma p y ) p ) log ma p y ) ) log ma ep 2 y A 2 2 min 2 y A 2 2. ) Recall that the OLS method assumes the observation vector with i.i.d Gaussian noise, the distance y and A can therefore be defined as, d OLS y; A) = min 2 y A 2 2 y A A ) A y. = ) Similarly, under the i.i.d Laplacian noise assumption, the distance between y and A can be defined as, d LAD y; A) = min y A, 3) which is difficult to be calculated as mentioned in Section 2.). In this work, we adopt the Gaussian-Laplacian distribution to model observation noise. Therefore, we define the distance between y and A under the i.i.d Gaussian- Laplacian noise assumption as, d LSS y; A) = min,s 2 y A s λ s. 4) As it is related to the Least Soft-thresold Squares LSS) regression, we denote it as the LSS distance. Figure 2 illustrates a toy eample of good and bad candidates for template matching with partial occlusion. For simplification, we merely use a single template shown in Figure 2a). Figure 2b) and c) show good and bad candidates shown in red and blue boes respectively). In Table 2, we report the OLS distance and the LSS distance between the template and different candidates. We denote the

5 Table 2. The OLS and LSS distances between the template and different candidates of Figure 2. d OLS d LSS d LSS λ = 0.05) λ = 0.) Good Candidate Bad Candidate template as t, the good candidate as y G and the bad candidate as y B respectively. In this eample, d OLS y B ; t) is smaller than d OLS y G ; t), which means the bad candidate is picked if the OLS distance is used. On the other hand, the good candidate is selected d LSS y G ; t) < d LSS y B ; t)) when the proposed LSS distance is used. Thus, we note that the proposed LSS distance is better than the OLS distance for handling outliers e.g., partial occlusion). where p t t ) is the motion model that describes the state transition between consecutive frames, and p y t t ) is the observation model that estimates the likelihood of an observed image patch belonging to the object class. The affine motion model is used in this work and the state transition is formulated by random walk, i.e., p t t ) = N t ; t, Σ), where Σ is a diagonal covariance matri that indicates the variances of affine parameters. Observation model: In this paper, we assume that the tracked target object is generated by a PCA subspace spanned by U and centered at µ) with i.i.d Gaussian- Laplacian noise y = µ + Uz + n + s, 7) where y denotes an observation vector, U represents a matri of column basis vectors, z indicates the coefficients of basis vectors, n is the Gaussian noise component and s is the Laplacian noise component. Based on the discussion in Section 2, under the i.i.d Gaussian-Laplacian noise assumption, the distance between the vector y and the subspace U, µ) is the least softthreshold squares distance, d y; U, µ) = min z,s 2 y Uz s λ s, 8) a) Template b) Good Candidate c) Bad Candidate Figure 2. A toy eample of good and bad candidates for template matching. 3. Least Soft-thresold Squares Tracking In this paper, visual tracking is treated as a dynamic Bayesian inference task with a hidden Markov model. Given a set of observed image vectors y :t = {y, y 2,..., y t } up to the t-th frame, the aim is to estimate the target state variable t by using the maimum a posteriori estimation, t = arg ma p i ) t y :t, 5) i t where i t indicates the i-th sample of the state t. Based on the Bayes theorem, the posterior distribution p t y :t ) can be estimated recursively by, p t y :t ) p y t t ) p t t ) p t y :t ) t, 6) where y = y µ. Thus, for each observation y i corresponding to a predicted state i, we firstly solve the following optimization problem, ] [ẑi, ŝi = arg min y z i,s i 2 i Uz i s i 2 + λ s i 2, 9) where i denotes the i-th sample of the state without loss of generality, we drop the frame inde t). As the PCA basis vectors U is orthogonal, the per-computed matri P can be simply set to U. After the optimal ẑi and ŝi are obtained, the least soft-threshold squares distance can be calculated by d y i ; U, µ ) = 2 y i 2 Uẑi ŝi + λ ŝi. Then 2 the observation likelihood can be measured by p y i i) = ep γd y i ; U, µ )), 20) where γ is a constant controlling the shape of the Gaussian kernel. Model Update: We note that the non-zero components in the Laplacian noise term can be used to identify outliers. Thus, we present a simple yet effective update scheme. After obtaining the best candidate state of each frame, [ we etract ] its corresponding observation vector y o = y [ o ; yo; 2...; y ] o d and infer the Laplacian noise term so = s o ; s 2 o...; s d o. Then we reconstruct the observation vector

6 by replacing the outliers with its corresponding parts of the mean vector µ, { yr i y i = o, s i o = 0 µ i, s i o 0, 2) where y r = [ y r; y 2 r;...; y d r ] denotes the reconstructed vector and µ = [ µ r; µ 2 r;...; µ d r] is the mean vector. The reconstructed sample is cumulated and then used to update the tracker i.e., PCA basis vectors U and the mean vector µ) by using an incremental principal component analysis PCA) method [20]. 4. Eperiments The proposed tracker is implemented in MATLAB and runs at 3.5 frames per second on a PC with Intel i CPU 3.4 GHz) with 32 GB memory. The regularization constant λ is set to 0. in all eperiments. For each sequence, the location of the tracked target is manually labeled in the first frame. We resize each image observation to piels and use 6 eigenvectors for PCA representation. As a trade-off between effectiveness and speed, 600 particles are adopted and our tracker is incrementally updated every 5 frames. The MATLAB source codes, datasets and supplementary materials are available on our websites ice.dlut.edu.cn/lu/publications.html, faculty.ucmerced.edu/mhyang/pubs.html). In this work, we use fifteen challenging image sequences from prior work [20, 9, 4, 5, 29] and the CAVIAR data set CAVIAR/CAVIARDATA/). The challenging factors of these sequences include partial occlusion, illumination variation, pose change, background clutter and motion blur. We evaluate the proposed tracker against ten state-of-theart algorithms, including the FragT [], IVT [20], MIL [4], VTD [5], TLD [4], APGL [5], MTT [3], LSAT [7], SCM [32], ASLSA [3] and OSPT [26] trackers. For fair evaluation, we use the source codes provided by the authors and run them with adjusted parameters. 4.. Quantitative Evaluation We evaluate the above-mentioned algorithms using two criteria: the center location error and the overlap rate. Table 3 reports the average center location errors in piels, where a smaller average error means a more accurate result. In addition, we use the segmentation criterion in the PAS- CAL VOC challenge [9] to evaluate the overlap rate. Given the tracking result bounding bo) of each frame R T and the corresponding ground truth bounding bo R G, the overlap. Table 4 reports the average overlap rates, where larger average scores mean more accurate results. score is defined as score = arear T R G ) arear T R G ) 4.2. Qualitative Evaluation Severe Occlusion: We test several sequences Occlusion, Occlusion2, Caviar, Caviar2, Caviar3, DavidOutdoor) with heavy or long-time partial occlusion, scale change and rotation. Figure 3 a-c) demonstrate that the proposed method performs well in terms of position, rotation and scale when the target undergoes severe occlusion. This can be attributes to two reasons: ) the proposed LSS distance takes outliers e.g., occlusion) into account eplicitly; and 2) the update scheme is able to avoid degrading the observation model by removing the outliers from new observed samples. In addition, the SCM and ASLAS methods also achieve good performance in most cases as both of them include part-based representations with overlapping patches. The IVT method is sensitive to partial occlusion Occlusion2, Caviar, Caviar3, DavidOutdoor) since the OLS distance is not effective to handle outliers. The MIL and TLD methods do not perform well when the target object is occluded by a similar object Caviar, Caviar2, Caviar3). This can be eplained by that the rectangle features they adopted generalized Haar-liker features or binary patterns) are less effective when similar objects occlude each other. Illumination Change: Figure 3 d) shows the tracking results in the sequences DavidIndoor, Car4, Singer) with significant illumination variation and both scale and pose change. We can see that the MIL and TLD methods are less effective in these cases e.g., DavidIndoor #0350 and Singer #0090). Due to the use of incremental PCA algorithm, the proposed tracker achieves good performance in dealing with the appearance change caused by light change. For the same reason, the IVT and ASLSA methods also perform well. Background Clutter: Figure 3 e) demonstrates the tracking results in the Car, Deer and Football sequences with background clutter. These videos also pose other challenging factors including illumination variation Car), fast motion Deer) and partial occlusion Football). As the proposed LSS distance encourages good matching results when outliers occur, our tracker performs better than other methods in these videos e.g., Deer #0052 and Football #035). Fast Motion: Figure 3 f) illustrates the tracking results on the Jumping, Owl and Face sequences. It is difficult to predict the locations of the tracked objects when they undergo abrupt motion. Furthermore, the appearance change caused by motion blur poses great challenges for capturing the tracked targets accurately and updating the observation models properly. We can see that the TLD and proposed methods perform better than other algorithms e.g., Owl #0248 and Face #0254). We note that the TLD method is equipped with a re-initialization mechanism which facilitates object tracking.

7 Table 3. Average center location error in piels). The best three results are shown in red, blue, and green fonts. Sequence FragT IVT MIL VTD TLD APGL MTT LSAT SCM ASLAS OSPT LSSTOurs) Occlusion Occlusion Caviar Caviar Caviar DavidOutdoor DavidIndoor Singer Car Car Deer Football Jumping Owl Face Average Table 4. Average overlap rate. The best three results are shown in red, blue, and green fonts. Sequence FragT IVT MIL VTD TLD APGL MTT LSAT SCM ASLAS OSPT LSSTOurs) Occlusion Occlusion Caviar Caviar Caviar DavidOutdoor DavidIndoor Singer Car Car Deer Football Jumping Owl Face Average Conclusion In this paper, we propose a Least Soft-thresold Squares LSS) regression method that assumes the noise is Gaussian-Laplacian distributed, and apply it to object tracking. We present an efficient iteration algorithm to solve the LSS problem, which achieves the global minimal solution. We derive a LSS distance to measure the difference between an observation sample and the dictionary. The LSS distance is effective in handling outliers and therefore provides an accurate match, which facilitates object tracking e.g., in dealing with partial occlusion). In addition, we develop a robust generative tracker based on the proposed LSS method and a simple update scheme. Both quantitative and qualitative evaluations on challenging image sequences show that the proposed tracker performs favorably against several state-of-the-art algorithms. In the future, we will etend the LSS method to solve other vision problems e.g., face recognition). Acknowledgements: D. Wang and H. Lu are supported by the National Natural Science Foundation of China # and # M.-H. Yang is supported by the US National Science Foundation CAREER Grant #49783 and IIS Grant # References [] A. Adam, E. Rivlin, and I. Shimshoni. Robust fragments-based tracking using the integral histogram. In CVPR, pages , [2] S. Avidan. Ensemble tracking. TPAMI, 292):26, [3] A. Ayvaci, M. Raptis, and S. Soatto. Occlusion detection and motion estimation with conve optimization. In NIPS, pages 00 08, 200. [4] B. Babenko, M.-H. Yang, and S. Belongie. Visual tracking with online multiple instance learning. In CVPR, pages , [5] C. Bao, Y. Wu, H. Ling, and H. Ji. Real time robust l tracker using accelerated proimal gradient approach. In CVPR, pages , 202. [6] I. Barrodale and F. D. K. Roberts. An improved algorithm for discrete l linear approimation. SIAM Journal on Numerical Analysis, 05): , 973. [7] M. J. Black. EigenTracking : Robust Matching and Tracking of Articulated Objects Using a View-Based Representation. IJCV, 26):63 84, 998. [8] Z. Chen, J. Wang, and Y. Wu. Decomposing and regularizing sparse/non-sparse components for motion field estimation. In CVPR, pages , 202. [9] M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn, and A. Zisserman. The pascal visual object classes voc) challenge. IJCV, 882): , 200. [0] H. Grabner and H. Bischof. On-line boosting and vision. In CVPR, pages , [] S. Hare, A. Saffari, and P. H. S. Torr. Struck: Structured output tracking with kernels. In ICCV, pages , 20. [2] W. Hu, X. Li, X. Zhang, X. Shi, S. J. Maybank, and Z. Zhang. Incremental tensor subspace learning and its applications to foreground segmentation and tracking. IJCV, 93): , 20. [3] X. Jia, H. Lu, and M.-H. Yang. Visual tracking via adaptive structural local sparse appearance model. In CVPR, pages , 202. [4] Z. Kalal, K. Mikolajczyk, and J. Matas. Tracking-learning-detection. TPAMI, 347): , 202.

8 a) Tracking results on sequence Occlusion and Occlusion2 with heavy occlusion and in-plane rotation. b) Tracking results on sequence Caviar and Caviar2 with partial occlusion and scale change. c) Tracking results on sequence Caviar3 and DavidOutdoor with severe occlusion, scale change and pose variation. d) Tracking results on sequence DavidIndoor, Car4 and Singer with illumination variation. e) Tracking results on sequence Car, Deer and Football with background clutter. f) Tracking results on sequence Jumping, Owl and Face with abrupt motion. IVT MIL TLD APGL SCM ASLAS Our Figure 3. Sample tracking results on fifteen challenging image sequences. This figure demonstrates the results of the IVT [20], MIL [4], TLD [4], APGL [5], SCM [32], ASLAS [3] and the proposed methods. More results can be found in the supplementary material. [5] J. Kwon and K. M. Lee. Visual tracking decomposition. In CVPR, pages , 200. [24] D. Wang and H. Lu. Object tracking via 2DPCA and ` -regularization. SPL, 9):7 74, 202. [6] X. Li, W. Hu, Z. Zhang, X. Zhang, M. Zhu, and J. Cheng. Visual tracking via incremental log-euclidean Riemannian subspace learning. In CVPR, pages 8, [25] D. Wang and H. Lu. On-line learning parts-based representation via incremental orthogonal projective non-negative matri factorization. SP, 93: , 203. [7] B. Liu, J. Huang, L. Yang, and C. A. Kulikowski. Robust tracking using local sparse appearance model and k-selection. In CVPR, pages , 20. [26] D. Wang, H. Lu, and M.-H. Yang. Online object tracking with sparse prototypes. TIP, 22):34 325, 203. [8] H. Lu, S. Lu, D. Wang, S. Wang, and H. Leung. Piel-wise spatial pyramidbased hybrid tracking. TCSVT, 229): , 202. [27] S. Wang, H. Lu, F. Yang, and M.-H. Yang. Superpiel tracking. In ICCV, pages , 20. [9] X. Mei and H. Ling. Robust visual tracking using ` minimization. In ICCV, pages , [28] J. Wright, A. Y. Yang, A. Ganesh, S. S. Sastry, and Y. Ma. Robust face recognition via sparse representation. TPAMI, 32):20 227, [20] D. Ross, J. Lim, R.-S. Lin, and M.-H. Yang. Incremental learning for robust visual tracking. IJCV, 77-3):25 4, [29] Y. Wu, H. Ling, J. Yu, F. Li, X. Mei, and E. Cheng. Blurred target tracking by blur-driven tracker. In ICCV, pages 00 07, 20. [2] J. Santner, C. Leistner, A. Saffari, T. Pock, and H. Bischof. PROST: Parallel robust online simple tracking. In CVPR, pages , 200. [30] K. Zhang, L. Zhang, and M.-H. Yang. Real-time compressive tracking. In ECCV, pages , 202. [22] E. J. Schlossmacher. An iterative technique for absolute deviations curve fitting. JOSA, 68344): , 973. [3] T. Zhang, B. Ghanem, S. Liu, and N. Ahuja. Robust visual tracking via multitask sparse learning. In CVPR, pages , 202. [23] I. W. Selesnick. The estimation of laplace random vectors in additive white gaussian noise. TSP, 568-): , [32] W. Zhong, H. Lu, and M.-H. Yang. Robust object tracking via sparsity-based collaborative model. In CVPR, pages , 202.

Structural Sparse Tracking

Structural Sparse Tracking Structural Sparse Tracking Tianzhu Zhang 1,2 Si Liu 3 Changsheng Xu 2,4 Shuicheng Yan 5 Bernard Ghanem 1,6 Narendra Ahuja 1,7 Ming-Hsuan Yang 8 1 Advanced Digital Sciences Center 2 Institute of Automation,

More information

Enhanced Laplacian Group Sparse Learning with Lifespan Outlier Rejection for Visual Tracking

Enhanced Laplacian Group Sparse Learning with Lifespan Outlier Rejection for Visual Tracking Enhanced Laplacian Group Sparse Learning with Lifespan Outlier Rejection for Visual Tracking Behzad Bozorgtabar 1 and Roland Goecke 1,2 1 Vision & Sensing, HCC Lab, ESTeM University of Canberra 2 IHCC,

More information

arxiv: v2 [cs.cv] 9 Sep 2016

arxiv: v2 [cs.cv] 9 Sep 2016 arxiv:1608.08171v2 [cs.cv] 9 Sep 2016 Tracking Completion Yao Sui 1, Guanghui Wang 1,4, Yafei Tang 2, Li Zhang 3 1 Dept. of EECS, University of Kansas, Lawrence, KS 66045, USA 2 China Unicom Research Institute,

More information

Online visual tracking by integrating spatio-temporal cues

Online visual tracking by integrating spatio-temporal cues Published in IET Computer Vision Received on 12th September 2013 Revised on 27th March 2014 Accepted on 14th May 2014 ISSN 1751-9632 Online visual tracking by integrating spatio-temporal cues Yang He,

More information

THE goal of object tracking is to estimate the states of

THE goal of object tracking is to estimate the states of 2356 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, NO. 5, MAY 2014 Robust Object Tracking via Sparse Collaborative Appearance Model Wei Zhong, Huchuan Lu, Senior Member, IEEE, and Ming-Hsuan Yang, Senior

More information

AS ONE of the fundamental problems in computer vision,

AS ONE of the fundamental problems in computer vision, 34 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL., NO., JANUARY 3 Online Object Tracking With Sparse Prototypes Dong Wang, Huchuan Lu, Member, IEEE, and Ming-Hsuan Yang, Senior Member, IEEE Abstract Online

More information

Superpixel Tracking. The detail of our motion model: The motion (or dynamical) model of our tracker is assumed to be Gaussian distributed:

Superpixel Tracking. The detail of our motion model: The motion (or dynamical) model of our tracker is assumed to be Gaussian distributed: Superpixel Tracking Shu Wang 1, Huchuan Lu 1, Fan Yang 1 abnd Ming-Hsuan Yang 2 1 School of Information and Communication Engineering, University of Technology, China 2 Electrical Engineering and Computer

More information

arxiv: v1 [cs.cv] 24 Feb 2014

arxiv: v1 [cs.cv] 24 Feb 2014 EXEMPLAR-BASED LINEAR DISCRIMINANT ANALYSIS FOR ROBUST OBJECT TRACKING Changxin Gao, Feifei Chen, Jin-Gang Yu, Rui Huang, Nong Sang arxiv:1402.5697v1 [cs.cv] 24 Feb 2014 Science and Technology on Multi-spectral

More information

Struck: Structured Output Tracking with Kernels. Presented by Mike Liu, Yuhang Ming, and Jing Wang May 24, 2017

Struck: Structured Output Tracking with Kernels. Presented by Mike Liu, Yuhang Ming, and Jing Wang May 24, 2017 Struck: Structured Output Tracking with Kernels Presented by Mike Liu, Yuhang Ming, and Jing Wang May 24, 2017 Motivations Problem: Tracking Input: Target Output: Locations over time http://vision.ucsd.edu/~bbabenko/images/fast.gif

More information

Visual Tracking Using Pertinent Patch Selection and Masking

Visual Tracking Using Pertinent Patch Selection and Masking Visual Tracking Using Pertinent Patch Selection and Masking Dae-Youn Lee, Jae-Young Sim, and Chang-Su Kim School of Electrical Engineering, Korea University, Seoul, Korea School of Electrical and Computer

More information

Tracking Completion. Institute of Automation, CAS, Beijing, China

Tracking Completion.  Institute of Automation, CAS, Beijing, China Tracking Completion Yao Sui 1(B), Guanghui Wang 1,4, Yafei Tang 2, and Li Zhang 3 1 Department of EECS, University of Kansas, Lawrence, KS 66045, USA suiyao@gmail.com, ghwang@ku.edu 2 China Unicom Research

More information

THE recent years have witnessed significant advances

THE recent years have witnessed significant advances IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, NO. 4, APRIL 2014 1639 Robust Superpixel Tracking Fan Yang, Student Member, IEEE, Huchuan Lu, Senior Member, IEEE, and Ming-Hsuan Yang, Senior Member, IEEE

More information

Tracking Using Online Feature Selection and a Local Generative Model

Tracking Using Online Feature Selection and a Local Generative Model Tracking Using Online Feature Selection and a Local Generative Model Thomas Woodley Bjorn Stenger Roberto Cipolla Dept. of Engineering University of Cambridge {tew32 cipolla}@eng.cam.ac.uk Computer Vision

More information

Keywords:- Object tracking, multiple instance learning, supervised learning, online boosting, ODFS tracker, classifier. IJSER

Keywords:- Object tracking, multiple instance learning, supervised learning, online boosting, ODFS tracker, classifier. IJSER International Journal of Scientific & Engineering Research, Volume 5, Issue 2, February-2014 37 Object Tracking via a Robust Feature Selection approach Prof. Mali M.D. manishamali2008@gmail.com Guide NBNSCOE

More information

Tracking. Hao Guan( 管皓 ) School of Computer Science Fudan University

Tracking. Hao Guan( 管皓 ) School of Computer Science Fudan University Tracking Hao Guan( 管皓 ) School of Computer Science Fudan University 2014-09-29 Multimedia Video Audio Use your eyes Video Tracking Use your ears Audio Tracking Tracking Video Tracking Definition Given

More information

Real-time Object Tracking via Online Discriminative Feature Selection

Real-time Object Tracking via Online Discriminative Feature Selection IEEE TRANSACTION ON IMAGE PROCESSING 1 Real-time Object Tracking via Online Discriminative Feature Selection Kaihua Zhang, Lei Zhang, and Ming-Hsuan Yang Abstract Most tracking-by-detection algorithms

More information

Multi-Cue Visual Tracking Using Robust Feature-Level Fusion Based on Joint Sparse Representation

Multi-Cue Visual Tracking Using Robust Feature-Level Fusion Based on Joint Sparse Representation Multi-Cue Visual Tracking Using Robust Feature-Level Fusion Based on Joint Sparse Representation Xiangyuan Lan Andy J Ma Pong C Yuen Department of Computer Science, Hong Kong Baptist University {xylan,

More information

2 Cascade detection and tracking

2 Cascade detection and tracking 3rd International Conference on Multimedia Technology(ICMT 213) A fast on-line boosting tracking algorithm based on cascade filter of multi-features HU Song, SUN Shui-Fa* 1, MA Xian-Bing, QIN Yin-Shi,

More information

Online Graph-Based Tracking

Online Graph-Based Tracking Online Graph-Based Tracking Hyeonseob Nam, Seunghoon Hong, and Bohyung Han Department of Computer Science and Engineering, POSTECH, Korea {namhs09,maga33,bhhan}@postech.ac.kr Abstract. Tracking by sequential

More information

Online Multiple Support Instance Tracking

Online Multiple Support Instance Tracking Online Multiple Support Instance Tracking Qiu-Hong Zhou Huchuan Lu Ming-Hsuan Yang School of Electronic and Information Engineering Electrical Engineering and Computer Science Dalian University of Technology

More information

Outline. Advanced Digital Image Processing and Others. Importance of Segmentation (Cont.) Importance of Segmentation

Outline. Advanced Digital Image Processing and Others. Importance of Segmentation (Cont.) Importance of Segmentation Advanced Digital Image Processing and Others Xiaojun Qi -- REU Site Program in CVIP (7 Summer) Outline Segmentation Strategies and Data Structures Algorithms Overview K-Means Algorithm Hidden Markov Model

More information

Tracking Algorithms. Lecture16: Visual Tracking I. Probabilistic Tracking. Joint Probability and Graphical Model. Deterministic methods

Tracking Algorithms. Lecture16: Visual Tracking I. Probabilistic Tracking. Joint Probability and Graphical Model. Deterministic methods Tracking Algorithms CSED441:Introduction to Computer Vision (2017F) Lecture16: Visual Tracking I Bohyung Han CSE, POSTECH bhhan@postech.ac.kr Deterministic methods Given input video and current state,

More information

Real Time Face Tracking By Using Naïve Bayes Classifier

Real Time Face Tracking By Using Naïve Bayes Classifier Real Time Face Tracking By Using Naïve Bayes Classifier Krishna Priya Arni 1, Satish Ammuloju 2, Jagan Mohan Vithalachari Gollapalli 3, Dinesh Bhogapurapu 4) Shaik Azeez 5 Student, Department of ECE, Lendi

More information

Efficient Visual Tracking via Low-Complexity Sparse Representation

Efficient Visual Tracking via Low-Complexity Sparse Representation Efficient Visual Tracking via Low-Complexity Sparse Representation Weizhi Lu, Jinglin Zhang, Kidiyo Kpalma, Joseph Ronsin To cite this version: Weizhi Lu, Jinglin Zhang, Kidiyo Kpalma, Joseph Ronsin. Efficient

More information

Blind Image Deblurring Using Dark Channel Prior

Blind Image Deblurring Using Dark Channel Prior Blind Image Deblurring Using Dark Channel Prior Jinshan Pan 1,2,3, Deqing Sun 2,4, Hanspeter Pfister 2, and Ming-Hsuan Yang 3 1 Dalian University of Technology 2 Harvard University 3 UC Merced 4 NVIDIA

More information

Object Tracking using HOG and SVM

Object Tracking using HOG and SVM Object Tracking using HOG and SVM Siji Joseph #1, Arun Pradeep #2 Electronics and Communication Engineering Axis College of Engineering and Technology, Ambanoly, Thrissur, India Abstract Object detection

More information

Transferring Visual Prior for Online Object Tracking

Transferring Visual Prior for Online Object Tracking IEEE TRANSACTIONS ON IMAGE PROCESSING 1 Transferring Visual Prior for Online Object Tracking Qing Wang Feng Chen Jimei Yang Wenli Xu Ming-Hsuan Yang Abstract Visual prior from generic real-world images

More information

Implementation of Robust Visual Tracker by using Weighted Spatio-Temporal Context Learning Algorithm

Implementation of Robust Visual Tracker by using Weighted Spatio-Temporal Context Learning Algorithm Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology ISSN 2320 088X IMPACT FACTOR: 5.258 IJCSMC,

More information

Structural Local Sparse Tracking Method Based on Multi-feature Fusion and Fractional Differential

Structural Local Sparse Tracking Method Based on Multi-feature Fusion and Fractional Differential Journal of Information Hiding and Multimedia Signal Processing c 28 ISSN 273-422 Ubiquitous International Volume 9, Number, January 28 Structural Local Sparse Tracking Method Based on Multi-feature Fusion

More information

Estimating Human Pose in Images. Navraj Singh December 11, 2009

Estimating Human Pose in Images. Navraj Singh December 11, 2009 Estimating Human Pose in Images Navraj Singh December 11, 2009 Introduction This project attempts to improve the performance of an existing method of estimating the pose of humans in still images. Tasks

More information

OBJECT tracking has been extensively studied in computer

OBJECT tracking has been extensively studied in computer IEEE TRANSAION ON IMAGE PROCESSING 1 Real-time Object Tracking via Online Discriminative Feature Selection Kaihua Zhang, Lei Zhang, and Ming-Hsuan Yang Abstract Most tracking-by-detection algorithms train

More information

An efficient face recognition algorithm based on multi-kernel regularization learning

An efficient face recognition algorithm based on multi-kernel regularization learning Acta Technica 61, No. 4A/2016, 75 84 c 2017 Institute of Thermomechanics CAS, v.v.i. An efficient face recognition algorithm based on multi-kernel regularization learning Bi Rongrong 1 Abstract. A novel

More information

Incremental Structured Dictionary Learning for Video Sensor-Based Object Tracking

Incremental Structured Dictionary Learning for Video Sensor-Based Object Tracking Sensors 2014, 14, 3130-3155; doi:10.3390/s140203130 OPEN ACCESS sensors ISSN 1424-8220 www.mdpi.com/journal/sensors Article Incremental Structured Dictionary Learning for Video Sensor-Based Object Tracking

More information

Efficient visual tracking via low-complexity sparse representation

Efficient visual tracking via low-complexity sparse representation Lu et al. EURASIP Journal on Advances in Signal Processing (2015) 2015:21 DOI 10.1186/s13634-015-0200-7 RESEARCH Efficient visual tracking via low-complexity sparse representation Weizhi Lu 1*, Jinglin

More information

Online Object Tracking: A Benchmark

Online Object Tracking: A Benchmark Online Object Tracking: A Benchmark Yi Wu, Ming-Hsuan Yang University of California at Merced {ywu29,mhyang}@ucmerced.edu Jongwoo Lim Hanyang University jlim@hanyang.ac.kr Abstract Object tracking is one

More information

A Comparative Study of Object Trackers for Infrared Flying Bird Tracking

A Comparative Study of Object Trackers for Infrared Flying Bird Tracking A Comparative Study of Object Trackers for Infrared Flying Bird Tracking Ying Huang, Hong Zheng, Haibin Ling 2, Erik Blasch 3, Hao Yang 4 School of Automation Science and Electrical Engineering, Beihang

More information

Deep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks

Deep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks Deep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks Si Chen The George Washington University sichen@gwmail.gwu.edu Meera Hahn Emory University mhahn7@emory.edu Mentor: Afshin

More information

Online Robust Non-negative Dictionary Learning for Visual Tracking

Online Robust Non-negative Dictionary Learning for Visual Tracking 213 IEEE International Conference on Computer Vision Online Robust Non-negative Dictionary Learning for Visual Tracking Naiyan Wang Jingdong Wang Dit-Yan Yeung Hong Kong University of Science and Technology

More information

Face detection and recognition. Detection Recognition Sally

Face detection and recognition. Detection Recognition Sally Face detection and recognition Detection Recognition Sally Face detection & recognition Viola & Jones detector Available in open CV Face recognition Eigenfaces for face recognition Metric learning identification

More information

Robust Face Recognition via Sparse Representation Authors: John Wright, Allen Y. Yang, Arvind Ganesh, S. Shankar Sastry, and Yi Ma

Robust Face Recognition via Sparse Representation Authors: John Wright, Allen Y. Yang, Arvind Ganesh, S. Shankar Sastry, and Yi Ma Robust Face Recognition via Sparse Representation Authors: John Wright, Allen Y. Yang, Arvind Ganesh, S. Shankar Sastry, and Yi Ma Presented by Hu Han Jan. 30 2014 For CSE 902 by Prof. Anil K. Jain: Selected

More information

Unsupervised Learning

Unsupervised Learning Unsupervised Learning Pierre Gaillard ENS Paris September 28, 2018 1 Supervised vs unsupervised learning Two main categories of machine learning algorithms: - Supervised learning: predict output Y from

More information

Segmentation and Tracking of Partial Planar Templates

Segmentation and Tracking of Partial Planar Templates Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract

More information

SUPPLEMENTARY MATERIAL

SUPPLEMENTARY MATERIAL SUPPLEMENTARY MATERIAL Zhiyuan Zha 1,3, Xin Liu 2, Ziheng Zhou 2, Xiaohua Huang 2, Jingang Shi 2, Zhenhong Shang 3, Lan Tang 1, Yechao Bai 1, Qiong Wang 1, Xinggan Zhang 1 1 School of Electronic Science

More information

Evaluation and comparison of interest points/regions

Evaluation and comparison of interest points/regions Introduction Evaluation and comparison of interest points/regions Quantitative evaluation of interest point/region detectors points / regions at the same relative location and area Repeatability rate :

More information

Deblurring Text Images via L 0 -Regularized Intensity and Gradient Prior

Deblurring Text Images via L 0 -Regularized Intensity and Gradient Prior Deblurring Text Images via L -Regularized Intensity and Gradient Prior Jinshan Pan, Zhe Hu, Zhixun Su, Ming-Hsuan Yang School of Mathematical Sciences, Dalian University of Technology Electrical Engineering

More information

How and what do we see? Segmentation and Grouping. Fundamental Problems. Polyhedral objects. Reducing the combinatorics of pose estimation

How and what do we see? Segmentation and Grouping. Fundamental Problems. Polyhedral objects. Reducing the combinatorics of pose estimation Segmentation and Grouping Fundamental Problems ' Focus of attention, or grouping ' What subsets of piels do we consider as possible objects? ' All connected subsets? ' Representation ' How do we model

More information

Face Detection and Recognition in an Image Sequence using Eigenedginess

Face Detection and Recognition in an Image Sequence using Eigenedginess Face Detection and Recognition in an Image Sequence using Eigenedginess B S Venkatesh, S Palanivel and B Yegnanarayana Department of Computer Science and Engineering. Indian Institute of Technology, Madras

More information

OBJECT tracking is an important problem in image

OBJECT tracking is an important problem in image 4454 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 21, NO. 1, OCTOBER 212 Object Tracking via Partial Least Squares Analysis Qing Wang, Student Member, IEEE, Feng Chen, Member, IEEE, Wenli Xu, and Ming-Hsuan

More information

Pairwise Threshold for Gaussian Mixture Classification and its Application on Human Tracking Enhancement

Pairwise Threshold for Gaussian Mixture Classification and its Application on Human Tracking Enhancement Pairwise Threshold for Gaussian Mixture Classification and its Application on Human Tracking Enhancement Daegeon Kim Sung Chun Lee Institute for Robotics and Intelligent Systems University of Southern

More information

OBJECT tracking has been an important and active research

OBJECT tracking has been an important and active research 3296 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 21, NO. 7, JULY 2012 Transferring Visual Prior for Online Object Tracking Qing Wang, Feng Chen, Jimei Yang, Wenli Xu, and Ming-Hsuan Yang Abstract Visual

More information

Simultaneous Appearance Modeling and Segmentation for Matching People under Occlusion

Simultaneous Appearance Modeling and Segmentation for Matching People under Occlusion Simultaneous Appearance Modeling and Segmentation for Matching People under Occlusion Zhe Lin, Larry S. Davis, David Doermann, and Daniel DeMenthon Institute for Advanced Computer Studies University of

More information

In Defense of Sparse Tracking: Circulant Sparse Tracker

In Defense of Sparse Tracking: Circulant Sparse Tracker In Defense of Sparse Tracking: Circulant Sparse Tracker Tianzhu Zhang,2, Adel Bibi, and Bernard Ghanem King Abdullah University of Science and Technology (KAUST), Saudi Arabia 2 Institute of Automation,

More information

Real-Time Compressive Tracking

Real-Time Compressive Tracking Real-Time Compressive Tracking Kaihua Zhang 1,LeiZhang 1, and Ming-Hsuan Yang 2 1 Depart. of Computing, Hong Kong Polytechnic University 2 Electrical Engineering and Computer Science, University of California

More information

Learning Data Terms for Non-blind Deblurring Supplemental Material

Learning Data Terms for Non-blind Deblurring Supplemental Material Learning Data Terms for Non-blind Deblurring Supplemental Material Jiangxin Dong 1, Jinshan Pan 2, Deqing Sun 3, Zhixun Su 1,4, and Ming-Hsuan Yang 5 1 Dalian University of Technology dongjxjx@gmail.com,

More information

Visual Motion Analysis and Tracking Part II

Visual Motion Analysis and Tracking Part II Visual Motion Analysis and Tracking Part II David J Fleet and Allan D Jepson CIAR NCAP Summer School July 12-16, 16, 2005 Outline Optical Flow and Tracking: Optical flow estimation (robust, iterative refinement,

More information

Key properties of local features

Key properties of local features Key properties of local features Locality, robust against occlusions Must be highly distinctive, a good feature should allow for correct object identification with low probability of mismatch Easy to etract

More information

Sparse Shape Registration for Occluded Facial Feature Localization

Sparse Shape Registration for Occluded Facial Feature Localization Shape Registration for Occluded Facial Feature Localization Fei Yang, Junzhou Huang and Dimitris Metaxas Abstract This paper proposes a sparsity driven shape registration method for occluded facial feature

More information

Efficient Visual Object Tracking with Online Nearest Neighbor Classifier

Efficient Visual Object Tracking with Online Nearest Neighbor Classifier Efficient Visual Object Tracking with Online Nearest Neighbor Classifier Steve Gu and Ying Zheng and Carlo Tomasi Department of Computer Science, Duke University Abstract. A tracking-by-detection framework

More information

Fast Visual Tracking via Dense Spatio-Temporal Context Learning

Fast Visual Tracking via Dense Spatio-Temporal Context Learning Fast Visual Tracking via Dense Spatio-Temporal Context Learning Kaihua Zhang 1, Lei Zhang 2, Qingshan Liu 1, David Zhang 2, and Ming-Hsuan Yang 3 1 S-mart Group, Nanjing University of Information Science

More information

Linearization to Nonlinear Learning for Visual Tracking

Linearization to Nonlinear Learning for Visual Tracking Linearization to Nonlinear Learning for Visual Tracking Bo Ma 1, Hongwei Hu 1, Jianbing Shen 1, Yuping Zhang 1, Fatih Porikli 2 1 Beijing Lab of Intelligent Information Technology, School of Computer Science,

More information

Online Discriminative Tracking with Active Example Selection

Online Discriminative Tracking with Active Example Selection 1 Online Discriminative Tracking with Active Example Selection Min Yang, Yuwei Wu, Mingtao Pei, Bo Ma and Yunde Jia, Member, IEEE Abstract Most existing discriminative tracking algorithms use a sampling-and-labeling

More information

AS a fundamental component in computer vision system,

AS a fundamental component in computer vision system, 18 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 7, NO. 3, MARCH 018 Exploiting Spatial-Temporal Locality of Tracking via Structured Dictionary Learning Yao Sui, Guanghui Wang, Senior Member, IEEE, Li Zhang,

More information

Real-Time Visual Tracking: Promoting the Robustness of Correlation Filter Learning

Real-Time Visual Tracking: Promoting the Robustness of Correlation Filter Learning Real-Time Visual Tracking: Promoting the Robustness of Correlation Filter Learning Yao Sui 1, Ziming Zhang 2, Guanghui Wang 1, Yafei Tang 3, Li Zhang 4 1 Dept. of EECS, University of Kansas, Lawrence,

More information

Robust Visual Tracking Via Consistent Low-Rank Sparse Learning

Robust Visual Tracking Via Consistent Low-Rank Sparse Learning DOI 10.1007/s11263-014-0738-0 Robust Visual Tracking Via Consistent Low-Rank Sparse Learning Tianzhu Zhang Si Liu Narendra Ahuja Ming-Hsuan Yang Bernard Ghanem Received: 15 November 2013 / Accepted: 3

More information

Co-training Framework of Generative and Discriminative Trackers with Partial Occlusion Handling

Co-training Framework of Generative and Discriminative Trackers with Partial Occlusion Handling Co-training Framework of Generative and Discriminative Trackers with Partial Occlusion Handling Thang Ba Dinh Gérard Medioni Institute of Robotics and Intelligent Systems University of Southern, Los Angeles,

More information

Efficient Detector Adaptation for Object Detection in a Video

Efficient Detector Adaptation for Object Detection in a Video 2013 IEEE Conference on Computer Vision and Pattern Recognition Efficient Detector Adaptation for Object Detection in a Video Pramod Sharma and Ram Nevatia Institute for Robotics and Intelligent Systems,

More information

Robust Visual Tracking with Deep Convolutional Neural Network based Object Proposals on PETS

Robust Visual Tracking with Deep Convolutional Neural Network based Object Proposals on PETS Robust Visual Tracking with Deep Convolutional Neural Network based Object Proposals on PETS Gao Zhu Fatih Porikli,2,3 Hongdong Li,3 Australian National University, NICTA 2 ARC Centre of Excellence for

More information

Computer Vision I - Filtering and Feature detection

Computer Vision I - Filtering and Feature detection Computer Vision I - Filtering and Feature detection Carsten Rother 30/10/2015 Computer Vision I: Basics of Image Processing Roadmap: Basics of Digital Image Processing Computer Vision I: Basics of Image

More information

Robust Incremental Subspace Learning for Object Tracking

Robust Incremental Subspace Learning for Object Tracking Robust Incremental Subspace Learning for Object Tracking Gang Yu, Zhiwei Hu, and Hongtao Lu MOE-Microsoft Laboratory for Intelligent Computing and Intelligent Systems, Department of Computer Science and

More information

Dynamic Shape Tracking via Region Matching

Dynamic Shape Tracking via Region Matching Dynamic Shape Tracking via Region Matching Ganesh Sundaramoorthi Asst. Professor of EE and AMCS KAUST (Joint work with Yanchao Yang) The Problem: Shape Tracking Given: exact object segmentation in frame1

More information

Multiple-Person Tracking by Detection

Multiple-Person Tracking by Detection http://excel.fit.vutbr.cz Multiple-Person Tracking by Detection Jakub Vojvoda* Abstract Detection and tracking of multiple person is challenging problem mainly due to complexity of scene and large intra-class

More information

Discrete Optimization of Ray Potentials for Semantic 3D Reconstruction

Discrete Optimization of Ray Potentials for Semantic 3D Reconstruction Discrete Optimization of Ray Potentials for Semantic 3D Reconstruction Marc Pollefeys Joined work with Nikolay Savinov, Christian Haene, Lubor Ladicky 2 Comparison to Volumetric Fusion Higher-order ray

More information

Fast trajectory matching using small binary images

Fast trajectory matching using small binary images Title Fast trajectory matching using small binary images Author(s) Zhuo, W; Schnieders, D; Wong, KKY Citation The 3rd International Conference on Multimedia Technology (ICMT 2013), Guangzhou, China, 29

More information

Biometrics Technology: Image Processing & Pattern Recognition (by Dr. Dickson Tong)

Biometrics Technology: Image Processing & Pattern Recognition (by Dr. Dickson Tong) Biometrics Technology: Image Processing & Pattern Recognition (by Dr. Dickson Tong) References: [1] http://homepages.inf.ed.ac.uk/rbf/hipr2/index.htm [2] http://www.cs.wisc.edu/~dyer/cs540/notes/vision.html

More information

A Generalized Search Method for Multiple Competing Hypotheses in Visual Tracking

A Generalized Search Method for Multiple Competing Hypotheses in Visual Tracking A Generalized Search Method for Multiple Competing Hypotheses in Visual Tracking Muhammad H. Khan, Michel F. Valstar, and Tony P. Pridmore Computer Vision Laboratory, School of Computer Science, University

More information

Robust Visual Tracking via Convolutional Networks

Robust Visual Tracking via Convolutional Networks SUBMITTED 1 Robust Visual Tracking via Convolutional Networks Kaihua Zhang, Qingshan Liu, Yi Wu, and Ming-Hsuan Yang arxiv:1501.04505v2 [cs.cv] 24 Aug 2015 Abstract Deep networks have been successfully

More information

Midterm Wed. Local features: detection and description. Today. Last time. Local features: main components. Goal: interest operator repeatability

Midterm Wed. Local features: detection and description. Today. Last time. Local features: main components. Goal: interest operator repeatability Midterm Wed. Local features: detection and description Monday March 7 Prof. UT Austin Covers material up until 3/1 Solutions to practice eam handed out today Bring a 8.5 11 sheet of notes if you want Review

More information

Object Recognition Using Pictorial Structures. Daniel Huttenlocher Computer Science Department. In This Talk. Object recognition in computer vision

Object Recognition Using Pictorial Structures. Daniel Huttenlocher Computer Science Department. In This Talk. Object recognition in computer vision Object Recognition Using Pictorial Structures Daniel Huttenlocher Computer Science Department Joint work with Pedro Felzenszwalb, MIT AI Lab In This Talk Object recognition in computer vision Brief definition

More information

VISUAL TRACKING has been one of the fundamental

VISUAL TRACKING has been one of the fundamental 242 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 24, NO. 2, FEBRUARY 2014 Robust Visual Tracking via Multiple Kernel Boosting With Affinity Constraints Fan Yang, Student Member,

More information

Inverse Sparse Group Lasso Model for Robust Object Tracking

Inverse Sparse Group Lasso Model for Robust Object Tracking 1 Inverse Sparse Group Lasso Model for Robust Object Tracking Yun Zhou, Jianghong Han, Xiaohui Yuan, Senior Member, IEEE, Zhenchun Wei, and Richang Hong Abstract Sparse representation has been applied

More information

Object Tracking using Superpixel Confidence Map in Centroid Shifting Method

Object Tracking using Superpixel Confidence Map in Centroid Shifting Method Indian Journal of Science and Technology, Vol 9(35), DOI: 10.17485/ijst/2016/v9i35/101783, September 2016 ISSN (Print) : 0974-6846 ISSN (Online) : 0974-5645 Object Tracking using Superpixel Confidence

More information

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu FMA901F: Machine Learning Lecture 3: Linear Models for Regression Cristian Sminchisescu Machine Learning: Frequentist vs. Bayesian In the frequentist setting, we seek a fixed parameter (vector), with value(s)

More information

Self Lane Assignment Using Smart Mobile Camera For Intelligent GPS Navigation and Traffic Interpretation

Self Lane Assignment Using Smart Mobile Camera For Intelligent GPS Navigation and Traffic Interpretation For Intelligent GPS Navigation and Traffic Interpretation Tianshi Gao Stanford University tianshig@stanford.edu 1. Introduction Imagine that you are driving on the highway at 70 mph and trying to figure

More information

Motion Tracking and Event Understanding in Video Sequences

Motion Tracking and Event Understanding in Video Sequences Motion Tracking and Event Understanding in Video Sequences Isaac Cohen Elaine Kang, Jinman Kang Institute for Robotics and Intelligent Systems University of Southern California Los Angeles, CA Objectives!

More information

Online Multi-Face Detection and Tracking using Detector Confidence and Structured SVMs

Online Multi-Face Detection and Tracking using Detector Confidence and Structured SVMs Online Multi-Face Detection and Tracking using Detector Confidence and Structured SVMs Francesco Comaschi 1, Sander Stuijk 1, Twan Basten 1,2 and Henk Corporaal 1 1 Eindhoven University of Technology,

More information

Robust and Fast Collaborative Tracking with Two Stage Sparse Optimization

Robust and Fast Collaborative Tracking with Two Stage Sparse Optimization Robust and Fast Collaborative Tracking with Two Stage Sparse Optimization Baiyang Liu 1,2, Lin Yang 2, Junzhou Huang 1, Peter Meer 3, Leiguang Gong 4, and Casimir Kulikowski 1 1 Department of Computer

More information

A novel template matching method for human detection

A novel template matching method for human detection University of Wollongong Research Online Faculty of Informatics - Papers (Archive) Faculty of Engineering and Information Sciences 2009 A novel template matching method for human detection Duc Thanh Nguyen

More information

Image Processing. Image Features

Image Processing. Image Features Image Processing Image Features Preliminaries 2 What are Image Features? Anything. What they are used for? Some statements about image fragments (patches) recognition Search for similar patches matching

More information

Gaussian Processes for Robotics. McGill COMP 765 Oct 24 th, 2017

Gaussian Processes for Robotics. McGill COMP 765 Oct 24 th, 2017 Gaussian Processes for Robotics McGill COMP 765 Oct 24 th, 2017 A robot must learn Modeling the environment is sometimes an end goal: Space exploration Disaster recovery Environmental monitoring Other

More information

Learning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009

Learning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009 Learning and Inferring Depth from Monocular Images Jiyan Pan April 1, 2009 Traditional ways of inferring depth Binocular disparity Structure from motion Defocus Given a single monocular image, how to infer

More information

Colored Point Cloud Registration Revisited Supplementary Material

Colored Point Cloud Registration Revisited Supplementary Material Colored Point Cloud Registration Revisited Supplementary Material Jaesik Park Qian-Yi Zhou Vladlen Koltun Intel Labs A. RGB-D Image Alignment Section introduced a joint photometric and geometric objective

More information

Binary Principal Component Analysis

Binary Principal Component Analysis 1 Binary Principal Component Analysis Feng Tang and Hai Tao University of California, Santa Cruz tang tao@soe.ucsc.edu Abstract Efficient and compact representation of images is a fundamental problem in

More information

Visual Tracking with Online Multiple Instance Learning

Visual Tracking with Online Multiple Instance Learning Visual Tracking with Online Multiple Instance Learning Boris Babenko University of California, San Diego bbabenko@cs.ucsd.edu Ming-Hsuan Yang University of California, Merced mhyang@ucmerced.edu Serge

More information

Action Recognition in Video by Sparse Representation on Covariance Manifolds of Silhouette Tunnels

Action Recognition in Video by Sparse Representation on Covariance Manifolds of Silhouette Tunnels Action Recognition in Video by Sparse Representation on Covariance Manifolds of Silhouette Tunnels Kai Guo, Prakash Ishwar, and Janusz Konrad Department of Electrical & Computer Engineering Motivation

More information

Robust Kernel Methods in Clustering and Dimensionality Reduction Problems

Robust Kernel Methods in Clustering and Dimensionality Reduction Problems Robust Kernel Methods in Clustering and Dimensionality Reduction Problems Jian Guo, Debadyuti Roy, Jing Wang University of Michigan, Department of Statistics Introduction In this report we propose robust

More information

Mixture Models and EM

Mixture Models and EM Mixture Models and EM Goal: Introduction to probabilistic mixture models and the expectationmaximization (EM) algorithm. Motivation: simultaneous fitting of multiple model instances unsupervised clustering

More information

Robotics Programming Laboratory

Robotics Programming Laboratory Chair of Software Engineering Robotics Programming Laboratory Bertrand Meyer Jiwon Shin Lecture 8: Robot Perception Perception http://pascallin.ecs.soton.ac.uk/challenges/voc/databases.html#caltech car

More information

Image Denoising based on Adaptive BM3D and Singular Value

Image Denoising based on Adaptive BM3D and Singular Value Image Denoising based on Adaptive BM3D and Singular Value Decomposition YouSai hang, ShuJin hu, YuanJiang Li Institute of Electronic and Information, Jiangsu University of Science and Technology, henjiang,

More information

Object tracking using a convolutional network and a structured output SVM

Object tracking using a convolutional network and a structured output SVM Computational Visual Media DOI 10.1007/s41095-017-0087-3 Vol. 3, No. 4, December 2017, 325 335 Research Article Object tracking using a convolutional network and a structured output SVM Junwei Li 1, Xiaolong

More information

Ensemble Tracking. Abstract. 1 Introduction. 2 Background

Ensemble Tracking. Abstract. 1 Introduction. 2 Background Ensemble Tracking Shai Avidan Mitsubishi Electric Research Labs 201 Broadway Cambridge, MA 02139 avidan@merl.com Abstract We consider tracking as a binary classification problem, where an ensemble of weak

More information