Pedestrian counting in video sequences using optical flow clustering

Similar documents
Pedestrian counting with grid-based binary sensors based on Monte Carlo method

An Adaptive Threshold LBP Algorithm for Face Recognition

Learning the Three Factors of a Non-overlapping Multi-camera Network Topology

Human Detection. A state-of-the-art survey. Mohammad Dorgham. University of Hamburg

A Novel Multi-Planar Homography Constraint Algorithm for Robust Multi-People Location with Severe Occlusion

An Approach for Real Time Moving Object Extraction based on Edge Region Determination

Occlusion Robust Multi-Camera Face Tracking

Multi-Camera Occlusion and Sudden-Appearance-Change Detection Using Hidden Markovian Chains

FAST HUMAN DETECTION USING TEMPLATE MATCHING FOR GRADIENT IMAGES AND ASC DESCRIPTORS BASED ON SUBTRACTION STEREO

A Street Scene Surveillance System for Moving Object Detection, Tracking and Classification

CS 565 Computer Vision. Nazar Khan PUCIT Lectures 15 and 16: Optic Flow

[2008] IEEE. Reprinted, with permission, from [Yan Chen, Qiang Wu, Xiangjian He, Wenjing Jia,Tom Hintz, A Modified Mahalanobis Distance for Human

PEOPLE IN SEATS COUNTING VIA SEAT DETECTION FOR MEETING SURVEILLANCE

Motion Estimation. There are three main types (or applications) of motion estimation:

Video Inter-frame Forgery Identification Based on Optical Flow Consistency

Aircraft Tracking Based on KLT Feature Tracker and Image Modeling

Color Content Based Image Classification

An ICA based Approach for Complex Color Scene Text Binarization

Optical Flow-Based Person Tracking by Multiple Cameras

Occlusion Detection of Real Objects using Contour Based Stereo Matching

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation

Improvement of SURF Feature Image Registration Algorithm Based on Cluster Analysis

Feature Tracking and Optical Flow

MULTI ORIENTATION PERFORMANCE OF FEATURE EXTRACTION FOR HUMAN HEAD RECOGNITION

Real-Time Document Image Retrieval for a 10 Million Pages Database with a Memory Efficient and Stability Improved LLAH

RESTORATION OF DEGRADED DOCUMENTS USING IMAGE BINARIZATION TECHNIQUE

International Journal of Advance Engineering and Research Development

arxiv: v1 [cs.cv] 19 May 2017

Content Based Image Retrieval Using Color Quantizes, EDBTC and LBP Features

Full-Motion Recovery from Multiple Video Cameras Applied to Face Tracking and Recognition

Real-Time Human Detection using Relational Depth Similarity Features

Production of Video Images by Computer Controlled Cameras and Its Application to TV Conference System

Moving Object Segmentation Method Based on Motion Information Classification by X-means and Spatial Region Segmentation

arxiv: v1 [cs.cv] 2 May 2016

Automatic Shadow Removal by Illuminance in HSV Color Space

Object detection using non-redundant local Binary Patterns

An Approach for Reduction of Rain Streaks from a Single Image

A Texture-based Method for Detecting Moving Objects

Distributed Clustering Method for Large-Scaled Wavelength Routed Networks

Visual Tracking (1) Tracking of Feature Points and Planar Rigid Objects

High-speed Three-dimensional Mapping by Direct Estimation of a Small Motion Using Range Images

Fast trajectory matching using small binary images

Face and Nose Detection in Digital Images using Local Binary Patterns

Ensemble of Bayesian Filters for Loop Closure Detection

Occluded Facial Expression Tracking

LOCAL FEATURE EXTRACTION METHODS FOR FACIAL EXPRESSION RECOGNITION

Detecting motion by means of 2D and 3D information

Face Tracking. Synonyms. Definition. Main Body Text. Amit K. Roy-Chowdhury and Yilei Xu. Facial Motion Estimation

Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera

Measurement of Pedestrian Groups Using Subtraction Stereo

Human Motion Detection and Tracking for Video Surveillance

An Acceleration Scheme to The Local Directional Pattern

Peripheral drift illusion

Illumination-Robust Face Recognition based on Gabor Feature Face Intrinsic Identity PCA Model

Scene Text Detection Using Machine Learning Classifiers

Measurement of 3D Foot Shape Deformation in Motion

Motion Analysis. Motion analysis. Now we will talk about. Differential Motion Analysis. Motion analysis. Difference Pictures

Subpixel Corner Detection Using Spatial Moment 1)

CS-465 Computer Vision

Texture Feature Extraction Using Improved Completed Robust Local Binary Pattern for Batik Image Retrieval

Visual Tracking (1) Feature Point Tracking and Block Matching

Robust color segmentation algorithms in illumination variation conditions

Semi-Supervised PCA-based Face Recognition Using Self-Training

Texture Sensitive Image Inpainting after Object Morphing

Face Recognition Using Long Haar-like Filters

Visual Tracking. Image Processing Laboratory Dipartimento di Matematica e Informatica Università degli studi di Catania.

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

Spatio temporal Segmentation using Laserscanner and Video Sequences

DYNAMIC BACKGROUND SUBTRACTION BASED ON SPATIAL EXTENDED CENTER-SYMMETRIC LOCAL BINARY PATTERN. Gengjian Xue, Jun Sun, Li Song

Practical Camera Auto-Calibration Based on Object Appearance and Motion for Traffic Scene Visual Surveillance

Short Survey on Static Hand Gesture Recognition

Using temporal seeding to constrain the disparity search range in stereo matching

Visual Attention Control by Sensor Space Segmentation for a Small Quadruped Robot based on Information Criterion

An Object Detection System using Image Reconstruction with PCA

Toward Part-based Document Image Decoding

Feature Tracking and Optical Flow

Survey on Extraction of Texture based Features using Local Binary Pattern

Chapter 9 Object Tracking an Overview

A Laplacian Based Novel Approach to Efficient Text Localization in Grayscale Images

This paper is part of the following report: UNCLASSIFIED

Flexible Calibration of a Portable Structured Light System through Surface Plane

MULTI-VIEW FACE DETECTION AND POSE ESTIMATION EMPLOYING EDGE-BASED FEATURE VECTORS

A Modified Mean Shift Algorithm for Visual Object Tracking

Motion Detection Algorithm

METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS

Color Local Texture Features Based Face Recognition

The Population Density of Early Warning System Based On Video Image

Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation

C. Premsai 1, Prof. A. Kavya 2 School of Computer Science, School of Computer Science Engineering, Engineering VIT Chennai, VIT Chennai

Outdoor Scene Reconstruction from Multiple Image Sequences Captured by a Hand-held Video Camera

CS 4495 Computer Vision Motion and Optic Flow

A Feature Point Matching Based Approach for Video Objects Segmentation

Expanding gait identification methods from straight to curved trajectories

Detecting and Identifying Moving Objects in Real-Time

Peg-Free Hand Geometry Verification System

arxiv: v1 [cs.cv] 28 Sep 2018

3D object recognition used by team robotto

Efficient Acquisition of Human Existence Priors from Motion Trajectories

SURVEY OF LOCAL AND GLOBAL OPTICAL FLOW WITH COARSE TO FINE METHOD

Background/Foreground Detection 1

Transcription:

Pedestrian counting in video sequences using optical flow clustering SHIZUKA FUJISAWA, GO HASEGAWA, YOSHIAKI TANIGUCHI, HIROTAKA NAKANO Graduate School of Information Science and Technology Osaka University 1-5 Yamadaoka, Suita, Osaka 565-0871 JAPAN {s-fujisw, hasegawa, y-tanigu, nakano}@ist.osaka-u.ac.jp Abstract: Existing systems for counting pedestrians in video sequences have the problem that counting accuracy is worse when many pedestrians are present and occlusion occurs frequently. In this paper, we introduce a method of clustering optical flows in video frames to improve the counting accuracy in cases where occlusion occurs. The proposed method counts the number of pedestrians by using pre-learned statistics, based on the strong correlation between the number of optical flow clusters detected by our method and the actual number of pedestrians. We evaluate the accuracy of the proposed method using several video sequences, focusing in particular on the effect of parameters for optical flow clustering. We find that the proposed method improves the counting accuracy by up to 25% as compared with a non-clustering method. We also report that using a clustering threshold of angles less than 1 is effective for enhancing counting accuracy. Key Words: pedestrian counting, video sequence, optical flow, clustering 1 Introduction Measuring traffic flow on streets and counting pedestrians at various locations are essential functions for access/traffic control, marketing surveys, and so on. Traditionally, pedestrian counting is conducted manually, which entails high cost and may introduce human error. Therefore, automatic pedestrian counting has received much attention. Numerous approaches to automatically counting moving objects have been reported in the literature. The most widely deployed methods utilize laser sensors [1] and infrared sensors [2, 3], both of which are inexpensive and robust against changes in the illumination condition due to fluctuations of lighting and weather. However, these methods sometimes fail to count pedestrians correctly, especially when the heights of co-existing pedestrians are similar or when the heights do not fall within the presumed range, because such methods depend on the difference in propagation delays of reflected laser pulses or infrared light. On the other hand, pedestrian counting methods using video-processing technology have some advantages including that the camera position is flexible, the video sequences can be retrieved remotely, and information such as the height and gender of pedestrians can be obtained, in addition to the number of pedestrians. However, the methods using video processing are affected by changes in illumination conditions and occlusion caused by many pedestrians begin present. The counting of pedestrians by video processing is being actively researched, and a contest for counting pedestrians by using video-processing technology is held at the annual IEEE workshop Performance Evaluation of Tracking and Surveillance (PETS) [4]. Texture-based methods [5, 6, 7] are robust to the occlusion of moving objects whose texture pattern is uniform. However, people generally have nonuniformity in their appearance. Model-based methods [8, 9] use a shape model to extract moving objects, and have the high counting accuracy in the presence of occlusions. However, the performance of these methods becomes considerably worse when the moving object is a non-rigid object such as a person. In [10], a camera is set directly above pedestrians to avoid occlusion. However, the camera position is restricted to locations such as the entrance of a building. The method in [11] uses multiple cameras which give the three-dimensional positions of pedestrians to identify their locations more accurately. However, computing three-dimensional positions from point correspondences over multiple images leads to high computational complexity. In this paper, we propose a method for counting the number of pedestrians in video sequences retrieved by a single stationary camera. We assume a situation such as a crowded street where the density of pedestrians in video frames is very high and occlusion occurs frequently. The proposed method is based on clustering optical flows of moving objects. The clustering utilizes the lengths, angles, and source locations of optical flows. The proposed method estimates the ISBN: 978-1-61804-074-9 51

2.1 Method Overview and Basic Idea The overview of the proposed method is shown in Fig. 1. We place a stationary camera at an arbitrary position and record video of the target area (e.g., a street or building entrance) where pedestrians are to be counted. Retrieved video sequences are sent to a server, which computes the estimate of the number of pedestrians. On the server, we first use the background difference method to identify the moving parts in each video frame. We next detect optical flows of moving parts by using a well-known method, followed by clustering them based on the lengths, angles, and source locations of the detected optical flows. Finally, the proposed method estimates the number of pedestrians, based on the pre-learned correlation between the number of optical flow clusters and the actual number of pedestrians. In what follows, we explain each step of the proposed method. Figure 1: Overview of the method number of pedestrians based on the strong correlation between the number of optical flow clusters detected by our method and the actual number of pedestrians. Since the degree of correlation depends on camera position and various other factors, we utilize statistical data to estimate the number of pedestrians. The effectiveness of the proposed method is evaluated through implementation experiments with several video sequences, including the ones used at the PETS workshop. We compare the performance of the proposed method with a non-clustering method, focusing in particular on the parameter settings for optical flow clustering. We also investigate the relationships between the optimal threshold of optical flow angles in clustering and the upper-limit of angle differentiation resolution of the method. The remainder of this paper is organized as follows. In Section II, we present the estimation method for the number of pedestrians in video sequences. The proposed method is evaluated by using several video sequences in Section III. Finally, we summarize our conclusions and discuss future work in Section IV. 2 Proposed Method 2.2 Extracting Optical Flows For the i-th frame (2 i F, where F is the total number of frames in the video sequence), denoted by f i, we identify the moving parts in the frame by using the background difference method. To mitigate the adverse effects of changes in the illumination conditions, we update the background frame dynamically by using the exponential weighted moving average calculation as follows: B i,x,y = (1 α) B i 1,x,y + α B i,x,y (1) Here, B i,x,y (B {R,G,B}) denotes the pixel value at coordinate (x, y) of frame f i, B i,x,y (B {R,G,B}) denotes the RGB pixel value at coordinate (x, y) of the background frame, and α is a weight parameter. We identify the moving parts in the frame by evaluating the difference between the pixel value of the current frame and that of the background frame. d i,x,y = Bi,x,y B i 1,x,y (2) { 1 (di,x,y > p O i,x,y = d ) 0 (d i,x,y p d ) (3) where O i,x,y indicates whether (x, y) is in moving parts of the frame. p d is the threshold value. We then extract optical flows at multiple points in each moving part of the frame. Optical flow is calculated by the Lucas-Kanade method [12]. Here, we denote the coordinate of the j-th point as (x j, y j ). The Lucas-Kanade method requires the following assumptions. The brightness of each moving part remains unchanged in frames f i 1 and f i. The distance traveled by a moving part between frames f i 1 and f i is within a certain range. Neighboring points, such as (x j, y j ) and (x j+1, y j+1 ) of frame f i 1, which belong to the same moving part, retain the same relative positions in frame f i. On the basis of the above assumptions, we detect a point of frame f i 1 which corresponds to point (x j, y j ) of frame f i as a result of movement. By denoting the coordinate of such a point as (x k, y k ), we determine the optical flow whose starting point is (x k, y k ) and whose ending point is (x j, y j ) of frame f i. To alleviate the effects of noises and changes in illumination conditions, we delete optical flows whose lengths do not fall into a certain range. That is, we utilize only the optical flows that satisfy the following condition, where l i,j is the length of an optical flow ISBN: 978-1-61804-074-9 52

whose starting point is (x j, y j ) of frame f i. l min < l i,j < l max (4) l min and l max are determined empirically by prelearning. 2.3 Clustering Optical Flows We assume that optical flows detected from a single pedestrian have similar lengths and angles, and that their starting points exist within a certain range. Therefore, optical flows are clustered based on the degree of similarity of their lengths, angles, and starting points. Here, a set of detected optical flows in frame f i is denoted by D i and the l-th optical flow in D i is denoted by o i,l ( D i ). We perform clustering all optical flows in D i. At first, we create a new cluster, denoted by G i,1, and add the optical flow o i,1 ( D i ) to G i,1. For the optical flow o i,l (l 2), we compare o i,l with all optical flows in each cluster G i,m, in terms of length, angle, and starting point. In comparing two optical flows o i,l and o i,s, we determine that they belong to the same cluster when the following ( conditions ) are satisfied. ( ) y e i,l yi,l s y e i,s yi,s s arctan x e i,l arctan xs i,l x e i,s xs i,s a th (5) l i,l l i,s l th (6) x s i,l xs i,s x th (7) yi,l s ys i,s y th (8) where (x s i,l, ys i,l ) and (xe i,l, ye i,l ) are the starting point and ending point of optical flow o i,l, respectively. a th, l th, x th, and y th are the respective thresholds of angle, length, x-coordinate, and y-coordinate for optical flows. When o i,l and all optical flows in G i,m satisfy equations (5) (8), we add o i,l to G i,m. When o i,l does not belong to any existing cluster, we create a new cluster for o i,l. After clustering, we assess the number of optical flows in each cluster. When the number is smaller than the threshold N d, we delete the cluster since we assume the cluster is caused by noise such as changes in the illumination conditions. 2.4 Estimating the Number of Pedestrians Although we assume that the number of clusters detected in a frame and the actual number of the pedestrians in the frame have a strong correlation, the degree of correlation depends on the camera position, the size of pedestrians in the frames, and various other factors. We then utilize the pre-learned statistical data to estimate the number of pedestrians. In detail, we utilize a part of frames in the video sequence for prelearning and estimate the number of pedestrians for all frames in the video sequence. Here, the number of clusters in frame f i is denoted by C i. For pre-learning, we calculate the average number of clusters in frames for pre-leaning as follows: C = ΣF l q=1 C q F l (9) where F l (< F ) is the number of frames for prelearning. The actual number of pedestrians in frame f i is denoted by P i. We obtain the average number of pedestrians in the frames for pre-learning as follows. P = ΣF l q=1 P q F l (10) Note that to calculate P, we must count the actual number of pedestrians in the part of video sequence, which means pre-learning. The average number of optical flows per person in pre-learned frames, denoted by C p, is calculated as follows. C p = C (11) P The estimated number of pedestrians in frame f i, denoted by Ci e, is calculated as C e i = C i C p (12) 3 Experiments and Evaluation 3.1 Experimental Environment We recorded a 30 [s] video (720 480 [pixels]) at the Toyonaka Campus of Osaka University, Japan at 29 [frames per second (fps)]. The proposed method was implemented on a PC with a 2.93 [GHz] Intel Core i7 CPU and 3.00 [GB] of memory running Microsoft Windows 7 Enterprise. Note that our proposed method can be executed in real time, meaning that the processing time for estimating the number of pedestrians in each frame is less than 1/29 [s]. Unless otherwise noted, we set the parameters of the proposed method as follows: α = 0.05, l min = 0.2, l max = 25.0, a th = 0.2, l th = 3.0, x th = 38.0, y th = 46.0, N d = 13, µ = 1, and F l = F. An example of the execution of the proposed method is shown in Fig. 2. The colored lines represent detected optical flows, and the optical flows that have the same color belong to the same cluster. In a performance evaluation, we compared the counting accuracy of the proposed method with a nonclustering method that estimates the number of pedestrians based on a pre-learned correlation between the ISBN: 978-1-61804-074-9 53

Estimated Number of Pedestrians 20 15 10 5 true number of pedestrians proposed method non-clustering method 0 0 100 200 300 400 500 600 700 800 900 Frame Figure 2: Snapshot of the proposed method Figure 4: Temporal changes in counting accuracy 400 1.1 Number of Clusters 350 300 250 200 150 100 Mean Absolute Error 1.05 1 0.95 50 0 0 5 10 15 20 Actual Number of Pedestrians 0.9 0.01 0.1 1 Angle Threshold [degree] Figure 3: Correlation between the number of clusters and the actual number of pedestrians total number of optical flows, not clusters, and the actual number of pedestrians in each frame. The other parts of the non-clustering method are the same as in the proposed method. 3.2 Results Figure 3 shows the relationship between the number of clusters detected by the proposed method and the actual number of pedestrians. Each plot shows the result for each frame. We observe a linear correlation between the number of clusters and the number of pedestrians. This is because we utilize the simple pre-learning algorithm described in Subsection 2.4. Table 1 summarizes the evaluation results of overall counting accuracy in terms of mean absolute error (MAE), mean relative error (MRE), and variance for the entire video sequence. The proposed method outperforms the non-clustering method, thus indicating means the effectiveness of the optical flow clustering proposed in this paper. Figure 4 shows the tempo- Figure 5: Effect of angle threshold on counting accuracy ral change in the counting accuracy of the proposed method and of the non-clustering method. The counting accuracy of both methods becomes worse, especially when the actual number of pedestrians increases or decreases largely. One possible reason is that the method described in Subsection 2.2 fails to detect optical flows correctly, especially when pedestrians enter or leave the target area. In such situations, a part of a pedestrian is often detected at the edge of the target area, which degrades the estimation accuracy based on pre-learning method. Other reasons for the lower counting accuracy include pedestrians that stop moving in the frame and the effect of shadow. 3.3 Effect of Angle Threshold on Optical Flow Clustering In this subsection, we focus on the angle threshold for optical flow clustering, since the angle threshold is important for distinguishing between multiple pedestrians who are walking together. Figure 5 shows the relationship between the angle threshold and the ISBN: 978-1-61804-074-9 54

Recent Researches in Applications of Electrical and Computer Engineering 0.7 degree=1.0 degree=0.5 degree=0.1 Mean Reletive Error 0.6 0.5 0.4 0.3 0.2 0.1 0 4 6 8 10 12 14 16 18 Actual Number of Pedestrians Figure 8: Snapshot from video sequence from the PETS workshop Figure 6: Effect of number of pedestrians on counting accuracy Table 1: Overall counting accuracy for the video sequence at Osaka University Method proposed non-clustering MRE 0.103 0.110 variance 1.211 1.284 number of pedestrians, a larger threshold is desirable, whereas a smaller threshold is preferable for detecting a larger number of pedestrians. These characteristics are caused by the effect of the pedestrians legs. When the number of pedestrians is small, less occlusion occurs and many optical flows are detected from the legs. Since the movement of the pedestrians legs is more complicated than that of other body parts, a large angle threshold is better for avoiding redundant and fluctuating segmentation of optical flow clusters. On the other hand, when there are many pedestrians, a substantial amount of occlusion can be observed and the legs of the pedestrian are often hidden. In such a case, we should utilize a small angle threshold to distinguish between multiple pedestrians who move together causing occlusion. Figure 7: Snapshot from video sequence taken in Tokyo MRE of the number of pedestrians estimated by proposed method. Note that the finest angle resolution of the proposed method is approximately 0.2, which is governed by the frame resolution, the average size, and velocity of pedestrians. From Fig. 5, we found that counting accuracy is good when the angle threshold is around 0.1, which is roughly the same as the method s upper limit of angle resolution. Note that by using such a small threshold, the proposed method detects multiple clusters for each pedestrian, as depicted in Fig. 2, where each pedestrian has optical flows with multiple colors. However, by utilizing the linear correlation exhibited in Fig. 3, we can increase the counting accuracy. Figure 6 shows the relationship between the actual number of pedestrians and the MRE of the number of pedestrians estimated by the proposed method, for various angle threshold values. The error bars in the graph indicate the 95% confidence interval. We can see from this figure that for detecting a small ISBN: 978-1-61804-074-9 MAE 0.952 0.989 3.4 Performance Evaluation with Various Video Sequences To confirming the generality of the proposed method, we evaluated its performance with two other video sequences. One was recorded at a pedestrian crossing in Tokyo, Japan; in the video, many pedestrians move in different directions. This video (320 240 [pixels]) is 20 [s] long and was recorded at 29 [fps]. The other is from the PETS workshop, which can be obtained from [13]. The PETS workshop video (768 576 [pixels]) is 34 [s] long and was recorded at 7 [fps]. A snap- 55

Table 2: Overall counting accuracy using the video sequence taken in Tokyo Method MAE MRE variance proposed 1.123 0.170 1.603 non-clustering 1.502 0.201 1.932 Table 3: Overall counting accuracy using the video sequence from the PETS workshop Method MAE MRE variance proposed 2.528 0.174 3.164 non-clustering 3.085 0.220 3.738 shot from each video sequence is shown in Fig. 7 and Fig. 8. Table 2 and Table 3 summarize the evaluation results of overall counting accuracy for the two video sequences. We can observe that the proposed method gives better counting accuracy than does the non-clustering method. For the video sequence of Tokyo in particular, the proposed method had counting accuracy that was about 25% higher than that of the non-clustering method. 4 Conclusion In this paper, we proposed an optical flow clustering method to improve the accuracy of counting pedestrians in video sequences. The proposed method counts the number of pedestrians using pre-learned statistics, based on the number of clusters detected by our method and the number of pedestrians having a strong and linear correlation. We evaluated the accuracy of the proposed method using several video sequences and showed that the counting accuracy using proposed method is up to 25% better than that of a nonclustering method. We also found that the clustering threshold of angles less than 1 is effective for enhancing counting accuracy. In the future, we plan to evaluate the proposed method using other video sequences. We will also improve the accuracy of the proposed method in order to resolve the problems explained in Subsections 3.1 and 3.2. References: [1] Q. Chen, M. Gao, J. Ma, D. Zhang, L. M. Ni, and Y. Liu, MOCUS: Moving object counting using ultrasonic sensor networks, International Journal of Sensor Networks, vol. 3, no. 1, pp. 55 65, December 2007. [2] A. Leykin and R. Hammoud, Robust multipedestrian tracking in thermal-visible surveillance videos, in Proceedings of Conference on CVPRW 2006, June 2006. [3] E. Goubet, J. Katz, and F. Porikli, Pedestrian tracking using thermal infrared imaging, in Proceedings of SPIE 2006, no. 6206, pp. 797 808, April 2006. [4] PETS2010. http://www.cvg.rdg.ac.uk/pets2010/. [5] T. Ojala, M. Pietikäinen, and T. Mäenpää, Multiresolution gray-scale and rotation invariant texture classification with local binary patterns, IEEE Transactions on Neural Networks, vol. 24, no. 7, pp. 971 989, July 2002. [6] S. Bileschi and L. Wolf, A unified system for object detection, texture recognition, and context analysis based on the standard model feature set, in Proceedings of BMVC 2005, September 2005. [7] M. Heeikkilä and M. Pietikäinen, A texturebased method for modeling the background and detection moving objects, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 28, no. 4, pp. 657 662, April 2006. [8] T. Zhao and R. Nevatia, Bayesian human segmentation in crowded situations, in Proceedings of CVPR 2003, June 2003. [9] T. Zhao and B. Wu, Segmentation and tracking of multiple humans in crowded environments, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 30, no. 7, pp. 1198 1211, July 2008. [10] R. Eshel and Y. Moses, Homography based multiple camera detection and tracking of people in a dense crowd, in Proceedings of CVPR 2008, June 2008. [11] V. Kettnaker and R. Zabih, Counting people from multiple cameras, in Proceedings of MCS 1999, July 1999. [12] B. D. Lucas and T. Kanade, An iterative image registration technique with an application to stereo vision, in Proceedings of the 1981 DARPA Image Understanding Workshop, pp. 121 130, April 1981. [13] The video sequence from PETS workshop. http://www.cvg.rdg.ac.uk/pets2010/a.html. ISBN: 978-1-61804-074-9 56