Context analysis: sky, water and motion
|
|
- Benjamin Perry
- 5 years ago
- Views:
Transcription
1 Context analysis: sky, water and motion Solmaz Javanbakhti Svitlana Zinger Peter H.N. de With Eindhoven University of Technology Electrical Engineering department P.O. Box 513, 5600 MB Eindhoven, The Netherlands Abstract Interpreting the events present in the video is a complex task, and the same gesture or motion can be understood in several ways depending on the context of the event and/or the scene. Therefore the context of the scene can contribute to the semantic understanding of the video. In this paper, we present our research on context analysis on video sequences. By context analysis we mean not only determining the general conditions such as daytime or nighttime, indoor or outdoor environments, but also region labeling [1] and motion analysis of the scene. This paper reports on our research results on sky and water labeling and on motion analysis for determining the context. Later, this can be extended with regions such as roads, greenery, buildings, etc. Experiments based on the above detection techniques show that we achieve results comparable with other state-of-the-art techniques for sky and water detection, although in our case the color information is poor. To evaluate results, we use the Coverability Rate (CR) which measures how much of the true sky or water is detected by the algorithm. The obtained average of CR for water detection is about 96.6% and for sky detection it is about 98%. 1 Introduction Content analysis of digital images and video sequences is widely used, with applications ranging from high-level image understanding and semantic-driven image- and video retrieval, to pixel-level applications like object recognition and local picture quality improvement. Sky is one of the important visual cues, which appears frequently in video and images [2]. Sky detection provides the context information for further image analysis, and it helps to extract information about weather and illumination conditions [3]. Accurate identification of the sky area also enhances advanced object recognition algorithms [3]. Due to sky variations, systems which restrict themselves to blue sky detection have limited practical value [3]. From [2] we learn that the color of the sky changes gradually from dark blue in the zenith to a bright, sometimes almost white, blue near the horizon. In this case, we can model the values in three color channels along straight lines from zenith to horizon as three one-dimensional functions for further analysis. Sky detection in [2] is based on calculating an initial sky belief map, followed by a selection of connected areas based on texture and color analysis and the fitting degree to a two-dimensional (2D) model [2]. An alternative approach is based on the analysis of color, position and shape properties of color-homogeneous spatially connected regions [3]. Another research direction for outdoor environment analysis is the localization of water regions. This analysis involves several influencing factors, like day/night time, reflection at the water surface, relative size of the water region, wave state and possible The research of this paper is part of the ITEA 2 ViCoMo research project.
2 occurrence of material at the water surface [4]. Matthies et al. [5] developed a color image classifier based on a mixture of Gaussians to exploit mean and standard deviation of brightness and saturation, and they trained this classifier on water regions in the RGB color space. In our research, we also perform context analysis based on motion, which can be used to annotate roads and to restrict the computationally heavy search for moving object to the areas where the motion is detected. Most motion analysis techniques are based on comparing the current video frame with a previous frame or with the background [6]. In [7], the motion is detected by comparing neighboring frames when a video stream is recorded in real-time. In this paper, we present and evaluate our approaches to sky and water detection when the color information is poor. The sky detection algorithm [2] that we apply has two phases: (1) training phase which defines the color model, texture properties at multi-resolution and vertical position, (2) detecting phase which adapts the color model, vertical position and calculates texture properties. The water detection algorithm that we present in this paper also consists of two parts: (1) graph-based image segmentation, which generates initial regions, and (2) SVM-based region recognition. Normalized RGB color information is used as a feature for SVM, and we evaluate the entropy of pixels as an additional metric. We also consider the location of pixels, but this feature reduces the flexibility. To analyze motion, we develop a heat map, which is a 2D histogram indicating main regions of motion activity [8], and identify the direction of the movement in the scene using optical flow [8]. The structure of this paper will start by presenting water, sky and motion detection algorithms. Sections 3 will show the evaluation method which was used in this work then the different results of this work. 2 Description of context detection algorithms 2.1 Sky detection algorithm Our sky detection algorithm is based on [9]. It assumes that sky regions are smooth and are found around the top of the image. An initial sky probability is calculated based on color, texture and vertical position, and the features of high probability areas are used to compute a final sky probability. The applied algorithm has clearly lower false detection/rejection rates compared to state-of-the-art algorithms. The improvements are primarily due to an extensive multi-scale texture analysis, adaptive thresholds and a spatially-adaptive color model. Next to position and color features, the key assumption of our system is that sky has a smooth texture and shows limited luminance and chrominance gradients, which are different in the horizontal and vertical directions. Using predefined settings for the color, vertical position, texture, and horizontal and vertical gradients, an initial sky probability is calculated for each image pixel by [9]: P sky = P color P position P texture Q sky, (1) where P color is computed by a three-dimensional Gaussian function in the YUV color space, having a fixed variance, and centered at a predetermined color. Parameter P position is defined by a function that emphasizes the upper parts of the image. For the texture probability P texture, a multi-scale analysis is performed on the image, using an analysis window of 5 5 pixels. Furthermore, we compute a confidence metric Q sky, that prevents small sky-blue objects from being accepted as sky, in images where large areas with high sky probability are absent [9]. 2.2 Water detection algorithm Finding water regions in the image provides a useful context for image understanding. For example, to improve the robustness of ship detection, it can be helpful to locate
3 the water region in the image. The idea is that the detected ship can be confirmed if its major part is included within the water region. Our system basically consists of three components. First, each image is segmented in terms of the color uniformity of a region. In order to choose a proper segmentation algorithm, several popular algorithms, including mean shift, graph cut, normalized cut and graph-based algorithms, are investigated and compared. The graph-based method is proven to be suited for our application due to both its real-time processing capabilities and sufficient accuracy. Let us now explain the first component - segmentation - of our water detection algorithm. The whole image is treated as a graph {V, E}. Each pixel is regarded as a node v i V, and for each adjacent pixel v i,j, there is an edge v i,j E. The weight of the edge is the Euclidean distance of the two nodes in color space, specified by: W i,j = (R i R j ) 2 + (G i G j ) 2 + (B i B j ) 2. (2) Here, two kinds of weights are defined for regions: inner-region weight and inter-region weight. the inner-region weight is defined as the maximum edge weight within a region, and the inter-region weight is specified as the minimum edge weight that connects two regions. The basic idea behind this graph-based method as discussed in [10] is that pixels within one region are closer in feature space (color space in this case) than pixels from different regions. The difference between inter-region weight and inner-region weight is not a fixed number. Instead, there are high variance regions and smooth regions, so that the image cannot be segmented simply according to the absolute edge weight. However, there should always be significant weight changes if we go from one region to another. Such a change depends also on the size of a region. We assume that a larger region has a bigger tolerance while a smaller region can hardly have a large variance. Or if a region is large, it typically incorporates more neighboring pixels. In this algorithm, the Euclidean distance in RGB space is used as a metric. However, for this application, when there is strong disturbance in the river, usually caused by the movement of ships, the river tends to be broken into different regions. Instead, the desired result is to segment the whole river as one region. The idea is to reduce the distance in gray scale and increase the distance in color scale. As an extreme, we could segment mainly according to its color. This is because one object usually has one color but its appearance may change because of the light, shadow and reflection conditions. Moreover, usually the color change caused by the environment does not have a high saturation. Therefore, we explored the normalized RGB color to segment the image. This leads to a weight W i,j = ( Ri R ) 2 j + L i L j ( Gi G ) 2 ( j Bi + B ) 2 j, (3) L i L j L i L j where L i is the brightness, L i = max{r i, G i, B i } in pixel i. It can be seen that the normalized RGB is less sensitive to the brightness, which is good for detecting river and mush regions. However, the sand bank and ships sometimes also merge with the river and sky. Therefore, we combine the normalized difference with the brightness as the weight of edge. This results in the final definition of the weight, giving W i,j = (1 α) ( Ri R ) 2 j + L i L j ( Gi G ) 2 ( j Bi + B ) 2 j + α. L i L j, (4) L i L j L i L j where α is used to adjust the ratio of brightness and normalized RGB. Graph-based segmentation can be performed with fast processing and its result is still acceptable. This explains why we finally adopt a graph-based segmentation technique. In the second processing stage of water detection, visual features of each region, such
4 as the color feature (RGB) and the texture feature (entropy) are extracted and analyzed. It is possible to sample the region and calculate the mean saturation in an HSV histogram to distinguish the regions. However, this method may not robust. Therefore we explore the entropy as a feature, which is defined by E = p log p, where p is the probability in the local histogram or occurrence matrix. Here we use 5 5 neighborhoods to calculate the entropy. Generally, the sky is a smooth area with a low entropy, mush has a large entropy, while water has a moderate entropy. We have found that the entropy of the pixels does not achieve much improvement in segmentation. For this reason, the final model only uses RGB as input vectors. Finally in the third stage, a Support Vector Machine (SVM) classifier is used to recognize the water region. The classifier is trained off-line, based on the image samples captured at the same harbor in different weather conditions. 2.3 Motion detection algorithm Context analysis based on motion is another goal of this work which can be used to annotate roads and to restrict the computationally heavy search for moving objects to the areas where the motion is detected. To analyze and visualized the motion, we apply the concept of a heat map. A heat map is a 2D histogram indicating main regions of motion activity [8]. We also identify the direction of the movement in the scene using optical flow [8]. Our approach is based on motion variation in regions of interest. First, a motion intensity heat map is computed. The motion heat map represents hot and cold areas on the basis of motion intensities. The hot areas are the zones of the scene where the motion is high. The cold areas are regions of the scene where the motion intensities are low. This histogram can be designed from the accumulation of binary blobs of moving objects, which are extracted following the background subtraction method [8]. The obtained heat map can be used as a mask to define regions of interest for a following step of the processing algorithm, such as change detection. The use of the heat map image improves the quality of the results and reduces processing time which is an important factor for real-time applications. We track a moving object over the succeeding frames using the optical flow technique. For this tracking, we have employed the Kanade-Lucas-Tomasi (KLT) feature tracker [8]. After matching an object between frames, the result is a set of vectors which indicate the direction of that moving object. Optical flow is an approximation of the local image motion, based upon local derivatives in a given sequence of images. Thus, the computation of differential optical flow is essentially a two-step procedure [11]. First, it involves a measurement of the spatio-temporal intensity derivatives, which is equivalent to measuring the velocities normal to the local intensity structures. Second, it includes the integration of normal velocities into full velocities, for example, either locally via a least squares calculation, or globally via a regularization. Assume I(x, y, t) is the center pixel in a n n neighborhood and moves by δx, δy in time δt to I(x + δx, y + δy, t + δt). Since I(x, y, t) and I(x + δx, y + δy, t + δt) are the images of the same point (and therefore identical) we conclude: I(x, y, t) = I(x + δx, y + δy, t + δt). (5) After performing a Taylor series expansion of I(x, y, z) in Equation (5) on n n n 3D block, we obtain a 3D motion constraint which is used in the KLT algorithm, specified by: I x V x + I y V y + I z V z = I t, (6) where I x,i y, I z and I t are the 3D intensity derivatives in an n n n neighborhood centered at a voxel (x, y, z), so that I x = δi δx, I y = δi δy, I t = δi δt. (7)
5 A constant velocity (V x, V y, V z ) in that neighborhood is solved by: v = [A T W 2 A] 1 A T WB, (8) where the diagonal elements of W are the N = n n n 3D Gaussian coefficients. The N rows of A consist of I x, I y and I z for each (x, y, z) position and the N rows of B consist of the I t values for those (x, y, z) positions, see [11] for further details. The previous considerations lead to the following matrices: A = [ I(x 1, y 1, z 1 ),..., I(x N, y N, z N )], W = diag[w (x 1, y 1, z 1 ),..., W (x N, y N, z N )], B = (I t (x 1, y 1, z 1 ),..., I t (x N, y N, z N )). (9) 3 Experimental results To express the performance of our evaluated results, we use the Coverability Rate (CR), which measures how much of the true sky or water is detected by the algorithm [3]. This rate is computed by O GT CR(O, GT ) =, (10) GT where Ground Truth (GT) is manually annotated water or sky, and O is the area detected as sky or water. For quality evaluation, we have manually annotated sky on 17 images and water on 15 images. Let us first discuss the sky detection results. Fig. 1 (left) shows the original image of our data set which is one frame of a video sequence. As can be observed, this data set does not provide sufficient information in terms of color. Fig. 1 (middle) visualizes the probability map of the sky region which is achieved by the sky detection algorithm on the original image, and Fig. 1 (right) shows the result of applying a threshold on the found probabilities shown on Fig. 1 (middle). For sky detection, the average CR is about 98%. For water detection, we use RGB colors as a feature. We omitted the pixel location as a feature, because this feature reduces the flexibility. Hence, the final model only uses normalized RGB as input vectors. In the region recognition stage, SVM is used to classify the regions using the normalized RGB feature. Fig. 2 shows the original image and the obtained detection results of the detection algorithm. Fig. 3 shows an alternative case, where a ship is entering the camera scene. In both cases the water region is correctly found and the obtained average of CR for water detection is about 96.6%. For motion analysis, we have applied background subtraction on our video sequence to compute a heat map. The threshold that we impose on the gray scale values during background subtraction is 120. The resulting heat map corresponding with our video sequence expressed in gray values, is shown in Fig. 4 (left). We have applied KLT technique to pairs of smoothed neighboring frames of our data set, after which velocity vectors of each pixel are computed. If velocity vectors are zero, we ignore them to achieve better visualization. The result of the motion detection is shown in Fig. 4 (rigth). At the right side of the Figure, we show the direction of the moving object. 4 Conclusions In this paper, we have presented our research on context analysis for video sequences. The purpose of a context detection system is to provide additional background in-
6 Figure 1: Original image (left). Probability of the sky region (middle). Result of the sky detection after thresholding (right). Figure 2: Original image (left). Result of water detection algorithm (right). Figure 3: Original water vision image with ship (left). Result of water detection algorithm (right).
7 Figure 4: Result of gray scale heat map on video of a moving ship ( The white area indicates region with highest movement) (left). Result of optic flow KLT on a moving ship, arrow shows the direction of ship s movement(right). formation to the already existing foreground object detection, to facilitate further interpretation about event classification in the scene. This view point is shared in the research program of the European ViCoMo project. Our research concentrates on sky and water labeling and on motion analysis for determining contextual information of the scene. For sky detection, we have adopted a detection algorithm from earlier work, based on a probability map that jointly uses a color model, texture properties at multi-resolution and the vertical position [9]. We have designed a supplementary water detection algorithm, combining segmentation and region recognition. Water segmentation is carried out at the first stage to increase the robustness of detection and to decrease the calculation complexity in region recognition. Besides this, the usage of the normalized color space and also an additional parameter measuring the intensity difference of two neighboring pixels are introduced to the graph-based segmentation, in order to improve the algorithm with respect to light reflections at the water surface and sub region scattering. In the region recognition stage, color is analyzed, and SVM is used to classify the regions using the RGB pixel values. To analyze motion, we compute a heat map. This information provides context for identifying regions in the scene where the motion occurs and it helps in fast searching of moving objects. As a result, less video processing is required when this context is used. We use a method for motion detection that is sensitive to velocity and direction. This method is based on optical flow and the KLT tracker. To evaluate the results, we have computed the Coverability Rate (CR). The obtained CR average for water detection is about 96.6% and for sky detection it is approximately 98%. The advantage of this work is reusing the several detectors in one system to achieve parallel context analysis. It will be interesting to explore our concept with another real object detection such as group detection to estimate the abnormality of the scene events.
8 References [1] J. Fan, Y. Gao, H. Luo, Multi-Level Annotation of Natural Scenes Using Dominant Image Components and Semantic Concepts, Multimedia 04: Proceedings of the 12th annual ACM international conference on Multimedia, New York, USA, October 2004 [2] B. Zafarifar and P. H. N. de With, Adaptive Modeling of Sky for Video Processing and Coding Applications, WIC, 2006 [3] F. Schmitt, L. Priese, Sky Detection in CSC-segmented Color Images, In fourth International Conference on Computer Vision Theory and Applications (VISAPP), Lisboa, Portugal, 2009 [4] M. Iqbal, O. Morel, F. Meriaudeau, A survey on outdoor water hazard detection, Communication Technology and Systems (ICTS), Indonesia, 2009 [5] L. Matthies, P. Bellutta and M. McHenry, Detecting water hazards for autonomous offroad navigation, In Proceedings of SPIE Conference 5083: Unmanned Ground Vehicle Technology V, 2003 [6] A. Kirillov, Motion Detection Algorithms, http : // video/motion D etection.aspx., last access 20 Jan 2011 [7] M. AzamQsman, A. Zawawi Talib, T. Kian Lam, W. Poh Lee, M. Sabudin Vehicle monitoring system using motion detection and character recognition algorithms for USM campus, Proceedings of the third IMT-GT Regional Conference on Mathematics, Statistics and Applications Universiti Sains Malaysia, Dec 2007 [8] N. Ihaddadene, C. Djeraba, Real-time Crowd Motion Analysis, 19th International Conference on Pattern Recognition (ICPR 2008) , Tampa, Florida - USA, December, [9] B. Zafarifar, P. H. N. de With, Blue Sky Detection for Content-based Television Picture Quality Enhancement, IEEE International Conference on Consumer Electronics (ICCE), Las Vegas, [10] S. Liu, J. Han, Camera-based water region detection in harbor monitoring, personal comunication, [11] Tutorial on Optical Flow, University of Manchester, last access 14 Jan 2011.
Blue Sky Detection for Picture Quality Enhancement
Blue Sky Detection for Picture Quality Enhancement Bahman Zafarifar 2,3 and Peter H. N. de With 1,2 1 Eindhoven University of Technology, PO Box 513, 5600 MB, The Netherlands, {B.Zafarifar, P.H.N.de.With}@tue.nl
More informationIntroduction to Medical Imaging (5XSA0) Module 5
Introduction to Medical Imaging (5XSA0) Module 5 Segmentation Jungong Han, Dirk Farin, Sveta Zinger ( s.zinger@tue.nl ) 1 Outline Introduction Color Segmentation region-growing region-merging watershed
More informationMULTI ORIENTATION PERFORMANCE OF FEATURE EXTRACTION FOR HUMAN HEAD RECOGNITION
MULTI ORIENTATION PERFORMANCE OF FEATURE EXTRACTION FOR HUMAN HEAD RECOGNITION Panca Mudjirahardjo, Rahmadwati, Nanang Sulistiyanto and R. Arief Setyawan Department of Electrical Engineering, Faculty of
More informationMotion Estimation. There are three main types (or applications) of motion estimation:
Members: D91922016 朱威達 R93922010 林聖凱 R93922044 謝俊瑋 Motion Estimation There are three main types (or applications) of motion estimation: Parametric motion (image alignment) The main idea of parametric motion
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 11 140311 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Motion Analysis Motivation Differential Motion Optical
More informationMULTIVIEW REPRESENTATION OF 3D OBJECTS OF A SCENE USING VIDEO SEQUENCES
MULTIVIEW REPRESENTATION OF 3D OBJECTS OF A SCENE USING VIDEO SEQUENCES Mehran Yazdi and André Zaccarin CVSL, Dept. of Electrical and Computer Engineering, Laval University Ste-Foy, Québec GK 7P4, Canada
More informationChapter 9 Object Tracking an Overview
Chapter 9 Object Tracking an Overview The output of the background subtraction algorithm, described in the previous chapter, is a classification (segmentation) of pixels into foreground pixels (those belonging
More informationAn ICA based Approach for Complex Color Scene Text Binarization
An ICA based Approach for Complex Color Scene Text Binarization Siddharth Kherada IIIT-Hyderabad, India siddharth.kherada@research.iiit.ac.in Anoop M. Namboodiri IIIT-Hyderabad, India anoop@iiit.ac.in
More informationFace Tracking : An implementation of the Kanade-Lucas-Tomasi Tracking algorithm
Face Tracking : An implementation of the Kanade-Lucas-Tomasi Tracking algorithm Dirk W. Wagener, Ben Herbst Department of Applied Mathematics, University of Stellenbosch, Private Bag X1, Matieland 762,
More informationShort Survey on Static Hand Gesture Recognition
Short Survey on Static Hand Gesture Recognition Huu-Hung Huynh University of Science and Technology The University of Danang, Vietnam Duc-Hoang Vo University of Science and Technology The University of
More informationAutomatic Tracking of Moving Objects in Video for Surveillance Applications
Automatic Tracking of Moving Objects in Video for Surveillance Applications Manjunath Narayana Committee: Dr. Donna Haverkamp (Chair) Dr. Arvin Agah Dr. James Miller Department of Electrical Engineering
More informationFeature Tracking and Optical Flow
Feature Tracking and Optical Flow Prof. D. Stricker Doz. G. Bleser Many slides adapted from James Hays, Derek Hoeim, Lana Lazebnik, Silvio Saverse, who in turn adapted slides from Steve Seitz, Rick Szeliski,
More informationAutomatic Shadow Removal by Illuminance in HSV Color Space
Computer Science and Information Technology 3(3): 70-75, 2015 DOI: 10.13189/csit.2015.030303 http://www.hrpub.org Automatic Shadow Removal by Illuminance in HSV Color Space Wenbo Huang 1, KyoungYeon Kim
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 14 130307 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Stereo Dense Motion Estimation Translational
More informationFeature Tracking and Optical Flow
Feature Tracking and Optical Flow Prof. D. Stricker Doz. G. Bleser Many slides adapted from James Hays, Derek Hoeim, Lana Lazebnik, Silvio Saverse, who 1 in turn adapted slides from Steve Seitz, Rick Szeliski,
More informationRegion-based Segmentation
Region-based Segmentation Image Segmentation Group similar components (such as, pixels in an image, image frames in a video) to obtain a compact representation. Applications: Finding tumors, veins, etc.
More informationChapter 3 Image Registration. Chapter 3 Image Registration
Chapter 3 Image Registration Distributed Algorithms for Introduction (1) Definition: Image Registration Input: 2 images of the same scene but taken from different perspectives Goal: Identify transformation
More informationMoving Object Tracking in Video Using MATLAB
Moving Object Tracking in Video Using MATLAB Bhavana C. Bendale, Prof. Anil R. Karwankar Abstract In this paper a method is described for tracking moving objects from a sequence of video frame. This method
More informationTexture Segmentation by Windowed Projection
Texture Segmentation by Windowed Projection 1, 2 Fan-Chen Tseng, 2 Ching-Chi Hsu, 2 Chiou-Shann Fuh 1 Department of Electronic Engineering National I-Lan Institute of Technology e-mail : fctseng@ccmail.ilantech.edu.tw
More informationInternational Journal of Advance Engineering and Research Development
Scientific Journal of Impact Factor (SJIF): 4.72 International Journal of Advance Engineering and Research Development Volume 4, Issue 11, November -2017 e-issn (O): 2348-4470 p-issn (P): 2348-6406 Comparative
More informationCoarse-to-fine image registration
Today we will look at a few important topics in scale space in computer vision, in particular, coarseto-fine approaches, and the SIFT feature descriptor. I will present only the main ideas here to give
More informationDigital Image Processing. Prof. P.K. Biswas. Department of Electronics & Electrical Communication Engineering
Digital Image Processing Prof. P.K. Biswas Department of Electronics & Electrical Communication Engineering Indian Institute of Technology, Kharagpur Image Segmentation - III Lecture - 31 Hello, welcome
More informationCorner Detection. Harvey Rhody Chester F. Carlson Center for Imaging Science Rochester Institute of Technology
Corner Detection Harvey Rhody Chester F. Carlson Center for Imaging Science Rochester Institute of Technology rhody@cis.rit.edu April 11, 2006 Abstract Corners and edges are two of the most important geometrical
More informationFish species recognition from video using SVM classifier
Fish species recognition from video using SVM classifier Katy Blanc, Diane Lingrand, Frédéric Precioso Univ. Nice Sophia Antipolis, I3S, UMR 7271, 06900 Sophia Antipolis, France CNRS, I3S, UMR 7271, 06900
More informationRange Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation
Obviously, this is a very slow process and not suitable for dynamic scenes. To speed things up, we can use a laser that projects a vertical line of light onto the scene. This laser rotates around its vertical
More informationDense Image-based Motion Estimation Algorithms & Optical Flow
Dense mage-based Motion Estimation Algorithms & Optical Flow Video A video is a sequence of frames captured at different times The video data is a function of v time (t) v space (x,y) ntroduction to motion
More informationMotion. 1 Introduction. 2 Optical Flow. Sohaib A Khan. 2.1 Brightness Constancy Equation
Motion Sohaib A Khan 1 Introduction So far, we have dealing with single images of a static scene taken by a fixed camera. Here we will deal with sequence of images taken at different time intervals. Motion
More informationAn Introduction to Content Based Image Retrieval
CHAPTER -1 An Introduction to Content Based Image Retrieval 1.1 Introduction With the advancement in internet and multimedia technologies, a huge amount of multimedia data in the form of audio, video and
More informationTexture. Frequency Descriptors. Frequency Descriptors. Frequency Descriptors. Frequency Descriptors. Frequency Descriptors
Texture The most fundamental question is: How can we measure texture, i.e., how can we quantitatively distinguish between different textures? Of course it is not enough to look at the intensity of individual
More informationDetecting and Identifying Moving Objects in Real-Time
Chapter 9 Detecting and Identifying Moving Objects in Real-Time For surveillance applications or for human-computer interaction, the automated real-time tracking of moving objects in images from a stationary
More informationTopic 4 Image Segmentation
Topic 4 Image Segmentation What is Segmentation? Why? Segmentation important contributing factor to the success of an automated image analysis process What is Image Analysis: Processing images to derive
More informationA Robust Wipe Detection Algorithm
A Robust Wipe Detection Algorithm C. W. Ngo, T. C. Pong & R. T. Chin Department of Computer Science The Hong Kong University of Science & Technology Clear Water Bay, Kowloon, Hong Kong Email: fcwngo, tcpong,
More informationText Information Extraction And Analysis From Images Using Digital Image Processing Techniques
Text Information Extraction And Analysis From Images Using Digital Image Processing Techniques Partha Sarathi Giri Department of Electronics and Communication, M.E.M.S, Balasore, Odisha Abstract Text data
More informationHuman Motion Detection and Tracking for Video Surveillance
Human Motion Detection and Tracking for Video Surveillance Prithviraj Banerjee and Somnath Sengupta Department of Electronics and Electrical Communication Engineering Indian Institute of Technology, Kharagpur,
More informationMotion Analysis. Motion analysis. Now we will talk about. Differential Motion Analysis. Motion analysis. Difference Pictures
Now we will talk about Motion Analysis Motion analysis Motion analysis is dealing with three main groups of motionrelated problems: Motion detection Moving object detection and location. Derivation of
More informationA Fast Moving Object Detection Technique In Video Surveillance System
A Fast Moving Object Detection Technique In Video Surveillance System Paresh M. Tank, Darshak G. Thakore, Computer Engineering Department, BVM Engineering College, VV Nagar-388120, India. Abstract Nowadays
More informationA new predictive image compression scheme using histogram analysis and pattern matching
University of Wollongong Research Online University of Wollongong in Dubai - Papers University of Wollongong in Dubai 00 A new predictive image compression scheme using histogram analysis and pattern matching
More informationSupervised texture detection in images
Supervised texture detection in images Branislav Mičušík and Allan Hanbury Pattern Recognition and Image Processing Group, Institute of Computer Aided Automation, Vienna University of Technology Favoritenstraße
More informationTraffic Signs Recognition using HP and HOG Descriptors Combined to MLP and SVM Classifiers
Traffic Signs Recognition using HP and HOG Descriptors Combined to MLP and SVM Classifiers A. Salhi, B. Minaoui, M. Fakir, H. Chakib, H. Grimech Faculty of science and Technology Sultan Moulay Slimane
More informationMOVING OBJECT DETECTION USING BACKGROUND SUBTRACTION ALGORITHM USING SIMULINK
MOVING OBJECT DETECTION USING BACKGROUND SUBTRACTION ALGORITHM USING SIMULINK Mahamuni P. D 1, R. P. Patil 2, H.S. Thakar 3 1 PG Student, E & TC Department, SKNCOE, Vadgaon Bk, Pune, India 2 Asst. Professor,
More informationCHAPTER 4 SEMANTIC REGION-BASED IMAGE RETRIEVAL (SRBIR)
63 CHAPTER 4 SEMANTIC REGION-BASED IMAGE RETRIEVAL (SRBIR) 4.1 INTRODUCTION The Semantic Region Based Image Retrieval (SRBIR) system automatically segments the dominant foreground region and retrieves
More informationChange detection using joint intensity histogram
Change detection using joint intensity histogram Yasuyo Kita National Institute of Advanced Industrial Science and Technology (AIST) Information Technology Research Institute AIST Tsukuba Central 2, 1-1-1
More informationSchedule for Rest of Semester
Schedule for Rest of Semester Date Lecture Topic 11/20 24 Texture 11/27 25 Review of Statistics & Linear Algebra, Eigenvectors 11/29 26 Eigenvector expansions, Pattern Recognition 12/4 27 Cameras & calibration
More informationAnalysis of Image and Video Using Color, Texture and Shape Features for Object Identification
IOSR Journal of Computer Engineering (IOSR-JCE) e-issn: 2278-0661,p-ISSN: 2278-8727, Volume 16, Issue 6, Ver. VI (Nov Dec. 2014), PP 29-33 Analysis of Image and Video Using Color, Texture and Shape Features
More informationCORRELATION BASED CAR NUMBER PLATE EXTRACTION SYSTEM
CORRELATION BASED CAR NUMBER PLATE EXTRACTION SYSTEM 1 PHYO THET KHIN, 2 LAI LAI WIN KYI 1,2 Department of Information Technology, Mandalay Technological University The Republic of the Union of Myanmar
More informationA Feature Point Matching Based Approach for Video Objects Segmentation
A Feature Point Matching Based Approach for Video Objects Segmentation Yan Zhang, Zhong Zhou, Wei Wu State Key Laboratory of Virtual Reality Technology and Systems, Beijing, P.R. China School of Computer
More informationMotion in 2D image sequences
Motion in 2D image sequences Definitely used in human vision Object detection and tracking Navigation and obstacle avoidance Analysis of actions or activities Segmentation and understanding of video sequences
More informationLearning a Sparse, Corner-based Representation for Time-varying Background Modelling
Learning a Sparse, Corner-based Representation for Time-varying Background Modelling Qiang Zhu 1, Shai Avidan 2, Kwang-Ting Cheng 1 1 Electrical & Computer Engineering Department University of California
More informationEvaluation of Moving Object Tracking Techniques for Video Surveillance Applications
International Journal of Current Engineering and Technology E-ISSN 2277 4106, P-ISSN 2347 5161 2015INPRESSCO, All Rights Reserved Available at http://inpressco.com/category/ijcet Research Article Evaluation
More informationBSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy
BSB663 Image Processing Pinar Duygulu Slides are adapted from Selim Aksoy Image matching Image matching is a fundamental aspect of many problems in computer vision. Object or scene recognition Solving
More informationShadow removal in indoor scenes
Shadow removal in indoor scenes A. T. Nghiem, F. Bremond, M. Thonnat Project Pulsar INRIA Sophia Antipolis France 2004 Route des Lucioles BP 93 06902 Sophia Antipolis France atnghiem@sophia.inria.fr, Francois.Bremond@sophia.inria.fr,
More informationLearning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009
Learning and Inferring Depth from Monocular Images Jiyan Pan April 1, 2009 Traditional ways of inferring depth Binocular disparity Structure from motion Defocus Given a single monocular image, how to infer
More informationMotion Detection Algorithm
Volume 1, No. 12, February 2013 ISSN 2278-1080 The International Journal of Computer Science & Applications (TIJCSA) RESEARCH PAPER Available Online at http://www.journalofcomputerscience.com/ Motion Detection
More informationTime Stamp Detection and Recognition in Video Frames
Time Stamp Detection and Recognition in Video Frames Nongluk Covavisaruch and Chetsada Saengpanit Department of Computer Engineering, Chulalongkorn University, Bangkok 10330, Thailand E-mail: nongluk.c@chula.ac.th
More informationCS 4495 Computer Vision Motion and Optic Flow
CS 4495 Computer Vision Aaron Bobick School of Interactive Computing Administrivia PS4 is out, due Sunday Oct 27 th. All relevant lectures posted Details about Problem Set: You may *not* use built in Harris
More informationVideo Inter-frame Forgery Identification Based on Optical Flow Consistency
Sensors & Transducers 24 by IFSA Publishing, S. L. http://www.sensorsportal.com Video Inter-frame Forgery Identification Based on Optical Flow Consistency Qi Wang, Zhaohong Li, Zhenzhen Zhang, Qinglong
More informationOptical flow and tracking
EECS 442 Computer vision Optical flow and tracking Intro Optical flow and feature tracking Lucas-Kanade algorithm Motion segmentation Segments of this lectures are courtesy of Profs S. Lazebnik S. Seitz,
More informationLocal Feature Detectors
Local Feature Detectors Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr Slides adapted from Cordelia Schmid and David Lowe, CVPR 2003 Tutorial, Matthew Brown,
More informationAutomatic Colorization of Grayscale Images
Automatic Colorization of Grayscale Images Austin Sousa Rasoul Kabirzadeh Patrick Blaes Department of Electrical Engineering, Stanford University 1 Introduction ere exists a wealth of photographic images,
More informationFinal Review CMSC 733 Fall 2014
Final Review CMSC 733 Fall 2014 We have covered a lot of material in this course. One way to organize this material is around a set of key equations and algorithms. You should be familiar with all of these,
More informationDetecting motion by means of 2D and 3D information
Detecting motion by means of 2D and 3D information Federico Tombari Stefano Mattoccia Luigi Di Stefano Fabio Tonelli Department of Electronics Computer Science and Systems (DEIS) Viale Risorgimento 2,
More informationColour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation
ÖGAI Journal 24/1 11 Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation Michael Bleyer, Margrit Gelautz, Christoph Rhemann Vienna University of Technology
More informationImage Analysis Lecture Segmentation. Idar Dyrdal
Image Analysis Lecture 9.1 - Segmentation Idar Dyrdal Segmentation Image segmentation is the process of partitioning a digital image into multiple parts The goal is to divide the image into meaningful
More informationFeature extraction. Bi-Histogram Binarization Entropy. What is texture Texture primitives. Filter banks 2D Fourier Transform Wavlet maxima points
Feature extraction Bi-Histogram Binarization Entropy What is texture Texture primitives Filter banks 2D Fourier Transform Wavlet maxima points Edge detection Image gradient Mask operators Feature space
More informationColumbia University High-Level Feature Detection: Parts-based Concept Detectors
TRECVID 2005 Workshop Columbia University High-Level Feature Detection: Parts-based Concept Detectors Dong-Qing Zhang, Shih-Fu Chang, Winston Hsu, Lexin Xie, Eric Zavesky Digital Video and Multimedia Lab
More informationModule 7 VIDEO CODING AND MOTION ESTIMATION
Module 7 VIDEO CODING AND MOTION ESTIMATION Lesson 20 Basic Building Blocks & Temporal Redundancy Instructional Objectives At the end of this lesson, the students should be able to: 1. Name at least five
More informationRobust Camera Pan and Zoom Change Detection Using Optical Flow
Robust Camera and Change Detection Using Optical Flow Vishnu V. Makkapati Philips Research Asia - Bangalore Philips Innovation Campus, Philips Electronics India Ltd. Manyata Tech Park, Nagavara, Bangalore
More informationPictures at an Exhibition
Pictures at an Exhibition Han-I Su Department of Electrical Engineering Stanford University, CA, 94305 Abstract We employ an image identification algorithm for interactive museum guide with pictures taken
More informationIMAGE SEGMENTATION. Václav Hlaváč
IMAGE SEGMENTATION Václav Hlaváč Czech Technical University in Prague Faculty of Electrical Engineering, Department of Cybernetics Center for Machine Perception http://cmp.felk.cvut.cz/ hlavac, hlavac@fel.cvut.cz
More informationRobust color segmentation algorithms in illumination variation conditions
286 CHINESE OPTICS LETTERS / Vol. 8, No. / March 10, 2010 Robust color segmentation algorithms in illumination variation conditions Jinhui Lan ( ) and Kai Shen ( Department of Measurement and Control Technologies,
More informationOptic Flow and Basics Towards Horn-Schunck 1
Optic Flow and Basics Towards Horn-Schunck 1 Lecture 7 See Section 4.1 and Beginning of 4.2 in Reinhard Klette: Concise Computer Vision Springer-Verlag, London, 2014 1 See last slide for copyright information.
More informationTEVI: Text Extraction for Video Indexing
TEVI: Text Extraction for Video Indexing Hichem KARRAY, Mohamed SALAH, Adel M. ALIMI REGIM: Research Group on Intelligent Machines, EIS, University of Sfax, Tunisia hichem.karray@ieee.org mohamed_salah@laposte.net
More information2 Proposed Methodology
3rd International Conference on Multimedia Technology(ICMT 2013) Object Detection in Image with Complex Background Dong Li, Yali Li, Fei He, Shengjin Wang 1 State Key Laboratory of Intelligent Technology
More informationBackground/Foreground Detection 1
Chapter 2 Background/Foreground Detection 1 2.1 Introduction With the acquisition of an image, the first step is to distinguish objects of interest from the background. In surveillance applications, those
More informationMotion and Optical Flow. Slides from Ce Liu, Steve Seitz, Larry Zitnick, Ali Farhadi
Motion and Optical Flow Slides from Ce Liu, Steve Seitz, Larry Zitnick, Ali Farhadi We live in a moving world Perceiving, understanding and predicting motion is an important part of our daily lives Motion
More informationExtensions of One-Dimensional Gray-level Nonlinear Image Processing Filters to Three-Dimensional Color Space
Extensions of One-Dimensional Gray-level Nonlinear Image Processing Filters to Three-Dimensional Color Space Orlando HERNANDEZ and Richard KNOWLES Department Electrical and Computer Engineering, The College
More informationObject Detection in Video Streams
Object Detection in Video Streams Sandhya S Deore* *Assistant Professor Dept. of Computer Engg., SRES COE Kopargaon *sandhya.deore@gmail.com ABSTRACT Object Detection is the most challenging area in video
More informationAn Edge-Based Approach to Motion Detection*
An Edge-Based Approach to Motion Detection* Angel D. Sappa and Fadi Dornaika Computer Vison Center Edifici O Campus UAB 08193 Barcelona, Spain {sappa, dornaika}@cvc.uab.es Abstract. This paper presents
More informationComputer Science Faculty, Bandar Lampung University, Bandar Lampung, Indonesia
Application Object Detection Using Histogram of Oriented Gradient For Artificial Intelegence System Module of Nao Robot (Control System Laboratory (LSKK) Bandung Institute of Technology) A K Saputra 1.,
More informationOperators-Based on Second Derivative double derivative Laplacian operator Laplacian Operator Laplacian Of Gaussian (LOG) Operator LOG
Operators-Based on Second Derivative The principle of edge detection based on double derivative is to detect only those points as edge points which possess local maxima in the gradient values. Laplacian
More informationA Background Modeling Approach Based on Visual Background Extractor Taotao Liu1, a, Lin Qi2, b and Guichi Liu2, c
4th International Conference on Mechatronics, Materials, Chemistry and Computer Engineering (ICMMCCE 2015) A Background Modeling Approach Based on Visual Background Extractor Taotao Liu1, a, Lin Qi2, b
More informationSmall-scale objects extraction in digital images
102 Int'l Conf. IP, Comp. Vision, and Pattern Recognition IPCV'15 Small-scale objects extraction in digital images V. Volkov 1,2 S. Bobylev 1 1 Radioengineering Dept., The Bonch-Bruevich State Telecommunications
More informationCOMPUTER VISION > OPTICAL FLOW UTRECHT UNIVERSITY RONALD POPPE
COMPUTER VISION 2017-2018 > OPTICAL FLOW UTRECHT UNIVERSITY RONALD POPPE OUTLINE Optical flow Lucas-Kanade Horn-Schunck Applications of optical flow Optical flow tracking Histograms of oriented flow Assignment
More informationVisual Tracking (1) Feature Point Tracking and Block Matching
Intelligent Control Systems Visual Tracking (1) Feature Point Tracking and Block Matching Shingo Kagami Graduate School of Information Sciences, Tohoku University swk(at)ic.is.tohoku.ac.jp http://www.ic.is.tohoku.ac.jp/ja/swk/
More informationVisual Tracking. Image Processing Laboratory Dipartimento di Matematica e Informatica Università degli studi di Catania.
Image Processing Laboratory Dipartimento di Matematica e Informatica Università degli studi di Catania 1 What is visual tracking? estimation of the target location over time 2 applications Six main areas:
More informationExample 2: Straight Lines. Image Segmentation. Example 3: Lines and Circular Arcs. Example 1: Regions
Image Segmentation Image segmentation is the operation of partitioning an image into a collection of connected sets of pixels. 1. into regions, which usually cover the image Example : Straight Lines. into
More informationExample 1: Regions. Image Segmentation. Example 3: Lines and Circular Arcs. Example 2: Straight Lines. Region Segmentation: Segmentation Criteria
Image Segmentation Image segmentation is the operation of partitioning an image into a collection of connected sets of pixels. 1. into regions, which usually cover the image Example 1: Regions. into linear
More informationCar tracking in tunnels
Czech Pattern Recognition Workshop 2000, Tomáš Svoboda (Ed.) Peršlák, Czech Republic, February 2 4, 2000 Czech Pattern Recognition Society Car tracking in tunnels Roman Pflugfelder and Horst Bischof Pattern
More informationHuman Detection. A state-of-the-art survey. Mohammad Dorgham. University of Hamburg
Human Detection A state-of-the-art survey Mohammad Dorgham University of Hamburg Presentation outline Motivation Applications Overview of approaches (categorized) Approaches details References Motivation
More informationPeripheral drift illusion
Peripheral drift illusion Does it work on other animals? Computer Vision Motion and Optical Flow Many slides adapted from J. Hays, S. Seitz, R. Szeliski, M. Pollefeys, K. Grauman and others Video A video
More informationMediaTek Video Face Beautify
MediaTek Video Face Beautify November 2014 2014 MediaTek Inc. Table of Contents 1 Introduction... 3 2 The MediaTek Solution... 4 3 Overview of Video Face Beautify... 4 4 Face Detection... 6 5 Skin Detection...
More informationSIMULATIVE ANALYSIS OF EDGE DETECTION OPERATORS AS APPLIED FOR ROAD IMAGES
SIMULATIVE ANALYSIS OF EDGE DETECTION OPERATORS AS APPLIED FOR ROAD IMAGES Sukhpreet Kaur¹, Jyoti Saxena² and Sukhjinder Singh³ ¹Research scholar, ²Professsor and ³Assistant Professor ¹ ² ³ Department
More informationCS 664 Segmentation. Daniel Huttenlocher
CS 664 Segmentation Daniel Huttenlocher Grouping Perceptual Organization Structural relationships between tokens Parallelism, symmetry, alignment Similarity of token properties Often strong psychophysical
More informationImage Analysis - Lecture 5
Texture Segmentation Clustering Review Image Analysis - Lecture 5 Texture and Segmentation Magnus Oskarsson Lecture 5 Texture Segmentation Clustering Review Contents Texture Textons Filter Banks Gabor
More informationA Texture-based Method for Detecting Moving Objects
A Texture-based Method for Detecting Moving Objects Marko Heikkilä University of Oulu Machine Vision Group FINLAND Introduction The moving object detection, also called as background subtraction, is one
More informationPerformance Evaluation Metrics and Statistics for Positional Tracker Evaluation
Performance Evaluation Metrics and Statistics for Positional Tracker Evaluation Chris J. Needham and Roger D. Boyle School of Computing, The University of Leeds, Leeds, LS2 9JT, UK {chrisn,roger}@comp.leeds.ac.uk
More information5. Feature Extraction from Images
5. Feature Extraction from Images Aim of this Chapter: Learn the Basic Feature Extraction Methods for Images Main features: Color Texture Edges Wie funktioniert ein Mustererkennungssystem Test Data x i
More informationComputer Vision I - Filtering and Feature detection
Computer Vision I - Filtering and Feature detection Carsten Rother 30/10/2015 Computer Vision I: Basics of Image Processing Roadmap: Basics of Digital Image Processing Computer Vision I: Basics of Image
More informationMulti-Camera Calibration, Object Tracking and Query Generation
MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Multi-Camera Calibration, Object Tracking and Query Generation Porikli, F.; Divakaran, A. TR2003-100 August 2003 Abstract An automatic object
More informationTraffic Signs Recognition Experiments with Transform based Traffic Sign Recognition System
Sept. 8-10, 010, Kosice, Slovakia Traffic Signs Recognition Experiments with Transform based Traffic Sign Recognition System Martin FIFIK 1, Ján TURÁN 1, Ľuboš OVSENÍK 1 1 Department of Electronics and
More information