Evaluation of Stereo Matching Algorithms for Occupant Detection
|
|
- Dora Magdalen Spencer
- 6 years ago
- Views:
Transcription
1 Evaluation of Stereo Matching Algorithms for Occupant Detection Sidharta Gautama TELIN-Universiteit Gent St.Pietersnieuwstr.41 B-9000 Gent (Belgium) Simon Lacroix, Michel Devy LAAS-CNRS 7 avenue Colonel Roche Toulouse Cedex (France) simon@laas.fr, devy@laas.fr Abstract Vision systems within vehicles offer new opportunities in the automobile industry. The detection and classification of passenger and driver seat occupancy open up new ways to improve the safety and comfort of the passengers. In this paper we present results of a stereo system designed for the observation of the cockpit scene in order to provide information about the passenger presence and location within the vehicle to improve the control of the airbag firing. We compare different techniques and examine the effect of random and systematic errors on the performance in precision, robustness and processing speed. These results establish a foundation for on-going work on occupant detection for vehicle safety. 1 Introduction Improvement of the safety and comfort of the vehicle occupant is an important research domain in the automobile industry. New sophisticated functions are being developed which improve and automate a variety of vehicle components (e.g. intelligent cruise control, airbag firing). The development of these functions is related to the introduction of new sensors in the automobile application field (radar, lidar, video sensing devices). These sensors will be able to provide information to decision algorithms which can process the raw data into an interpretable analysis of the observed scene. In this paper we discuss occupant detection using stereo vision, related to safety issues. The massive introduction of airbags in cars has generated new problems: indeed, airbag systems are operating in open loop conditions; when an impact is detected, the airbag is automatically This work is supported by the PREDIT program of the french Ministry of National Education and Research. inflated without any feedback about the nature of the seat occupant. This mode is not without problems and has caused a number of dramatic accidents (e.g. passengers in unusual situations are severely injured by the airbag, babies, installed on a child seat in reversed position, are thrown to the back of the vehicle when the airbag inflates). The determination of the presence of a person or object on the seat, and the evaluation of the position of the passenger within the cockpit can be used to improve the operational conditions of the airbag. All these developments are supported by the improvement of the sensor technology and the cost reduction of the electronic components. In order to analyze the seat occupancy, we developed a 3D vision system based on a pair of stereo cameras. Stereo vision offers the potential to produce detailed results within real-time constraints and is suited for irregular environments. The application of occupant detection demands a level of accuracy in addition to the mere detection of presence. A fine location of the boundaries is desired to accurately position the person within the vehicle. This puts strong constraints on the performance of the 3D vision system: 1. Real-time demands: in order for the system to react quickly enough in the case of an accident, a 3D reconstruction at 10 Hz or higher is required; 2. Dense reconstruction: the large variety of situations (e.g. person with child, person reversed, different objects) puts high demands on the density of the 3D reconstruction to be able to sufficiently characterize the form in later stages; 3. Accurate reconstruction: Of importance is the positioning of the head within the cockpit. The application aims at an accurate positioning within centimeter range. In this paper, we evaluate the performance of different stereo algorithms within this context. Section 2 introduces
2 area-based matching. Matching is a fundamental problem in stereo vision, where the aim is to identify corresponding points in left and right image. We examine two approaches to this problem. In section 3, the performance in precision, robustness and computational complexity for these two approaches is examined. Section 4 outlines the application of the system to head detection. 2 Area-based matching Area-based matching algorithms in stereo vision rely on the comparison of regularly sized regions of pixel values in the two images, in order to find the best match. Matching measures are used to provide a numerical measure of the similarity between a window of pixels in one image and a window in another image. Based on these measures an optimal match is determined. 2.1 Zero mean Normalised Cross Correlation (ZNCC) ZNCC belongs to the class of correlation-based measures. These measures are based on the cross-correlation function R between two X images I 1 and I 2 : R(x y)= I 1(u v)i 2(x + u y + v) u v2w where I 1 (x y) and I 2 (x y) denote the intensity values at pixel (x y). ZNCC deals with a possible gain factor and bias between the two images (radiometric distortion) by resp. dividing by the variances of each correlation window, and by first substracting the mean of the correlation window from each pixel value: ScoreZNCC(x y) = P u v2w (I1(u v) ; I 1):(I2(x + u y + v) ; I 2) q P u v2w (I1(u v) ; 2.2 Census Transform (CT) P : I 1)) 2 :( u v2w (I2(u v) ; I 2)) 2 CT belongs to the class of non-parametric techniques [1] [2], which are based on the relative ordering of pixel intensities within a window, rather than the intensity values themselves. Consequently, these techniques are robust with respect to radiometric distortion, since differences in gain and bias between two images will not affect the ordering of pixels within a window. The CT maps the window surrounding the centre pixel to a bit string. If a particular pixel s value is less than the centre pixel then the corresponding position in the bit string will be set to 1, otherwise it is set to 0. Two census transformed images are compared using a similarity measure based on the Hamming distance, i.e. the number of bits that differ in the two bit strings: ScoreCT (x y) = X u v2w Hamming(I1 0 (u v) I0 2 (x + u y + v)) where I 0 1 and I 0 2 represent the census transform of I 1 and I 2. We have introduced a modification within the census transform to make it more robust to image noise and rectification errors. In correspondence with ZNCC, we perform a census transform by comparing each pixel s value with the mean of the correlation window instead of the intensity of the centre pixel. 3 Experimental evaluation For stereo vision, performance depends on many factors, including the contrast and spatial distribution of texture in the intensity signal, the noise level in the images, artifacts introduced by the matching algorithms, and the true range to the objects in the scene [3]. In our experiments, we characterize both random errors caused by noise and the systematic errors resulting from artifacts. To evaluate the performance of the ZNCC and CT algorithms within the framework of occupant detection, we collected different sets of stereo pairs of a flat board, with and without an object of known geometry, from an optical bench. We used a set cameras (2.6 mm objectives) with a 5 cm separation (stereo baseline). A third order distortion model is used to calibrate the cameras and estimate the epipolar geometry. The range of interest is limited to 2 meter. Sets of 50 stereo pairs were taken under constant illumination and processed at resolution to generate sample statistics (fig. 1). Besides these images, we mounted an experimental system within a vehicle cockpit which is used to evaluate the algorithms within the context of occupant detection. 3.1 Analysis of random errors At subpixel resolution, the precision (variance) of a estimate d reflects statistical fluctuations around the sub-pixel mean. This precision depends on the variance of noise in the image, the image intensity in the image and the matching algorithm used. We are interested in the uncertainty in 3-D measurements computed from. If the estimate d follows a Gaussian distribution (which has been verified by Matthies [3]), it has been shown that the estimate of the range Z can be modelled as Gaussian with mean Z = k= d and variance 2 Z ( Z d )2 = k 2 d 4 :2 d (1)
3 (a) Flat Board (b) Book (a) ZNCC Figure 1. Test Images where k is a calibration constant. Using equation (1), the standard deviation on the range can be plotted out against the range (fig. 2). We have analysed this for a set of 50 images, taken from the same scene (a flat board, fig 1(a)). All images were processed with ZNCC and both CT algorithms using a correlation window of 99 pixels and statistics collected. Plotting the standard deviation on the range, we see that the precision remains within the centimeter range for both ZNCC and CT algorithms. ZNCC provides a slightly more stable result, which is due to the higher smoothing effect established by the correlation window. The modification on the CT algorithm introduces no effect on the precision over the original CT. To allow a qualitative evaluation of the robustness of the different algorithms, we averaged the number of matched pixels found for each image pair over the total dataset. This comparison shows the immediate benefit of the modification made to the original CT. Introducing an averaging process in combination with the census transform stabilises the transform for image noise and errors due to the rectification process (fig. 3). Table 1 summarizes the statistics. Based on these results, we opted to continue with the modified version of the CT algorithm. In the following discussion CT stands for the modified census transform. ZNCC CT CT (modified) % % % Table 1. Matched pixels per image pair, averaged over the total dataset. 100% = total image 3.2 Quality of the estimate The at each pixel is estimated using area-based matching with a correlation window around each pixel. (b) CT (c) CT mod Figure 2. Std.deviation on range estimates Several criteria can be used to filter false matchings: strength, uniqueness and form of the correlation peak (fig. 4), and inverse validition. This last test inverses the role of left and right image and considers as valid only those matches for which the reverse correlation has fallen on the initial point in the left image [4]. Sub-pixel estimates are obtained by fitting a parabole to the three correlation values surrounding the optimum. We asses the quality not only on the correctness of the estimate, but also on the ability to filter out false matchings using the validity criteria. To evaluate the matching algorithms, we examined the estimate at various points within the images
4 (a) (b) (c) Figure 3. Disparity images of a sample: (a) ZNCC, (b) CT, (c) CT modified correlation signal difference range variance strength Figure 4. Thresholds on correlation peak shown in fig 5. The dots show the locations where samples were taken. Some example textures are shown for a correlation window of pixels. For these textures, the matching strength of the texture over a given range is plotted out in fig. 6, comparing the correlation signal for ZNCC and CT. (a) (b) (c) Figure 5. Texture samples validation has been plotted as a grey bar when the inverse test is validated (i.e. the reciprocal match falls within one pixel of the original point). This test proves to be the most reliable. Estimates which fall outside the inverse validity test are false or ambiguous. However, the inverse does not hold and false estimates can still filter through. To be able to use the other measures in conjunction with the inverse validity test, the plot for these measures should show a monotonously increasing curve. For the different points we examined, only the difference measure proves to be useful. The peak value and the variance generally show a fluctuating behaviour. We examine the evolution of the peak value, the difference with the second largest peak, the variance and the inverse validation, which are used by the filter criteria, and this for varying correlation window sizes. Since small correlation windows tend to give an unreliable estimate, this gives a rough idea of which criteria are useful to filter out unwanted responses. Fig. 7 shows the plot of these measures for a given point (texture b, fig. 5). The inverse The difference measure proves to be a useful attribute with CT in contrast with ZNCC. To show this, we classified the estimates in our sample points as reliable or unreliable. This was done by regarding each estimate within the observed scene and see if the estimate was consistent with the scene or not. Only the points which passed the inverse validity test are regarded. ZNCC and CT give the same consistency evaluation for each point (i.e. both techniques give approximately the same estimate). However using the difference measure, CT is
5 (a) Texture a (a) ZNCC (b) Texture b (c) Texture c Figure 6. Correlation signals: (right) CT (left) ZNCC, (b) CT Figure 7. Evolution of validity measures able to filter out a large part of the inconsistent points (cfr. fig. 8). With CT, the difference measures generally shows a high respons ( unique peak) with the good estimates and a low respons ( ambiguous peak) with the unreliable estimates. ZNCC on the other hand tends to give a strong respons even in the case of unreliable estimates. 3.3 Analysis of systematic errors Systematic errors are due to processing artifacts. One of these artifacts is the so-called object growing effect. This growing effect is a bias introduced by the finite size of the correlation window. When the correlation window encounters an object edge, the mean calculated within this window (as does ZNCC) reflects both the high and low contrast texture. This causes a confusion in the matching measure and results in a bias towards high contrast. A halo is seen around the object within which the range estimates are nearly the same as the range to the object instead of the range to the background. The size of the halo is half the size of the correlation window. In 3D, this size grows with the distance to the cameras. The problem has been investigated by a number of people (cfr. [5]), but the proposed solutions are not always well adapted to real-time systems. To characterize the object growing effect, we take the 3D reconstruction of a book (fig. 1(b)) and transform the object into a frame following the principal axes of the book. After this transformation, the 3D points are thresholded by keeping all the points above a certain height off the board plane. In this way we are able to segment the points belonging to the object. Using this segmentation, we take the histogram of the 3D points along the horizontal axis. This allows us to estimate the absolute horizontal precision of the system. The book has a width of 22 cm. Both ZNCC and CT give an overestimation of the width, but CT proves to be more accurate (i.e cm compared to 24.9 cm, fig. 9(a)). CT avoids the object growing effect not by a better estimate but by a more efficient filtering, as
6 (a) ZNCC (a) Positional accuracy (b) CT Figure 8. Difference and peak values shown in the previous section. This answer of unknown is much prefered over erroneous estimations. The size of the correlation window also influences the directional accuracy. Increasing the correlation window smooths out the normal directions of the 3D points. This is shown in fig. 9(b). The normal direction is computed in each 3D point using ~n = ~x u ~xv j ~xu ~xvj (2) where ~xu and ~xv are the partial derivatives calculated by differencing in a 4-neighbourhood. Taking the histogram of the normal direction along the range axis, which should be a delta peak in the ideal case (= perfectly flat surface). We note a better performance by the CT algorithm. 3.4 Performance within cockpit model A prototype system, consisting of a pair of stereo cameras combined with infrared illumination, is being developed for installation within a cockpit model. The (b) Directional accuracy Figure 9. Accuracy of range estimates images produced by the cameras are subsampled to a size of pixels. Table 2 shows the timing results for the matching and reconstruction process, performed on an SUN Ultra Sparc 10. The range which is scanned for each pixel is 1 of the image width. The correlation window size is 4 taken pixels and 5 5 pixels for resp. resolution and ZNCC CT 3D Rec ms 550 ms 20 ms ms 100 ms 10 ms Table 2. Timing results on an Ultra Sparc 10 Figure 11 shows an extract of the reconstructed scene. It illustrates the performance of CT compared to ZNCC with respect to the characterisation of form and the object growing effect.
7 4 Head detection using data fusion Although the prototype system s main goal is the classification of the seat occupancy (e.g. child seat, object, passenger), the results obtained with the system allow the exploration of head detection inside the vehicle. Within the context of passenger safety, the head plays a central role. Locating the head with respect to the dashboard is an important issue and fast, robust techniques need to be developed. We are currently exploring a system which fuses 3D and intensity information. Since the vehicle cockpit offers a well-defined environment, a narrow region of interest (ROI) can be defined in which head detection based on skin tone is performed. To eliminate false alarms, we envisage a curvature estimation on the detected points. For this phase, we rely on the better results of CT over ZNCC in precision which makes a finer characterisation of form possible. The flow chart of the system is outlined in fig. 10. Skin tone Left Camera rectification 5 Conclusions Disparity estimation 3D reconstruction ROI Data fusion Right Camera rectification Figure 10. Flow chart Shape In this paper, we compared ZNCC and CT matching algorithms for stereo vision. The data indicates precision within centimeter range for a range up to 2 metres and processing times of 1-10 Hz on our system. CT performs better in terms of accuracy and speed. We examined different validity tests to filter out false matches. The inverse validity test proved to be the most efficient, but does not filter out all unreliable matches. While examining the signal (peak, difference and variance) offers no further improvement of ZNCÇ the difference test shows to be reliable to further suppress unwanted matches in the case of CT. More particularly, the effect of object growing is handled much better in this case. CT can avoid the object growing effect not by a better estimate but by a more efficient filtering. The respons of unknown is much prefered over erroneous estimations. Compared with other solutions [5], CT requires no extra processing time and is well suited for real-time applications. A strong interest of car manufacturers exists in the concept of occupant detection using video sensors. Smart airbag concepts with partial inflation capabilities should take into account information provided by such a perception system. Our prototype system shows encouraging results for the problem of head detection within the vehicle cockpit. Future work is planned to further exploit the results of stereo using CT in order to accurately position the passengers head with respect to the airbag. References [1] R.Zabih, J.Woodfill, Non-parametric Local Transforms for Computing Visual Correspondence, Third European Conference on Computer Vision, Stockholm, Sweden, [2] J.Banks, M.Bennamoun, P.Corke, Fast and robust stereo matching algorithms for mining automation, Proc. of the Int. Workshop on Image Analysis and Information Fusion, Adelaide, Australia, [3] L.Matthies, P.Grandjean, Stochastic Performance Modeling and Evaluation of Obstacle Detectability with Imaging Range Sensors, IEEE Trans. Robotics and Automation, Vol.10, No.6, pp , [4] P.Lassere, P.Grandjean, Stereo Vision Improvement, 7th Int. Conf. on Advanced Robotics, Guixols, Spain, [5] L.Robert, R.Deriche, Dense Depth Map Reconstruction: A Minimization and Regularization Approach which Preserves Discontinuities, Fourth European Conference on Computer Vision, Cambridge, UK, 1996.
8 (a) Disparity map ZNCC (b) Correlated pixels ZNCC (c) Disparity map CT (d) Correlated pixels CT (e) ZNCC (f) CT Figure 11. Reconstruction examples
Transactions on Information and Communications Technologies vol 16, 1996 WIT Press, ISSN
ransactions on Information and Communications echnologies vol 6, 996 WI Press, www.witpress.com, ISSN 743-357 Obstacle detection using stereo without correspondence L. X. Zhou & W. K. Gu Institute of Information
More informationSpatio-Temporal Stereo Disparity Integration
Spatio-Temporal Stereo Disparity Integration Sandino Morales and Reinhard Klette The.enpeda.. Project, The University of Auckland Tamaki Innovation Campus, Auckland, New Zealand pmor085@aucklanduni.ac.nz
More informationMobile Human Detection Systems based on Sliding Windows Approach-A Review
Mobile Human Detection Systems based on Sliding Windows Approach-A Review Seminar: Mobile Human detection systems Njieutcheu Tassi cedrique Rovile Department of Computer Engineering University of Heidelberg
More informationMultiple View Geometry
Multiple View Geometry CS 6320, Spring 2013 Guest Lecture Marcel Prastawa adapted from Pollefeys, Shah, and Zisserman Single view computer vision Projective actions of cameras Camera callibration Photometric
More informationFundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision
Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision What Happened Last Time? Human 3D perception (3D cinema) Computational stereo Intuitive explanation of what is meant by disparity Stereo matching
More informationTyzx DeepSea High Speed Stereo Vision System
Tyzx DeepSea High Speed Stereo Vision System John Iselin Woodfill, Gaile Gordon, Ron Buck Tyzx, Inc. 3885 Bohannon Drive Menlo Park, CA 9425 {Woodfill, Gaile, Ron}@tyzx.com Abstract This paper describes
More information2 OVERVIEW OF RELATED WORK
Utsushi SAKAI Jun OGATA This paper presents a pedestrian detection system based on the fusion of sensors for LIDAR and convolutional neural network based image classification. By using LIDAR our method
More informationFundamentals of Digital Image Processing
\L\.6 Gw.i Fundamentals of Digital Image Processing A Practical Approach with Examples in Matlab Chris Solomon School of Physical Sciences, University of Kent, Canterbury, UK Toby Breckon School of Engineering,
More informationCalibration of a rotating multi-beam Lidar
The 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems October 18-22, 2010, Taipei, Taiwan Calibration of a rotating multi-beam Lidar Naveed Muhammad 1,2 and Simon Lacroix 1,2 Abstract
More informationEXAM SOLUTIONS. Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006,
School of Computer Science and Communication, KTH Danica Kragic EXAM SOLUTIONS Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006, 14.00 19.00 Grade table 0-25 U 26-35 3 36-45
More informationData Term. Michael Bleyer LVA Stereo Vision
Data Term Michael Bleyer LVA Stereo Vision What happened last time? We have looked at our energy function: E ( D) = m( p, dp) + p I < p, q > N s( p, q) We have learned about an optimization algorithm that
More informationChapter 3 Image Registration. Chapter 3 Image Registration
Chapter 3 Image Registration Distributed Algorithms for Introduction (1) Definition: Image Registration Input: 2 images of the same scene but taken from different perspectives Goal: Identify transformation
More informationAdaptative Elimination of False Edges for First Order Detectors
Adaptative Elimination of False Edges for First Order Detectors Djemel ZIOU and Salvatore TABBONE D~partement de math~matiques et d'informatique, universit~ de Sherbrooke, Qc, Canada, J1K 2R1 Crin/Cnrs
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 14 130307 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Stereo Dense Motion Estimation Translational
More informationProf. Fanny Ficuciello Robotics for Bioengineering Visual Servoing
Visual servoing vision allows a robotic system to obtain geometrical and qualitative information on the surrounding environment high level control motion planning (look-and-move visual grasping) low level
More informationDepth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth
Common Classification Tasks Recognition of individual objects/faces Analyze object-specific features (e.g., key points) Train with images from different viewing angles Recognition of object classes Analyze
More informationMicro-scale Stereo Photogrammetry of Skin Lesions for Depth and Colour Classification
Micro-scale Stereo Photogrammetry of Skin Lesions for Depth and Colour Classification Tim Lukins Institute of Perception, Action and Behaviour 1 Introduction The classification of melanoma has traditionally
More informationLaserscanner Based Cooperative Pre-Data-Fusion
Laserscanner Based Cooperative Pre-Data-Fusion 63 Laserscanner Based Cooperative Pre-Data-Fusion F. Ahlers, Ch. Stimming, Ibeo Automobile Sensor GmbH Abstract The Cooperative Pre-Data-Fusion is a novel
More informationTowards the completion of assignment 1
Towards the completion of assignment 1 What to do for calibration What to do for point matching What to do for tracking What to do for GUI COMPSCI 773 Feature Point Detection Why study feature point detection?
More informationDepth Measurement and 3-D Reconstruction of Multilayered Surfaces by Binocular Stereo Vision with Parallel Axis Symmetry Using Fuzzy
Depth Measurement and 3-D Reconstruction of Multilayered Surfaces by Binocular Stereo Vision with Parallel Axis Symmetry Using Fuzzy Sharjeel Anwar, Dr. Shoaib, Taosif Iqbal, Mohammad Saqib Mansoor, Zubair
More information3D DEFORMATION MEASUREMENT USING STEREO- CORRELATION APPLIED TO EXPERIMENTAL MECHANICS
3D DEFORMATION MEASUREMENT USING STEREO- CORRELATION APPLIED TO EXPERIMENTAL MECHANICS Dorian Garcia, Jean-José Orteu École des Mines d Albi, F-81013 ALBI CT Cedex 09, France Dorian.Garcia@enstimac.fr,
More informationAccurate and Dense Wide-Baseline Stereo Matching Using SW-POC
Accurate and Dense Wide-Baseline Stereo Matching Using SW-POC Shuji Sakai, Koichi Ito, Takafumi Aoki Graduate School of Information Sciences, Tohoku University, Sendai, 980 8579, Japan Email: sakai@aoki.ecei.tohoku.ac.jp
More informationAuto-focusing Technique in a Projector-Camera System
2008 10th Intl. Conf. on Control, Automation, Robotics and Vision Hanoi, Vietnam, 17 20 December 2008 Auto-focusing Technique in a Projector-Camera System Lam Bui Quang, Daesik Kim and Sukhan Lee School
More informationA Statistical Consistency Check for the Space Carving Algorithm.
A Statistical Consistency Check for the Space Carving Algorithm. A. Broadhurst and R. Cipolla Dept. of Engineering, Univ. of Cambridge, Cambridge, CB2 1PZ aeb29 cipolla @eng.cam.ac.uk Abstract This paper
More informationEffects Of Shadow On Canny Edge Detection through a camera
1523 Effects Of Shadow On Canny Edge Detection through a camera Srajit Mehrotra Shadow causes errors in computer vision as it is difficult to detect objects that are under the influence of shadows. Shadow
More informationA threshold decision of the object image by using the smart tag
A threshold decision of the object image by using the smart tag Chang-Jun Im, Jin-Young Kim, Kwan Young Joung, Ho-Gil Lee Sensing & Perception Research Group Korea Institute of Industrial Technology (
More informationStereo Vision. MAN-522 Computer Vision
Stereo Vision MAN-522 Computer Vision What is the goal of stereo vision? The recovery of the 3D structure of a scene using two or more images of the 3D scene, each acquired from a different viewpoint in
More informationLearning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009
Learning and Inferring Depth from Monocular Images Jiyan Pan April 1, 2009 Traditional ways of inferring depth Binocular disparity Structure from motion Defocus Given a single monocular image, how to infer
More informationCS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching
Stereo Matching Fundamental matrix Let p be a point in left image, p in right image l l Epipolar relation p maps to epipolar line l p maps to epipolar line l p p Epipolar mapping described by a 3x3 matrix
More informationMeasurement of Pedestrian Groups Using Subtraction Stereo
Measurement of Pedestrian Groups Using Subtraction Stereo Kenji Terabayashi, Yuki Hashimoto, and Kazunori Umeda Chuo University / CREST, JST, 1-13-27 Kasuga, Bunkyo-ku, Tokyo 112-8551, Japan terabayashi@mech.chuo-u.ac.jp
More informationФУНДАМЕНТАЛЬНЫЕ НАУКИ. Информатика 9 ИНФОРМАТИКА MOTION DETECTION IN VIDEO STREAM BASED ON BACKGROUND SUBTRACTION AND TARGET TRACKING
ФУНДАМЕНТАЛЬНЫЕ НАУКИ Информатика 9 ИНФОРМАТИКА UDC 6813 OTION DETECTION IN VIDEO STREA BASED ON BACKGROUND SUBTRACTION AND TARGET TRACKING R BOGUSH, S ALTSEV, N BROVKO, E IHAILOV (Polotsk State University
More informationRange Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation
Obviously, this is a very slow process and not suitable for dynamic scenes. To speed things up, we can use a laser that projects a vertical line of light onto the scene. This laser rotates around its vertical
More informationDepth Sensors Kinect V2 A. Fornaser
Depth Sensors Kinect V2 A. Fornaser alberto.fornaser@unitn.it Vision Depth data It is not a 3D data, It is a map of distances Not a 3D, not a 2D it is a 2.5D or Perspective 3D Complete 3D - Tomography
More informationDetecting motion by means of 2D and 3D information
Detecting motion by means of 2D and 3D information Federico Tombari Stefano Mattoccia Luigi Di Stefano Fabio Tonelli Department of Electronics Computer Science and Systems (DEIS) Viale Risorgimento 2,
More informationPedestrian Detection Using Correlated Lidar and Image Data EECS442 Final Project Fall 2016
edestrian Detection Using Correlated Lidar and Image Data EECS442 Final roject Fall 2016 Samuel Rohrer University of Michigan rohrer@umich.edu Ian Lin University of Michigan tiannis@umich.edu Abstract
More informationA New Approach to Computation of Curvature Scale Space Image for Shape Similarity Retrieval
A New Approach to Computation of Curvature Scale Space Image for Shape Similarity Retrieval Farzin Mokhtarian, Sadegh Abbasi and Josef Kittler Centre for Vision Speech and Signal Processing Department
More informationSuRVoS Workbench. Super-Region Volume Segmentation. Imanol Luengo
SuRVoS Workbench Super-Region Volume Segmentation Imanol Luengo Index - The project - What is SuRVoS - SuRVoS Overview - What can it do - Overview of the internals - Current state & Limitations - Future
More informationFlexible Calibration of a Portable Structured Light System through Surface Plane
Vol. 34, No. 11 ACTA AUTOMATICA SINICA November, 2008 Flexible Calibration of a Portable Structured Light System through Surface Plane GAO Wei 1 WANG Liang 1 HU Zhan-Yi 1 Abstract For a portable structured
More informationDense 3D Reconstruction. Christiano Gava
Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Today: dense 3D reconstruction The matching problem
More informationEpipolar geometry-based ego-localization using an in-vehicle monocular camera
Epipolar geometry-based ego-localization using an in-vehicle monocular camera Haruya Kyutoku 1, Yasutomo Kawanishi 1, Daisuke Deguchi 1, Ichiro Ide 1, Hiroshi Murase 1 1 : Nagoya University, Japan E-mail:
More informationUsing temporal seeding to constrain the disparity search range in stereo matching
Using temporal seeding to constrain the disparity search range in stereo matching Thulani Ndhlovu Mobile Intelligent Autonomous Systems CSIR South Africa Email: tndhlovu@csir.co.za Fred Nicolls Department
More informationChapter 9 Object Tracking an Overview
Chapter 9 Object Tracking an Overview The output of the background subtraction algorithm, described in the previous chapter, is a classification (segmentation) of pixels into foreground pixels (those belonging
More informationCS4495/6495 Introduction to Computer Vision
CS4495/6495 Introduction to Computer Vision 9C-L1 3D perception Some slides by Kelsey Hawkins Motivation Why do animals, people & robots need vision? To detect and recognize objects/landmarks Is that a
More informationOn-line and Off-line 3D Reconstruction for Crisis Management Applications
On-line and Off-line 3D Reconstruction for Crisis Management Applications Geert De Cubber Royal Military Academy, Department of Mechanical Engineering (MSTA) Av. de la Renaissance 30, 1000 Brussels geert.de.cubber@rma.ac.be
More informationFast, Unconstrained Camera Motion Estimation from Stereo without Tracking and Robust Statistics
Fast, Unconstrained Camera Motion Estimation from Stereo without Tracking and Robust Statistics Heiko Hirschmüller, Peter R. Innocent and Jon M. Garibaldi Centre for Computational Intelligence, De Montfort
More informationthen assume that we are given the image of one of these textures captured by a camera at a different (longer) distance and with unknown direction of i
Image Texture Prediction using Colour Photometric Stereo Xavier Lladó 1, Joan Mart 1, and Maria Petrou 2 1 Institute of Informatics and Applications, University of Girona, 1771, Girona, Spain fllado,joanmg@eia.udg.es
More informationStructured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov
Structured Light II Johannes Köhler Johannes.koehler@dfki.de Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Introduction Previous lecture: Structured Light I Active Scanning Camera/emitter
More informationModel-Based Stereo. Chapter Motivation. The modeling system described in Chapter 5 allows the user to create a basic model of a
96 Chapter 7 Model-Based Stereo 7.1 Motivation The modeling system described in Chapter 5 allows the user to create a basic model of a scene, but in general the scene will have additional geometric detail
More informationTowards Autonomous Vehicle. What is an autonomous vehicle? Vehicle driving on its own with zero mistakes How? Using sensors
7 May 2017 Disclaimer Towards Autonomous Vehicle What is an autonomous vehicle? Vehicle driving on its own with zero mistakes How? Using sensors Why Vision Sensors? Humans use both eyes as main sense
More information3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller
3D Computer Vision Depth Cameras Prof. Didier Stricker Oliver Wasenmüller Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de
More informationCentre for Digital Image Measurement and Analysis, School of Engineering, City University, Northampton Square, London, ECIV OHB
HIGH ACCURACY 3-D MEASUREMENT USING MULTIPLE CAMERA VIEWS T.A. Clarke, T.J. Ellis, & S. Robson. High accuracy measurement of industrially produced objects is becoming increasingly important. The techniques
More informationCOMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION
COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION Mr.V.SRINIVASA RAO 1 Prof.A.SATYA KALYAN 2 DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING PRASAD V POTLURI SIDDHARTHA
More informationA virtual tour of free viewpoint rendering
A virtual tour of free viewpoint rendering Cédric Verleysen ICTEAM institute, Université catholique de Louvain, Belgium cedric.verleysen@uclouvain.be Organization of the presentation Context Acquisition
More informationColorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.
Professor William Hoff Dept of Electrical Engineering &Computer Science http://inside.mines.edu/~whoff/ 1 Stereo Vision 2 Inferring 3D from 2D Model based pose estimation single (calibrated) camera > Can
More informationRevising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History
Revising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History Simon Thompson and Satoshi Kagami Digital Human Research Center National Institute of Advanced
More informationOutline. ETN-FPI Training School on Plenoptic Sensing
Outline Introduction Part I: Basics of Mathematical Optimization Linear Least Squares Nonlinear Optimization Part II: Basics of Computer Vision Camera Model Multi-Camera Model Multi-Camera Calibration
More informationparco area delle Scienze, 181A via Ferrata, , Parma 27100, Pavia
Proceedings of the IEEE Intelligent Vehicles Symposium 2000 Dearbon (MI), USA October 3-5, 2000 Stereo Vision-based Vehicle Detection M. Bertozzi 1 A. Broggi 2 A. Fascioli 1 S. Nichele 2 1 Dipartimento
More informationCHAPTER 3 DISPARITY AND DEPTH MAP COMPUTATION
CHAPTER 3 DISPARITY AND DEPTH MAP COMPUTATION In this chapter we will discuss the process of disparity computation. It plays an important role in our caricature system because all 3D coordinates of nodes
More informationRobust color segmentation algorithms in illumination variation conditions
286 CHINESE OPTICS LETTERS / Vol. 8, No. / March 10, 2010 Robust color segmentation algorithms in illumination variation conditions Jinhui Lan ( ) and Kai Shen ( Department of Measurement and Control Technologies,
More informationCS 4495/7495 Computer Vision Frank Dellaert, Fall 07. Dense Stereo Some Slides by Forsyth & Ponce, Jim Rehg, Sing Bing Kang
CS 4495/7495 Computer Vision Frank Dellaert, Fall 07 Dense Stereo Some Slides by Forsyth & Ponce, Jim Rehg, Sing Bing Kang Etymology Stereo comes from the Greek word for solid (στερεο), and the term can
More informationDigital Image Processing COSC 6380/4393
Digital Image Processing COSC 6380/4393 Lecture 21 Nov 16 th, 2017 Pranav Mantini Ack: Shah. M Image Processing Geometric Transformation Point Operations Filtering (spatial, Frequency) Input Restoration/
More informationLecture 6: Edge Detection
#1 Lecture 6: Edge Detection Saad J Bedros sbedros@umn.edu Review From Last Lecture Options for Image Representation Introduced the concept of different representation or transformation Fourier Transform
More informationEdge detection. Goal: Identify sudden. an image. Ideal: artist s line drawing. object-level knowledge)
Edge detection Goal: Identify sudden changes (discontinuities) in an image Intuitively, most semantic and shape information from the image can be encoded in the edges More compact than pixels Ideal: artist
More informationRecognition of Object Contours from Stereo Images: an Edge Combination Approach
Recognition of Object Contours from Stereo Images: an Edge Combination Approach Margrit Gelautz and Danijela Markovic Institute for Software Technology and Interactive Systems, Vienna University of Technology
More informationEmbedded real-time stereo estimation via Semi-Global Matching on the GPU
Embedded real-time stereo estimation via Semi-Global Matching on the GPU Daniel Hernández Juárez, Alejandro Chacón, Antonio Espinosa, David Vázquez, Juan Carlos Moure and Antonio M. López Computer Architecture
More informationCS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching
Stereo Matching Fundamental matrix Let p be a point in left image, p in right image l l Epipolar relation p maps to epipolar line l p maps to epipolar line l p p Epipolar mapping described by a 3x3 matrix
More informationVehicle Dimensions Estimation Scheme Using AAM on Stereoscopic Video
Workshop on Vehicle Retrieval in Surveillance (VRS) in conjunction with 2013 10th IEEE International Conference on Advanced Video and Signal Based Surveillance Vehicle Dimensions Estimation Scheme Using
More informationDISTANCE MEASUREMENT USING STEREO VISION
DISTANCE MEASUREMENT USING STEREO VISION Sheetal Nagar 1, Jitendra Verma 2 1 Department of Electronics and Communication Engineering, IIMT, Greater Noida (India) 2 Department of computer science Engineering,
More informationStereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman
Stereo 11/02/2012 CS129, Brown James Hays Slides by Kristen Grauman Multiple views Multi-view geometry, matching, invariant features, stereo vision Lowe Hartley and Zisserman Why multiple views? Structure
More informationComputer Vision Lecture 17
Computer Vision Lecture 17 Epipolar Geometry & Stereo Basics 13.01.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Announcements Seminar in the summer semester
More informationProceedings of the 6th Int. Conf. on Computer Analysis of Images and Patterns. Direct Obstacle Detection and Motion. from Spatio-Temporal Derivatives
Proceedings of the 6th Int. Conf. on Computer Analysis of Images and Patterns CAIP'95, pp. 874-879, Prague, Czech Republic, Sep 1995 Direct Obstacle Detection and Motion from Spatio-Temporal Derivatives
More informationComputer Vision Lecture 17
Announcements Computer Vision Lecture 17 Epipolar Geometry & Stereo Basics Seminar in the summer semester Current Topics in Computer Vision and Machine Learning Block seminar, presentations in 1 st week
More information3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.
3D Computer Vision Structured Light II Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de 1 Introduction
More informationAccurate 3D Face and Body Modeling from a Single Fixed Kinect
Accurate 3D Face and Body Modeling from a Single Fixed Kinect Ruizhe Wang*, Matthias Hernandez*, Jongmoo Choi, Gérard Medioni Computer Vision Lab, IRIS University of Southern California Abstract In this
More informationKinect Cursor Control EEE178 Dr. Fethi Belkhouche Christopher Harris Danny Nguyen I. INTRODUCTION
Kinect Cursor Control EEE178 Dr. Fethi Belkhouche Christopher Harris Danny Nguyen Abstract: An XBOX 360 Kinect is used to develop two applications to control the desktop cursor of a Windows computer. Application
More informationImage Processing Fundamentals. Nicolas Vazquez Principal Software Engineer National Instruments
Image Processing Fundamentals Nicolas Vazquez Principal Software Engineer National Instruments Agenda Objectives and Motivations Enhancing Images Checking for Presence Locating Parts Measuring Features
More informationComplex Sensors: Cameras, Visual Sensing. The Robotics Primer (Ch. 9) ECE 497: Introduction to Mobile Robotics -Visual Sensors
Complex Sensors: Cameras, Visual Sensing The Robotics Primer (Ch. 9) Bring your laptop and robot everyday DO NOT unplug the network cables from the desktop computers or the walls Tuesday s Quiz is on Visual
More informationNEW CONCEPT FOR JOINT DISPARITY ESTIMATION AND SEGMENTATION FOR REAL-TIME VIDEO PROCESSING
NEW CONCEPT FOR JOINT DISPARITY ESTIMATION AND SEGMENTATION FOR REAL-TIME VIDEO PROCESSING Nicole Atzpadin 1, Serap Askar, Peter Kauff, Oliver Schreer Fraunhofer Institut für Nachrichtentechnik, Heinrich-Hertz-Institut,
More informationMinimizing Noise and Bias in 3D DIC. Correlated Solutions, Inc.
Minimizing Noise and Bias in 3D DIC Correlated Solutions, Inc. Overview Overview of Noise and Bias Digital Image Correlation Background/Tracking Function Minimizing Noise Focus Contrast/Lighting Glare
More informationLocal Image preprocessing (cont d)
Local Image preprocessing (cont d) 1 Outline - Edge detectors - Corner detectors - Reading: textbook 5.3.1-5.3.5 and 5.3.10 2 What are edges? Edges correspond to relevant features in the image. An edge
More informationCompression of RADARSAT Data with Block Adaptive Wavelets Abstract: 1. Introduction
Compression of RADARSAT Data with Block Adaptive Wavelets Ian Cumming and Jing Wang Department of Electrical and Computer Engineering The University of British Columbia 2356 Main Mall, Vancouver, BC, Canada
More informationRobotics Programming Laboratory
Chair of Software Engineering Robotics Programming Laboratory Bertrand Meyer Jiwon Shin Lecture 8: Robot Perception Perception http://pascallin.ecs.soton.ac.uk/challenges/voc/databases.html#caltech car
More informationSensor Modalities. Sensor modality: Different modalities:
Sensor Modalities Sensor modality: Sensors which measure same form of energy and process it in similar ways Modality refers to the raw input used by the sensors Different modalities: Sound Pressure Temperature
More informationReal-time Stereo Vision for Urban Traffic Scene Understanding
Proceedings of the IEEE Intelligent Vehicles Symposium 2000 Dearborn (MI), USA October 3-5, 2000 Real-time Stereo Vision for Urban Traffic Scene Understanding U. Franke, A. Joos DaimlerChrylser AG D-70546
More informationDigital Image Forgery Detection Based on GLCM and HOG Features
Digital Image Forgery Detection Based on GLCM and HOG Features Liya Baby 1, Ann Jose 2 Department of Electronics and Communication, Ilahia College of Engineering and Technology, Muvattupuzha, Ernakulam,
More informationStructured Light. Tobias Nöll Thanks to Marc Pollefeys, David Nister and David Lowe
Structured Light Tobias Nöll tobias.noell@dfki.de Thanks to Marc Pollefeys, David Nister and David Lowe Introduction Previous lecture: Dense reconstruction Dense matching of non-feature pixels Patch-based
More informationHand-Eye Calibration from Image Derivatives
Hand-Eye Calibration from Image Derivatives Abstract In this paper it is shown how to perform hand-eye calibration using only the normal flow field and knowledge about the motion of the hand. The proposed
More informationOne category of visual tracking. Computer Science SURJ. Michael Fischer
Computer Science Visual tracking is used in a wide range of applications such as robotics, industrial auto-control systems, traffic monitoring, and manufacturing. This paper describes a new algorithm for
More informationSolution: filter the image, then subsample F 1 F 2. subsample blur subsample. blur
Pyramids Gaussian pre-filtering Solution: filter the image, then subsample blur F 0 subsample blur subsample * F 0 H F 1 F 1 * H F 2 { Gaussian pyramid blur F 0 subsample blur subsample * F 0 H F 1 F 1
More informationDense 3D Reconstruction. Christiano Gava
Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Wide baseline matching (SIFT) Today: dense 3D reconstruction
More informationKeywords: Thresholding, Morphological operations, Image filtering, Adaptive histogram equalization, Ceramic tile.
Volume 3, Issue 7, July 2013 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Blobs and Cracks
More informationOn Board 6D Visual Sensors for Intersection Driving Assistance Systems
On Board 6D Visual Sensors for Intersection Driving Assistance Systems S. Nedevschi, T. Marita, R. Danescu, F. Oniga, S. Bota, I. Haller, C. Pantilie, M. Drulea, C. Golban Sergiu.Nedevschi@cs.utcluj.ro
More informationTo Appear in IEEE International Symposium on Intelligent Vehicles in Parma, Italy, June 2004.
To Appear in IEEE International Symposium on Intelligent Vehicles in Parma, Italy, June 2004. Occupant Posture Analysis using Reflectance and Stereo Images for Smart Airbag Deployment Stephen J. Krotosky
More informationEstimating the wavelength composition of scene illumination from image data is an
Chapter 3 The Principle and Improvement for AWB in DSC 3.1 Introduction Estimating the wavelength composition of scene illumination from image data is an important topics in color engineering. Solutions
More informationMotion Estimation. There are three main types (or applications) of motion estimation:
Members: D91922016 朱威達 R93922010 林聖凱 R93922044 謝俊瑋 Motion Estimation There are three main types (or applications) of motion estimation: Parametric motion (image alignment) The main idea of parametric motion
More information3D Modeling of Objects Using Laser Scanning
1 3D Modeling of Objects Using Laser Scanning D. Jaya Deepu, LPU University, Punjab, India Email: Jaideepudadi@gmail.com Abstract: In the last few decades, constructing accurate three-dimensional models
More informationOverview. Related Work Tensor Voting in 2-D Tensor Voting in 3-D Tensor Voting in N-D Application to Vision Problems Stereo Visual Motion
Overview Related Work Tensor Voting in 2-D Tensor Voting in 3-D Tensor Voting in N-D Application to Vision Problems Stereo Visual Motion Binary-Space-Partitioned Images 3-D Surface Extraction from Medical
More informationPresented at the FIG Congress 2018, May 6-11, 2018 in Istanbul, Turkey
Presented at the FIG Congress 2018, May 6-11, 2018 in Istanbul, Turkey Evangelos MALTEZOS, Charalabos IOANNIDIS, Anastasios DOULAMIS and Nikolaos DOULAMIS Laboratory of Photogrammetry, School of Rural
More informationHuman Motion Detection and Tracking for Video Surveillance
Human Motion Detection and Tracking for Video Surveillance Prithviraj Banerjee and Somnath Sengupta Department of Electronics and Electrical Communication Engineering Indian Institute of Technology, Kharagpur,
More informationImage Processing
Image Processing 159.731 Canny Edge Detection Report Syed Irfanullah, Azeezullah 00297844 Danh Anh Huynh 02136047 1 Canny Edge Detection INTRODUCTION Edges Edges characterize boundaries and are therefore
More information