Evaluation of Stereo Matching Algorithms for Occupant Detection

Size: px
Start display at page:

Download "Evaluation of Stereo Matching Algorithms for Occupant Detection"

Transcription

1 Evaluation of Stereo Matching Algorithms for Occupant Detection Sidharta Gautama TELIN-Universiteit Gent St.Pietersnieuwstr.41 B-9000 Gent (Belgium) Simon Lacroix, Michel Devy LAAS-CNRS 7 avenue Colonel Roche Toulouse Cedex (France) simon@laas.fr, devy@laas.fr Abstract Vision systems within vehicles offer new opportunities in the automobile industry. The detection and classification of passenger and driver seat occupancy open up new ways to improve the safety and comfort of the passengers. In this paper we present results of a stereo system designed for the observation of the cockpit scene in order to provide information about the passenger presence and location within the vehicle to improve the control of the airbag firing. We compare different techniques and examine the effect of random and systematic errors on the performance in precision, robustness and processing speed. These results establish a foundation for on-going work on occupant detection for vehicle safety. 1 Introduction Improvement of the safety and comfort of the vehicle occupant is an important research domain in the automobile industry. New sophisticated functions are being developed which improve and automate a variety of vehicle components (e.g. intelligent cruise control, airbag firing). The development of these functions is related to the introduction of new sensors in the automobile application field (radar, lidar, video sensing devices). These sensors will be able to provide information to decision algorithms which can process the raw data into an interpretable analysis of the observed scene. In this paper we discuss occupant detection using stereo vision, related to safety issues. The massive introduction of airbags in cars has generated new problems: indeed, airbag systems are operating in open loop conditions; when an impact is detected, the airbag is automatically This work is supported by the PREDIT program of the french Ministry of National Education and Research. inflated without any feedback about the nature of the seat occupant. This mode is not without problems and has caused a number of dramatic accidents (e.g. passengers in unusual situations are severely injured by the airbag, babies, installed on a child seat in reversed position, are thrown to the back of the vehicle when the airbag inflates). The determination of the presence of a person or object on the seat, and the evaluation of the position of the passenger within the cockpit can be used to improve the operational conditions of the airbag. All these developments are supported by the improvement of the sensor technology and the cost reduction of the electronic components. In order to analyze the seat occupancy, we developed a 3D vision system based on a pair of stereo cameras. Stereo vision offers the potential to produce detailed results within real-time constraints and is suited for irregular environments. The application of occupant detection demands a level of accuracy in addition to the mere detection of presence. A fine location of the boundaries is desired to accurately position the person within the vehicle. This puts strong constraints on the performance of the 3D vision system: 1. Real-time demands: in order for the system to react quickly enough in the case of an accident, a 3D reconstruction at 10 Hz or higher is required; 2. Dense reconstruction: the large variety of situations (e.g. person with child, person reversed, different objects) puts high demands on the density of the 3D reconstruction to be able to sufficiently characterize the form in later stages; 3. Accurate reconstruction: Of importance is the positioning of the head within the cockpit. The application aims at an accurate positioning within centimeter range. In this paper, we evaluate the performance of different stereo algorithms within this context. Section 2 introduces

2 area-based matching. Matching is a fundamental problem in stereo vision, where the aim is to identify corresponding points in left and right image. We examine two approaches to this problem. In section 3, the performance in precision, robustness and computational complexity for these two approaches is examined. Section 4 outlines the application of the system to head detection. 2 Area-based matching Area-based matching algorithms in stereo vision rely on the comparison of regularly sized regions of pixel values in the two images, in order to find the best match. Matching measures are used to provide a numerical measure of the similarity between a window of pixels in one image and a window in another image. Based on these measures an optimal match is determined. 2.1 Zero mean Normalised Cross Correlation (ZNCC) ZNCC belongs to the class of correlation-based measures. These measures are based on the cross-correlation function R between two X images I 1 and I 2 : R(x y)= I 1(u v)i 2(x + u y + v) u v2w where I 1 (x y) and I 2 (x y) denote the intensity values at pixel (x y). ZNCC deals with a possible gain factor and bias between the two images (radiometric distortion) by resp. dividing by the variances of each correlation window, and by first substracting the mean of the correlation window from each pixel value: ScoreZNCC(x y) = P u v2w (I1(u v) ; I 1):(I2(x + u y + v) ; I 2) q P u v2w (I1(u v) ; 2.2 Census Transform (CT) P : I 1)) 2 :( u v2w (I2(u v) ; I 2)) 2 CT belongs to the class of non-parametric techniques [1] [2], which are based on the relative ordering of pixel intensities within a window, rather than the intensity values themselves. Consequently, these techniques are robust with respect to radiometric distortion, since differences in gain and bias between two images will not affect the ordering of pixels within a window. The CT maps the window surrounding the centre pixel to a bit string. If a particular pixel s value is less than the centre pixel then the corresponding position in the bit string will be set to 1, otherwise it is set to 0. Two census transformed images are compared using a similarity measure based on the Hamming distance, i.e. the number of bits that differ in the two bit strings: ScoreCT (x y) = X u v2w Hamming(I1 0 (u v) I0 2 (x + u y + v)) where I 0 1 and I 0 2 represent the census transform of I 1 and I 2. We have introduced a modification within the census transform to make it more robust to image noise and rectification errors. In correspondence with ZNCC, we perform a census transform by comparing each pixel s value with the mean of the correlation window instead of the intensity of the centre pixel. 3 Experimental evaluation For stereo vision, performance depends on many factors, including the contrast and spatial distribution of texture in the intensity signal, the noise level in the images, artifacts introduced by the matching algorithms, and the true range to the objects in the scene [3]. In our experiments, we characterize both random errors caused by noise and the systematic errors resulting from artifacts. To evaluate the performance of the ZNCC and CT algorithms within the framework of occupant detection, we collected different sets of stereo pairs of a flat board, with and without an object of known geometry, from an optical bench. We used a set cameras (2.6 mm objectives) with a 5 cm separation (stereo baseline). A third order distortion model is used to calibrate the cameras and estimate the epipolar geometry. The range of interest is limited to 2 meter. Sets of 50 stereo pairs were taken under constant illumination and processed at resolution to generate sample statistics (fig. 1). Besides these images, we mounted an experimental system within a vehicle cockpit which is used to evaluate the algorithms within the context of occupant detection. 3.1 Analysis of random errors At subpixel resolution, the precision (variance) of a estimate d reflects statistical fluctuations around the sub-pixel mean. This precision depends on the variance of noise in the image, the image intensity in the image and the matching algorithm used. We are interested in the uncertainty in 3-D measurements computed from. If the estimate d follows a Gaussian distribution (which has been verified by Matthies [3]), it has been shown that the estimate of the range Z can be modelled as Gaussian with mean Z = k= d and variance 2 Z ( Z d )2 = k 2 d 4 :2 d (1)

3 (a) Flat Board (b) Book (a) ZNCC Figure 1. Test Images where k is a calibration constant. Using equation (1), the standard deviation on the range can be plotted out against the range (fig. 2). We have analysed this for a set of 50 images, taken from the same scene (a flat board, fig 1(a)). All images were processed with ZNCC and both CT algorithms using a correlation window of 99 pixels and statistics collected. Plotting the standard deviation on the range, we see that the precision remains within the centimeter range for both ZNCC and CT algorithms. ZNCC provides a slightly more stable result, which is due to the higher smoothing effect established by the correlation window. The modification on the CT algorithm introduces no effect on the precision over the original CT. To allow a qualitative evaluation of the robustness of the different algorithms, we averaged the number of matched pixels found for each image pair over the total dataset. This comparison shows the immediate benefit of the modification made to the original CT. Introducing an averaging process in combination with the census transform stabilises the transform for image noise and errors due to the rectification process (fig. 3). Table 1 summarizes the statistics. Based on these results, we opted to continue with the modified version of the CT algorithm. In the following discussion CT stands for the modified census transform. ZNCC CT CT (modified) % % % Table 1. Matched pixels per image pair, averaged over the total dataset. 100% = total image 3.2 Quality of the estimate The at each pixel is estimated using area-based matching with a correlation window around each pixel. (b) CT (c) CT mod Figure 2. Std.deviation on range estimates Several criteria can be used to filter false matchings: strength, uniqueness and form of the correlation peak (fig. 4), and inverse validition. This last test inverses the role of left and right image and considers as valid only those matches for which the reverse correlation has fallen on the initial point in the left image [4]. Sub-pixel estimates are obtained by fitting a parabole to the three correlation values surrounding the optimum. We asses the quality not only on the correctness of the estimate, but also on the ability to filter out false matchings using the validity criteria. To evaluate the matching algorithms, we examined the estimate at various points within the images

4 (a) (b) (c) Figure 3. Disparity images of a sample: (a) ZNCC, (b) CT, (c) CT modified correlation signal difference range variance strength Figure 4. Thresholds on correlation peak shown in fig 5. The dots show the locations where samples were taken. Some example textures are shown for a correlation window of pixels. For these textures, the matching strength of the texture over a given range is plotted out in fig. 6, comparing the correlation signal for ZNCC and CT. (a) (b) (c) Figure 5. Texture samples validation has been plotted as a grey bar when the inverse test is validated (i.e. the reciprocal match falls within one pixel of the original point). This test proves to be the most reliable. Estimates which fall outside the inverse validity test are false or ambiguous. However, the inverse does not hold and false estimates can still filter through. To be able to use the other measures in conjunction with the inverse validity test, the plot for these measures should show a monotonously increasing curve. For the different points we examined, only the difference measure proves to be useful. The peak value and the variance generally show a fluctuating behaviour. We examine the evolution of the peak value, the difference with the second largest peak, the variance and the inverse validation, which are used by the filter criteria, and this for varying correlation window sizes. Since small correlation windows tend to give an unreliable estimate, this gives a rough idea of which criteria are useful to filter out unwanted responses. Fig. 7 shows the plot of these measures for a given point (texture b, fig. 5). The inverse The difference measure proves to be a useful attribute with CT in contrast with ZNCC. To show this, we classified the estimates in our sample points as reliable or unreliable. This was done by regarding each estimate within the observed scene and see if the estimate was consistent with the scene or not. Only the points which passed the inverse validity test are regarded. ZNCC and CT give the same consistency evaluation for each point (i.e. both techniques give approximately the same estimate). However using the difference measure, CT is

5 (a) Texture a (a) ZNCC (b) Texture b (c) Texture c Figure 6. Correlation signals: (right) CT (left) ZNCC, (b) CT Figure 7. Evolution of validity measures able to filter out a large part of the inconsistent points (cfr. fig. 8). With CT, the difference measures generally shows a high respons ( unique peak) with the good estimates and a low respons ( ambiguous peak) with the unreliable estimates. ZNCC on the other hand tends to give a strong respons even in the case of unreliable estimates. 3.3 Analysis of systematic errors Systematic errors are due to processing artifacts. One of these artifacts is the so-called object growing effect. This growing effect is a bias introduced by the finite size of the correlation window. When the correlation window encounters an object edge, the mean calculated within this window (as does ZNCC) reflects both the high and low contrast texture. This causes a confusion in the matching measure and results in a bias towards high contrast. A halo is seen around the object within which the range estimates are nearly the same as the range to the object instead of the range to the background. The size of the halo is half the size of the correlation window. In 3D, this size grows with the distance to the cameras. The problem has been investigated by a number of people (cfr. [5]), but the proposed solutions are not always well adapted to real-time systems. To characterize the object growing effect, we take the 3D reconstruction of a book (fig. 1(b)) and transform the object into a frame following the principal axes of the book. After this transformation, the 3D points are thresholded by keeping all the points above a certain height off the board plane. In this way we are able to segment the points belonging to the object. Using this segmentation, we take the histogram of the 3D points along the horizontal axis. This allows us to estimate the absolute horizontal precision of the system. The book has a width of 22 cm. Both ZNCC and CT give an overestimation of the width, but CT proves to be more accurate (i.e cm compared to 24.9 cm, fig. 9(a)). CT avoids the object growing effect not by a better estimate but by a more efficient filtering, as

6 (a) ZNCC (a) Positional accuracy (b) CT Figure 8. Difference and peak values shown in the previous section. This answer of unknown is much prefered over erroneous estimations. The size of the correlation window also influences the directional accuracy. Increasing the correlation window smooths out the normal directions of the 3D points. This is shown in fig. 9(b). The normal direction is computed in each 3D point using ~n = ~x u ~xv j ~xu ~xvj (2) where ~xu and ~xv are the partial derivatives calculated by differencing in a 4-neighbourhood. Taking the histogram of the normal direction along the range axis, which should be a delta peak in the ideal case (= perfectly flat surface). We note a better performance by the CT algorithm. 3.4 Performance within cockpit model A prototype system, consisting of a pair of stereo cameras combined with infrared illumination, is being developed for installation within a cockpit model. The (b) Directional accuracy Figure 9. Accuracy of range estimates images produced by the cameras are subsampled to a size of pixels. Table 2 shows the timing results for the matching and reconstruction process, performed on an SUN Ultra Sparc 10. The range which is scanned for each pixel is 1 of the image width. The correlation window size is 4 taken pixels and 5 5 pixels for resp. resolution and ZNCC CT 3D Rec ms 550 ms 20 ms ms 100 ms 10 ms Table 2. Timing results on an Ultra Sparc 10 Figure 11 shows an extract of the reconstructed scene. It illustrates the performance of CT compared to ZNCC with respect to the characterisation of form and the object growing effect.

7 4 Head detection using data fusion Although the prototype system s main goal is the classification of the seat occupancy (e.g. child seat, object, passenger), the results obtained with the system allow the exploration of head detection inside the vehicle. Within the context of passenger safety, the head plays a central role. Locating the head with respect to the dashboard is an important issue and fast, robust techniques need to be developed. We are currently exploring a system which fuses 3D and intensity information. Since the vehicle cockpit offers a well-defined environment, a narrow region of interest (ROI) can be defined in which head detection based on skin tone is performed. To eliminate false alarms, we envisage a curvature estimation on the detected points. For this phase, we rely on the better results of CT over ZNCC in precision which makes a finer characterisation of form possible. The flow chart of the system is outlined in fig. 10. Skin tone Left Camera rectification 5 Conclusions Disparity estimation 3D reconstruction ROI Data fusion Right Camera rectification Figure 10. Flow chart Shape In this paper, we compared ZNCC and CT matching algorithms for stereo vision. The data indicates precision within centimeter range for a range up to 2 metres and processing times of 1-10 Hz on our system. CT performs better in terms of accuracy and speed. We examined different validity tests to filter out false matches. The inverse validity test proved to be the most efficient, but does not filter out all unreliable matches. While examining the signal (peak, difference and variance) offers no further improvement of ZNCÇ the difference test shows to be reliable to further suppress unwanted matches in the case of CT. More particularly, the effect of object growing is handled much better in this case. CT can avoid the object growing effect not by a better estimate but by a more efficient filtering. The respons of unknown is much prefered over erroneous estimations. Compared with other solutions [5], CT requires no extra processing time and is well suited for real-time applications. A strong interest of car manufacturers exists in the concept of occupant detection using video sensors. Smart airbag concepts with partial inflation capabilities should take into account information provided by such a perception system. Our prototype system shows encouraging results for the problem of head detection within the vehicle cockpit. Future work is planned to further exploit the results of stereo using CT in order to accurately position the passengers head with respect to the airbag. References [1] R.Zabih, J.Woodfill, Non-parametric Local Transforms for Computing Visual Correspondence, Third European Conference on Computer Vision, Stockholm, Sweden, [2] J.Banks, M.Bennamoun, P.Corke, Fast and robust stereo matching algorithms for mining automation, Proc. of the Int. Workshop on Image Analysis and Information Fusion, Adelaide, Australia, [3] L.Matthies, P.Grandjean, Stochastic Performance Modeling and Evaluation of Obstacle Detectability with Imaging Range Sensors, IEEE Trans. Robotics and Automation, Vol.10, No.6, pp , [4] P.Lassere, P.Grandjean, Stereo Vision Improvement, 7th Int. Conf. on Advanced Robotics, Guixols, Spain, [5] L.Robert, R.Deriche, Dense Depth Map Reconstruction: A Minimization and Regularization Approach which Preserves Discontinuities, Fourth European Conference on Computer Vision, Cambridge, UK, 1996.

8 (a) Disparity map ZNCC (b) Correlated pixels ZNCC (c) Disparity map CT (d) Correlated pixels CT (e) ZNCC (f) CT Figure 11. Reconstruction examples

Transactions on Information and Communications Technologies vol 16, 1996 WIT Press, ISSN

Transactions on Information and Communications Technologies vol 16, 1996 WIT Press,   ISSN ransactions on Information and Communications echnologies vol 6, 996 WI Press, www.witpress.com, ISSN 743-357 Obstacle detection using stereo without correspondence L. X. Zhou & W. K. Gu Institute of Information

More information

Spatio-Temporal Stereo Disparity Integration

Spatio-Temporal Stereo Disparity Integration Spatio-Temporal Stereo Disparity Integration Sandino Morales and Reinhard Klette The.enpeda.. Project, The University of Auckland Tamaki Innovation Campus, Auckland, New Zealand pmor085@aucklanduni.ac.nz

More information

Mobile Human Detection Systems based on Sliding Windows Approach-A Review

Mobile Human Detection Systems based on Sliding Windows Approach-A Review Mobile Human Detection Systems based on Sliding Windows Approach-A Review Seminar: Mobile Human detection systems Njieutcheu Tassi cedrique Rovile Department of Computer Engineering University of Heidelberg

More information

Multiple View Geometry

Multiple View Geometry Multiple View Geometry CS 6320, Spring 2013 Guest Lecture Marcel Prastawa adapted from Pollefeys, Shah, and Zisserman Single view computer vision Projective actions of cameras Camera callibration Photometric

More information

Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision

Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision What Happened Last Time? Human 3D perception (3D cinema) Computational stereo Intuitive explanation of what is meant by disparity Stereo matching

More information

Tyzx DeepSea High Speed Stereo Vision System

Tyzx DeepSea High Speed Stereo Vision System Tyzx DeepSea High Speed Stereo Vision System John Iselin Woodfill, Gaile Gordon, Ron Buck Tyzx, Inc. 3885 Bohannon Drive Menlo Park, CA 9425 {Woodfill, Gaile, Ron}@tyzx.com Abstract This paper describes

More information

2 OVERVIEW OF RELATED WORK

2 OVERVIEW OF RELATED WORK Utsushi SAKAI Jun OGATA This paper presents a pedestrian detection system based on the fusion of sensors for LIDAR and convolutional neural network based image classification. By using LIDAR our method

More information

Fundamentals of Digital Image Processing

Fundamentals of Digital Image Processing \L\.6 Gw.i Fundamentals of Digital Image Processing A Practical Approach with Examples in Matlab Chris Solomon School of Physical Sciences, University of Kent, Canterbury, UK Toby Breckon School of Engineering,

More information

Calibration of a rotating multi-beam Lidar

Calibration of a rotating multi-beam Lidar The 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems October 18-22, 2010, Taipei, Taiwan Calibration of a rotating multi-beam Lidar Naveed Muhammad 1,2 and Simon Lacroix 1,2 Abstract

More information

EXAM SOLUTIONS. Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006,

EXAM SOLUTIONS. Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006, School of Computer Science and Communication, KTH Danica Kragic EXAM SOLUTIONS Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006, 14.00 19.00 Grade table 0-25 U 26-35 3 36-45

More information

Data Term. Michael Bleyer LVA Stereo Vision

Data Term. Michael Bleyer LVA Stereo Vision Data Term Michael Bleyer LVA Stereo Vision What happened last time? We have looked at our energy function: E ( D) = m( p, dp) + p I < p, q > N s( p, q) We have learned about an optimization algorithm that

More information

Chapter 3 Image Registration. Chapter 3 Image Registration

Chapter 3 Image Registration. Chapter 3 Image Registration Chapter 3 Image Registration Distributed Algorithms for Introduction (1) Definition: Image Registration Input: 2 images of the same scene but taken from different perspectives Goal: Identify transformation

More information

Adaptative Elimination of False Edges for First Order Detectors

Adaptative Elimination of False Edges for First Order Detectors Adaptative Elimination of False Edges for First Order Detectors Djemel ZIOU and Salvatore TABBONE D~partement de math~matiques et d'informatique, universit~ de Sherbrooke, Qc, Canada, J1K 2R1 Crin/Cnrs

More information

EE795: Computer Vision and Intelligent Systems

EE795: Computer Vision and Intelligent Systems EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 14 130307 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Stereo Dense Motion Estimation Translational

More information

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing Visual servoing vision allows a robotic system to obtain geometrical and qualitative information on the surrounding environment high level control motion planning (look-and-move visual grasping) low level

More information

Depth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth

Depth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth Common Classification Tasks Recognition of individual objects/faces Analyze object-specific features (e.g., key points) Train with images from different viewing angles Recognition of object classes Analyze

More information

Micro-scale Stereo Photogrammetry of Skin Lesions for Depth and Colour Classification

Micro-scale Stereo Photogrammetry of Skin Lesions for Depth and Colour Classification Micro-scale Stereo Photogrammetry of Skin Lesions for Depth and Colour Classification Tim Lukins Institute of Perception, Action and Behaviour 1 Introduction The classification of melanoma has traditionally

More information

Laserscanner Based Cooperative Pre-Data-Fusion

Laserscanner Based Cooperative Pre-Data-Fusion Laserscanner Based Cooperative Pre-Data-Fusion 63 Laserscanner Based Cooperative Pre-Data-Fusion F. Ahlers, Ch. Stimming, Ibeo Automobile Sensor GmbH Abstract The Cooperative Pre-Data-Fusion is a novel

More information

Towards the completion of assignment 1

Towards the completion of assignment 1 Towards the completion of assignment 1 What to do for calibration What to do for point matching What to do for tracking What to do for GUI COMPSCI 773 Feature Point Detection Why study feature point detection?

More information

Depth Measurement and 3-D Reconstruction of Multilayered Surfaces by Binocular Stereo Vision with Parallel Axis Symmetry Using Fuzzy

Depth Measurement and 3-D Reconstruction of Multilayered Surfaces by Binocular Stereo Vision with Parallel Axis Symmetry Using Fuzzy Depth Measurement and 3-D Reconstruction of Multilayered Surfaces by Binocular Stereo Vision with Parallel Axis Symmetry Using Fuzzy Sharjeel Anwar, Dr. Shoaib, Taosif Iqbal, Mohammad Saqib Mansoor, Zubair

More information

3D DEFORMATION MEASUREMENT USING STEREO- CORRELATION APPLIED TO EXPERIMENTAL MECHANICS

3D DEFORMATION MEASUREMENT USING STEREO- CORRELATION APPLIED TO EXPERIMENTAL MECHANICS 3D DEFORMATION MEASUREMENT USING STEREO- CORRELATION APPLIED TO EXPERIMENTAL MECHANICS Dorian Garcia, Jean-José Orteu École des Mines d Albi, F-81013 ALBI CT Cedex 09, France Dorian.Garcia@enstimac.fr,

More information

Accurate and Dense Wide-Baseline Stereo Matching Using SW-POC

Accurate and Dense Wide-Baseline Stereo Matching Using SW-POC Accurate and Dense Wide-Baseline Stereo Matching Using SW-POC Shuji Sakai, Koichi Ito, Takafumi Aoki Graduate School of Information Sciences, Tohoku University, Sendai, 980 8579, Japan Email: sakai@aoki.ecei.tohoku.ac.jp

More information

Auto-focusing Technique in a Projector-Camera System

Auto-focusing Technique in a Projector-Camera System 2008 10th Intl. Conf. on Control, Automation, Robotics and Vision Hanoi, Vietnam, 17 20 December 2008 Auto-focusing Technique in a Projector-Camera System Lam Bui Quang, Daesik Kim and Sukhan Lee School

More information

A Statistical Consistency Check for the Space Carving Algorithm.

A Statistical Consistency Check for the Space Carving Algorithm. A Statistical Consistency Check for the Space Carving Algorithm. A. Broadhurst and R. Cipolla Dept. of Engineering, Univ. of Cambridge, Cambridge, CB2 1PZ aeb29 cipolla @eng.cam.ac.uk Abstract This paper

More information

Effects Of Shadow On Canny Edge Detection through a camera

Effects Of Shadow On Canny Edge Detection through a camera 1523 Effects Of Shadow On Canny Edge Detection through a camera Srajit Mehrotra Shadow causes errors in computer vision as it is difficult to detect objects that are under the influence of shadows. Shadow

More information

A threshold decision of the object image by using the smart tag

A threshold decision of the object image by using the smart tag A threshold decision of the object image by using the smart tag Chang-Jun Im, Jin-Young Kim, Kwan Young Joung, Ho-Gil Lee Sensing & Perception Research Group Korea Institute of Industrial Technology (

More information

Stereo Vision. MAN-522 Computer Vision

Stereo Vision. MAN-522 Computer Vision Stereo Vision MAN-522 Computer Vision What is the goal of stereo vision? The recovery of the 3D structure of a scene using two or more images of the 3D scene, each acquired from a different viewpoint in

More information

Learning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009

Learning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009 Learning and Inferring Depth from Monocular Images Jiyan Pan April 1, 2009 Traditional ways of inferring depth Binocular disparity Structure from motion Defocus Given a single monocular image, how to infer

More information

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching Stereo Matching Fundamental matrix Let p be a point in left image, p in right image l l Epipolar relation p maps to epipolar line l p maps to epipolar line l p p Epipolar mapping described by a 3x3 matrix

More information

Measurement of Pedestrian Groups Using Subtraction Stereo

Measurement of Pedestrian Groups Using Subtraction Stereo Measurement of Pedestrian Groups Using Subtraction Stereo Kenji Terabayashi, Yuki Hashimoto, and Kazunori Umeda Chuo University / CREST, JST, 1-13-27 Kasuga, Bunkyo-ku, Tokyo 112-8551, Japan terabayashi@mech.chuo-u.ac.jp

More information

ФУНДАМЕНТАЛЬНЫЕ НАУКИ. Информатика 9 ИНФОРМАТИКА MOTION DETECTION IN VIDEO STREAM BASED ON BACKGROUND SUBTRACTION AND TARGET TRACKING

ФУНДАМЕНТАЛЬНЫЕ НАУКИ. Информатика 9 ИНФОРМАТИКА MOTION DETECTION IN VIDEO STREAM BASED ON BACKGROUND SUBTRACTION AND TARGET TRACKING ФУНДАМЕНТАЛЬНЫЕ НАУКИ Информатика 9 ИНФОРМАТИКА UDC 6813 OTION DETECTION IN VIDEO STREA BASED ON BACKGROUND SUBTRACTION AND TARGET TRACKING R BOGUSH, S ALTSEV, N BROVKO, E IHAILOV (Polotsk State University

More information

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation Obviously, this is a very slow process and not suitable for dynamic scenes. To speed things up, we can use a laser that projects a vertical line of light onto the scene. This laser rotates around its vertical

More information

Depth Sensors Kinect V2 A. Fornaser

Depth Sensors Kinect V2 A. Fornaser Depth Sensors Kinect V2 A. Fornaser alberto.fornaser@unitn.it Vision Depth data It is not a 3D data, It is a map of distances Not a 3D, not a 2D it is a 2.5D or Perspective 3D Complete 3D - Tomography

More information

Detecting motion by means of 2D and 3D information

Detecting motion by means of 2D and 3D information Detecting motion by means of 2D and 3D information Federico Tombari Stefano Mattoccia Luigi Di Stefano Fabio Tonelli Department of Electronics Computer Science and Systems (DEIS) Viale Risorgimento 2,

More information

Pedestrian Detection Using Correlated Lidar and Image Data EECS442 Final Project Fall 2016

Pedestrian Detection Using Correlated Lidar and Image Data EECS442 Final Project Fall 2016 edestrian Detection Using Correlated Lidar and Image Data EECS442 Final roject Fall 2016 Samuel Rohrer University of Michigan rohrer@umich.edu Ian Lin University of Michigan tiannis@umich.edu Abstract

More information

A New Approach to Computation of Curvature Scale Space Image for Shape Similarity Retrieval

A New Approach to Computation of Curvature Scale Space Image for Shape Similarity Retrieval A New Approach to Computation of Curvature Scale Space Image for Shape Similarity Retrieval Farzin Mokhtarian, Sadegh Abbasi and Josef Kittler Centre for Vision Speech and Signal Processing Department

More information

SuRVoS Workbench. Super-Region Volume Segmentation. Imanol Luengo

SuRVoS Workbench. Super-Region Volume Segmentation. Imanol Luengo SuRVoS Workbench Super-Region Volume Segmentation Imanol Luengo Index - The project - What is SuRVoS - SuRVoS Overview - What can it do - Overview of the internals - Current state & Limitations - Future

More information

Flexible Calibration of a Portable Structured Light System through Surface Plane

Flexible Calibration of a Portable Structured Light System through Surface Plane Vol. 34, No. 11 ACTA AUTOMATICA SINICA November, 2008 Flexible Calibration of a Portable Structured Light System through Surface Plane GAO Wei 1 WANG Liang 1 HU Zhan-Yi 1 Abstract For a portable structured

More information

Dense 3D Reconstruction. Christiano Gava

Dense 3D Reconstruction. Christiano Gava Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Today: dense 3D reconstruction The matching problem

More information

Epipolar geometry-based ego-localization using an in-vehicle monocular camera

Epipolar geometry-based ego-localization using an in-vehicle monocular camera Epipolar geometry-based ego-localization using an in-vehicle monocular camera Haruya Kyutoku 1, Yasutomo Kawanishi 1, Daisuke Deguchi 1, Ichiro Ide 1, Hiroshi Murase 1 1 : Nagoya University, Japan E-mail:

More information

Using temporal seeding to constrain the disparity search range in stereo matching

Using temporal seeding to constrain the disparity search range in stereo matching Using temporal seeding to constrain the disparity search range in stereo matching Thulani Ndhlovu Mobile Intelligent Autonomous Systems CSIR South Africa Email: tndhlovu@csir.co.za Fred Nicolls Department

More information

Chapter 9 Object Tracking an Overview

Chapter 9 Object Tracking an Overview Chapter 9 Object Tracking an Overview The output of the background subtraction algorithm, described in the previous chapter, is a classification (segmentation) of pixels into foreground pixels (those belonging

More information

CS4495/6495 Introduction to Computer Vision

CS4495/6495 Introduction to Computer Vision CS4495/6495 Introduction to Computer Vision 9C-L1 3D perception Some slides by Kelsey Hawkins Motivation Why do animals, people & robots need vision? To detect and recognize objects/landmarks Is that a

More information

On-line and Off-line 3D Reconstruction for Crisis Management Applications

On-line and Off-line 3D Reconstruction for Crisis Management Applications On-line and Off-line 3D Reconstruction for Crisis Management Applications Geert De Cubber Royal Military Academy, Department of Mechanical Engineering (MSTA) Av. de la Renaissance 30, 1000 Brussels geert.de.cubber@rma.ac.be

More information

Fast, Unconstrained Camera Motion Estimation from Stereo without Tracking and Robust Statistics

Fast, Unconstrained Camera Motion Estimation from Stereo without Tracking and Robust Statistics Fast, Unconstrained Camera Motion Estimation from Stereo without Tracking and Robust Statistics Heiko Hirschmüller, Peter R. Innocent and Jon M. Garibaldi Centre for Computational Intelligence, De Montfort

More information

then assume that we are given the image of one of these textures captured by a camera at a different (longer) distance and with unknown direction of i

then assume that we are given the image of one of these textures captured by a camera at a different (longer) distance and with unknown direction of i Image Texture Prediction using Colour Photometric Stereo Xavier Lladó 1, Joan Mart 1, and Maria Petrou 2 1 Institute of Informatics and Applications, University of Girona, 1771, Girona, Spain fllado,joanmg@eia.udg.es

More information

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Structured Light II Johannes Köhler Johannes.koehler@dfki.de Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Introduction Previous lecture: Structured Light I Active Scanning Camera/emitter

More information

Model-Based Stereo. Chapter Motivation. The modeling system described in Chapter 5 allows the user to create a basic model of a

Model-Based Stereo. Chapter Motivation. The modeling system described in Chapter 5 allows the user to create a basic model of a 96 Chapter 7 Model-Based Stereo 7.1 Motivation The modeling system described in Chapter 5 allows the user to create a basic model of a scene, but in general the scene will have additional geometric detail

More information

Towards Autonomous Vehicle. What is an autonomous vehicle? Vehicle driving on its own with zero mistakes How? Using sensors

Towards Autonomous Vehicle. What is an autonomous vehicle? Vehicle driving on its own with zero mistakes How? Using sensors 7 May 2017 Disclaimer Towards Autonomous Vehicle What is an autonomous vehicle? Vehicle driving on its own with zero mistakes How? Using sensors Why Vision Sensors? Humans use both eyes as main sense

More information

3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller

3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller 3D Computer Vision Depth Cameras Prof. Didier Stricker Oliver Wasenmüller Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de

More information

Centre for Digital Image Measurement and Analysis, School of Engineering, City University, Northampton Square, London, ECIV OHB

Centre for Digital Image Measurement and Analysis, School of Engineering, City University, Northampton Square, London, ECIV OHB HIGH ACCURACY 3-D MEASUREMENT USING MULTIPLE CAMERA VIEWS T.A. Clarke, T.J. Ellis, & S. Robson. High accuracy measurement of industrially produced objects is becoming increasingly important. The techniques

More information

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION Mr.V.SRINIVASA RAO 1 Prof.A.SATYA KALYAN 2 DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING PRASAD V POTLURI SIDDHARTHA

More information

A virtual tour of free viewpoint rendering

A virtual tour of free viewpoint rendering A virtual tour of free viewpoint rendering Cédric Verleysen ICTEAM institute, Université catholique de Louvain, Belgium cedric.verleysen@uclouvain.be Organization of the presentation Context Acquisition

More information

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science. Professor William Hoff Dept of Electrical Engineering &Computer Science http://inside.mines.edu/~whoff/ 1 Stereo Vision 2 Inferring 3D from 2D Model based pose estimation single (calibrated) camera > Can

More information

Revising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History

Revising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History Revising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History Simon Thompson and Satoshi Kagami Digital Human Research Center National Institute of Advanced

More information

Outline. ETN-FPI Training School on Plenoptic Sensing

Outline. ETN-FPI Training School on Plenoptic Sensing Outline Introduction Part I: Basics of Mathematical Optimization Linear Least Squares Nonlinear Optimization Part II: Basics of Computer Vision Camera Model Multi-Camera Model Multi-Camera Calibration

More information

parco area delle Scienze, 181A via Ferrata, , Parma 27100, Pavia

parco area delle Scienze, 181A via Ferrata, , Parma 27100, Pavia Proceedings of the IEEE Intelligent Vehicles Symposium 2000 Dearbon (MI), USA October 3-5, 2000 Stereo Vision-based Vehicle Detection M. Bertozzi 1 A. Broggi 2 A. Fascioli 1 S. Nichele 2 1 Dipartimento

More information

CHAPTER 3 DISPARITY AND DEPTH MAP COMPUTATION

CHAPTER 3 DISPARITY AND DEPTH MAP COMPUTATION CHAPTER 3 DISPARITY AND DEPTH MAP COMPUTATION In this chapter we will discuss the process of disparity computation. It plays an important role in our caricature system because all 3D coordinates of nodes

More information

Robust color segmentation algorithms in illumination variation conditions

Robust color segmentation algorithms in illumination variation conditions 286 CHINESE OPTICS LETTERS / Vol. 8, No. / March 10, 2010 Robust color segmentation algorithms in illumination variation conditions Jinhui Lan ( ) and Kai Shen ( Department of Measurement and Control Technologies,

More information

CS 4495/7495 Computer Vision Frank Dellaert, Fall 07. Dense Stereo Some Slides by Forsyth & Ponce, Jim Rehg, Sing Bing Kang

CS 4495/7495 Computer Vision Frank Dellaert, Fall 07. Dense Stereo Some Slides by Forsyth & Ponce, Jim Rehg, Sing Bing Kang CS 4495/7495 Computer Vision Frank Dellaert, Fall 07 Dense Stereo Some Slides by Forsyth & Ponce, Jim Rehg, Sing Bing Kang Etymology Stereo comes from the Greek word for solid (στερεο), and the term can

More information

Digital Image Processing COSC 6380/4393

Digital Image Processing COSC 6380/4393 Digital Image Processing COSC 6380/4393 Lecture 21 Nov 16 th, 2017 Pranav Mantini Ack: Shah. M Image Processing Geometric Transformation Point Operations Filtering (spatial, Frequency) Input Restoration/

More information

Lecture 6: Edge Detection

Lecture 6: Edge Detection #1 Lecture 6: Edge Detection Saad J Bedros sbedros@umn.edu Review From Last Lecture Options for Image Representation Introduced the concept of different representation or transformation Fourier Transform

More information

Edge detection. Goal: Identify sudden. an image. Ideal: artist s line drawing. object-level knowledge)

Edge detection. Goal: Identify sudden. an image. Ideal: artist s line drawing. object-level knowledge) Edge detection Goal: Identify sudden changes (discontinuities) in an image Intuitively, most semantic and shape information from the image can be encoded in the edges More compact than pixels Ideal: artist

More information

Recognition of Object Contours from Stereo Images: an Edge Combination Approach

Recognition of Object Contours from Stereo Images: an Edge Combination Approach Recognition of Object Contours from Stereo Images: an Edge Combination Approach Margrit Gelautz and Danijela Markovic Institute for Software Technology and Interactive Systems, Vienna University of Technology

More information

Embedded real-time stereo estimation via Semi-Global Matching on the GPU

Embedded real-time stereo estimation via Semi-Global Matching on the GPU Embedded real-time stereo estimation via Semi-Global Matching on the GPU Daniel Hernández Juárez, Alejandro Chacón, Antonio Espinosa, David Vázquez, Juan Carlos Moure and Antonio M. López Computer Architecture

More information

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching Stereo Matching Fundamental matrix Let p be a point in left image, p in right image l l Epipolar relation p maps to epipolar line l p maps to epipolar line l p p Epipolar mapping described by a 3x3 matrix

More information

Vehicle Dimensions Estimation Scheme Using AAM on Stereoscopic Video

Vehicle Dimensions Estimation Scheme Using AAM on Stereoscopic Video Workshop on Vehicle Retrieval in Surveillance (VRS) in conjunction with 2013 10th IEEE International Conference on Advanced Video and Signal Based Surveillance Vehicle Dimensions Estimation Scheme Using

More information

DISTANCE MEASUREMENT USING STEREO VISION

DISTANCE MEASUREMENT USING STEREO VISION DISTANCE MEASUREMENT USING STEREO VISION Sheetal Nagar 1, Jitendra Verma 2 1 Department of Electronics and Communication Engineering, IIMT, Greater Noida (India) 2 Department of computer science Engineering,

More information

Stereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman

Stereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman Stereo 11/02/2012 CS129, Brown James Hays Slides by Kristen Grauman Multiple views Multi-view geometry, matching, invariant features, stereo vision Lowe Hartley and Zisserman Why multiple views? Structure

More information

Computer Vision Lecture 17

Computer Vision Lecture 17 Computer Vision Lecture 17 Epipolar Geometry & Stereo Basics 13.01.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Announcements Seminar in the summer semester

More information

Proceedings of the 6th Int. Conf. on Computer Analysis of Images and Patterns. Direct Obstacle Detection and Motion. from Spatio-Temporal Derivatives

Proceedings of the 6th Int. Conf. on Computer Analysis of Images and Patterns. Direct Obstacle Detection and Motion. from Spatio-Temporal Derivatives Proceedings of the 6th Int. Conf. on Computer Analysis of Images and Patterns CAIP'95, pp. 874-879, Prague, Czech Republic, Sep 1995 Direct Obstacle Detection and Motion from Spatio-Temporal Derivatives

More information

Computer Vision Lecture 17

Computer Vision Lecture 17 Announcements Computer Vision Lecture 17 Epipolar Geometry & Stereo Basics Seminar in the summer semester Current Topics in Computer Vision and Machine Learning Block seminar, presentations in 1 st week

More information

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University. 3D Computer Vision Structured Light II Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de 1 Introduction

More information

Accurate 3D Face and Body Modeling from a Single Fixed Kinect

Accurate 3D Face and Body Modeling from a Single Fixed Kinect Accurate 3D Face and Body Modeling from a Single Fixed Kinect Ruizhe Wang*, Matthias Hernandez*, Jongmoo Choi, Gérard Medioni Computer Vision Lab, IRIS University of Southern California Abstract In this

More information

Kinect Cursor Control EEE178 Dr. Fethi Belkhouche Christopher Harris Danny Nguyen I. INTRODUCTION

Kinect Cursor Control EEE178 Dr. Fethi Belkhouche Christopher Harris Danny Nguyen I. INTRODUCTION Kinect Cursor Control EEE178 Dr. Fethi Belkhouche Christopher Harris Danny Nguyen Abstract: An XBOX 360 Kinect is used to develop two applications to control the desktop cursor of a Windows computer. Application

More information

Image Processing Fundamentals. Nicolas Vazquez Principal Software Engineer National Instruments

Image Processing Fundamentals. Nicolas Vazquez Principal Software Engineer National Instruments Image Processing Fundamentals Nicolas Vazquez Principal Software Engineer National Instruments Agenda Objectives and Motivations Enhancing Images Checking for Presence Locating Parts Measuring Features

More information

Complex Sensors: Cameras, Visual Sensing. The Robotics Primer (Ch. 9) ECE 497: Introduction to Mobile Robotics -Visual Sensors

Complex Sensors: Cameras, Visual Sensing. The Robotics Primer (Ch. 9) ECE 497: Introduction to Mobile Robotics -Visual Sensors Complex Sensors: Cameras, Visual Sensing The Robotics Primer (Ch. 9) Bring your laptop and robot everyday DO NOT unplug the network cables from the desktop computers or the walls Tuesday s Quiz is on Visual

More information

NEW CONCEPT FOR JOINT DISPARITY ESTIMATION AND SEGMENTATION FOR REAL-TIME VIDEO PROCESSING

NEW CONCEPT FOR JOINT DISPARITY ESTIMATION AND SEGMENTATION FOR REAL-TIME VIDEO PROCESSING NEW CONCEPT FOR JOINT DISPARITY ESTIMATION AND SEGMENTATION FOR REAL-TIME VIDEO PROCESSING Nicole Atzpadin 1, Serap Askar, Peter Kauff, Oliver Schreer Fraunhofer Institut für Nachrichtentechnik, Heinrich-Hertz-Institut,

More information

Minimizing Noise and Bias in 3D DIC. Correlated Solutions, Inc.

Minimizing Noise and Bias in 3D DIC. Correlated Solutions, Inc. Minimizing Noise and Bias in 3D DIC Correlated Solutions, Inc. Overview Overview of Noise and Bias Digital Image Correlation Background/Tracking Function Minimizing Noise Focus Contrast/Lighting Glare

More information

Local Image preprocessing (cont d)

Local Image preprocessing (cont d) Local Image preprocessing (cont d) 1 Outline - Edge detectors - Corner detectors - Reading: textbook 5.3.1-5.3.5 and 5.3.10 2 What are edges? Edges correspond to relevant features in the image. An edge

More information

Compression of RADARSAT Data with Block Adaptive Wavelets Abstract: 1. Introduction

Compression of RADARSAT Data with Block Adaptive Wavelets Abstract: 1. Introduction Compression of RADARSAT Data with Block Adaptive Wavelets Ian Cumming and Jing Wang Department of Electrical and Computer Engineering The University of British Columbia 2356 Main Mall, Vancouver, BC, Canada

More information

Robotics Programming Laboratory

Robotics Programming Laboratory Chair of Software Engineering Robotics Programming Laboratory Bertrand Meyer Jiwon Shin Lecture 8: Robot Perception Perception http://pascallin.ecs.soton.ac.uk/challenges/voc/databases.html#caltech car

More information

Sensor Modalities. Sensor modality: Different modalities:

Sensor Modalities. Sensor modality: Different modalities: Sensor Modalities Sensor modality: Sensors which measure same form of energy and process it in similar ways Modality refers to the raw input used by the sensors Different modalities: Sound Pressure Temperature

More information

Real-time Stereo Vision for Urban Traffic Scene Understanding

Real-time Stereo Vision for Urban Traffic Scene Understanding Proceedings of the IEEE Intelligent Vehicles Symposium 2000 Dearborn (MI), USA October 3-5, 2000 Real-time Stereo Vision for Urban Traffic Scene Understanding U. Franke, A. Joos DaimlerChrylser AG D-70546

More information

Digital Image Forgery Detection Based on GLCM and HOG Features

Digital Image Forgery Detection Based on GLCM and HOG Features Digital Image Forgery Detection Based on GLCM and HOG Features Liya Baby 1, Ann Jose 2 Department of Electronics and Communication, Ilahia College of Engineering and Technology, Muvattupuzha, Ernakulam,

More information

Structured Light. Tobias Nöll Thanks to Marc Pollefeys, David Nister and David Lowe

Structured Light. Tobias Nöll Thanks to Marc Pollefeys, David Nister and David Lowe Structured Light Tobias Nöll tobias.noell@dfki.de Thanks to Marc Pollefeys, David Nister and David Lowe Introduction Previous lecture: Dense reconstruction Dense matching of non-feature pixels Patch-based

More information

Hand-Eye Calibration from Image Derivatives

Hand-Eye Calibration from Image Derivatives Hand-Eye Calibration from Image Derivatives Abstract In this paper it is shown how to perform hand-eye calibration using only the normal flow field and knowledge about the motion of the hand. The proposed

More information

One category of visual tracking. Computer Science SURJ. Michael Fischer

One category of visual tracking. Computer Science SURJ. Michael Fischer Computer Science Visual tracking is used in a wide range of applications such as robotics, industrial auto-control systems, traffic monitoring, and manufacturing. This paper describes a new algorithm for

More information

Solution: filter the image, then subsample F 1 F 2. subsample blur subsample. blur

Solution: filter the image, then subsample F 1 F 2. subsample blur subsample. blur Pyramids Gaussian pre-filtering Solution: filter the image, then subsample blur F 0 subsample blur subsample * F 0 H F 1 F 1 * H F 2 { Gaussian pyramid blur F 0 subsample blur subsample * F 0 H F 1 F 1

More information

Dense 3D Reconstruction. Christiano Gava

Dense 3D Reconstruction. Christiano Gava Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Wide baseline matching (SIFT) Today: dense 3D reconstruction

More information

Keywords: Thresholding, Morphological operations, Image filtering, Adaptive histogram equalization, Ceramic tile.

Keywords: Thresholding, Morphological operations, Image filtering, Adaptive histogram equalization, Ceramic tile. Volume 3, Issue 7, July 2013 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Blobs and Cracks

More information

On Board 6D Visual Sensors for Intersection Driving Assistance Systems

On Board 6D Visual Sensors for Intersection Driving Assistance Systems On Board 6D Visual Sensors for Intersection Driving Assistance Systems S. Nedevschi, T. Marita, R. Danescu, F. Oniga, S. Bota, I. Haller, C. Pantilie, M. Drulea, C. Golban Sergiu.Nedevschi@cs.utcluj.ro

More information

To Appear in IEEE International Symposium on Intelligent Vehicles in Parma, Italy, June 2004.

To Appear in IEEE International Symposium on Intelligent Vehicles in Parma, Italy, June 2004. To Appear in IEEE International Symposium on Intelligent Vehicles in Parma, Italy, June 2004. Occupant Posture Analysis using Reflectance and Stereo Images for Smart Airbag Deployment Stephen J. Krotosky

More information

Estimating the wavelength composition of scene illumination from image data is an

Estimating the wavelength composition of scene illumination from image data is an Chapter 3 The Principle and Improvement for AWB in DSC 3.1 Introduction Estimating the wavelength composition of scene illumination from image data is an important topics in color engineering. Solutions

More information

Motion Estimation. There are three main types (or applications) of motion estimation:

Motion Estimation. There are three main types (or applications) of motion estimation: Members: D91922016 朱威達 R93922010 林聖凱 R93922044 謝俊瑋 Motion Estimation There are three main types (or applications) of motion estimation: Parametric motion (image alignment) The main idea of parametric motion

More information

3D Modeling of Objects Using Laser Scanning

3D Modeling of Objects Using Laser Scanning 1 3D Modeling of Objects Using Laser Scanning D. Jaya Deepu, LPU University, Punjab, India Email: Jaideepudadi@gmail.com Abstract: In the last few decades, constructing accurate three-dimensional models

More information

Overview. Related Work Tensor Voting in 2-D Tensor Voting in 3-D Tensor Voting in N-D Application to Vision Problems Stereo Visual Motion

Overview. Related Work Tensor Voting in 2-D Tensor Voting in 3-D Tensor Voting in N-D Application to Vision Problems Stereo Visual Motion Overview Related Work Tensor Voting in 2-D Tensor Voting in 3-D Tensor Voting in N-D Application to Vision Problems Stereo Visual Motion Binary-Space-Partitioned Images 3-D Surface Extraction from Medical

More information

Presented at the FIG Congress 2018, May 6-11, 2018 in Istanbul, Turkey

Presented at the FIG Congress 2018, May 6-11, 2018 in Istanbul, Turkey Presented at the FIG Congress 2018, May 6-11, 2018 in Istanbul, Turkey Evangelos MALTEZOS, Charalabos IOANNIDIS, Anastasios DOULAMIS and Nikolaos DOULAMIS Laboratory of Photogrammetry, School of Rural

More information

Human Motion Detection and Tracking for Video Surveillance

Human Motion Detection and Tracking for Video Surveillance Human Motion Detection and Tracking for Video Surveillance Prithviraj Banerjee and Somnath Sengupta Department of Electronics and Electrical Communication Engineering Indian Institute of Technology, Kharagpur,

More information

Image Processing

Image Processing Image Processing 159.731 Canny Edge Detection Report Syed Irfanullah, Azeezullah 00297844 Danh Anh Huynh 02136047 1 Canny Edge Detection INTRODUCTION Edges Edges characterize boundaries and are therefore

More information