DYNAMIC STEREO VISION FOR INTERSECTION ASSISTANCE

Size: px
Start display at page:

Download "DYNAMIC STEREO VISION FOR INTERSECTION ASSISTANCE"

Transcription

1 FISITA 2008 World Automotive Congress Munich, Germany, September DYNAMIC STEREO VISION FOR INTERSECTION ASSISTANCE 1 Franke, Uwe *, 2 Rabe, Clemens, 1 Gehrig, Stefan, 3 Badino, Hernan, 1 Barth, Alexander 1 Daimler AG, Group Research, Germany, 2 University of Kiel, 3 University of Frankfurt KEYWORDS environment perception, driver assistance, intersection assistance, stereo vision, space-time stereo. ABSTRACT More than one third of all traffic accidents with injuries occur in urban areas, especially at intersections. Therefore, a driver assistance system supporting the driver in cities is highly desirable and has a tremendous potential of reducing the number of collisions at intersections. A suitable system for such complex situations requires a comprehensive understanding of the scene. This implies a precise estimation of the free space and the reliable detection and tracking of other moving traffic participants. Since the goal of accident free traffic requires a sensor with high spatial and temporal resolution, stereo vision will play an important role in future driver assistance systems. Most known stereo systems concentrate on single image pairs. However, in intelligent vehicle applications image sequences have to be analyzed. The contribution shows that a smart fusion of stereo vision and motion analysis (optical flow) gives much better results than classical frame-by-frame reconstructions. The basic idea is to track points with depth known from stereo vision over two and more consecutive frames and to fuse the spatial and temporal information using Kalman filters. The result is an improved accuracy of the 3D-position and an estimation of the 3D-motion of the considered point at the same time. This approach, called 6D Vision, enables a detection of moving objects even if they are partially hidden. From static points very accurate occupancy grids are built. A global optimization technique delivers a robust estimation of the free space. Pixels moving in the world are clustered to objects which are then tracked over time in order to estimate their motion state and to predict their paths. This allows for powerful collision avoidance systems: pedestrians crossing the street are detected before they enter the lane; the same holds for vehicles from the sides which are not detectable by common radar systems. Since we are able to estimate the yaw rate of oncoming traffic, the prediction is not restricted to straight motion but can detect potential collisions with turning traffic, especially at intersections. Urban vision asks for a large field of view. Within the German project AKTIV a fisheye stereo camera system is under development with a field of view of up to 150 degrees. If the 6D-Vision principle is applied to these images, laterally entering vehicles are also detectable.

2 INTRODUCTION Stereo vision is a research topic with a long history. See (1) for an overview. For a long time, correlation based techniques were commonly used. They deliver precise and reliable measurements in real-time on a PC or on dedicated hardware. Recently, much progress has been achieved in dense stereo. Especially the work of Hirschmueller (2) paves the road towards real-time solutions. His so called Semi-Global-Matching algorithms deliver nearoptimum solutions on the computational expense of a classical correlation scheme. New subpixel algorithms (3) reduce the distance noise significantly and further push the limits of stereo for a given camera system. Fig. 1 compares the results obtained by a common correlation based scheme with a modern dense stereo algorithm. The colors encode distance, the warmer the color the closer the point. The results do not only differ in density. Note the differences in low contrast areas such as the building and the road surface. Using stereo vision, the three-dimensional structure of the scene is easily obtained. The standard approach for free space analysis and obstacle detection is as follows: After rectification, the stereo correspondences are computed. Then, all 3D points are projected onto an occupancy grid. In a third step, this grid is segmented and potential obstacles are tracked over time in order to verify their existence and to estimate their motion state. This strategy ignores the strong correlation of successive frames and the information contained within. This paper describes an efficient exploitation of the correlation in time. It leads to more precise and stable results, and allows estimating the motion state of single image points even before the objects are detected. This "track-before-detect" approach distinguishes between static and moving pixels before any segmentation has been performed. Using static points, very accurate occupancy grids are generated while moving points can be easily grouped. The paper is organized as follows: First we sketch the problems in stereo vision and show that the uncertainties of occupancy grids are significantly reduced if the stereo information is integrated over time. Then, we introduce a Kalman filter based integration of stereo and optical flow allowing for the direct measurement of 3D-position and 3D-motion of all tracked image points (6D-Vision). The following section describes the motion state estimation of oncoming vehicles at intersections. Finally, we highlight the potential of fisheye cameras for intersection assistance and give results. Fig. 1: Correlation based stereo (left) vs. dense stereo (right). Red encodes close, green encodes far points. Note the higher density especially in low-contrast areas like the road or the building on the right side.

3 STEREO VISION AND FREE SPACE ANALYSIS Given a carefully rectified stereo image pair (i.e. all lens distortions have been corrected and the epipolar lines coincide with the image rows), stereo vision aims to find corresponding features in the left and right image pair along the epipolar lines. From the disparities, i.e. the distance between corresponding points, the world position can be easily derived. Nevertheless, the task is not as simple as it sounds: periodic structures can cause false correspondences, hidden points are hard to identify, areas with low or even no contrast are difficult to evaluate and illumination differences ask for robustness of the used similarity measure. Besides the mentioned epipolar constraint, other constraints like the ordering constraint, the uniqueness constraint, the smoothness constraint or the recently introduced gravitational constraint (3) help to solve those problems. Since the relative orientation of a stereo camera system cannot be assumed to be constant over time, a slow on-line calibration is necessary. Recently, Dang (4) proposed a scheme that solves this task robustly. As mentioned in the introduction, it is common to accumulate all 3D points above ground in a stochastic occupancy grid. Figure 2 shows such a grid obtained for the urban situation considered in the sequel. The origin of the coordinate system is centered in our own vehicle. Our standard stereo camera system has a base line of 30cm and an angle of view of 42 deg. The imagers have VGA resolution. Fig. 2: Occupancy grids of an urban situation. Left: stereo image pair with enlarged bicyclist. Middle: the stochastic occupancy grid based one single image pair. Right: the improved accuracy obtained by the procedure. Note the decreased uncertainty especially at larger distances. It becomes obvious that the uncertainty of stereo depth measurements increases quadratically with distance. Therefore, the bicyclist (zoomed out in the left image) at around 60m is highly blurred in the occupancy grid. Free space analysis of those occupancy grids is not very reliable, thus we are looking for strategies to reduce the uncertainty.

4 One way to reduce the disparity noise is the tracking of features in the images over multiple frames. If the disparity measurements are uncorrelated, the variance decreases with 1/N, if N is the number of images. The 6D-Vision algorithm described below exploits this fact. Fortunately, tracking becomes redundant in static scenes when the ego-motion of the camera is known a priori. This is beneficial since it allows working with dense stereo disparity maps despite the real-time constraint. Disparity measurements which are consistent over time are considered as belonging to the same world point, and therefore, disparity variance is reduced accordingly. This stereo integration requires three main steps: Prediction: the current integrated disparity and a variance image are predicted. This is equivalent to computing the expected optical flow and disparity based on ego-motion. Our prediction of the variance image includes the addition of a driving noise parameter that models the uncertainties of the system, such as ego-motion inaccuracy. Measurement: disparity and variance images are computed based on the current left and right images. Update: if the measured disparity confirms its prediction, then both are fused together reducing the variance of the estimation. The verification of the disparity consistency is performed using a standard 3-sigma test. Figure 2 shows an example of the improvement achieved. The occupancy grid shown at the right was computed with an integrated disparity image. Note the significantly reduced uncertainties of the registered 3D points. A bicyclist at approximately 60 meters away is marked in the images. The occupancy grids shown above are in Cartesian coordinates. However, Cartesian space is not a suitable space to compute the free space because the search must be done in the direction of rays leaving the camera. The set of rays must span the whole grid. This leads to discretization problems. A more appropriate space is the polar space. In polar coordinates every grid column is, by definition, already in the direction of a ray. Therefore, searching for obstacles in the ray direction is straightforward. For the computation of free space the first step is to transform the Cartesian grid to a polar grid by applying a remapping operation. The polar representation we use is a Column/Disparity occupancy grid, for a detailed discussion see (5). A result of this is shown in the middle image of Figure 3. Fig. 3: Free space computation. The green carpet shows the computed available free space. The free space is obtained applying dynamic programming on a Column/Disparity occupancy grid, which is as a remapping of the Cartesian depth map, shown at the right. The free space resulting from the dynamic programming is shown over the grids.

5 In polar representation, the task is to find the first visible obstacle in the positive direction of depth. All the space found in front of occupied cell is considered free space. The desired solution forms a path from left to right segmenting the polar grid into two regions. Instead of simply thresholding each column, dynamic programming is used. The method based on dynamic programming has the following properties: Global optimization: every row is not considered independently, but as a part of a global optimization problem which is optimally solved. Spatial and temporal smoothness of the solution: the spatial smoothness is imposed by penalizing depth discontinuities while temporal smoothness is imposed by penalizing the deviation of the current solution from the prediction. Preservation of spatial and temporal discontinuities: the saturation of the spatial and temporal costs allows the preservation of discontinuities. Figure 3 shows the result of the dynamic programming applied to the considered scene. For more details on this analysis see (6). 6D-VISION Until now, we assumed the world to be static and showed how to combine successive stereo image pairs to reduce the variance of the free space estimation. This information can be used for obstacle detection and obstacle avoidance in a straight-forward manner, since all non-free space is considered an obstacle. However, the world is not completely static and a system for obstacle detection has to cope with moving objects and precisely estimate their movements to predict potential collisions. A common approach is to analyze the occupancy grid and to track isolated objects over time. The major disadvantage of this algorithm is that the segmentation of isolated objects is difficult in scenes consisting of multiple nearby objects. Fig. 4: Dangerous traffic scene. The left image shows a pedestrian appearing behind a standing car. The corresponding stereo reconstruction is shown in the center image. Red encodes close, green encodes far points. The optical flow field is shown in the right image. Here red lines encode large image displacements, green small displacements. This problem is illustrated in Figure 4: Here the pedestrian appears behind the standing car and runs towards the street. In the center image, the reconstructed stereo information is shown using the red to green color encoding scheme. Here, the points belonging to the pedestrian are hardly distinguished from the points on the standing car. A segmentation based on this

6 information only will therefore merge the pedestrian and the standing car into a single static object. In the right image, the optical flow between the last and the current frame is shown. The color encodes the length of the displacement vector: red encodes large image displacements, green small displacements. Here the pedestrian and the standing car can easily be distinguished. This leads to the main idea of the 6D-Vision algorithm: Track an image point in one camera from frame to frame and calculate its stereo disparity. Together with the known motion of the ego-vehicle, the movement of the corresponding world point can be calculated. In practice, a direct motion calculation based on two consecutive frames is extremely noisy. Therefore the obtained measurements are filtered by a Kalman filter. Since we allow the observer to move, we fix the origin of the coordinate system to the car. The state vector of the Kalman filter consists of the world point in the car coordinate system, and its corresponding velocity vector. The six-dimensional state vector ( X, Y, Z, X&, Y&, Z& ) gives this algorithm its name: 6D Vision. The mathematical details are given in (7).,, with u and v being the current image coordinates of the tracked image point and d its corresponding disparity. As the perspective projection formulae are non-linear, we have to apply the Extended Kalman filter. The measurement vector used in the update step of the Kalman filter is ( u v, d ) Fig. 5: 6D-Vision block diagram A block diagram of the algorithm is shown in Figure 5. In every cycle, a new stereo image pair is obtained. In the left image, appropriate features (e.g. edges, corners) are detected and tracked over time. In the current application we use a version of the Kanade-Lucas-Tomasi tracker (8) which provides sub-pixel accuracy and tracks the features robustly for a long image sequence. The disparities for all tracked features are determined in the stereo module. After this step the estimated 3D-position of each feature is known. Together with the ego-motion the measurements of the tracking and the stereo modules are given to the Kalman filter system that updates the state estimation. For the next image pair analysis, the acquired 6D information is used to predict the image position of the tracked features. This yields a better tracking performance with respect to speed and robustness. In addition, the predicted depth information is used to improve the stereo calculation. The motion of a vehicle is not at all straight but exhibits strong pitch and roll motion. In order to compensate for these disturbances, a precise ego-motion analysis is advisable. If stereo tracks are available, the full motion state (6 degrees of freedom) can be obtained from vision. The powerful real-time algorithm we use is described in (9).

7 Fig. 6: Estimation results for the pedestrian from Figure 5. The time between the images is (from left to right) 0, 80, 160 and 240 ms. The vectors point to the predicted position of the corresponding world point in 0.5s. The color encodes the distance of the points. The result of this algorithm is shown in Figure 6. From left to right the estimation results for the pedestrian from Figure 5 are shown at 0, 80, 160 and 240 ms relative to the first appearance of the pedestrian. The estimated velocity vectors point to the predicted position of the corresponding world point in 0.5s. The colors encode the distance of the points. It can be seen, that this rich information helps to detect the moving pedestrian and provides a first prediction of its movement at the same time. OBJECT TRACKING 6D-Vision is a powerful method to extract linear point motion in the 3D world. A group of 6D vectors corresponding to adjacent 3D points with similar 3D motion vector is likely to belong to the same object and, thus, can be used to generate object hypothesis. However, due to the linear motion model of the single points, predicting the motion of such object hypothesis without any further constraints is also limited to linear motion. With respect to vehicles, especially at turning maneuvers, the prediction of the driving path can not be very precise and may lead to misinterpretations. In (10), a vision-based approach for estimating the nonlinear motion state of vehicles from a moving platform is proposed. Objects are represented by a 3D point cloud combined with a state vector including object pose and dynamics. It is assumed that it is possible to distinguish vehicles from other moving objects such as pedestrians, e.g. based on dimension and velocity of a cluster of 6D vectors. Fig. 7: Turning vehicle at an intersection. The orange box indicates the estimated position and orientation. The red arrow indicates the predicted driving path assuming constant motion. The dynamics of a vehicle is approximated by a coordinated turn motion model, which restricts lateral movements to a circular path based on velocity and yaw rate. Moving the point cloud in the world induces changes in the image plane and is observed in terms of optical flow and disparity changes. An Extended Kalman Filter is used to solve for the

8 inverse problem, i.e. relating these observations in the image to a movement of the point cloud in the world. All points are referred to a local object coordinate system defined for each tracked vehicle. It is assumed that the real position of a point within the object coordinate system does not change over time (rigid body assumption) and that the object's structure is well described by the point cloud. In practice one has to deal with noisy observations of these points and the real position of a point is not known as is the overall structure. However, it is possible to refine the object point cloud over time based on a number of noisy observations of the single points. Fig. 7 shows a typical situation at an intersection. The orange box shows the current position of the oncoming car. The complete motion state of the turning vehicle has been estimated based on the stereo tracks. Assuming constant motion, the green arrow in front of the car indicates the expected circular driving path for the next second. The proposed system is able to estimate the motion state of vehicles at urban intersections including velocity, yaw rate, and acceleration as well as position and orientation, and runs currently at 25Hz on our demonstrator car UTA. The filter can be easily extended by adding additional measurements, for example radar sensor information such as relative velocity or distance. FISHEYE STEREO FOR INTERSECTION ASSISTANCE Common stereo camera systems have opening angles around 40 degrees. Simple investigations reveal that this angle must be increased to about 150 degrees if dangerous situations at intersections should be recognized, e.g. vehicles coming from the side. Fisheye lenses in contrast to standard wide angle lenses have the advantage of a constant resolution over the whole field of view. Currently, we use 150 degree lenses. A typical image is shown in fig. 8. The computation can be limited to 400 lines of the 1628x1236 imager. In the first step, the images are rectified, based on the data of calibration performed in an offline process. For details see (11). This allows using the free space analysis and 6D-Vision as described above without any changes. The rectification step works with a cylindrical camera model as opposed to the pinhole model in order to obtain a bounded image size. Figure 8 shows a situation at a pedestrian crossing. The computed free space is overlaid in green. Fig. 8: Free space analysis for the pedestrian crossing situation. Figure 9 shows a second intersection scene where a vehicle approaches quickly from the right having the right-of-way. Note the position of the vehicle at initial detection. It is first detected

9 at 15m longitudinal and 22m lateral distance, yielding 26m Euclidian distance. An earlier detection was impossible due to occlusion of a wall visible at the right edge. Fig. 9: 6D Vision result for a scene with a vehicle approaching fast from the right side having the right-of-way. The significant lateral motion is detected within 4 frames. The arrow length shows the predicted position in 0.5s. The arrow color encodes distance - red is near and green is far away Fig. 10: Situation shown in fig. 8 two seconds later after the stop of our own vehicle. The actual object detection is done via direction and position analysis of the 6D vectors (see previous section). Figure 10 shows the same scene two seconds later. The ego-vehicle almost stopped while the vehicle from the right was able to continue. SUMMARY Vehicles acting in a dynamic environment must be able to detect any static or moving obstacle. This implies that an optimal stereo vision algorithm has to seek for an optimal exploitation of spatial and temporal information contained in the image sequence. As shown in the paper, precision and robustness of 3D reconstructions are significantly improved if the stereo information is appropriately integrated over time. This requires knowledge of the ego-motion, which in turn can be efficiently computed from 3D-tracks. It turns out that the obtained ego-motion data outperforms commonly used inertial sensors. The obtained depth maps show less noise and uncertainties than those generated by simple frameby-frame analysis. A dynamic programming approach allows for determining the free space without any susceptible obstacle threshold. The algorithm runs in real-time on a PC and has proven robustness in daily traffic including night-time driving and heavy rain. The request to detect small or partly hidden moving objects from a moving observer asks for fusion of stereo and optical flow. This leads to the 6D-Vision approach that allows to simultaneously estimate position and motion of each observed image point. Since the fusion is based on Kalman filters, the information contained in multiple frames is integrated. This leads to a more robust and precise estimation than differential approaches like pure evaluation of the optical flow on consecutive image pairs. Grouping this 6D-information is very reliable

10 and enables fast detection of moving objects which can be further tracked using appropriate dynamic models. The same concept is applied to cameras with fisheye lenses. Practical tests confirm that a crossing cyclist at an intersection is detected within 4-5 frames. The implementation on a 3.2GHz Pentium 4 proves real-time capability. Currently, we select and track about 2000 image points at 25Hz (the images have VGA resolution). REFERENCES (1) D.Scharstein, R.Szeliski: A taxonomy and evaluation of dense two-frame stereo correspondence algorithms. IJCV 47(1) (2002) pp (2) H.Hirschmueller: Accurate and efficient stereo processing by semi-global matching and mutual information. CVPR 2005, San Diego, CA. Volume 2. (June 2005) pp (3) S.Gehrig, U.Franke: Improving Stereo Sub-Pixel Accuracy for Long Range Stereo, Workshop on Virtual Representations and Modeling of Large-Scale Environments VRML@ ICCV 07, Rio, (4) T.Dang, C.Hoffmann: Tracking Camera Parameters of an Active Stereo Rig. In 28th Annual Symposium of the German Association for Pattern Recognition (DAGM 2006), Berlin, September (5) H.Badino, U.Franke, and R.Mester: Free space computation using stotchastic occupancy grids and dynamic programming. Workshop on Dynamical Vision, ICCV 07, Rio, (6) U.Franke, S.Gehrig, H.Badino, C.Rabe: Towards Optimal Stereo Analysis of Image Sequences, RobotVision 2008, February 2008, Auckland. (7) U.Franke, C.Rabe, H.Badino, S.Gehrig: 6D-Vision: Fusion of Stereo and Motion for Robust Environment Perception, 27 th DAGM Symposium 2005, pp ISBN (8) J.Shi and C.Tomasi, Good Features to Track. IEEE Conference on Computer Vision and Pattern Recognition, pages , (9) H.Badino, U.Franke, C.Rabe, S.Gehrig, Stereo-vision based detection of moving objects under strong camera motion, VisApp, Setubal (Portugal), February (10) A.Barth, U.Franke: Where will the Oncoming Vehicle be the Next Second?, IEEE Intelligent Vehicles Symposium IV 2008, Eindhoven, June (11) S.Gehrig, C. Rabe, L. Krüger, 6D Vision Goes Fisheye for Intersection Assistance, Canadian Robot Vision, Windsor, May 2008.

6D-Vision: Fusion of Stereo and Motion for Robust Environment Perception

6D-Vision: Fusion of Stereo and Motion for Robust Environment Perception 6D-Vision: Fusion of Stereo and Motion for Robust Environment Perception Uwe Franke, Clemens Rabe, Hernán Badino, and Stefan Gehrig DaimlerChrysler AG, 70546 Stuttgart, Germany {uwe.franke,clemens.rabe,hernan.badino,stefan.gehrig}@daimlerchrysler.com

More information

Spatio-Temporal Stereo Disparity Integration

Spatio-Temporal Stereo Disparity Integration Spatio-Temporal Stereo Disparity Integration Sandino Morales and Reinhard Klette The.enpeda.. Project, The University of Auckland Tamaki Innovation Campus, Auckland, New Zealand pmor085@aucklanduni.ac.nz

More information

The Stixel World - A Compact Medium Level Representation of the 3D-World

The Stixel World - A Compact Medium Level Representation of the 3D-World The Stixel World - A Compact Medium Level Representation of the 3D-World Hernán Badino 1, Uwe Franke 2, and David Pfeiffer 2 hbadino@cs.cmu.edu, {uwe.franke,david.pfeiffer}@daimler.com 1 Goethe University

More information

Spatio temporal Segmentation using Laserscanner and Video Sequences

Spatio temporal Segmentation using Laserscanner and Video Sequences Spatio temporal Segmentation using Laserscanner and Video Sequences Nico Kaempchen, Markus Zocholl and Klaus C.J. Dietmayer Department of Measurement, Control and Microtechnology University of Ulm, D 89081

More information

Real-time Stereo Vision for Urban Traffic Scene Understanding

Real-time Stereo Vision for Urban Traffic Scene Understanding Proceedings of the IEEE Intelligent Vehicles Symposium 2000 Dearborn (MI), USA October 3-5, 2000 Real-time Stereo Vision for Urban Traffic Scene Understanding U. Franke, A. Joos DaimlerChrylser AG D-70546

More information

On Board 6D Visual Sensors for Intersection Driving Assistance Systems

On Board 6D Visual Sensors for Intersection Driving Assistance Systems On Board 6D Visual Sensors for Intersection Driving Assistance Systems S. Nedevschi, T. Marita, R. Danescu, F. Oniga, S. Bota, I. Haller, C. Pantilie, M. Drulea, C. Golban Sergiu.Nedevschi@cs.utcluj.ro

More information

W4. Perception & Situation Awareness & Decision making

W4. Perception & Situation Awareness & Decision making W4. Perception & Situation Awareness & Decision making Robot Perception for Dynamic environments: Outline & DP-Grids concept Dynamic Probabilistic Grids Bayesian Occupancy Filter concept Dynamic Probabilistic

More information

On-line and Off-line 3D Reconstruction for Crisis Management Applications

On-line and Off-line 3D Reconstruction for Crisis Management Applications On-line and Off-line 3D Reconstruction for Crisis Management Applications Geert De Cubber Royal Military Academy, Department of Mechanical Engineering (MSTA) Av. de la Renaissance 30, 1000 Brussels geert.de.cubber@rma.ac.be

More information

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation Obviously, this is a very slow process and not suitable for dynamic scenes. To speed things up, we can use a laser that projects a vertical line of light onto the scene. This laser rotates around its vertical

More information

Dense 3D Reconstruction. Christiano Gava

Dense 3D Reconstruction. Christiano Gava Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Today: dense 3D reconstruction The matching problem

More information

Egomotion Estimation by Point-Cloud Back-Mapping

Egomotion Estimation by Point-Cloud Back-Mapping Egomotion Estimation by Point-Cloud Back-Mapping Haokun Geng, Radu Nicolescu, and Reinhard Klette Department of Computer Science, University of Auckland, New Zealand hgen001@aucklanduni.ac.nz Abstract.

More information

Dense 3D Reconstruction. Christiano Gava

Dense 3D Reconstruction. Christiano Gava Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Wide baseline matching (SIFT) Today: dense 3D reconstruction

More information

EE795: Computer Vision and Intelligent Systems

EE795: Computer Vision and Intelligent Systems EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 14 130307 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Stereo Dense Motion Estimation Translational

More information

Stereo Scene Flow for 3D Motion Analysis

Stereo Scene Flow for 3D Motion Analysis Stereo Scene Flow for 3D Motion Analysis Andreas Wedel Daniel Cremers Stereo Scene Flow for 3D Motion Analysis Dr. Andreas Wedel Group Research Daimler AG HPC 050 G023 Sindelfingen 71059 Germany andreas.wedel@daimler.com

More information

arxiv: v1 [cs.cv] 28 Sep 2018

arxiv: v1 [cs.cv] 28 Sep 2018 Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,

More information

LIGHT STRIPE PROJECTION-BASED PEDESTRIAN DETECTION DURING AUTOMATIC PARKING OPERATION

LIGHT STRIPE PROJECTION-BASED PEDESTRIAN DETECTION DURING AUTOMATIC PARKING OPERATION F2008-08-099 LIGHT STRIPE PROJECTION-BASED PEDESTRIAN DETECTION DURING AUTOMATIC PARKING OPERATION 1 Jung, Ho Gi*, 1 Kim, Dong Suk, 1 Kang, Hyoung Jin, 2 Kim, Jaihie 1 MANDO Corporation, Republic of Korea,

More information

Depth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth

Depth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth Common Classification Tasks Recognition of individual objects/faces Analyze object-specific features (e.g., key points) Train with images from different viewing angles Recognition of object classes Analyze

More information

Transactions on Information and Communications Technologies vol 16, 1996 WIT Press, ISSN

Transactions on Information and Communications Technologies vol 16, 1996 WIT Press,   ISSN ransactions on Information and Communications echnologies vol 6, 996 WI Press, www.witpress.com, ISSN 743-357 Obstacle detection using stereo without correspondence L. X. Zhou & W. K. Gu Institute of Information

More information

1. Introduction. A CASE STUDY Dense Image Matching Using Oblique Imagery Towards All-in- One Photogrammetry

1. Introduction. A CASE STUDY Dense Image Matching Using Oblique Imagery Towards All-in- One Photogrammetry Submitted to GIM International FEATURE A CASE STUDY Dense Image Matching Using Oblique Imagery Towards All-in- One Photogrammetry Dieter Fritsch 1, Jens Kremer 2, Albrecht Grimm 2, Mathias Rothermel 1

More information

Tracking Oncoming and Turning Vehicles at Intersections

Tracking Oncoming and Turning Vehicles at Intersections Tracking Oncoming and Turning Vehicles at Intersections Alexander Barth and Uwe Franke Abstract This article addresses the reliable tracking of oncoming traffic at urban intersections from a moving platform

More information

Product information. Hi-Tech Electronics Pte Ltd

Product information. Hi-Tech Electronics Pte Ltd Product information Introduction TEMA Motion is the world leading software for advanced motion analysis. Starting with digital image sequences the operator uses TEMA Motion to track objects in images,

More information

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science. Professor William Hoff Dept of Electrical Engineering &Computer Science http://inside.mines.edu/~whoff/ 1 Stereo Vision 2 Inferring 3D from 2D Model based pose estimation single (calibrated) camera > Can

More information

Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision

Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision What Happened Last Time? Human 3D perception (3D cinema) Computational stereo Intuitive explanation of what is meant by disparity Stereo matching

More information

Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation

Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation ÖGAI Journal 24/1 11 Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation Michael Bleyer, Margrit Gelautz, Christoph Rhemann Vienna University of Technology

More information

Stereovision-Based Sensor for Intersection Assistance

Stereovision-Based Sensor for Intersection Assistance Stereovision-Based Sensor for Intersection Assistance S. Nedevschi, R. Danescu, T. Marita, F. Oniga, C. Pocol, S. Bota, Technical University of Cluj-Napoca, M.-M. Meinecke, M. A. Obojski, Volkswagen AG

More information

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing Visual servoing vision allows a robotic system to obtain geometrical and qualitative information on the surrounding environment high level control motion planning (look-and-move visual grasping) low level

More information

Automatic Tracking of Moving Objects in Video for Surveillance Applications

Automatic Tracking of Moving Objects in Video for Surveillance Applications Automatic Tracking of Moving Objects in Video for Surveillance Applications Manjunath Narayana Committee: Dr. Donna Haverkamp (Chair) Dr. Arvin Agah Dr. James Miller Department of Electrical Engineering

More information

3D Computer Vision. Dense 3D Reconstruction II. Prof. Didier Stricker. Christiano Gava

3D Computer Vision. Dense 3D Reconstruction II. Prof. Didier Stricker. Christiano Gava 3D Computer Vision Dense 3D Reconstruction II Prof. Didier Stricker Christiano Gava Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de

More information

Pedestrian Detection Using Multi-layer LIDAR

Pedestrian Detection Using Multi-layer LIDAR 1 st International Conference on Transportation Infrastructure and Materials (ICTIM 2016) ISBN: 978-1-60595-367-0 Pedestrian Detection Using Multi-layer LIDAR Mingfang Zhang 1, Yuping Lu 2 and Tong Liu

More information

2 OVERVIEW OF RELATED WORK

2 OVERVIEW OF RELATED WORK Utsushi SAKAI Jun OGATA This paper presents a pedestrian detection system based on the fusion of sensors for LIDAR and convolutional neural network based image classification. By using LIDAR our method

More information

Application questions. Theoretical questions

Application questions. Theoretical questions The oral exam will last 30 minutes and will consist of one application question followed by two theoretical questions. Please find below a non exhaustive list of possible application questions. The list

More information

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science. Professor William Hoff Dept of Electrical Engineering &Computer Science http://inside.mines.edu/~whoff/ 1 Stereo Vision 2 Inferring 3D from 2D Model based pose estimation single (calibrated) camera Stereo

More information

Laserscanner Based Cooperative Pre-Data-Fusion

Laserscanner Based Cooperative Pre-Data-Fusion Laserscanner Based Cooperative Pre-Data-Fusion 63 Laserscanner Based Cooperative Pre-Data-Fusion F. Ahlers, Ch. Stimming, Ibeo Automobile Sensor GmbH Abstract The Cooperative Pre-Data-Fusion is a novel

More information

Stereo Vision Based Advanced Driver Assistance System

Stereo Vision Based Advanced Driver Assistance System Stereo Vision Based Advanced Driver Assistance System Ho Gi Jung, Yun Hee Lee, Dong Suk Kim, Pal Joo Yoon MANDO Corp. 413-5,Gomae-Ri, Yongin-Si, Kyongi-Do, 449-901, Korea Phone: (82)31-0-5253 Fax: (82)31-0-5496

More information

Flow Estimation. Min Bai. February 8, University of Toronto. Min Bai (UofT) Flow Estimation February 8, / 47

Flow Estimation. Min Bai. February 8, University of Toronto. Min Bai (UofT) Flow Estimation February 8, / 47 Flow Estimation Min Bai University of Toronto February 8, 2016 Min Bai (UofT) Flow Estimation February 8, 2016 1 / 47 Outline Optical Flow - Continued Min Bai (UofT) Flow Estimation February 8, 2016 2

More information

Clemens Rabe Detection of Moving Objects by Spatio-Temporal Motion Analysis

Clemens Rabe Detection of Moving Objects by Spatio-Temporal Motion Analysis Clemens Rabe Detection of Moving Objects by Spatio-Temporal Motion Analysis Detection of Moving Objects by Spatio-Temporal Motion Analysis Dissertation zur Erlangung des akademischen Grades Doktor der

More information

An Evaluation Framework for Stereo-Based Driver Assistance

An Evaluation Framework for Stereo-Based Driver Assistance An Evaluation Framework for Stereo-Based Driver Assistance Nicolai Schneider 1, Stefan Gehrig 2, David Pfeiffer 2, and Konstantinos Banitsas 3 1 IT-Designers GmbH, Esslingen, Germany 2 Daimler AG, Team

More information

Creating a distortion characterisation dataset for visual band cameras using fiducial markers.

Creating a distortion characterisation dataset for visual band cameras using fiducial markers. Creating a distortion characterisation dataset for visual band cameras using fiducial markers. Robert Jermy Council for Scientific and Industrial Research Email: rjermy@csir.co.za Jason de Villiers Council

More information

Vehicle Occupant Posture Analysis Using Voxel Data

Vehicle Occupant Posture Analysis Using Voxel Data Ninth World Congress on Intelligent Transport Systems, Chicago, Illinois, October Vehicle Occupant Posture Analysis Using Voxel Data Ivana Mikic, Mohan Trivedi Computer Vision and Robotics Research Laboratory

More information

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION Mr.V.SRINIVASA RAO 1 Prof.A.SATYA KALYAN 2 DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING PRASAD V POTLURI SIDDHARTHA

More information

Self-calibration of a pair of stereo cameras in general position

Self-calibration of a pair of stereo cameras in general position Self-calibration of a pair of stereo cameras in general position Raúl Rojas Institut für Informatik Freie Universität Berlin Takustr. 9, 14195 Berlin, Germany Abstract. This paper shows that it is possible

More information

Feature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies

Feature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies Feature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies M. Lourakis, S. Tzurbakis, A. Argyros, S. Orphanoudakis Computer Vision and Robotics Lab (CVRL) Institute of

More information

CS 4758: Automated Semantic Mapping of Environment

CS 4758: Automated Semantic Mapping of Environment CS 4758: Automated Semantic Mapping of Environment Dongsu Lee, ECE, M.Eng., dl624@cornell.edu Aperahama Parangi, CS, 2013, alp75@cornell.edu Abstract The purpose of this project is to program an Erratic

More information

Sensory Augmentation for Increased Awareness of Driving Environment

Sensory Augmentation for Increased Awareness of Driving Environment Sensory Augmentation for Increased Awareness of Driving Environment Pranay Agrawal John M. Dolan Dec. 12, 2014 Technologies for Safe and Efficient Transportation (T-SET) UTC The Robotics Institute Carnegie

More information

Pedestrian Detection with Radar and Computer Vision

Pedestrian Detection with Radar and Computer Vision Pedestrian Detection with Radar and Computer Vision camera radar sensor Stefan Milch, Marc Behrens, Darmstadt, September 25 25 / 26, 2001 Pedestrian accidents and protection systems Impact zone: 10% opposite

More information

Introduction.

Introduction. Product information Image Systems AB Main office: Ågatan 40, SE-582 22 Linköping Phone +46 13 200 100, fax +46 13 200 150 info@imagesystems.se, Introduction TEMA Automotive is the world leading system

More information

Real-Time Detection of Road Markings for Driving Assistance Applications

Real-Time Detection of Road Markings for Driving Assistance Applications Real-Time Detection of Road Markings for Driving Assistance Applications Ioana Maria Chira, Ancuta Chibulcutean Students, Faculty of Automation and Computer Science Technical University of Cluj-Napoca

More information

Motion Analysis. Motion analysis. Now we will talk about. Differential Motion Analysis. Motion analysis. Difference Pictures

Motion Analysis. Motion analysis. Now we will talk about. Differential Motion Analysis. Motion analysis. Difference Pictures Now we will talk about Motion Analysis Motion analysis Motion analysis is dealing with three main groups of motionrelated problems: Motion detection Moving object detection and location. Derivation of

More information

An Interactive Technique for Robot Control by Using Image Processing Method

An Interactive Technique for Robot Control by Using Image Processing Method An Interactive Technique for Robot Control by Using Image Processing Method Mr. Raskar D. S 1., Prof. Mrs. Belagali P. P 2 1, E&TC Dept. Dr. JJMCOE., Jaysingpur. Maharashtra., India. 2 Associate Prof.

More information

Accurate 3D Face and Body Modeling from a Single Fixed Kinect

Accurate 3D Face and Body Modeling from a Single Fixed Kinect Accurate 3D Face and Body Modeling from a Single Fixed Kinect Ruizhe Wang*, Matthias Hernandez*, Jongmoo Choi, Gérard Medioni Computer Vision Lab, IRIS University of Southern California Abstract In this

More information

SLAM with SIFT (aka Mobile Robot Localization and Mapping with Uncertainty using Scale-Invariant Visual Landmarks ) Se, Lowe, and Little

SLAM with SIFT (aka Mobile Robot Localization and Mapping with Uncertainty using Scale-Invariant Visual Landmarks ) Se, Lowe, and Little SLAM with SIFT (aka Mobile Robot Localization and Mapping with Uncertainty using Scale-Invariant Visual Landmarks ) Se, Lowe, and Little + Presented by Matt Loper CS296-3: Robot Learning and Autonomy Brown

More information

Final Exam Study Guide

Final Exam Study Guide Final Exam Study Guide Exam Window: 28th April, 12:00am EST to 30th April, 11:59pm EST Description As indicated in class the goal of the exam is to encourage you to review the material from the course.

More information

3D Model Acquisition by Tracking 2D Wireframes

3D Model Acquisition by Tracking 2D Wireframes 3D Model Acquisition by Tracking 2D Wireframes M. Brown, T. Drummond and R. Cipolla {96mab twd20 cipolla}@eng.cam.ac.uk Department of Engineering University of Cambridge Cambridge CB2 1PZ, UK Abstract

More information

Performance Evaluation Metrics and Statistics for Positional Tracker Evaluation

Performance Evaluation Metrics and Statistics for Positional Tracker Evaluation Performance Evaluation Metrics and Statistics for Positional Tracker Evaluation Chris J. Needham and Roger D. Boyle School of Computing, The University of Leeds, Leeds, LS2 9JT, UK {chrisn,roger}@comp.leeds.ac.uk

More information

PERFORMANCE CAPTURE FROM SPARSE MULTI-VIEW VIDEO

PERFORMANCE CAPTURE FROM SPARSE MULTI-VIEW VIDEO Stefan Krauß, Juliane Hüttl SE, SoSe 2011, HU-Berlin PERFORMANCE CAPTURE FROM SPARSE MULTI-VIEW VIDEO 1 Uses of Motion/Performance Capture movies games, virtual environments biomechanics, sports science,

More information

Ruch (Motion) Rozpoznawanie Obrazów Krzysztof Krawiec Instytut Informatyki, Politechnika Poznańska. Krzysztof Krawiec IDSS

Ruch (Motion) Rozpoznawanie Obrazów Krzysztof Krawiec Instytut Informatyki, Politechnika Poznańska. Krzysztof Krawiec IDSS Ruch (Motion) Rozpoznawanie Obrazów Krzysztof Krawiec Instytut Informatyki, Politechnika Poznańska 1 Krzysztof Krawiec IDSS 2 The importance of visual motion Adds entirely new (temporal) dimension to visual

More information

Map Guided Lane Detection Alexander Döbert 1,2, Andre Linarth 1,2, Eva Kollorz 2

Map Guided Lane Detection Alexander Döbert 1,2, Andre Linarth 1,2, Eva Kollorz 2 Map Guided Lane Detection Alexander Döbert 1,2, Andre Linarth 1,2, Eva Kollorz 2 1 Elektrobit Automotive GmbH, Am Wolfsmantel 46, 91058 Erlangen, Germany {AndreGuilherme.Linarth, Alexander.Doebert}@elektrobit.com

More information

Stereo and Epipolar geometry

Stereo and Epipolar geometry Previously Image Primitives (feature points, lines, contours) Today: Stereo and Epipolar geometry How to match primitives between two (multiple) views) Goals: 3D reconstruction, recognition Jana Kosecka

More information

Stochastic Road Shape Estimation, B. Southall & C. Taylor. Review by: Christopher Rasmussen

Stochastic Road Shape Estimation, B. Southall & C. Taylor. Review by: Christopher Rasmussen Stochastic Road Shape Estimation, B. Southall & C. Taylor Review by: Christopher Rasmussen September 26, 2002 Announcements Readings for next Tuesday: Chapter 14-14.4, 22-22.5 in Forsyth & Ponce Main Contributions

More information

On Road Vehicle Detection using Shadows

On Road Vehicle Detection using Shadows On Road Vehicle Detection using Shadows Gilad Buchman Grasp Lab, Department of Computer and Information Science School of Engineering University of Pennsylvania, Philadelphia, PA buchmag@seas.upenn.edu

More information

STEREO-VISION SYSTEM PERFORMANCE ANALYSIS

STEREO-VISION SYSTEM PERFORMANCE ANALYSIS STEREO-VISION SYSTEM PERFORMANCE ANALYSIS M. Bertozzi, A. Broggi, G. Conte, and A. Fascioli Dipartimento di Ingegneria dell'informazione, Università di Parma Parco area delle Scienze, 181A I-43100, Parma,

More information

Lecture 14: Computer Vision

Lecture 14: Computer Vision CS/b: Artificial Intelligence II Prof. Olga Veksler Lecture : Computer Vision D shape from Images Stereo Reconstruction Many Slides are from Steve Seitz (UW), S. Narasimhan Outline Cues for D shape perception

More information

FAB verses tradition camera-based motion capture systems

FAB verses tradition camera-based motion capture systems FAB verses tradition camera-based motion capture systems The advent of micromachined inertial sensors, such as rate gyroscopes and accelerometers, has made new navigation and tracking technologies possible.

More information

Measurement of Pedestrian Groups Using Subtraction Stereo

Measurement of Pedestrian Groups Using Subtraction Stereo Measurement of Pedestrian Groups Using Subtraction Stereo Kenji Terabayashi, Yuki Hashimoto, and Kazunori Umeda Chuo University / CREST, JST, 1-13-27 Kasuga, Bunkyo-ku, Tokyo 112-8551, Japan terabayashi@mech.chuo-u.ac.jp

More information

Car tracking in tunnels

Car tracking in tunnels Czech Pattern Recognition Workshop 2000, Tomáš Svoboda (Ed.) Peršlák, Czech Republic, February 2 4, 2000 Czech Pattern Recognition Society Car tracking in tunnels Roman Pflugfelder and Horst Bischof Pattern

More information

Mini Survey Paper (Robotic Mapping) Ryan Hamor CPRE 583 September 2011

Mini Survey Paper (Robotic Mapping) Ryan Hamor CPRE 583 September 2011 Mini Survey Paper (Robotic Mapping) Ryan Hamor CPRE 583 September 2011 Introduction The goal of this survey paper is to examine the field of robotic mapping and the use of FPGAs in various implementations.

More information

Finally: Motion and tracking. Motion 4/20/2011. CS 376 Lecture 24 Motion 1. Video. Uses of motion. Motion parallax. Motion field

Finally: Motion and tracking. Motion 4/20/2011. CS 376 Lecture 24 Motion 1. Video. Uses of motion. Motion parallax. Motion field Finally: Motion and tracking Tracking objects, video analysis, low level motion Motion Wed, April 20 Kristen Grauman UT-Austin Many slides adapted from S. Seitz, R. Szeliski, M. Pollefeys, and S. Lazebnik

More information

Revising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History

Revising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History Revising Stereo Vision Maps in Particle Filter Based SLAM using Localisation Confidence and Sample History Simon Thompson and Satoshi Kagami Digital Human Research Center National Institute of Advanced

More information

Human Detection. A state-of-the-art survey. Mohammad Dorgham. University of Hamburg

Human Detection. A state-of-the-art survey. Mohammad Dorgham. University of Hamburg Human Detection A state-of-the-art survey Mohammad Dorgham University of Hamburg Presentation outline Motivation Applications Overview of approaches (categorized) Approaches details References Motivation

More information

Three-Dimensional Sensors Lecture 2: Projected-Light Depth Cameras

Three-Dimensional Sensors Lecture 2: Projected-Light Depth Cameras Three-Dimensional Sensors Lecture 2: Projected-Light Depth Cameras Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inria.fr http://perception.inrialpes.fr/ Outline The geometry of active stereo.

More information

Large-Scale Traffic Sign Recognition based on Local Features and Color Segmentation

Large-Scale Traffic Sign Recognition based on Local Features and Color Segmentation Large-Scale Traffic Sign Recognition based on Local Features and Color Segmentation M. Blauth, E. Kraft, F. Hirschenberger, M. Böhm Fraunhofer Institute for Industrial Mathematics, Fraunhofer-Platz 1,

More information

Stereo imaging ideal geometry

Stereo imaging ideal geometry Stereo imaging ideal geometry (X,Y,Z) Z f (x L,y L ) f (x R,y R ) Optical axes are parallel Optical axes separated by baseline, b. Line connecting lens centers is perpendicular to the optical axis, and

More information

Minimizing Noise and Bias in 3D DIC. Correlated Solutions, Inc.

Minimizing Noise and Bias in 3D DIC. Correlated Solutions, Inc. Minimizing Noise and Bias in 3D DIC Correlated Solutions, Inc. Overview Overview of Noise and Bias Digital Image Correlation Background/Tracking Function Minimizing Noise Focus Contrast/Lighting Glare

More information

THE POSITION AND ORIENTATION MEASUREMENT OF GONDOLA USING A VISUAL CAMERA

THE POSITION AND ORIENTATION MEASUREMENT OF GONDOLA USING A VISUAL CAMERA THE POSITION AND ORIENTATION MEASUREMENT OF GONDOLA USING A VISUAL CAMERA Hwadong Sun 1, Dong Yeop Kim 1 *, Joon Ho Kwon 2, Bong-Seok Kim 1, and Chang-Woo Park 1 1 Intelligent Robotics Research Center,

More information

차세대지능형자동차를위한신호처리기술 정호기

차세대지능형자동차를위한신호처리기술 정호기 차세대지능형자동차를위한신호처리기술 008.08. 정호기 E-mail: hgjung@mando.com hgjung@yonsei.ac.kr 0 . 지능형자동차의미래 ) 단위 system functions 운전자상황인식 얼굴방향인식 시선방향인식 졸음운전인식 운전능력상실인식 차선인식, 전방장애물검출및분류 Lane Keeping System + Adaptive Cruise

More information

Video Georegistration: Key Challenges. Steve Blask Harris Corporation GCSD Melbourne, FL 32934

Video Georegistration: Key Challenges. Steve Blask Harris Corporation GCSD Melbourne, FL 32934 Video Georegistration: Key Challenges Steve Blask sblask@harris.com Harris Corporation GCSD Melbourne, FL 32934 Definitions Registration: image to image alignment Find pixel-to-pixel correspondences between

More information

LUMS Mine Detector Project

LUMS Mine Detector Project LUMS Mine Detector Project Using visual information to control a robot (Hutchinson et al. 1996). Vision may or may not be used in the feedback loop. Visual (image based) features such as points, lines

More information

From Orientation to Functional Modeling for Terrestrial and UAV Images

From Orientation to Functional Modeling for Terrestrial and UAV Images From Orientation to Functional Modeling for Terrestrial and UAV Images Helmut Mayer 1 Andreas Kuhn 1, Mario Michelini 1, William Nguatem 1, Martin Drauschke 2 and Heiko Hirschmüller 2 1 Visual Computing,

More information

What have we leaned so far?

What have we leaned so far? What have we leaned so far? Camera structure Eye structure Project 1: High Dynamic Range Imaging What have we learned so far? Image Filtering Image Warping Camera Projection Model Project 2: Panoramic

More information

Advanced Vision Guided Robotics. David Bruce Engineering Manager FANUC America Corporation

Advanced Vision Guided Robotics. David Bruce Engineering Manager FANUC America Corporation Advanced Vision Guided Robotics David Bruce Engineering Manager FANUC America Corporation Traditional Vision vs. Vision based Robot Guidance Traditional Machine Vision Determine if a product passes or

More information

Camera Calibration for a Robust Omni-directional Photogrammetry System

Camera Calibration for a Robust Omni-directional Photogrammetry System Camera Calibration for a Robust Omni-directional Photogrammetry System Fuad Khan 1, Michael Chapman 2, Jonathan Li 3 1 Immersive Media Corporation Calgary, Alberta, Canada 2 Ryerson University Toronto,

More information

Locating ego-centers in depth for hippocampal place cells

Locating ego-centers in depth for hippocampal place cells 204 5th Joint Symposium on Neural Computation Proceedings UCSD (1998) Locating ego-centers in depth for hippocampal place cells Kechen Zhang,' Terrence J. Sejeowski112 & Bruce L. ~cnau~hton~ 'Howard Hughes

More information

3D DEFORMATION MEASUREMENT USING STEREO- CORRELATION APPLIED TO EXPERIMENTAL MECHANICS

3D DEFORMATION MEASUREMENT USING STEREO- CORRELATION APPLIED TO EXPERIMENTAL MECHANICS 3D DEFORMATION MEASUREMENT USING STEREO- CORRELATION APPLIED TO EXPERIMENTAL MECHANICS Dorian Garcia, Jean-José Orteu École des Mines d Albi, F-81013 ALBI CT Cedex 09, France Dorian.Garcia@enstimac.fr,

More information

Module 7 VIDEO CODING AND MOTION ESTIMATION

Module 7 VIDEO CODING AND MOTION ESTIMATION Module 7 VIDEO CODING AND MOTION ESTIMATION Lesson 20 Basic Building Blocks & Temporal Redundancy Instructional Objectives At the end of this lesson, the students should be able to: 1. Name at least five

More information

Monitoring surrounding areas of truck-trailer combinations

Monitoring surrounding areas of truck-trailer combinations Monitoring surrounding areas of truck-trailer combinations Tobias Ehlgen 1 and Tomas Pajdla 2 1 Daimler-Chrysler Research and Technology, Ulm tobias.ehlgen@daimlerchrysler.com 2 Center of Machine Perception,

More information

3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller

3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller 3D Computer Vision Depth Cameras Prof. Didier Stricker Oliver Wasenmüller Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de

More information

Fundamental Technologies Driving the Evolution of Autonomous Driving

Fundamental Technologies Driving the Evolution of Autonomous Driving 426 Hitachi Review Vol. 65 (2016), No. 9 Featured Articles Fundamental Technologies Driving the Evolution of Autonomous Driving Takeshi Shima Takeshi Nagasaki Akira Kuriyama Kentaro Yoshimura, Ph.D. Tsuneo

More information

Pedestrian counting in video sequences using optical flow clustering

Pedestrian counting in video sequences using optical flow clustering Pedestrian counting in video sequences using optical flow clustering SHIZUKA FUJISAWA, GO HASEGAWA, YOSHIAKI TANIGUCHI, HIROTAKA NAKANO Graduate School of Information Science and Technology Osaka University

More information

Non-flat Road Detection Based on A Local Descriptor

Non-flat Road Detection Based on A Local Descriptor Non-flat Road Detection Based on A Local Descriptor Kangru Wang, Lei Qu, Lili Chen, Yuzhang Gu, Xiaolin Zhang Abstrct The detection of road surface and free space remains challenging for non-flat plane,

More information

APPLICATION OF AERIAL VIDEO FOR TRAFFIC FLOW MONITORING AND MANAGEMENT

APPLICATION OF AERIAL VIDEO FOR TRAFFIC FLOW MONITORING AND MANAGEMENT Pitu Mirchandani, Professor, Department of Systems and Industrial Engineering Mark Hickman, Assistant Professor, Department of Civil Engineering Alejandro Angel, Graduate Researcher Dinesh Chandnani, Graduate

More information

Kinect Cursor Control EEE178 Dr. Fethi Belkhouche Christopher Harris Danny Nguyen I. INTRODUCTION

Kinect Cursor Control EEE178 Dr. Fethi Belkhouche Christopher Harris Danny Nguyen I. INTRODUCTION Kinect Cursor Control EEE178 Dr. Fethi Belkhouche Christopher Harris Danny Nguyen Abstract: An XBOX 360 Kinect is used to develop two applications to control the desktop cursor of a Windows computer. Application

More information

Ground Plane Motion Parameter Estimation For Non Circular Paths

Ground Plane Motion Parameter Estimation For Non Circular Paths Ground Plane Motion Parameter Estimation For Non Circular Paths G.J.Ellwood Y.Zheng S.A.Billings Department of Automatic Control and Systems Engineering University of Sheffield, Sheffield, UK J.E.W.Mayhew

More information

Topics to be Covered in the Rest of the Semester. CSci 4968 and 6270 Computational Vision Lecture 15 Overview of Remainder of the Semester

Topics to be Covered in the Rest of the Semester. CSci 4968 and 6270 Computational Vision Lecture 15 Overview of Remainder of the Semester Topics to be Covered in the Rest of the Semester CSci 4968 and 6270 Computational Vision Lecture 15 Overview of Remainder of the Semester Charles Stewart Department of Computer Science Rensselaer Polytechnic

More information

Stereo vision. Many slides adapted from Steve Seitz

Stereo vision. Many slides adapted from Steve Seitz Stereo vision Many slides adapted from Steve Seitz What is stereo vision? Generic problem formulation: given several images of the same object or scene, compute a representation of its 3D shape What is

More information

Motion-based obstacle detection and tracking for car driving assistance

Motion-based obstacle detection and tracking for car driving assistance Motion-based obstacle detection and tracking for car driving assistance G. Lefaix, E. Marchand, Patrick Bouthemy To cite this version: G. Lefaix, E. Marchand, Patrick Bouthemy. Motion-based obstacle detection

More information

Training models for road scene understanding with automated ground truth Dan Levi

Training models for road scene understanding with automated ground truth Dan Levi Training models for road scene understanding with automated ground truth Dan Levi With: Noa Garnett, Ethan Fetaya, Shai Silberstein, Rafi Cohen, Shaul Oron, Uri Verner, Ariel Ayash, Kobi Horn, Vlad Golder,

More information

Motion estimation of unmanned marine vehicles Massimo Caccia

Motion estimation of unmanned marine vehicles Massimo Caccia Motion estimation of unmanned marine vehicles Massimo Caccia Consiglio Nazionale delle Ricerche Istituto di Studi sui Sistemi Intelligenti per l Automazione Via Amendola 122 D/O, 70126, Bari, Italy massimo.caccia@ge.issia.cnr.it

More information

People detection and tracking using stereo vision and color

People detection and tracking using stereo vision and color People detection and tracking using stereo vision and color Rafael Munoz-Salinas, Eugenio Aguirre, Miguel Garcia-Silvente. In Image and Vision Computing Volume 25 Issue 6 (2007) 995-1007. Presented by

More information

Solid State LiDAR for Ubiquitous 3D Sensing

Solid State LiDAR for Ubiquitous 3D Sensing April 6, 2016 Solid State LiDAR for Ubiquitous 3D Sensing Louay Eldada, Ph.D. CEO, Co-founder Quanergy Systems New Paradigm in 3D Sensing Disruptive Technologies: Solid State 3D LiDAR sensors Embedded

More information

SURVEY OF LOCAL AND GLOBAL OPTICAL FLOW WITH COARSE TO FINE METHOD

SURVEY OF LOCAL AND GLOBAL OPTICAL FLOW WITH COARSE TO FINE METHOD SURVEY OF LOCAL AND GLOBAL OPTICAL FLOW WITH COARSE TO FINE METHOD M.E-II, Department of Computer Engineering, PICT, Pune ABSTRACT: Optical flow as an image processing technique finds its applications

More information