Gradient-Based time to contact on paracatadioptric camera

Similar documents
HySCaS: Hybrid Stereoscopic Calibration Software

An adapted Lucas-Kanade s method for optical flow estimation in catadioptric images

Motion-based obstacle detection and tracking for car driving assistance

Scale Invariant Detection and Tracking of Elongated Structures

Moveability and Collision Analysis for Fully-Parallel Manipulators

Kernel-Based Laplacian Smoothing Method for 3D Mesh Denoising

Setup of epiphytic assistance systems with SEPIA

BoxPlot++ Zeina Azmeh, Fady Hamoui, Marianne Huchard. To cite this version: HAL Id: lirmm

A Method for Interactive 3D Reconstruction of Piecewise Planar Objects from Single Images

Squaring the Circle with Weak Mobile Robots

Multi-atlas labeling with population-specific template and non-local patch-based label fusion

SDLS: a Matlab package for solving conic least-squares problems

Lossless and Lossy Minimal Redundancy Pyramidal Decomposition for Scalable Image Compression Technique

Light field video dataset captured by a R8 Raytrix camera (with disparity maps)

Tacked Link List - An Improved Linked List for Advance Resource Reservation

DSM GENERATION FROM STEREOSCOPIC IMAGERY FOR DAMAGE MAPPING, APPLICATION ON THE TOHOKU TSUNAMI

The optimal routing of augmented cubes.

ROBUST MOTION SEGMENTATION FOR HIGH DEFINITION VIDEO SEQUENCES USING A FAST MULTI-RESOLUTION MOTION ESTIMATION BASED ON SPATIO-TEMPORAL TUBES

Real-time FEM based control of soft surgical robots

Branch-and-price algorithms for the Bi-Objective Vehicle Routing Problem with Time Windows

Fault-Tolerant Storage Servers for the Databases of Redundant Web Servers in a Computing Grid

Continuous Control of Lagrangian Data

Fuzzy sensor for the perception of colour

Malware models for network and service management

Blind Browsing on Hand-Held Devices: Touching the Web... to Understand it Better

Comparison of spatial indexes

Plane-Based Calibration for Linear Cameras

How to simulate a volume-controlled flooding with mathematical morphology operators?

Fast and precise kinematic skeleton extraction of 3D dynamic meshes

DANCer: Dynamic Attributed Network with Community Structure Generator

Real-Time Collision Detection for Dynamic Virtual Environments

Comparison of radiosity and ray-tracing methods for coupled rooms

Multimedia CTI Services for Telecommunication Systems

NP versus PSPACE. Frank Vega. To cite this version: HAL Id: hal

An Experimental Assessment of the 2D Visibility Complex

Workspace and joint space analysis of the 3-RPS parallel robot

SLMRACE: A noise-free new RACE implementation with reduced computational time

Kinematic skeleton extraction based on motion boundaries for 3D dynamic meshes

Regularization parameter estimation for non-negative hyperspectral image deconvolution:supplementary material

Traffic Grooming in Bidirectional WDM Ring Networks

Structuring the First Steps of Requirements Elicitation

Study on Feebly Open Set with Respect to an Ideal Topological Spaces

An Efficient Numerical Inverse Scattering Algorithm for Generalized Zakharov-Shabat Equations with Two Potential Functions

Real-Time and Resilient Intrusion Detection: A Flow-Based Approach

Time-to-Contact from Image Intensity

The Proportional Colouring Problem: Optimizing Buffers in Radio Mesh Networks

From medical imaging to numerical simulations

LaHC at CLEF 2015 SBS Lab

Shape from Probability Maps with Image-Adapted Voxelization

Time To Contact (TTC) Berthold K.P. Horn, Yajun Fang, Ichiro Masaki

An efficient alternative approach for home furniture detection and localization by an autonomous mobile robot

A Fuzzy Approach for Background Subtraction

An FCA Framework for Knowledge Discovery in SPARQL Query Answers

QuickRanking: Fast Algorithm For Sorting And Ranking Data

A novel coupled transmission-reflection tomography and the V-line Radon transform

Preliminary analysis of the drive system of the CTA LST Telescope and its integration in the whole PLC architecture

Relabeling nodes according to the structure of the graph

Sliding HyperLogLog: Estimating cardinality in a data stream

DETERMINATION OF THE TRANSDUCER VELOCITIES IN A SONAR ARRAY USING DIGITAL ACOUSTICAL HOLOGRAPHY

Service Reconfiguration in the DANAH Assistive System

THE COVERING OF ANCHORED RECTANGLES UP TO FIVE POINTS

Inverting the Reflectance Map with Binary Search

A Practical Approach for 3D Model Indexing by combining Local and Global Invariants

Very Tight Coupling between LTE and WiFi: a Practical Analysis

Simulations of VANET Scenarios with OPNET and SUMO

Paris-Lille-3D: A Point Cloud Dataset for Urban Scene Segmentation and Classification

Comparator: A Tool for Quantifying Behavioural Compatibility

TEXTURE SIMILARITY METRICS APPLIED TO HEVC INTRA PREDICTION

A case-based reasoning approach for unknown class invoice processing

Linked data from your pocket: The Android RDFContentProvider

Deformetrica: a software for statistical analysis of anatomical shapes

FIT IoT-LAB: The Largest IoT Open Experimental Testbed

Prototype Selection Methods for On-line HWR

Free-Form Mirror Design Inspired by Photometric Stereo

Representation of Finite Games as Network Congestion Games

IMPLEMENTATION OF MOTION ESTIMATION BASED ON HETEROGENEOUS PARALLEL COMPUTING SYSTEM WITH OPENC

Formal modelling of ontologies within Event-B

Spectral Active Clustering of Remote Sensing Images

Scalewelis: a Scalable Query-based Faceted Search System on Top of SPARQL Endpoints

Taking Benefit from the User Density in Large Cities for Delivering SMS

Linux: Understanding Process-Level Power Consumption

A Multi-purpose Objective Quality Metric for Image Watermarking

Syrtis: New Perspectives for Semantic Web Adoption

A Laser Speckle Method for Measuring Displacement Field. Application to Resistance Heating Tensile Test on Steel

A Practical Evaluation Method of Network Traffic Load for Capacity Planning

Texture classification based on the generalized gamma distribution and the dual tree complex wavelet transform

A Voronoi-Based Hybrid Meshing Method

Approximate RBF Kernel SVM and Its Applications in Pedestrian Classification

lambda-min Decoding Algorithm of Regular and Irregular LDPC Codes

Assisted Policy Management for SPARQL Endpoints Access Control

A million pixels, a million polygons: which is heavier?

Real-time tracking of multiple persons by Kalman filtering and face pursuit for multimedia applications

Accurate Conversion of Earth-Fixed Earth-Centered Coordinates to Geodetic Coordinates

RETIN AL: An Active Learning Strategy for Image Category Retrieval

The Connectivity Order of Links

Skyline Matching: A robust registration method between Video and GIS

A Methodology for Improving Software Design Lifecycle in Embedded Control Systems

Quasi-tilings. Dominique Rossin, Daniel Krob, Sebastien Desreux

Natural Language Based User Interface for On-Demand Service Composition

Mokka, main guidelines and future

Transcription:

Gradient-Based time to contact on paracatadioptric camera Fatima Zahra Benamar, Sanaa El Fkihi, Cédric Demonceaux, E. Mouaddib, Driss Aboutajdine To cite this version: Fatima Zahra Benamar, Sanaa El Fkihi, Cédric Demonceaux, E. Mouaddib, Driss Aboutajdine. Gradient-Based time to contact on paracatadioptric camera. IEEE International Conference on Image Processing, ICIP 2013, Sep 2013, Melbourne, Australia. <hal-00830883> HAL Id: hal-00830883 https://hal.archives-ouvertes.fr/hal-00830883 Submitted on 16 Sep 2013 HAL is a multi-disciplinary open access archive for the deposit and dissemination of scientific research documents, whether they are published or not. The documents may come from teaching and research institutions in France or abroad, or from public or private research centers. L archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d enseignement et de recherche français ou étrangers, des laboratoires publics ou privés.

GRADIENT-BASED TIME TO CONTACT ON PARACATADIOPTRIC CAMERA F. Benamar 1,4 S. El Fkihi 2,1 C. Demonceaux 3 E. Mouaddib 4 D. Aboutajdine 1 1 LRIT, Unité associée au CNRST, URAC 29, FSR, Mohammed V-Agdal University, Rabat, Morocco 2 RIITM, ENSIAS, Mohammed V-Souissi University, Rabat, Morocco 3 Le2i, UMR CNRS 6306 University of Burgundy, Le Creusot, France 4 MIS, University of Picardie Jules Verne, Amiens, France fatima.zahra.benamar @ u-picardie.fr ABSTRACT The problem of time to contact or time to collision (TTC) estimation is largely discussed in perspective images. However, a few works have dealt with images of catadioptric sensors despite of their utility in robotics applications. The objective of this paper is to develop a novel model for estimating TTC with catadioptric images relative to a planar surface, and to demonstrate that TTC can be estimated only with derivative brightness and image coordinates. This model, called gradient based time to contact, does not need high processing such as explicit estimation of optical flow and feature detection/or tracking. The proposed method allows to estimate TTC and gives additional information about the orientation of planar surface. It was tested on simulated and real datasets. Index Terms Time to contact, omnidirectional vision, obstacle avoidance, collision detection, mobile robotic. 1. INTRODUCTION The time to contact or time to collision (TTC) is the time available to a robot before reaching an object. When a robot has a translation motion along X-axis, TTC equation is: T T C = X dx dt where X is the distance between the sensor and the obstacle, and dx dt is the sensor velocity. An advantage of the TTC formulation is that it can be obtained from the image plane, without measuring real quantities (distance, velocity,...). Many approaches have been adopted for computing TTC in perspective vision. In optical flow based methods, we cite [9] which estimated TTC using particular sensor based on polar and log-polar representation. Besides, to avoid optical flow explicit computation problems, methods based closed contours have been proposed. Thus, in [1], TTC is estimated by tracking active contours, it requires segmentation, which is a quite difficult task. Another TTC category approaches (1) This work was supported by STIC program and CNRST. is gradient based TTC [4]. They use spatial and temporal image brightness; and avoids high level processing (segmentation, explicit optical flow estimation,...). In this work, we propose to estimate TTC using a catadioptric camera. We may classify the existing works in two axes: (1) TTC estimated on the catadioptric image plane. In our previous work [2], we proposed a TTC formula at each point in the image plane using parabolic projection and adapted optical flow [8]. (2) TTC estimated on the unit sphere. For instance, [3] describes a closed form bound of TTC from local motion field in the case of rigid motion and surface orientation with an arbitrary wide FOV. Recently, [5] estimates TTC using the maximum flow field divergence on the view sphere. In our paper, we propose to compute the TTC directly in the image plane using gradient based TTC. In section 2, we explain the discussed problem. Then, section 3 presents a new formula of TTC for omnidirectional images while the experimental results with synthetic and real sequences are given in section 4. Finally, conclusion is drawn in section 5. 2. MATHEMATICAL FORMULATION The traditional gradient based TTC method of [4] is put forward in case of translation motion along the optical axis towards an arbitrary orientation surface and extended on the geometric modeling in [6]. A planar surface (using perspective projection) in image coordinates (x, y) is: Z 0 Z(x, y) = 1 a x f b y f with f: focal length, Z 0 : distance to the surface along Z-axis (optical axis), and (a, b): slopes in (X, Y ) directions. The brightness constancy constraint equation (BCCE) can be written: UE u + V E v + E t = 0 (3) with (E u, E v, E t ) are the brightness derivatives w.r.t (x, y, t). dx (U, V ) = (α u dt, α v dy dt ) is the optical flow in the image and: (2)

dx dt = xż Z ; dy dt = y Ż (4) Z while (α u, α v ) is a scale conversion mm to pixel in the (x, y) direction. Using eq.(2), eq.(3), eq.(4). The TTC is solution of: G(C + P x + Qy) + E t = 0 (5) where, C = Ż/Z 0 is the inverse of TTC, P = ( a f )C, Q = ( b f )C, G is the radial gradient (α uxe u + α v ye v ). Nonetheless, using paracatadioptric camera, the mirror introduces additional geometric transformations. Indeed, scene points are projected onto the image plane according to a nonlinear model. So, eq.(2), eq.(4) and eq.(5) are not valid. We have to introduce parabolic projection into the surface equation to express the depth function in the image coordinates. Let us consider the case of paracatadioptric sensor; the equation of the parabolic mirror is given by: z p = x2 p + yp 2 h 2. (6) 2h with h is a mirror parameter. Let P (X, Y, Z) be a 3D point. P is projected onto the mirror on the point m p (x p, y p, z p ). Then, m p is projected on the camera plane on m(x, y) (see figure 1). In pixel coordinates, the image point m(u, v) is given by eq.(8) ( (u 0, v 0 ): a principal point coordinates). h X x p = = x X2 +Y 2 +Z 2 Z h Y y p = = y X (7) 2 +Y 2 +Z 2 Z z p = h Z X 2 +Y 2 +Z 2 Z u = α u x + u 0 ; v = α v y + v 0 (8) To define the depth of a surface in a paracatadioptric image, we use eq.(6), eq.(7), and eq.(9). So, a planar surface can be expressed as a function of the image coordinates (x, y): X(x, y) = X 0 1 ( ay x + b(x2 +y 2 h 2 ) )) (10) In the next section, we will define a new equation which links TTC to m, to take into account the mirror distortion. 3. A NEW FRAMEWORK TO ESTIMATE TTC We consider paracatadioptric sensor moving along the X-axis towards an inclined surface. Let P (X, Y, Z) be a fixed obstacle in the environment space R 3. In the case of the perspective camera, TTC is defined by the time it would take for the object to cross the image plane. But, in the case of paracatadioptric camera, we define the collision condition when the obstacle crosses the parabolic mirror focus. The contact point on the plane surface is (X 0, 0, 0) T and its TTC equation is: T T C = X 0 Ẋ We set C X = Ẋ X. Eq.(11) can be written as: (11) T T C = 1 C = X 0 X.Ẋ X = X 0 X. 1 C X (12) Hence, from eq.(10) and eq.(12), we obtain: C X = C.(1 (a y x + + y 2 h 2 bx2 )) (13) The key idea of the new method is the exploitation of the BCCE. We compute the image coordinate derivatives x and y from eq.(7), we consider α u = α v = α, and we set: A = X 2 + Y 2 + Z 2. Thus, we obtain: { ẋ = xcx ( x h (XA 1/2 ) 1) ẏ = C X ( xy (14) h (XA 1/2 )) From eq.(6) and eq.(7); and z p + h > 0, we have: XA 1/2 = x (z p + h) = x 2 + y 2 + h 2 (15) Therefore, using (13), ẋ and ẏ are written as: Fig. 1. Configuration of paracatadioptric projection A given planar surface (see figure 1) is defined by: X(Y, Z) = X 0 + ay + bz (9) where a and b give the planar surface slopes with respect to X-axis, while X 0 is the distance robot-surface along X-axis. { ẋ = xc.(1 (a y x + b x2 +y 2 h 2 2x ))( 2 x 2 +y 2 +h 1) 2 ẏ = yc.(1 (a y x + b x2 +y 2 h 2 2x ))( 2 x 2 +y 2 +h ) 2 (16) From eq.(3), eq.(8) and eq.(16), we obtain the following equation in pixel coordinates: 2(u u 0 ) 2 S = C[ (u u 0 ) 2 + (v v 0 ) 2 + (αh) 2 G (u u 0 )E u ]+

2(u u 0 )(v v 0 ) P [ (u u 0 ) 2 + (v v 0 ) 2 + (αh) 2 G (v v 0 )E u ]+ +Q[ (u u 0) (u u 0 ) 2 + (v v 0 ) 2 (αh) 2 α.h (u u 0 ) 2 + (v v 0 ) 2 + (αh) 2 G (u u 0 ) 2 + (v v 0 ) 2 (αh) 2 E u ] + E t = 0 (17) 2.α.h with P = ac, Q = bc and G = ((u u 0 )E u + (v v 0 )E v ). Eq.(17) has to estimate: C (inverse of TTC), P and Q. To solve this equation, we proceed as follows: We consider a region of interest in the image. For all points in, we write its corresponding eq.(17); 4.1. Synthetic sequence Our synthetic sequence is generated by Pov-ray software 1. Virtual paracatadioptric sensor is moving with uniform unidirectional motion towards X-axis (see figure 2) and generates images of 200 200 pixels. For all sequences, we have: α=40, h=2.5 and (u 0, v 0 )=(100, 100). We denote by (β, θ) the planar surface inclination angles around (Z-axis,Y-axis); where and b= tan(θ). Figure 3 shows TTCs (red line) of four sequences with inclination angles (β, θ) = (0, 0 ), ( 10, 5 ), ( 20, 10 ) and ( 30, 20 ). The ground truth is illustrated in dashed green line. In this figure, it is shown that our algorithm gives interesting results, TTC values dropped linearly as expected. a= tan(β) cos(θ) We construct the system of the obtained equations composed of as many equations as points in the ; We solve this linear system by the least squares method. We can rewrite eq.(17) as: HΘ = B (18) Fig. 2. Synthetic image of inclined planar surface with β = 30 and θ = 20 ; (left to right) frame n 1, 100, and 170. where Θ = C P Q ; H = [H 1 H 2 H 3 ]; B = [ E t ] (19) H 1 = H 2 = 2(u u 0 ) 2 (u u 0 ) 2 + (v v 0 ) 2 + (αh) 2 G (u u 0 )E u (20) 2(u u 0 )(v v 0 ) (u u0) 2 + (v v 0 ) 2 + (αh) 2 G (v v 0 )E u H 3 = (u u 0) (u u 0 ) 2 + (v v 0 ) 2 (αh) 2 α.h (u u 0 ) 2 + (v v 0 ) 2 + (αh) 2 G (21) (u u 0 ) 2 + (v v 0 ) 2 (αh) 2 E u (22) 2.α.h Remark, if a = b = 0, that is, if the planar surface is perpendicular to the X-Axis, the eq.(17) becomes: CH 1 = E t (23) Based on spatial gradients, temporal gradient, image coordinates and constants α, h and (u 0, v 0 ), our approach allows to estimate TTC and the orientations of a planar surface a and b. 4. EXPERIMENTAL RESULTS Here, we will evaluate the new TTC proposal for perpendicular and inclined surfaces, in both of synthetic and real images. Fig. 3. Estimation of TTC on synthetic sequences for different inclinations of the planar surface. The commonly used performance measure between an estimated vector (V e ) and a real one (V r is angular error (AE): AE = arccos( < V e, V r > V e V r ) (24) Let us consider the vector ( 1, a, b), figure 4 shows the evaluation results of a and b for the considered synthetic sequences above. Because of the bias introduced by b, AE is larger at the beginning of the sequence. We can note some perturbations at the beginning and at the end of the sequence. This phenomena is well known in optical flow estimation and due to temporal aliasing. The latter can be overcome using a multiresolution framework [4] after alteration for paracatadioptric images[7]. 1 www.povray.com

stop before collision at a security distance of 21 frames. Fig. 4. Angular error (AE) in degree for synthetic sequence. From left to right: AE corresponds to sequence with (β, θ) = (0, 0 ), ( 10, 5 ), ( 20, 10 ) and ( 30, 20 ). 4.2. Real sequence We consider a paracatadioptric sensor embedded on the Pioneer robot. It generates images of 480 600 pixels for all sequences. The camera calibration was conducted with the Hyscac toolbox 2 and the intrinsic parameters are: (u 0, v 0 )=(237.13, 316.04), h=1, and α=156.39. The sensor moves along X-axis towards a perpendicular wall (rigid surface) (see figures 5.(a) and (b)). It has also acquired sequences with inclined wall (see figures 5.(c) and (d)). An online demo of this work is available at 3. Fig. 6. Estimation of TTC (red line) on real sequence in the case of perpendicular planar surface. Fig. 7. Estimation of T T C (red line) on real sequence in case of two different inclined surfaces. Fig. 5. Images of two real sequences; (a) and (c) are the first frames while (b) and (d) are the last ones. (a) and (b): Images with perpendicular planar surface. (c) and (d): Images with inclined planar surface. Figures 6 and 7 show the TTC computed with our proposal algorithm. As the robot speed is constant during the experiment, the TTC should decrease linearly (dashed green line in figures 6 and 7). TTC curve decreases linearly from the first to the last frame as expected. The robot is able to 2 http://www.hyscas.com 3 http://home.mis.u-picardie.fr/ fatima 5. CONCLUSION AND FUTURE WORK In this paper we proposed a novel framework to estimate TTC in omnidirectional images for the planar surface of an arbitrary orientation. The introduced approach gives additional information about the planar surface. Focus was on gradient based methods because they are simple and fast, and they avoid high level processing in computer vision, thus making them a good candidate for real time implementations. The illustrated results show that our approach has the potential to be effective in complex real-world scenarios. In fact, the work can be expanded in several directions: it could be interesting to generalize it to a general motion towards X and Y axes.

6. REFERENCES [1] G. Alenyà, A. Nègre, and J. L. Crowley. A comparison of three methods for measure of time to contact. In Proceedings of the 2009 IEEE/RSJ international conference on Intelligent robots and systems, IROS 09, pages 4565 4570, Piscataway, NJ, USA, 2009. IEEE Press. [2] F.Z. Benamar, C. Demonceaux, S. EL. Fkihi, E. Mouaddib, and D. Aboutajdine. Time to contact estimation on paracatadioptric cameras. In Pattern Recognition (ICPR), 2012 21th International Conference on, number 3602-3605, November 2012. [3] C. Colombo. Time to collision from first-order spherical image motion. Robotics and Autonomous Systems, 31(1-2):5 15, 2000. [4] B.K.P. Horn, Y. Fang, and I. Masaki. Hierarchical framework for direct gradient-based time-to-contact estimation. In In IEEE Intelligent Vehicles Symposium, pages 1394 1400, Shaanxi, China, June 2009. [5] C. McCarthy and N. Barnes. A unified strategy for landing and docking using spherical flow divergence. IEEE Transactions on Pattern Analysis and Machine Intelligence, 34:1024 1031, 2012. [6] F. G. Meyer. Time-to-collision from first-order models of the motion field - robotics and automation. IEEE Transactions on, 10(6), 1994. [7] Y. Mochizuki and A. Imiya. Multiresolution optical flow computation of spherical images. In Computer Analysis of Images and Patterns, pages 348 355, January 2011. [8] A. Radgui, C. Demonceaux, E. Mouaddib, M. Rziza, and D. Aboutajdine. An adapted lucas-kanade s method for optical flow estimation in catadioptric images. In In OM- NIVIS 2008, the Eighth Workshop on Omnidirectional Vision, Camera Networks and Non-classical Cameras, Marseille, France, October 2008. [9] M. Tistarelli and G. Sandini. On the advantages of polar and log-polar mapping for direct estimation of time-toimpact from optical flow. IEEE Trans. on PAMI, 15:401 410, 1992.