Investigation of Methods for Target State Estimation Using Vision Sensors

Size: px
Start display at page:

Download "Investigation of Methods for Target State Estimation Using Vision Sensors"

Transcription

1 AIAA Guidance, Navigation, and Control Conference and Exhibit 5-8 August 25, San Francisco, California AIAA Investigation of Methods for Target State Estimation Using Vision Sensors Gregory F. Ivey and Eric N. Johnson Georgia Institute of Technology, Atlanta, GA, 3332 Identifying object location and orientation are important tasks in target tracking and obstacle avoidance. This paper proposes an algorithm for identifying target location, size, and orientation without any prior knowledge based on machine vision. The algorithm uses target measurements taken from a video camera to update estimates of the target. Two forms of the Kalman filter were compared as methods for estimating the targets states. The first, the extended Kalman filter (EKF), uses a Taylor series expansion about the current state to estimate the nonlinear measurement update. The second, the square-root unscented Kalman filter (SRUKF), a form of the unscented Kalman filter (UKF), uses perturbed sigma points to estimate the mean and covariance of the measurement update. The algorithm was tested using periodic vehicle motion and random target location, orientation, and area. Both Kalman filter methods converged on the targets 3-D position, orientation, and area but required an irregular trajectory to accurately estimate the orientation and area. The estimators were simulated using identical parameters to compare them based on equivalent conditions. This produced similar results from the EKF and SRUKF and therefore neither showed a significant improvement over the other. Nomenclature B Body fixed reference frame C Camera fixed reference frame I Inertial reference frame x process state z measurement state C vector from inertial frame to camera, ft D vector from camera to center of target, [d x, d y, d z ] T, ft n target unit normal vector, [n x, n y, n z ] T, ft T vector from inertial frame to center of target, [X, Y, Z] T, ft A Area of target, ft 2 f focal length of the camera P A area of target in image plane, px 2 P Y horizontal position of target center in image plane, px P Z vertical position of target center in image plane, px P n pseudo-measurement of normal vector s magnitude px pixels Subscripts i k x y i th target k th iteration x-coordinate y-coordinate Graduate Research Assistant, Student Member AIAA, gregory ivey@ae.gatech.edu. Lockheed Martin Assistant Professor of Avionics Integration, Member AIAA, Eric.Johnson@aerospace.gatech.edu. of 2 Copyright 25 by the authors. Published by the American American Institute of Aeronautics Institute of and Aeronautics Astronautics, and Inc., Astronautics with permission.

2 z Symbols ν ζ z-coordinate zero mean measurement noise vector zero mean process noise vector I. Introduction Target tracking and obstacle avoidance have been prime tasks for autonomous unmanned aerial vehicles (UAV s). The mission capabilities of autonomous UAV s depend on how well they maintain surveillance on a desired target while avoiding surrounding objects. Vehicle attitude and location can be determined using an inertial measurement unit (IMU) and global positioning system (GPS) respectively, but determining the location and orientation of the vehicle in relation to its environment must be accomplished with other sensors. Machine vision has been used as a viable method to sense the environment around the UAV. 3 Monocular machine vision is the base of the algorithm presented in this paper and will be used for detecting the environment for either obstacle avoidance or target tracking. Exact knowledge of the surrounding objects location, size, and orientation allows the UAV to not only determine its position but also any new object s position and size relative to the known surroundings. An algorithm can be proposed that would constantly update a dynamic database of object information as the UAV moves through an environment. This algorithm is founded on the ability of the image processor and target estimator to create such a database. New advances in machine vision and image processing allow framegrabbers to capture still images from the camera and send them to the image processor, which then can determine the targets in each image. 4 The area and location of the center of each target in the image plane can be calculated and a database of all the targets information compiled. Figure displays the process of grabbing an image frame from the video camera, sending it to the image processor, and creating a database of target measurements. With a database of the targets sizes and locations in the 2-D image plane, an a priori estimate of the orientation, size, and location of the targets relative to the camera can be made. Kalman filtering can be used with an initial estimate of the Camera Framegrabber Image at time k Image Processor Identified Targets 2... i... N Database Figure. Framegrabber and image processing sequence distance from the camera to converge on the actual size, location, and orientation of each target. Since the relationship between the 2-D and the 3-D image coordinates is nonlinear, this paper investigates the extend Kalman filter (EKF) and the square-root unscented Kalman filter (SRUKF), a form of the unscented Kalman filter (UKF), as methods for estimation. The next section describes the system and measurement models used with the filtering approaches. Section III and IV detail the nuances of the EKF and SRUKF estimators, respectively. The design of the algorithm is outlined in section V with explanations on the solution of the correspondence problem. Simulations and the results are discussed in section VI. P Y P Z P A II. System Model The process model for the system is based on seven states for the i th target: target location (X i,y i,z i ), target orientation (n xi,n yi,n zi ), and target area (A i ). The orientation of the target is based on a unit vector normal to the plane of the target. Since the target is in view of the camera, the normal vector is assumed 2 of 2

3 to be pointing towards the camera. The complete process and measurement states are listed as Xi Yi PYi Zi PZi xi = and zi =, nxi PAi ny i P~ni n zi Ai () where PAi is the area and (PY i,pz i ) is the coordinate of the center of each target in the image plane in pixels. The image processor calculates these values for each object and compiles a database. The fourth measurement, P~ni, is a construct of the normal vector s magnitude. This pseudo-measurement is computed during estimation and it ensures that the magnitude remains constant. During calculation of these states, the position vectors of the vehicle and camera are assumed to be constant for each iteration of the algorithm, that way the vehicle state dynamics can be modeled separately from the dynamics of the targets states. For simplicity, the targets being tracked are assumed to be stationary. Since the states being estimated, target position, orientation, and size, are constant in the inertial reference frame, the process model updating these states is an autonomous one of the form (2) xk = xk. The measurements, however, are not related to the states by such a simple linear equation. Equations (3) (7) describe the nonlinear relationship between the measurement values and the state variables. ~i D PYi PZ i PAi P~ni ~ = T~i C f dy = dxi i f = dz dxi i 2 f = Ai dxi = (3) (4) (5) ~i ~ni D ~ ik k~ni k kd (6) 2 nxi + n2yi + n2zi 2 A negative sign is placed in front of ~ni in Eq. (6) to obtain the acute angle between the target normal vector (~ni ) and the vector from the camera to the center of the target ~ i ), since these vectors point in opposite directions. The (D curvature of the camera lens is assumed to be negligible so targets are projected onto a flat image plane. Figure 2 shows an example of the camera in view of two targets and the vectors used in defining the measurement equations. III.! (7) B C C I Extended Kalman Filter The extended Kalman filter is a linearization approach based on a Taylor series expansion about the current state.5 The expansion requires determining the derivatives of the nonlinear equations and formulating Jacobian matrices from the derivatives. The process state equations at any instance, given in Eq. (2), are linear, whereas the measurements are governed by nonlinear equations of the form zk = h(xk, vk ), given in Eqs. (3) - (7). n D D2 Target T T2 n2 Target 2 Figure 2. Diagram showing position vectors of the targets (green), position vector of the camera (blue), and vectors from the camera to the center of the targets (red). I, B, and C are the inertial, body, and camera reference frames, respectively. 3 of 2

4 A. Process Update Since the process update is autonomous, no derivatives need be calculated for the process model. Equations (8) and (9) describe the process model. B. Measurement Update ˆx k = ˆx k (8) P k = P k + Q k (9) The measurement update requires the determination of the Jacobian, H k = h (Ref. 6), () x k xk =x k in every iteration. The measurement update can be derived as follows: 5 K k = P k H T k (H k P k H T k + R k ) () ˆx k = ˆx k + K k(z k h(ˆx k )) (2) P k = (I K k H k )P k (3) The measurement and process noise covariances, R k and Q k respectively, are assumed to be constant over all the iterations and are given an initial value at the beginning of the algorithm. C. Variable Process and Measurement Noise Covariances The measurement and process noise covariances may also be dynamic to help improve the convergence properties of the EKF. In Ref. 7, Boutayeb and Aubry developed an extended Kalman observer which alters R k and Q k based upon Lyapunov theory. Their paper may be used as a method for improving the convergence of the target tracking EKF algorithm. They utilize the measurement residual e k = z k h(ˆx k ) to adjust the process noise covariance. They propose choices of and Q k = γe T ke k I 7 + δi 7 (4) R k+ = λh k+ P k+ H T k+ + δi 3 (5) where γ, δ, and λ are constants chosen such that γ, δ >, and λ >. The equations for the measurement and process noise covariances are derived by determining bounds on linear matrix inequalities which satisfy a decreasing Lyapunov function. Proof of these choices for R and Q can be found in Ref. 7. IV. Unscented Kalman Filter The UKF was found by Julier, Uhlmann, and Durrant-Whyte in Ref. 8 to be easier to implement than the EKF because it does not require calculation of Jacobian matrices. It was also found by Wan and van der Merwe in Ref. 9 to have better accuracy than the EKF and equal complexity. These results prompted investigation into its use for the target tracking algorithm. The unscented Kalman filter is based on the unscented transform (UT) and does not require linearization to handle nonlinear equations. The UT approximates the state estimate and covariance by creating weighted points and applying them to the nonlinear function to create the measurement estimate and covariance. The sigma points are chosen such that the their sample mean and covariance are equal to the state estimate and covariance. 8 A more computationally efficient form of the UKF is the Square-Root UKF (SRUKF). The SRUKF is fundamentally the same as the UKF but it uses a Cholesky factorization of the error covariance matrix to eliminate some of the computations. In Ref., van der Merwe and Wan found the SRUKF guarantees a non-negative definite state covariance matrix and requires less numerical computations. The SRUKF was implemented as a method of estimation to compare with the EKF. 4 of 2

5 A. Sigma Point Weights An outline of the SRUKF algorithm is described in Eqs. (6) - (32). These equations are based on the SRUKF algorithm presented in Ref.. An augmented state vector and covariance matrix are used in the SRUKF algorithm where the initial augmented state vector is given by Eq. (6) and the corresponding augmented covariance matrix is given by Eq. (7). The process noise mean vector (ζ) and the measurement noise mean vector (ν) are concatenated to the state vector. The lower-triangular Cholesky factors of the covariance matrices P, Q, and R are grouped into an augmented matrix, S a, where P = S x Sx T, Q = S ζ Sζ T, and R = S ν Sν T. ˆx a = [ˆx ζ ν] T (6) S x S a = S ζ (7) S ν Wan and van der Merwe propose in Ref. 9 the following equations for choosing the sigma point weights. The constant L is the dimension of the augmented state vector and λ is found using λ = α 2 (L + κ) L, where α is chosen to be small and positive and κ is usually chosen equal to. The optimal value of 2 is usually assigned to β. W (m) = λ/(l + λ) W (c) = λ/(l + λ) + ( α 2 + β) W (m) (8) j = /(2(L + λ)) j =,..., 2L W (c) j = /(2(L + λ)) j =,..., 2L B. Process Update The process update equations for the SRUKF are given in Eqs. (9) - (23). The 2L + sigma points are computed in Eq. (9), where η = L + λ. The augmented sigma point matrix is formed by the concatenation of the state sigma point matrix (X x ), the process noise sigma point matrix (X ζ ), and the measurement noise sigma point matrix (X ν ), such that X a = [ (X x ) T (X ζ ) T (X ν ) T] T. Xk a = [ˆx a k ˆx a k + ηsk a ˆx a k ηsk a ] X x (9) k = X x k + X ζ k (2) ˆx k = 2L j= {[ Sx k = qr W (m) j Xj,k x (2) W (c) S x k = cholupdate (X x ) ]} :2L,k ˆx k (22) } {S xk, (X,k x ˆx (c) k ), W (23) The functions qr{ } and cholupdate{ } are linear algebra operators that correspond to a QR decomposition and a rank Cholesky update, respectively. qr{a} returns the lower triangular part of R from the economy QR decomposition, A = QR. cholupdate{s, U, ±ω} returns the rank update or downdate (depending on the sign of ω) of the lower triangular Cholesky factor S by ωu. If U is a matrix, then N consecutive rank updates or downdates are performed by the columns of U, where N is the number of columns of U. C. Measurement Update The measurement update equations are given by Eqs. (24) - (32). The nonlinear measurement equations, h( ), are given in Eqs. (3) - (7). The linear algebra operator, /, is a pseudoinverse operation that uses pivoting. Z k = h ) (Xk x + Xk ν (24) 5 of 2

6 ẑ k = 2L j= W (m) j Z j,k (25) {[ ( ) ]} S zk = qr W (c) Z :2L,k ẑ k (26) { ( ) } S zk = cholupdate S zk, Z,k ẑ k, W (c) (27) P xk z k = 2L j= ) ( ) W (c) j (Xj,k x ˆx k Z j,k T ẑ k (28) K k = ( P xk y k /S T z k ) /Szk (29) ˆx k = ˆx k + K k(z k ẑ k ) (3) U = K k S zk (3) S xk = cholupdate { S x k, U, } (32) V. Algorithm Design The target tracking algorithm begins by taking the database of target information from the image processor and then using the data to create an initial estimate of the 3-D position of each target. The initial estimates are based on fixed values for X, n x, n y, and n z assigned to every new target. The other three state variables are calculated using these values and the measurement data from the database with Eqs. (3) - (6). This is a crude method and a thorough investigation into determining the optimal initial estimates will need to be done. The initial estimates are then sent to the estimator which will use either the standard EKF or the SRUKF to update the states. The next iteration of the algorithm receives updated measurements from the image processor. These measurements are correlated to the correct target estimates with the use of the statistical z-test. 2, 3 If the measurements correspond to an existing estimate, then the measurements and a priori state estimates are sent to the estimator to be updated. If a set of measurements does not match any of the estimates, then it is assumed that they are for a new target. A state estimate is added for the new target and computed with Eqs. (3) - (6) based on the fixed values for X, n x, n y, n z, and the measurements. A block diagram outlining the algorithm is shown in figure 3. The statistical z-test was implemented to check the correlation between the target s measurements and the state estimates stored in the database. The test can be defined with residuals as where the linear form of the residual is defined as 2... i... N Database at k Initialize First Image Xi P Y P Z P A Z i k _ Xi k = Xi k- _ Z k Yes _ Xi k i = h( ) _ Z i k Correspondence z-test Is target New? Z i k No _ Xi k Estimator Xi k Figure 3. Block diagram of the target tracking algorithm. Z = e T (E[e e T ]) e (33) e = z H ˆx. (34) Knowledge of the estimation and measurement errors given respectively in Eqs. (35) and (36), can be used to determine the known covariances R and P : x = ˆx x (35) z = z Hx, (36) R = E[ z z T ] (37) 6 of 2

7 P = E[ x x T ]. (38) The error equations, (35) and (36), can be substituted in Eq. (34) and the expected value, E[e e T ], transformed using (38) and (37) into E[e e T ] = R + HP H T. (39) Equation (39) is substituted into (33) to obtain the Z value Z = e T (R + HP H T ) e. (4) This equation is easily implemented in the EKF estimator and an equivalent form for the SRUKF estimator can be developed likewise. The Kalman gain is defined for the EKF in Eq. () as K = P H(HP H T + R) and can be compared to the Kalman gain defined for the UKF in Ref. 9 as K = P xz Pz. Based on this comparison, the Z value for the UKF can be assumed to be Z = e T (P z ) e. (4) The lower triangular Cholesky factorization, P z = S z S T z, and the pseudo-inverse operator, /, can be used with Eq. (4) to efficiently compute the Z value for the SRUKF. The SRUKF Z value equation is given by Z = e T [(I/S T z )/S z ] e. (42) The magnitude of Z depends on the covariance matrices and will be small when P and R are large. Therefore, even if the residual is large, great uncertainties in the estimates and measurements will cause the value of Z to be small. The z-test correlates the measurements and estimates by comparing the magnitude of Z to a critical value. If Z is larger than the critical value then they do not correspond. VI. Simulations and Results The simulations were performed with the expectation that later flight tests on the Georgia Tech GTMax research UAV 4 would be performed. Some base assumptions were made to simplify the simulations. The GTMax has a Hz processing speed, so the algorithm was simulated at Hz to replicate this. The camera was assumed to be rigidly attached to the nose of the vehicle and the camera, body, and inertial frames all aligned. The vehicle was simulated to perform periodic motions with little body roll, which was found to incur faster estimator convergence. No rolling, yawing, or pitching was assumed so the reference frames all remained aligned and any transformation to account for these rotations could be neglected. The vehicle started at coordinate (,, -3) ft in the inertial reference frame and traveled on a circular path that spanned all three directions. The trajectory had an amplitude of 2 ft in the x, y, and z directions and a periodic speed of.5π rad/s. This would result in the vehicle moving on the circular path at 27 ft/s. The vehicle started on one circular path for 2 seconds then changed direction and continued on a similar path for 7 seconds. The trajectory of the 9 second flight is shown in Figure 4. Zero mean noise with a standard deviation of. ft was added to the x, y, and z positions of the vehicle to simulate position disturbances. A set of targets with random locations, orientations, and areas was created and their image plane measurements simulated. A zero mean, Gaussian measurement noise vector was applied to the measurement calculations. The noise for P Y, P Z, and P A had standard deviations of 3 px, 3 px, and. px 2, respectively. For parity, the same noise vector was used in each estimator. The image plane was assumed to have a width of 74 px and a height of 48 px. The image plane was normalized by half the width so the corners of the image plane would have the following coordinates: ( , ), ( , ), ( , ), and ( 352 ) 352, 24. (43) 352 A 7 degree field of view was assumed for the camera and the image plane focal length was calculated using f = tan(35 ) (Ref. 4). (44) The algorithm initialized each target estimate with X =, n x = 3, n y =, and n z =. These states along with the initial measurements (P Y, P Z, P A ) were used to calculate the other states (Y, Z, A) 7 of 2

8 using Eqs. (3) - (6). The EKF and the SRUKF process noise and measurement noise covariance matrices, Q and R, were set to the values shown in Eqs. (45) and (46), respectively. The same initial estimate error covariance matrix, P, shown in Eq. (47), was also used for both the SRUKF and EKF. The filters were tuned with identical parameters for an equal comparison. Q = diag { } (45) { ( ) 2 ( ) 2 ( ) } R = diag (.) 2 (46) P = diag { } (47) The results for three random targets are shown in the Appendix in Figures 5-4. The actual values for the states of each target are shown Table. Figures 5-7 show the EKF and SRUKF estimates of the states compared to the actual values for each target. The error of the SRUKF estimate and EKF estimate for each state are shown in Figures 8-4. Plus and minus two standard deviation bounds of the state error are also included for comparison. The EKF and SRUKF filters converged on the X, Y, and Z positions of the targets with similar performances. Figures 8 - show the estimator errors compared to two standard deviations of the error. The SRUKF error for X, Y, and Z remained within two standard deviations. The EKF error for X, Y, and Z deviated from the two standard deviation bounds but always returned and was identical to the SRUKF estimate by the end of the simulation. Estimating the attitude and area required the vehicle view the targets from different locations and the change in circular trajectories at 2 seconds helped the filters converge on the correct values. This is apparent in Figure 6, where the estimates of n x, n y, n z, and A changed dramatically after 2 seconds when the vehicle changed trajectory. The EKF error for n x, n y, and n z in Figures, 2, and 3 remained within the two standard deviation bounds after the trajectory changed at 2 seconds. The SRUKF error for n y and n z had similar behavior, but the error for the n x state of Target 3 in Figure had divergent behavior and exceeded the two standard deviation bound at about 4 seconds. The SRUKF error of the area for Target and 2 in Figure 4 stayed within the two standard deviation bounds but diverged for Target 3. The EKF error of A for Target 2 and 3 converged into the two standard deviation bounds but the error for Target did not make it under the bounds by the end of the simulation. Table. Actual state values for three random targets given in the inertial reference frame. Target X (ft) Y (ft) Z (ft) n x (ft) n y (ft) n z (ft) A (ft 2 ) After the values are assigned, the normal vector is normalized so that it is a unit normal vector. VII. Conclusions The target tracking algorithm can quickly and accurately estimate the position of a target. It, however, has difficulty converging on the orientation and area and requires the vehicle obtain different views of each target and navigate an irregular trajectory. Further research needs to be done on the vehicle motion and how this affects the convergence of the algorithm. This could be used to command the vehicle to perform specific motions to help the algorithm estimate the correct values. A thorough study of the optimal initial state estimates also needs to be done. The SRUKF and EKF estimators performed similarly when given identical parameters. Both converged on the states but each had a problem estimating the attitude and area of the targets. The SRUKF estimator error tended to diverge for the states A and n x, whereas the EKF took longer than the SRUKF to converge on the area state for Target. It is possible to independently tune each filter for better estimator performances, but the identical tuning parameters chosen caused similar performances and the results show no significant improvement of one estimation method over the other. 8 of 2

9 Appendix (a) (b) Figure 4. Plots of the camera trajectory. (a) The X,Y, and Z camera position versus time. (b) The 3-D view of the camera trajectory (red) and the target locations (blue s). The start position is the green square and the end position is the magenta square. The vehicle travels on circular paths and changes its trajectory at 2 seconds, resulting in the two circles. 9 of 2

10 X Position, ft Y Position, ft Z Position, ft n x Component, ft n y Component, ft n z Component, ft 2 Target Area, ft Time, s Figure 5. Plots showing the states of target. The Actual state value is the solid red line, the EKF estimate is the dashed blue line, and the SRUKF estimate is the solid green line. of 2

11 X Position, ft Z Position, ft n x Component, ft n y Component, ft n z Component, ft Y Position, ft 9 Target Area, ft Time, s Figure 6. Plots showing the states of target 2. The Actual state value is the solid red line, the EKF estimate is the dashed blue line, and the SRUKF estimate is the solid green line. of 2

12 X Position, ft Y Position, ft Z Position, ft n x Component, ft n y Component, ft n z Component, ft 3 2 Target Area, ft Time, s Figure 7. Plots showing the states of target 3. The Actual state value is the solid red line, the EKF estimate is the dashed blue line, and the SRUKF estimate is the solid green line. 2 of 2

13 2 Target 2 X Position Error and Two Standard Deviations, ft 2 Target Target 3 2 Time, s Figure 8. Plots showing, for the target X position, EKF estimate error (dashed blue line), SRUKF estimate error (solid green line), plus and minus two standard deviations of the EKF error (dashed magenta line), and plus and minus two standard deviations of the SRUKF error (solid cyan line) for Targets, 2, and 3. 3 of 2

14 Target.5.5 Y Position Error and Two Standard Deviations, ft.5.5 Target 2 Target Time, s Figure 9. Plots showing, for the target Y position, EKF estimate error (dashed blue line), SRUKF estimate error (solid green line), plus and minus two standard deviations of the EKF error (dashed magenta line), and plus and minus two standard deviations of the SRUKF error (solid cyan line) for Targets, 2, and 3. 4 of 2

15 Target.5.5 Z Position Error and Two Standard Deviations, ft.5.5 Target 2 Target Time, s Figure. Plots showing, for the target Z position, EKF estimate error (dashed blue line), SRUKF estimate error (solid green line), plus and minus two standard deviations of the EKF error (dashed magenta line), and plus and minus two standard deviations of the SRUKF error (solid cyan line) for Targets, 2, and 3. 5 of 2

16 .2 Target.. Normal Vector X Component Error and Two Standard Deviations, ft Target Target Time, s Figure. Plots showing, for the target normal vector X component (n x), EKF estimate error (dashed blue line), SRUKF estimate error (solid green line), plus and minus two standard deviations of the EKF error (dashed magenta line), and plus and minus two standard deviations of the SRUKF error (solid cyan line) for Targets, 2, and 3. 6 of 2

17 .4 Target.2.2 Normal Vector Y Component Error and Two Standard Deviations, ft Target Target Time, s Figure 2. Plots showing, for the target normal vector Y component (n y), EKF estimate error (dashed blue line), SRUKF estimate error (solid green line), plus and minus two standard deviations of the EKF error (dashed magenta line), and plus and minus two standard deviations of the SRUKF error (solid cyan line) for Targets, 2, and 3. 7 of 2

18 .4 Target.2.2 Normal Vector Z Component Error and Two Standard Deviations, ft Target Target Time, s Figure 3. Plots showing, for the target normal vector Z component (n z), EKF estimate error (dashed blue line), SRUKF estimate error (solid green line), plus and minus two standard deviations of the EKF error (dashed magenta line), and plus and minus two standard deviations of the SRUKF error (solid cyan line) for Targets, 2, and 3. 8 of 2

19 2 Target 2 Area Error and Two Standard Deviations, ft 2 Target Target 3 2 Time, s Figure 4. Plots showing, for the target area, EKF estimate error (dashed blue line), SRUKF estimate error (solid green line), plus and minus two standard deviations of the EKF error (dashed magenta line), and plus and minus two standard deviations of the SRUKF error (solid cyan line) for Targets, 2, and 3. 9 of 2

20 References Saripalli, S., Montgomery, J. F., and Sukhatme, G. S., Vision-based Autonomous Landing of an Unmanned Aerial Vehicle, Proceedings of the 22 IEEE International Conference on Robotics & Automation, Washington, DC, May 22, pp Langelaan, J. and Rock, S., Navigation of Small UAVs Operating in Forests, AIAA Guidance, Navigation, and Control Conference and Exhibit, Providence, Rhode Island, August Sinopoli, B., Micheli, M., Donato, G., and Koo, T. J., Vision Based Navigation for an Unmanned Aerial Vehicle, Proceedings of the 2 IEEE International Conference on Robotics & Automation, Seoul, Korea, May 2, pp De Wagter, C., Proctor, A. A., and Johnson, E. N., Vision-Only Aircraft Flight Control, Digital Avionics Systems Conference, 23. DASC 3. The 22nd, Vol. 2, October 23, pp. 8.B Welch, G. and Bishop, G., An Introduction to the Kalman Filter, Tr 95-4, Computer Science, University of North Carolina at Chapel Hill, Chapel Hill, NC, Watanabe, Y., Johnson, E. N., and Calise, A. J., Optimal 3-D Guidance from a 2-D Vision Sensor, AIAA Guidance, Navigation, and Control Conference and Exhibit, Providence, Rhode Island, August Boutayeb, M. and Aubry, D., A Strong Tracking Extended Kalman Observer for Nonlinear Discrete-Time Systems, IEEE Transactions on Automatic Control, Vol. 44, No. 8, 999, pp Julier, S., Uhlmann, J., and Durrant-Whyte, H. F., A New Method for the Nonlinear Transformation of Means and Covariances in Filters and Estimators, IEEE Transactions on Automatic Control, Vol. 45, No. 3, March 2, pp Wan, E. A. and van der Merwe, R., The Unscented Kalman Filter for Nonlinear Estimation, Adaptive Systems for Signal Processing, Communications,and Control Symposium 2. AS-SPCC. The IEEE 2, October 2, pp van der Merwe, R. and Wan, E., The Square-Root Unscented Kalman Filters for State and Parameter-Estimation, 2 IEEE International Conference on Acoustics, Speech, and Signal Processing, 2.Proceedings.(ICASSP )., Salt Lake City, Utah, May 2. van der Merwe, R., Wan, E., and Julier, S., Sigma-Point Kalman Filters for Nonlinear Estimation and Sensor-Fusion - Applications to Integrated Navigation, AIAA Guidance, Navigation, and Control Conference and Exhibit, Providence, Rhode Island, August Hayter, A. J., Probability and Statistics For Engineers and Scientists, Duxbury, Thomas Learning, Inc., 2nd ed., Hines, W. W., Montgomery, D. C., Goldsman, D. M., and Borror, C. M., Probability and Statistics in Engineering, John Wiley and Sons, Inc., 4th ed., Johnson, E. N. and Schrage, D. P., The Georgia Tech Unmanned Aerial Research Vehicle: GTMax, AIAA Guidance, Navigation, and Control Conference and Exhibit, Austin, Texas, August of 2

POTENTIAL ACTIVE-VISION CONTROL SYSTEMS FOR UNMANNED AIRCRAFT

POTENTIAL ACTIVE-VISION CONTROL SYSTEMS FOR UNMANNED AIRCRAFT 26 TH INTERNATIONAL CONGRESS OF THE AERONAUTICAL SCIENCES POTENTIAL ACTIVE-VISION CONTROL SYSTEMS FOR UNMANNED AIRCRAFT Eric N. Johnson* *Lockheed Martin Associate Professor of Avionics Integration, Georgia

More information

Vision-Aided Inertial Navigation for Flight Control

Vision-Aided Inertial Navigation for Flight Control JOURNAL OF AEROSPACE COMPUTING, INFORMATION, AND COMMUNICATION Vol. 2, September 2005 Vision-Aided Inertial Navigation for Flight Control Allen D. Wu, Eric N. Johnson, and Alison A. Proctor Georgia Institute

More information

Vision-Aided Inertial Navigation for Flight Control

Vision-Aided Inertial Navigation for Flight Control AIAA Guidance, Navigation, and Control Conference and Exhibit 15-18 August 5, San Francisco, California AIAA 5-5998 Vision-Aided Inertial Navigation for Flight Control Allen D. Wu, Eric N. Johnson, and

More information

Unscented Kalman Filter for Vision Based Target Localisation with a Quadrotor

Unscented Kalman Filter for Vision Based Target Localisation with a Quadrotor Unscented Kalman Filter for Vision Based Target Localisation with a Quadrotor Jos Alejandro Dena Ruiz, Nabil Aouf Centre of Electronic Warfare, Defence Academy of the United Kingdom Cranfield University,

More information

Methods for Localization and Mapping Using Vision and Inertial Sensors

Methods for Localization and Mapping Using Vision and Inertial Sensors AIAA Guidance, Navigation and Control Conference and Exhibit 8 - August 8, Honolulu, Hawaii AIAA 8-744 Methods for Localization and Mapping Using Vision and Inertial Sensors Allen D. Wu and Eric N. Johnson

More information

VISION-ONLY AIRCRAFT FLIGHT CONTROL

VISION-ONLY AIRCRAFT FLIGHT CONTROL VISION-ONLY AIRCRAFT FLIGHT CONTROL Christophe De Wagter Delft University of Technology, Delft, The Netherlands Alison A. Proctor 2 and Eric N. Johnson 3 Georgia Institute of Technology, Atlanta, GA, 30332

More information

Parameterized Optimal Trajectory Generation for Target Localization

Parameterized Optimal Trajectory Generation for Target Localization Guidance, Navigation and Control Conference, August 6-9 7, Hilton Head, South Carolina Parameterized Optimal Trajectory Generation for Target Localization Jeffrey B. Corbets Jack W. Langelaan The Pennsylvania

More information

Geometrical Feature Extraction Using 2D Range Scanner

Geometrical Feature Extraction Using 2D Range Scanner Geometrical Feature Extraction Using 2D Range Scanner Sen Zhang Lihua Xie Martin Adams Fan Tang BLK S2, School of Electrical and Electronic Engineering Nanyang Technological University, Singapore 639798

More information

Estimation of Altitude and Vertical Velocity for Multirotor Aerial Vehicle using Kalman Filter

Estimation of Altitude and Vertical Velocity for Multirotor Aerial Vehicle using Kalman Filter Estimation of Altitude and Vertical Velocity for Multirotor Aerial Vehicle using Kalman Filter Przemys law G asior, Stanis law Gardecki, Jaros law Gośliński and Wojciech Giernacki Poznan University of

More information

Leaderless Formation Control for Multiple Autonomous Vehicles. Wei Ren

Leaderless Formation Control for Multiple Autonomous Vehicles. Wei Ren AIAA Guidance, Navigation, and Control Conference and Exhibit - 4 August 6, Keystone, Colorado AIAA 6-669 Leaderless Formation Control for Multiple Autonomous Vehicles Wei Ren Department of Electrical

More information

Autonomous Landing of an Unmanned Aerial Vehicle

Autonomous Landing of an Unmanned Aerial Vehicle Autonomous Landing of an Unmanned Aerial Vehicle Joel Hermansson, Andreas Gising Cybaero AB SE-581 12 Linköping, Sweden Email: {joel.hermansson, andreas.gising}@cybaero.se Martin Skoglund and Thomas B.

More information

Mobile Robotics. Mathematics, Models, and Methods. HI Cambridge. Alonzo Kelly. Carnegie Mellon University UNIVERSITY PRESS

Mobile Robotics. Mathematics, Models, and Methods. HI Cambridge. Alonzo Kelly. Carnegie Mellon University UNIVERSITY PRESS Mobile Robotics Mathematics, Models, and Methods Alonzo Kelly Carnegie Mellon University HI Cambridge UNIVERSITY PRESS Contents Preface page xiii 1 Introduction 1 1.1 Applications of Mobile Robots 2 1.2

More information

Tightly-Integrated Visual and Inertial Navigation for Pinpoint Landing on Rugged Terrains

Tightly-Integrated Visual and Inertial Navigation for Pinpoint Landing on Rugged Terrains Tightly-Integrated Visual and Inertial Navigation for Pinpoint Landing on Rugged Terrains PhD student: Jeff DELAUNE ONERA Director: Guy LE BESNERAIS ONERA Advisors: Jean-Loup FARGES Clément BOURDARIAS

More information

(1) and s k ωk. p k vk q

(1) and s k ωk. p k vk q Sensing and Perception: Localization and positioning Isaac Sog Project Assignment: GNSS aided INS In this project assignment you will wor with a type of navigation system referred to as a global navigation

More information

Autonomous on-orbit Calibration Of Star Trackers

Autonomous on-orbit Calibration Of Star Trackers Autonomous on-orbit Calibration Of Star Trackers Malak Samaan 1, Todd Griffith 2,Puneet Singla 3, and John L. Junkins 4 Department of Aerospace Engineering, Texas A&M University, College Station, TX 77843-3141

More information

Motion estimation of unmanned marine vehicles Massimo Caccia

Motion estimation of unmanned marine vehicles Massimo Caccia Motion estimation of unmanned marine vehicles Massimo Caccia Consiglio Nazionale delle Ricerche Istituto di Studi sui Sistemi Intelligenti per l Automazione Via Amendola 122 D/O, 70126, Bari, Italy massimo.caccia@ge.issia.cnr.it

More information

Quaternion-Based Tracking of Multiple Objects in Synchronized Videos

Quaternion-Based Tracking of Multiple Objects in Synchronized Videos Quaternion-Based Tracking of Multiple Objects in Synchronized Videos Quming Zhou 1, Jihun Park 2, and J.K. Aggarwal 1 1 Department of Electrical and Computer Engineering The University of Texas at Austin

More information

Video integration in a GNSS/INS hybridization architecture for approach and landing

Video integration in a GNSS/INS hybridization architecture for approach and landing Author manuscript, published in "IEEE/ION PLANS 2014, Position Location and Navigation Symposium, Monterey : United States (2014)" Video integration in a GNSS/INS hybridization architecture for approach

More information

EFFECT OF YAW-TILTED HINGE AXIS ON DEPLOYMENT ROBUSTNESS OF MARS AIRPLANE

EFFECT OF YAW-TILTED HINGE AXIS ON DEPLOYMENT ROBUSTNESS OF MARS AIRPLANE EFFET OF YAW-TILTED HINGE AXIS ON DEPLOYMENT ROBUSTNESS OF MARS AIRPLANE Koji Fujita* Hiroki Nagai** and Akira Oyama* * Institute of Space and Astronautical Science Japan Aerospace Exploration Agency --

More information

Overview. EECS 124, UC Berkeley, Spring 2008 Lecture 23: Localization and Mapping. Statistical Models

Overview. EECS 124, UC Berkeley, Spring 2008 Lecture 23: Localization and Mapping. Statistical Models Introduction ti to Embedded dsystems EECS 124, UC Berkeley, Spring 2008 Lecture 23: Localization and Mapping Gabe Hoffmann Ph.D. Candidate, Aero/Astro Engineering Stanford University Statistical Models

More information

State Estimation for Continuous-Time Systems with Perspective Outputs from Discrete Noisy Time-Delayed Measurements

State Estimation for Continuous-Time Systems with Perspective Outputs from Discrete Noisy Time-Delayed Measurements State Estimation for Continuous-Time Systems with Perspective Outputs from Discrete Noisy Time-Delayed Measurements António Pedro Aguiar aguiar@ece.ucsb.edu João Pedro Hespanha hespanha@ece.ucsb.edu Dept.

More information

Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview

Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview Arun Das 05/09/2017 Arun Das Waterloo Autonomous Vehicles Lab Introduction What s in a name? Arun Das Waterloo Autonomous

More information

Dynamical Modeling and Controlof Quadrotor

Dynamical Modeling and Controlof Quadrotor Dynamical Modeling and Controlof Quadrotor Faizan Shahid NUST PNEC Pakistan engr.faizan_shahid@hotmail.com Muhammad Bilal Kadri, Nasir Aziz Jumani, Zaid Pirwani PAF KIET Pakistan bilal.kadri@pafkiet.edu.pk

More information

Planar-Based Visual Inertial Navigation

Planar-Based Visual Inertial Navigation Planar-Based Visual Inertial Navigation Ghaaleh Panahandeh and Peter Händel KTH Royal Institute of Technology, ACCESS Linnaeus Center, Stockholm, Sweden Volvo Car Corporation, Gothenburg, Sweden KTH Royal

More information

Low Cost solution for Pose Estimation of Quadrotor

Low Cost solution for Pose Estimation of Quadrotor Low Cost solution for Pose Estimation of Quadrotor mangal@iitk.ac.in https://www.iitk.ac.in/aero/mangal/ Intelligent Guidance and Control Laboratory Indian Institute of Technology, Kanpur Mangal Kothari

More information

Autonomous Navigation in Complex Indoor and Outdoor Environments with Micro Aerial Vehicles

Autonomous Navigation in Complex Indoor and Outdoor Environments with Micro Aerial Vehicles Autonomous Navigation in Complex Indoor and Outdoor Environments with Micro Aerial Vehicles Shaojie Shen Dept. of Electrical and Systems Engineering & GRASP Lab, University of Pennsylvania Committee: Daniel

More information

EKF Localization and EKF SLAM incorporating prior information

EKF Localization and EKF SLAM incorporating prior information EKF Localization and EKF SLAM incorporating prior information Final Report ME- Samuel Castaneda ID: 113155 1. Abstract In the context of mobile robotics, before any motion planning or navigation algorithm

More information

Vision-Only Aircraft Flight Control Methods and Test Results

Vision-Only Aircraft Flight Control Methods and Test Results AIAA Guidance, Navigation, and Control Conference and Exhibit 16-19 August 24, Providence, Rhode Island AIAA 24-5351 Vision-Only Aircraft Flight Control Methods and Test Results Alison A. Proctor and Eric

More information

Data Association for SLAM

Data Association for SLAM CALIFORNIA INSTITUTE OF TECHNOLOGY ME/CS 132a, Winter 2011 Lab #2 Due: Mar 10th, 2011 Part I Data Association for SLAM 1 Introduction For this part, you will experiment with a simulation of an EKF SLAM

More information

Unmanned Aerial Vehicles

Unmanned Aerial Vehicles Unmanned Aerial Vehicles Embedded Control Edited by Rogelio Lozano WILEY Table of Contents Chapter 1. Aerodynamic Configurations and Dynamic Models 1 Pedro CASTILLO and Alejandro DZUL 1.1. Aerodynamic

More information

MULTI-ROBOT research has gained a broad attention. A Novel Way to Implement Self-localization in a Multi-robot Experimental Platform

MULTI-ROBOT research has gained a broad attention. A Novel Way to Implement Self-localization in a Multi-robot Experimental Platform 21 American Control Conference Marriott Waterfront, Baltimore, MD, USA June 3-July 2, 21 FrC16.5 A Novel Way to Implement Self-localization in a Multi-robot Experimental Platform Sheng Zhao and Manish

More information

Motion. 1 Introduction. 2 Optical Flow. Sohaib A Khan. 2.1 Brightness Constancy Equation

Motion. 1 Introduction. 2 Optical Flow. Sohaib A Khan. 2.1 Brightness Constancy Equation Motion Sohaib A Khan 1 Introduction So far, we have dealing with single images of a static scene taken by a fixed camera. Here we will deal with sequence of images taken at different time intervals. Motion

More information

Calibration of Inertial Measurement Units Using Pendulum Motion

Calibration of Inertial Measurement Units Using Pendulum Motion Technical Paper Int l J. of Aeronautical & Space Sci. 11(3), 234 239 (2010) DOI:10.5139/IJASS.2010.11.3.234 Calibration of Inertial Measurement Units Using Pendulum Motion Keeyoung Choi* and Se-ah Jang**

More information

Autonomous Mobile Robot Design

Autonomous Mobile Robot Design Autonomous Mobile Robot Design Topic: EKF-based SLAM Dr. Kostas Alexis (CSE) These slides have partially relied on the course of C. Stachniss, Robot Mapping - WS 2013/14 Autonomous Robot Challenges Where

More information

SAE Aerospace Control & Guidance Systems Committee #97 March 1-3, 2006 AFOSR, AFRL. Georgia Tech, MIT, UCLA, Virginia Tech

SAE Aerospace Control & Guidance Systems Committee #97 March 1-3, 2006 AFOSR, AFRL. Georgia Tech, MIT, UCLA, Virginia Tech Systems for Aircraft SAE Aerospace Control & Guidance Systems Committee #97 March 1-3, 2006 AFOSR, AFRL Georgia Tech, MIT, UCLA, Virginia Tech controls.ae.gatech.edu/avcs Systems Systems MURI Development

More information

Incremental Light Bundle Adjustment for Robotics Navigation

Incremental Light Bundle Adjustment for Robotics Navigation Incremental Light Bundle Adjustment for Robotics Vadim Indelman, Andrew Melim, Frank Dellaert Robotics and Intelligent Machines (RIM) Center College of Computing Georgia Institute of Technology Introduction

More information

Marker Based Localization of a Quadrotor. Akshat Agarwal & Siddharth Tanwar

Marker Based Localization of a Quadrotor. Akshat Agarwal & Siddharth Tanwar Marker Based Localization of a Quadrotor Akshat Agarwal & Siddharth Tanwar Objective Introduction Objective: To implement a high level control pipeline on a quadrotor which could autonomously take-off,

More information

Exploitation of GPS-Control Points in low-contrast IR-imagery for homography estimation

Exploitation of GPS-Control Points in low-contrast IR-imagery for homography estimation Exploitation of GPS-Control Points in low-contrast IR-imagery for homography estimation Patrick Dunau 1 Fraunhofer-Institute, of Optronics, Image Exploitation and System Technologies (IOSB), Gutleuthausstr.

More information

COMBINED BUNDLE BLOCK ADJUSTMENT VERSUS DIRECT SENSOR ORIENTATION ABSTRACT

COMBINED BUNDLE BLOCK ADJUSTMENT VERSUS DIRECT SENSOR ORIENTATION ABSTRACT COMBINED BUNDLE BLOCK ADJUSTMENT VERSUS DIRECT SENSOR ORIENTATION Karsten Jacobsen Institute for Photogrammetry and Engineering Surveys University of Hannover Nienburger Str.1 D-30167 Hannover, Germany

More information

Towards Autonomous UAV Flight in Forests

Towards Autonomous UAV Flight in Forests Guidance, Navigation and Control Conference, August 15-18 5, San Francisco, California Towards Autonomous UAV Flight in Forests Jack Langelaan and Steve Rock Stanford University, Stanford, CA 935, USA

More information

Mobile robot localisation and navigation using multi-sensor fusion via interval analysis and UKF

Mobile robot localisation and navigation using multi-sensor fusion via interval analysis and UKF Mobile robot localisation and navigation using multi-sensor fusion via interval analysis and UKF Immanuel Ashokaraj, Antonios Tsourdos, Peter Silson and Brian White. Department of Aerospace, Power and

More information

Non-symmetric membership function for Fuzzy-based visual servoing onboard a UAV

Non-symmetric membership function for Fuzzy-based visual servoing onboard a UAV 1 Non-symmetric membership function for Fuzzy-based visual servoing onboard a UAV M. A. Olivares-Méndez and P. Campoy and C. Martínez and I. F. Mondragón B. Computer Vision Group, DISAM, Universidad Politécnica

More information

Driven Cavity Example

Driven Cavity Example BMAppendixI.qxd 11/14/12 6:55 PM Page I-1 I CFD Driven Cavity Example I.1 Problem One of the classic benchmarks in CFD is the driven cavity problem. Consider steady, incompressible, viscous flow in a square

More information

GPS denied Navigation Solutions

GPS denied Navigation Solutions GPS denied Navigation Solutions Krishnraj Singh Gaur and Mangal Kothari ksgaur@iitk.ac.in, mangal@iitk.ac.in https://www.iitk.ac.in/aero/mangal/ Intelligent Guidance and Control Laboratory Indian Institute

More information

Optimization of the Simultaneous Localization and Map-Building Algorithm for Real-Time Implementation

Optimization of the Simultaneous Localization and Map-Building Algorithm for Real-Time Implementation 242 IEEE TRANSACTIONS ON ROBOTICS AND AUTOMATION, VOL. 17, NO. 3, JUNE 2001 Optimization of the Simultaneous Localization and Map-Building Algorithm for Real-Time Implementation José E. Guivant and Eduardo

More information

Nonlinear Filtering with IMM Algorithm for Coastal Radar Target Tracking System

Nonlinear Filtering with IMM Algorithm for Coastal Radar Target Tracking System TELKOMNIKA, Vol.13, No.1, March 2015, pp. 211~220 ISSN: 1693-6930, accredited A by DIKTI, Decree No: 58/DIKTI/Kep/2013 DOI: 10.12928/TELKOMNIKA.v13i1.791 211 Nonlinear Filtering with IMM Algorithm for

More information

Efficient Covariance Intersection of Attitude Estimates Using a Local Error Representation

Efficient Covariance Intersection of Attitude Estimates Using a Local Error Representation Efficient Covariance Intersection of Attitude Estimates Using a Local Error Representation Christopher K. Nebelecky, John L. Crassidis University at Buffalo, State University of New York, Amherst, NY 14260-4400

More information

Monocular SLAM for a Small-Size Humanoid Robot

Monocular SLAM for a Small-Size Humanoid Robot Tamkang Journal of Science and Engineering, Vol. 14, No. 2, pp. 123 129 (2011) 123 Monocular SLAM for a Small-Size Humanoid Robot Yin-Tien Wang*, Duen-Yan Hung and Sheng-Hsien Cheng Department of Mechanical

More information

An Experimental Study of the Autonomous Helicopter Landing Problem

An Experimental Study of the Autonomous Helicopter Landing Problem An Experimental Study of the Autonomous Helicopter Landing Problem Srikanth Saripalli 1, Gaurav S. Sukhatme 1, and James F. Montgomery 2 1 Department of Computer Science, University of Southern California,

More information

Dealing with Scale. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

Dealing with Scale. Stephan Weiss Computer Vision Group NASA-JPL / CalTech Dealing with Scale Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline Why care about size? The IMU as scale provider: The

More information

Localization of Multiple Robots with Simple Sensors

Localization of Multiple Robots with Simple Sensors Proceedings of the IEEE International Conference on Mechatronics & Automation Niagara Falls, Canada July 2005 Localization of Multiple Robots with Simple Sensors Mike Peasgood and Christopher Clark Lab

More information

Ground Plane Motion Parameter Estimation For Non Circular Paths

Ground Plane Motion Parameter Estimation For Non Circular Paths Ground Plane Motion Parameter Estimation For Non Circular Paths G.J.Ellwood Y.Zheng S.A.Billings Department of Automatic Control and Systems Engineering University of Sheffield, Sheffield, UK J.E.W.Mayhew

More information

Particle Filtering for Attitude Estimation Using a Minimal. Local-Error Representation: A Revisit

Particle Filtering for Attitude Estimation Using a Minimal. Local-Error Representation: A Revisit Particle Filtering for Attitude Estimation Using a Minimal Local-Error Representation: A Revisit Lubin Chang Department of navigation engineering, Naval University of Engineering, China For spacecraft

More information

Dynamic IBVS Control of an Underactuated UAV

Dynamic IBVS Control of an Underactuated UAV Proceedings of the 212 IEEE International Conference on Robotics and Biomimetics December 11-14, 212, Guangzhou, China Dynamic IBVS Control of an Underactuated UAV Hamed Jabbari 1,2, Giuseppe Oriolo 2

More information

Distributed Vision-Aided Cooperative Navigation Based on Three-View Geometry

Distributed Vision-Aided Cooperative Navigation Based on Three-View Geometry Distributed Vision-Aided Cooperative Navigation Based on hree-view Geometry Vadim Indelman, Pini Gurfil Distributed Space Systems Lab, Aerospace Engineering, echnion Ehud Rivlin Computer Science, echnion

More information

Unscented Kalman Filtering for Attitude Determination Using MEMS Sensors

Unscented Kalman Filtering for Attitude Determination Using MEMS Sensors Journal of Applied Science and Engineering, Vol. 16, No. 2, pp. 165 176 (2013) DOI: 10.6180/jase.2013.16.2.08 Unscented Kalman Filtering for Attitude Determination Using MEMS Sensors Jaw-Kuen Shiau* and

More information

Satellite and Inertial Navigation and Positioning System

Satellite and Inertial Navigation and Positioning System Satellite and Inertial Navigation and Positioning System Project Proposal By: Luke Pfister Dan Monroe Project Advisors: Dr. In Soo Ahn Dr. Yufeng Lu EE 451 Senior Capstone Project December 10, 2009 PROJECT

More information

Evaluation of Moving Object Tracking Techniques for Video Surveillance Applications

Evaluation of Moving Object Tracking Techniques for Video Surveillance Applications International Journal of Current Engineering and Technology E-ISSN 2277 4106, P-ISSN 2347 5161 2015INPRESSCO, All Rights Reserved Available at http://inpressco.com/category/ijcet Research Article Evaluation

More information

VISION-BASED UAV FLIGHT CONTROL AND OBSTACLE AVOIDANCE. Zhihai He, Ram Venkataraman Iyer, and Phillip R. Chandler

VISION-BASED UAV FLIGHT CONTROL AND OBSTACLE AVOIDANCE. Zhihai He, Ram Venkataraman Iyer, and Phillip R. Chandler VISION-BASED UAV FLIGHT CONTROL AND OBSTACLE AVOIDANCE Zhihai He, Ram Venkataraman Iyer, and Phillip R Chandler ABSTRACT In this work, we explore various ideas and approaches to deal with the inherent

More information

Cooperative Vision Based Estimation and Tracking Using Multiple UAVs

Cooperative Vision Based Estimation and Tracking Using Multiple UAVs Cooperative Vision Based Estimation and Tracking Using Multiple UAVs Brett Bethke, Mario Valenti, and Jonathan How Massachusetts Institute of Technology, Cambridge, MA {bbethke, valenti, jhow}@mit.edu

More information

PLEASE DO NOT REMOVE THIS PAGE

PLEASE DO NOT REMOVE THIS PAGE Thank you for downloading this document from the RMIT ResearchR Repository Citation: Sabatini, R, Cappello, F, Ramasamy, S, Gardi, A and Clothier, R 2015, 'An innovative navigation and guidance system

More information

Sensor fusion methods for indoor navigation using UWB radio aided INS/DR

Sensor fusion methods for indoor navigation using UWB radio aided INS/DR Sensor fusion methods for indoor navigation using UWB radio aided INS/DR JOSÉ BORRÀS SILLERO Master s Degree Project Stockholm, Sweden July 2012 XR-EE-SB 2012:015 Abstract Some applications such as industrial

More information

Chapter 13. Vision Based Guidance. Beard & McLain, Small Unmanned Aircraft, Princeton University Press, 2012,

Chapter 13. Vision Based Guidance. Beard & McLain, Small Unmanned Aircraft, Princeton University Press, 2012, Chapter 3 Vision Based Guidance Beard & McLain, Small Unmanned Aircraft, Princeton University Press, 22, Chapter 3: Slide Architecture w/ Camera targets to track/avoid vision-based guidance waypoints status

More information

Spline Mapping to Maximize Energy Exploitation of Non-Uniform Thermals

Spline Mapping to Maximize Energy Exploitation of Non-Uniform Thermals Spline to Maximize Energy Exploitation of Non-Uniform Thermals John J. Bird, Jack W. Langelaan Aerospace Engineering, The Pennsylvania State University This paper describes a method for modeling and maximizing

More information

INCREMENTAL DISPLACEMENT ESTIMATION METHOD FOR VISUALLY SERVOED PARIED STRUCTURED LIGHT SYSTEM (ViSP)

INCREMENTAL DISPLACEMENT ESTIMATION METHOD FOR VISUALLY SERVOED PARIED STRUCTURED LIGHT SYSTEM (ViSP) Blucher Mechanical Engineering Proceedings May 2014, vol. 1, num. 1 www.proceedings.blucher.com.br/evento/10wccm INCREMENAL DISPLACEMEN ESIMAION MEHOD FOR VISUALLY SERVOED PARIED SRUCURED LIGH SYSEM (ViSP)

More information

Tracking Multiple Mobile Targets Using Cooperative Unmanned Aerial Vehicles

Tracking Multiple Mobile Targets Using Cooperative Unmanned Aerial Vehicles 215 International Conference on Unmanned Aircraft Systems (ICUAS) Denver Marriott Tech Center Denver, Colorado, USA, June 9-12, 215 Tracking Multiple Mobile Targets Using Cooperative Unmanned Aerial Vehicles

More information

Sensory Augmentation for Increased Awareness of Driving Environment

Sensory Augmentation for Increased Awareness of Driving Environment Sensory Augmentation for Increased Awareness of Driving Environment Pranay Agrawal John M. Dolan Dec. 12, 2014 Technologies for Safe and Efficient Transportation (T-SET) UTC The Robotics Institute Carnegie

More information

Trajectory Generation for Constant Velocity Target Motion Estimation Using Monocular Vision

Trajectory Generation for Constant Velocity Target Motion Estimation Using Monocular Vision Trajectory Generation for Constant Velocity Targ Motion Estimation Using Monocular Vision Eric W. Frew, Stephen M. Rock Aerospace Robotics Laboratory Department of Aeronautics and Astronautics Stanford

More information

Midterm Exam Solutions

Midterm Exam Solutions Midterm Exam Solutions Computer Vision (J. Košecká) October 27, 2009 HONOR SYSTEM: This examination is strictly individual. You are not allowed to talk, discuss, exchange solutions, etc., with other fellow

More information

USING THE ORTHOGONAL PROJECTION FOR PARAMETER INITIALIZATION IN THE 3D RECONSTRUCTION OF DISTANT OBJECTS INTRODUCTION

USING THE ORTHOGONAL PROJECTION FOR PARAMETER INITIALIZATION IN THE 3D RECONSTRUCTION OF DISTANT OBJECTS INTRODUCTION USING THE ORTHOGONAL PROJECTION FOR PARAMETER INITIALIZATION IN THE 3D RECONSTRUCTION OF DISTANT OBJECTS Keith F. Blonquist, Junior Project Engineer Lidar Pacific Corporation Logan, UT 84341 kfblonquist@lidarpacific.com

More information

Take Home Exam # 2 Machine Vision

Take Home Exam # 2 Machine Vision 1 Take Home Exam # 2 Machine Vision Date: 04/26/2018 Due : 05/03/2018 Work with one awesome/breathtaking/amazing partner. The name of the partner should be clearly stated at the beginning of your report.

More information

MODIFIED KALMAN FILTER BASED METHOD FOR TRAINING STATE-RECURRENT MULTILAYER PERCEPTRONS

MODIFIED KALMAN FILTER BASED METHOD FOR TRAINING STATE-RECURRENT MULTILAYER PERCEPTRONS MODIFIED KALMAN FILTER BASED METHOD FOR TRAINING STATE-RECURRENT MULTILAYER PERCEPTRONS Deniz Erdogmus, Justin C. Sanchez 2, Jose C. Principe Computational NeuroEngineering Laboratory, Electrical & Computer

More information

A Comparison of Unscented and Extended Kalman Filtering for Estimating Quaternion Motion

A Comparison of Unscented and Extended Kalman Filtering for Estimating Quaternion Motion A Comparison of Unscented and Extended Kalman Filtering for Estimating Quaternion Motion Joseph J. LaViola Jr. Brown University Technology Center for Advanced Scientific Computing and Visualization PO

More information

Information Fusion in Navigation Systems via Factor Graph Based Incremental Smoothing

Information Fusion in Navigation Systems via Factor Graph Based Incremental Smoothing Information Fusion in Navigation Systems via Factor Graph Based Incremental Smoothing Vadim Indelman a, Stephen Williams a, Michael Kaess b, Frank Dellaert a a College of Computing, Georgia Institute of

More information

Towards Gaussian Multi-Robot SLAM for Underwater Robotics

Towards Gaussian Multi-Robot SLAM for Underwater Robotics Towards Gaussian Multi-Robot SLAM for Underwater Robotics Dave Kroetsch davek@alumni.uwaterloo.ca Christoper Clark cclark@mecheng1.uwaterloo.ca Lab for Autonomous and Intelligent Robotics University of

More information

Stable Trajectory Design for Highly Constrained Environments using Receding Horizon Control

Stable Trajectory Design for Highly Constrained Environments using Receding Horizon Control Stable Trajectory Design for Highly Constrained Environments using Receding Horizon Control Yoshiaki Kuwata and Jonathan P. How Space Systems Laboratory Massachusetts Institute of Technology {kuwata,jhow}@mit.edu

More information

Robotic Perception and Action: Vehicle SLAM Assignment

Robotic Perception and Action: Vehicle SLAM Assignment Robotic Perception and Action: Vehicle SLAM Assignment Mariolino De Cecco Mariolino De Cecco, Mattia Tavernini 1 CONTENTS Vehicle SLAM Assignment Contents Assignment Scenario 3 Odometry Localization...........................................

More information

Aberrations in Holography

Aberrations in Holography Aberrations in Holography D Padiyar, J Padiyar 1070 Commerce St suite A, San Marcos, CA 92078 dinesh@triple-take.com joy@triple-take.com Abstract. The Seidel aberrations are described as they apply to

More information

Factorization-Based Calibration Method for MEMS Inertial Measurement Unit

Factorization-Based Calibration Method for MEMS Inertial Measurement Unit Factorization-Based Calibration Method for MEMS Inertial Measurement Unit Myung Hwangbo Takeo Kanade The Robotics Institute, Carnegie Mellon University 5000 Forbes Avenue, Pittsburgh, PA, 15213, USA {myung,

More information

The Performance Evaluation of the Integration of Inertial Navigation System and Global Navigation Satellite System with Analytic Constraints

The Performance Evaluation of the Integration of Inertial Navigation System and Global Navigation Satellite System with Analytic Constraints Journal of Environmental Science and Engineering A 6 (2017) 313-319 doi:10.17265/2162-5298/2017.06.005 D DAVID PUBLISHING The Performance Evaluation of the Integration of Inertial Navigation System and

More information

Computationally Efficient Visual-inertial Sensor Fusion for GPS-denied Navigation on a Small Quadrotor

Computationally Efficient Visual-inertial Sensor Fusion for GPS-denied Navigation on a Small Quadrotor Computationally Efficient Visual-inertial Sensor Fusion for GPS-denied Navigation on a Small Quadrotor Chang Liu & Stephen D. Prior Faculty of Engineering and the Environment, University of Southampton,

More information

Exterior Orientation Parameters

Exterior Orientation Parameters Exterior Orientation Parameters PERS 12/2001 pp 1321-1332 Karsten Jacobsen, Institute for Photogrammetry and GeoInformation, University of Hannover, Germany The georeference of any photogrammetric product

More information

AUTONOMOUS PLANETARY ROVER CONTROL USING INVERSE SIMULATION

AUTONOMOUS PLANETARY ROVER CONTROL USING INVERSE SIMULATION AUTONOMOUS PLANETARY ROVER CONTROL USING INVERSE SIMULATION Kevin Worrall (1), Douglas Thomson (1), Euan McGookin (1), Thaleia Flessa (1) (1)University of Glasgow, Glasgow, G12 8QQ, UK, Email: kevin.worrall@glasgow.ac.uk

More information

Adversarial Attacks on Image Recognition*

Adversarial Attacks on Image Recognition* Adversarial Attacks on Image Recognition* Masha Itkina, Yu Wu, and Bahman Bahmani 3 Abstract This project extends the work done by Papernot et al. in [4] on adversarial attacks in image recognition. We

More information

Chapter 7: Computation of the Camera Matrix P

Chapter 7: Computation of the Camera Matrix P Chapter 7: Computation of the Camera Matrix P Arco Nederveen Eagle Vision March 18, 2008 Arco Nederveen (Eagle Vision) The Camera Matrix P March 18, 2008 1 / 25 1 Chapter 7: Computation of the camera Matrix

More information

High-precision, consistent EKF-based visual-inertial odometry

High-precision, consistent EKF-based visual-inertial odometry High-precision, consistent EKF-based visual-inertial odometry Mingyang Li and Anastasios I. Mourikis, IJRR 2013 Ao Li Introduction What is visual-inertial odometry (VIO)? The problem of motion tracking

More information

Robust Controller Design for an Autonomous Underwater Vehicle

Robust Controller Design for an Autonomous Underwater Vehicle DRC04 Robust Controller Design for an Autonomous Underwater Vehicle Pakpong Jantapremjit 1, * 1 Department of Mechanical Engineering, Faculty of Engineering, Burapha University, Chonburi, 20131 * E-mail:

More information

Adaptive tracking control scheme for an autonomous underwater vehicle subject to a union of boundaries

Adaptive tracking control scheme for an autonomous underwater vehicle subject to a union of boundaries Indian Journal of Geo-Marine Sciences Vol. 42 (8), December 2013, pp. 999-1005 Adaptive tracking control scheme for an autonomous underwater vehicle subject to a union of boundaries Zool Hilmi Ismail 1

More information

Zürich. Roland Siegwart Margarita Chli Martin Rufli Davide Scaramuzza. ETH Master Course: L Autonomous Mobile Robots Summary

Zürich. Roland Siegwart Margarita Chli Martin Rufli Davide Scaramuzza. ETH Master Course: L Autonomous Mobile Robots Summary Roland Siegwart Margarita Chli Martin Rufli Davide Scaramuzza ETH Master Course: 151-0854-00L Autonomous Mobile Robots Summary 2 Lecture Overview Mobile Robot Control Scheme knowledge, data base mission

More information

IMPROVING QUADROTOR 3-AXES STABILIZATION RESULTS USING EMPIRICAL RESULTS AND SYSTEM IDENTIFICATION

IMPROVING QUADROTOR 3-AXES STABILIZATION RESULTS USING EMPIRICAL RESULTS AND SYSTEM IDENTIFICATION IMPROVING QUADROTOR 3-AXES STABILIZATION RESULTS USING EMPIRICAL RESULTS AND SYSTEM IDENTIFICATION Övünç Elbir & Electronics Eng. oelbir@etu.edu.tr Anıl Ufuk Batmaz & Electronics Eng. aubatmaz@etu.edu.tr

More information

Analysis of Euler Angles in a Simple Two-Axis Gimbals Set

Analysis of Euler Angles in a Simple Two-Axis Gimbals Set Vol:5, No:9, 2 Analysis of Euler Angles in a Simple Two-Axis Gimbals Set Ma Myint Myint Aye International Science Index, Mechanical and Mechatronics Engineering Vol:5, No:9, 2 waset.org/publication/358

More information

Simultaneous Planning, Localization, and Mapping in a Camera Sensor Network

Simultaneous Planning, Localization, and Mapping in a Camera Sensor Network Simultaneous Planning, Localization, and Mapping in a Camera Sensor Network David Meger 1, Ioannis Rekleitis 2, and Gregory Dudek 1 1 McGill University, Montreal, Canada [dmeger,dudek]@cim.mcgill.ca 2

More information

The Pennsylvania State University The Graduate School LOCAL TERRAIN MAPPING FOR OBSTACLE AVOIDANCE USING MONOCULAR VISION

The Pennsylvania State University The Graduate School LOCAL TERRAIN MAPPING FOR OBSTACLE AVOIDANCE USING MONOCULAR VISION The Pennsylvania State University The Graduate School LOCAL TERRAIN MAPPING FOR OBSTACLE AVOIDANCE USING MONOCULAR VISION A Thesis in Aerospace Engineering by Sean Quinn Marlow c 29 Sean Quinn Marlow Submitted

More information

Spacecraft Actuation Using CMGs and VSCMGs

Spacecraft Actuation Using CMGs and VSCMGs Spacecraft Actuation Using CMGs and VSCMGs Arjun Narayanan and Ravi N Banavar (ravi.banavar@gmail.com) 1 1 Systems and Control Engineering, IIT Bombay, India Research Symposium, ISRO-IISc Space Technology

More information

ROBOT TEAMS CH 12. Experiments with Cooperative Aerial-Ground Robots

ROBOT TEAMS CH 12. Experiments with Cooperative Aerial-Ground Robots ROBOT TEAMS CH 12 Experiments with Cooperative Aerial-Ground Robots Gaurav S. Sukhatme, James F. Montgomery, and Richard T. Vaughan Speaker: Jeff Barnett Paper Focus Heterogeneous Teams for Surveillance

More information

Visual Servoing for Tracking Features in Urban Areas Using an Autonomous Helicopter

Visual Servoing for Tracking Features in Urban Areas Using an Autonomous Helicopter Visual Servoing for Tracking Features in Urban Areas Using an Autonomous Helicopter Abstract The use of Unmanned Aerial Vehicles (UAVs) in civilian and domestic applications is highly demanding, requiring

More information

2005 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media,

2005 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, 25 IEEE Personal use of this material is permitted Permission from IEEE must be obtained for all other uses in any current or future media including reprinting/republishing this material for advertising

More information

Quaternion to Euler Angle Conversion for Arbitrary Rotation Sequence Using Geometric Methods

Quaternion to Euler Angle Conversion for Arbitrary Rotation Sequence Using Geometric Methods uaternion to Euler Angle Conversion for Arbitrary Rotation Sequence Using Geometric Methods ê = normalized Euler ation axis i Noel H. Hughes Nomenclature = indices of first, second and third Euler

More information

SIMPLE FORMATION CONTROL SCHEME TOLERANT TO COMMUNICATION FAILURES FOR SMALL UNMANNED AIR VEHICLES

SIMPLE FORMATION CONTROL SCHEME TOLERANT TO COMMUNICATION FAILURES FOR SMALL UNMANNED AIR VEHICLES SIMPLE FORMATION CONTROL SCHEME TOLERANT TO COMMUNICATION FAILURES FOR SMALL UNMANNED AIR VEHICLES Takuma Hino *Dept. of Aeronautics and Astronautics, University of Tokyo Keywords: Small UAV, Formation

More information

Factor Graph Based Incremental Smoothing in Inertial Navigation Systems

Factor Graph Based Incremental Smoothing in Inertial Navigation Systems Factor Graph Based Incremental in Inertial Navigation Systems Vadim Indelman, Stephen Williams, Michael Kaess and Frank Dellaert College of Computing, Georgia Institute of Technology, Atlanta, GA 3332,

More information