State-space models for 3D visual tracking

Size: px
Start display at page:

Download "State-space models for 3D visual tracking"

Transcription

1 SA-1 State-space models for 3D visual tracking Doz. G. Bleser Prof. Stricker Computer Vision: Object and People Tracking

2 Example applications Head tracking, e.g. for mobile seating buck Object tracking, e.g. for visual grasping in robotics Camera tracking, e.g. for Augmented Reality Human motion capture 2

3 Outline Preliminaries: Recap: camera geometry, 3D reconstruction vs. pose estimation 3D transformations, rigid body motion, rotation parametrisations Multi-purpose (non)linear Gaussian state-space models for 3D visual tracking x t = g x t 1, u t, ε t, ε t ~N(0, R t ) z t = h x t, δ t, δ t ~N(0, Q t ) This is the most general statespace model form Example applications 3

4 Typical estimation problems Pose estimation in 3D. The state contains: pose of moving object observed by stationary camera relative to static world frame pose of moving camera observing stationary features in the world pose of moving object observed by moving camera poses of an articulated body world Similar estimation problems (it doesn t matter, whether object or camera are moving) camera object 4

5 Recap: 3D reconstruction (triangulation) Estimation of 3D points given corresponding 2D coordinates and camera poses. 3D feature location 2D feature location (from image processing) Camera pose 5

6 Recap: camera pose estimation Camera pose estimation based on corresponding 2D and 3D coordinates. Tracking of landmarks/features in the world. 3D feature location 2D feature location (from image processing) Camera pose 6

7 Object pose estimation Stationary camera, moving object (essentially the same problem) y o x c z c z o x o y c Object pose {R oc, o c } t 7

8 Recap: world/object-to-camera-mapping Rigid transformation: world/object to camera space (3D-3D) (to be estimated) Camera projection (3D-2D): intrinsic parameters (assumed calibrated) m w Image formation (perfect pinhole): m p = Π K (R m w +t) m p 3D rotation Translation y T c Π x y z T x z z R, t w Perspective projection 8

9 3D articulated body Transformation hierarchy: rigid segments (bones) and joints with varying DoF Estimation parameters: (relative) poses of each joint 3 2 Kinematic hand model Stick-figure (kinematic model) 9

10 SA-1 Kinematics Rigid body motion, rotation representation

11 How can we parametrise pose? R = r 11 r 12 r 13 r 21 r 22 r 23 R 3 3 t = r 31 r 32 r 33 t x t y t z R 3 Rows of R are the coordinate axes of the new frame, given in the old frame 11

12 Properties of rotation matrices Rotation: Special-Orthogonal group in 3D SO 3 Preserves angles and distances (orthogonal) v T w R T R 1 T Rv Rw R I R 1 v T R R T T Rw Columns have all unit length T T T R 1R 1 1, R 2R 2 1, R 3R 3 1 and are orthogonal to each T T T other R R 0, R R 0, R R matrix elements, 6 constraints 3 degrees of freedom (DoF) 12

13 Rotation 1: Euler angles Parametrizing 3D rotations: Euler angles (x-y-z) R x α = cos (α) sin α 0 sin (α) cos (α) R y β = R z γ = cos (β) 0 sin (β) sin (β) 0 cos (β) cos (γ) sin (γ) 0 sin (γ) cos (γ) Sequence of 3 elementary rotations around (x, y, z): R α, β, γ = R x α R y β R z γ Order not standardized! 13

14 Rotation 1: Euler angles Easy to understand Minimal number of 3 parameters Intuitive geometric interpretation Periodicity of angles needs to be handled Singularity for some particular configurations, e.g. gimbal lock: loss of one DoF, if two gimbals are in the same plane R α, π 2, γ 14

15 Rotation 2: exponential map (axis angle) Axis-angle: w, θ, w = 1 Definition: rotation vector ρ ρ = w θ = ρ x ρ y ρ z θ = ρ Theorem: for any rotation R SO 3, there exists a vector ρ such that: R = exp(ρ) Cross product matrix: 0 ρ z ρ y ρ = ρ z 0 ρ x ρ y ρ x 0 Matrix exponential 15

16 Rotation 2: exponential map (axis angle) From ρ = w θ to R (Rodrigues formula): R = exp ρ = I 3 + w sin θ + w 2 (1 cos(θ)) Minimal number of 3 parameters No singularities Complex formulae Ambiguity for w, if θ = 180 Periodicity of angle needs to be handled 16

17 Rotation 3: Quaternions Quaternions = hyper-complex numbers with 4 components q = a + bi + cj + dk = a, w Scalar Vector Hamilton axioms i i = j j = k k = i j k = j k i = k i j = 1 Quaternion product q 1 q 2 = a 1, w 1 a 2, w 2 = a 1 a 2 w 1 T w 2, a 1 w 2 + a 2 w 1 + w 1 w 2 17

18 Rotation 3: Quaternions Unit quaternions: q = a 2 + b 2 + c 2 + d 2 =1 Quaternion conjugate: q = a, w q = a, w q 1 = q q = 1 can be written as q = cos θ 2, w sin θ 2 Represents a rotation about axis w by angle θ In order to rotate a 3D point X: Write it as a quaternion q X = 0, X Compute the product q X = q q X q Get back the rotated point: q X = 0, X 18

19 Rotation 3: Quaternions Equivalence between unit quaternion and rotation matrix q X = q q X q X = R q X R q = a 2 + b 2 c 2 d 2 2bc 2ad 2ac + 2bd 2ad + 2bc a 2 b 2 + c 2 d 2 2cd 2ab 2bd 2ac 2ab + 2cd a 2 b 2 c 2 + d 2 Question: What is R 1 q? R 1 q = R(a, w) conjugate=inverse in case of a unit quaternion 19

20 Rotation 3: Quaternions No singularities Simple expression (polynomial) for rotation Ambiguity: Rot q = Rot( q) Redundant parameters (4 instead of 3) needs an additional, nonlinear constraint: q = 1 Methods to enforce the constraint: Normalization after each measurement update step (brute-force ) Add a penalty term: (1 q 2 ) during optimization 20

21 How can we represent rigid pose Translation: t = R(x) t x t y t z R 3 The chosen representation provides the parameters to be estimated in the state Rotation: choose appropriate representation, x, (rule-of-thumbs) Linear least squares estimation: rotation matrix Nonlinear least squares estimation: Euler angles, axis angle (minimal parametrization) Interpolation: quaternions Extended Kalman filter: often quaternions (also other representations) 21

22 Kinematics: translational motion Velocity: Acceleration: v p = p t a v = p = v t = 2 p 2 t Assuming constant velocity: v p = v t = p 0 + [vt] Assuming constant acceleration: p = v t = a 2 t = p at2 This holds for n-dimensional vectors 22

23 Kinematics: rotational motion Complicated in 3D! Angular velocity (cf. exponential map): ω = wθ Instantaneous rotation axis Angular acceleration: Differential equation: ω = ω t R = ωr Integration yields (rectangular rule): R = exp ωt R Rotation speed around this axis r ω ω r Rodrigues formula 23

24 Kinematics: rotational motion Differential equation based on quaternions: q = 1 ω q 2 Integration yields (rectangular rule): t q = exp ω q 2 Quaternion exponential: Closed-form solution available Model can be simplified with small-angle approximation r ω ω r Bibliography: Kleppner/Kolvenkow: An introduction to mechanics Shuster: A Survey of Attitude Representations 24

25 SA-1 Multi-purpose (non)linear Gaussian statespace models for 3D visual tracking All noises are assumed zero-mean Gaussian distributed Some slides are based on G. Panin

26 Recap: motion model Motion/Dynamical model: a probabilistic model describing the state evolution in time, possibly based on a control input x t = g x t 1, u t, ε t deterministic function of x t 1 and u t v t 1 + p t 1 ε t v t random component g(x t 1, u t, 0) pt Deterministic term: in absence of pertubation, gives predicted state Process noise ε: gives the uncertainty about the prediction 26

27 Translational motion (linear) Unconstrained models without control input (holds for n = 1 3D case) Notation: x, p, v, a, j, ε n-vectors and I, 0, T ΔtI n n diagonal matrices State Motion model Assumptions x = p x t+δt = x t + Tε t (also: Brownian motion) Const. position, white noise velocity ε t = v t x = p v x t+δt = I T 0 I x t + T 2 2 T ε t Const. velocity, white noise acceleration ε t = a t x = p v a x t+δt = I T T2 2 0 I T 0 0 I x t + T 3 3 T 2 2 T ε t Const. acceleration, white noise jerk ε t = j t 27

28 Translational motion (linear) Unconstrained models with control input State Motion model Assumptions x = p x t+δt = x t + T(u t + ε t ) Const. velocity, velocity as noisy input v t = u t + ε t x = p v x t+δt = I T 0 I x t + T 2 2 T (u t + ε t ) Const. acceleration, acceleration as noisy input a t = u t + ε t Smaller state, more efficient Control input enters estimate immediately, not filtered 28

29 Translational motion (linear) Ballistic trajectory (gravity modelled as noisy control input) State Motion model Assumptions x = p v x t+δt = I T 0 I x t + T 2 2 T (u t +ε t ) Const. acceleration due to gravity, u t = g, with white noise acceleration, ε t The model doesn t have to be physically motivated Basically, we are free to model whatever we want Constrained Brownian motion (in 1D) State Motion model Assumptions x = p x t+δt = ax t + bε t a 1 Random walk in position 29

30 Exercise: translational motion in 2D 1. Implement the different motion models in 2D 2. Choose noise settings 3. Simulate a trajectory 4. Plot the results 30

31 Rotational motion in 3D Linear (analogous to translational case) in case of Euler angles/rates/accelerations representation Often, quaternions,q, and angular velocity, ω, are used nonlinear State Motion model Assumptions x = q ω x t+δt = exp Δt 2 ω t + ε t q t ω t + ε t Const. angular velocity, random walk in angular velocity ε t Possible model variations: Constant orientation/acceleration assumptions can be made Angular velocity can be modelled as control input 31

32 Rigid body motion Translational and rotational models can be combined For articulated bodies, several (relative) poses are combined in one state vector (more in the next lecture) 32

33 Recap: measurement model Measurement/observation model: a probabilistic model describing, how measurements are generated from the state Explicit form: (with additive noise for simplicity) z t = h x t + δ t ~N(0,Q t ) deterministic function of x t + random component Implicit form: p z t x t = p z t h x t N(z t h x t ;0,Q t ) z t h x t Measurement likelihood given the state Deterministic term: gives expected measurement Measurement noise δ: imprecise model, noisy data, 33

34 Pixel level measurement The measurement is a set of pixels (e.g. in a texture patch), or a segmentation map can be computationally intense h x t z t Textured 3D model for predicting feature appearance image Feature points tracked based on surrounding texture patches using optical flow (with brightness model) 34

35 Feature level measurement The measurement is a set of matched/tracked feature points (their 2D pixel positions) Typical model for a perspective camera (assume perfect pinhole): Pose with quaternion m w z = h x q t + δ m p = Π K (R q m w +t) + δ p h x t As before: perspective projection R(q), t w c z t = m p 35

36 Object level measurement The measurement is the camera/object pose The measurement model seems trivial: What is the covariance of the pose? z = h x + δ q t = q t + δ Pose can t be measured directly estimated from a set of features, e.g. using least squares estimation Expected pose q t Question: Which other problem do we have here? Measured pose q t 36

37 SA-1 Example applications

38 Example: ball trajectory 38

39 Example: ball tracking Assumptions: calibrated camera setup State (per ball): Motion model: Measurement model: 39

40 Example: ball tracking Assumptions: calibrated camera setup State (per ball): position, velocity of ball in world frame Motion model: ballistic trajectory Measurement model: Feature level: detected balls in each image Object level: triangulated 3D coordinate Which problems could violate the Markov assumption? How is the state initialized? 40

41 Application: augmented reality Camera on HMD to track head Tracking based on CAD model and optical flow, online reconstruction of feature points 41

42 Application: mobile seating buck augment real car with virtual parts (e.g. design review) Requirements: Easy-to-setup head tracking solution Minimal delay (for see-through HMD) High precision Overcome occlusions Wide field of view (fov) Approach here: Perspective camera mounted on the head tracks several artificial landmarks in the car 42

43 Application: mobile seating buck Advantages: Wide fox (up to 180 horizontal) Easy set-up (just 1 marker) Stable pose (less features required due to IMU) High update rates (e.g. 100 Hz) Modeling of head motion possible in EKF framework Modified approach: Camera-IMU (inertial measurement unit) system with fisheye lens mounted on the head tracks one single marker visualized on the navigation display 43

44 Fisheye lenses Perspective lens Small field of view: ~45 Fisheye lens Large field of view: ~180 44

45 3D gyroscope: Inertial sensors Measures angular velocity in sensor frame, s, with respect to global frame, g Typical model: Slowly varying bias (quasi-static) z ω = ω s + bω s + δ ω 3D accelerometer: Measures linear acceleration in sensor frame, s, with respect to global frame, g Typical model: Acceleration due to gravity z a = s s g s + ba s + δ a z a = R sg (sg g g ) + ba s + δ a 45

46 Why visual-inertial sensor fusion? Cameras: accurate but slow (e.g. 25 Hz) Inertial sensors: fast (e.g. 100 Hz) but not accurate Camera-IMUs: Fast and accurate (complementary sensors) 46

47 47

48 Application: mobile seating buck Assumptions/boundary conditions: More rotational than translational head motion Camera and inertial sensor frame identical: c = s Global and object/world frame identical: w = g Calibrated fisheye camera (each pixel, m p, corresponds to a normalized ray, r c, in the camera frame) 48

49 Visual-inertial head tracking (model gravity) State: x = c w cw q cw T Position, velocity, orientation of camera in world frame Motion model: constant velocity, constant angular velocity c w cw q cw t+δt = c w + Tcw + T2 2 εc cw + Tε c exp T 2 ω c + ε ω q t t Standard const. vel. model (better damped model) Angular velocity from gyroscopes as control input Possible variants: Angular velocity in state, gyroscope measurement Gyroscope bias can also be estimated as a quasi-static parameter 49

50 Visual-inertial head tracking (model gravity) Measurement model (vision): feature-level λr c = R q cw m w + c w + δ r 0 = λr c R q cw m w + c w + δ r Measurement model (accelerometer, used as inclinometer): z a = R q cw cw g w + δ a =0 Constant velocity assumption z a = R q cw g w + δ a Provides information about 2 angles Variants: constant acceleration assumption, requires acceleration in state 50

51 Exercise: visual-inertial head tracking Write the state-space model assuming constant acceleration 1. Variant 1: Model accelerometer data as control inputs 2. Variant 2: Model accelerometer data as measurements 3. Discuss the advantages/disadvantages of both approaches 4. How could the constrained translational motion be incorporated (the head does not move outside the car)? 51

52 Head tracking for Augmented Reality Goal: reliable camera tracking in a larger volume with very agile camera motions Given: textured CAD model [Bleser and Stricker, 2008] 52

53 Head tracking for Augmented Reality Image processing approach: Feature points tracked based on surrounding texture patches Appearance predicted with help of CAD model (needs good pose prediction, otherwise tracking loss) Optical flow tracking (with brightness model) Estimation approach: Visual-inertial sensor fusion as before Const. acceleration, const. angular velocity model Inertial measurements as control inputs [Bleser and Stricker, 2008] 53

54 Simultaneous localisation and mapping Goal: tracking in partially known environments Image processing approach: Initialisation with CAD model (edge-based) Natural feature tracking with optical flow Estimation approach: 3D feature positions estimated in separate 3D Extended Kalman filters Robust weighted least squares for camera pose [Bleser et al., 2007] 54

55 Visual-inertial human motion capture Goal: track upper body motions and locate body with respect to workspace [Bleser et al., 2011] 55

56 Visual-inertial human motion capture Kinematic body model with rigid segments and joints (varying DoF) State: Euler angles/rates/accelerations of each joint (hierarchical), global position of root Motion model: Joints: constant angular acceleration Root: constant position Measurements: Inertial measurements Egocentric camera measurements: 2D wrist positions in image Camera pose with respect to workspace (marker-based) 56

57 Exercise: human motion capture (outside-in) Suggest a state-space model for visual outside-in tracking Assumption: markers exactly in joints Kinematic model 57

58 Application: hand pose tracking Kinematic hand model with rigid segments and joints (varying DoF) State: joint orientations as axis angle (hierarchical), global position of root Measurements: Twin edges (indicate fingers) Skin colour to remove outliers End points of fingers By Nils Petersen 58

59 Summary Various tracking problems can be solved in the Bayesian tracking framework Basic toolkit of multi-purpose state-space models The tracking solution is only as good as your model! Modeling skills need to be trained! start with some of the suggested exercises Videos and papers describing the presented works: 59

60 Organisation Exercise: 3 rd exercise will be uploaded today Next lectures: People tracking based on RGB-D data (by Markus Weber) Discussion of 3 rd exercise (as last lecture) 60

Lecture 13 Visual Inertial Fusion

Lecture 13 Visual Inertial Fusion Lecture 13 Visual Inertial Fusion Davide Scaramuzza Course Evaluation Please fill the evaluation form you received by email! Provide feedback on Exercises: good and bad Course: good and bad How to improve

More information

Robotics (Kinematics) Winter 1393 Bonab University

Robotics (Kinematics) Winter 1393 Bonab University Robotics () Winter 1393 Bonab University : most basic study of how mechanical systems behave Introduction Need to understand the mechanical behavior for: Design Control Both: Manipulators, Mobile Robots

More information

Dealing with Scale. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

Dealing with Scale. Stephan Weiss Computer Vision Group NASA-JPL / CalTech Dealing with Scale Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline Why care about size? The IMU as scale provider: The

More information

Inertial Measurement Units II!

Inertial Measurement Units II! ! Inertial Measurement Units II! Gordon Wetzstein! Stanford University! EE 267 Virtual Reality! Lecture 10! stanford.edu/class/ee267/!! wikipedia! Polynesian Migration! Lecture Overview! short review of

More information

Parameter estimation. Christiano Gava Gabriele Bleser

Parameter estimation. Christiano Gava Gabriele Bleser Parameter estimation Christiano Gava Christiano.Gava@dfki.de Gabriele Bleser gabriele.bleser@dfki.de Introduction Previous lectures: P-matrix 2D projective transformations Estimation (direct linear transform)

More information

Autonomous Mobile Robot Design

Autonomous Mobile Robot Design Autonomous Mobile Robot Design Topic: EKF-based SLAM Dr. Kostas Alexis (CSE) These slides have partially relied on the course of C. Stachniss, Robot Mapping - WS 2013/14 Autonomous Robot Challenges Where

More information

CS 445 / 645 Introduction to Computer Graphics. Lecture 21 Representing Rotations

CS 445 / 645 Introduction to Computer Graphics. Lecture 21 Representing Rotations CS 445 / 645 Introduction to Computer Graphics Lecture 21 Representing Rotations Parameterizing Rotations Straightforward in 2D A scalar, θ, represents rotation in plane More complicated in 3D Three scalars

More information

Selection and Integration of Sensors Alex Spitzer 11/23/14

Selection and Integration of Sensors Alex Spitzer 11/23/14 Selection and Integration of Sensors Alex Spitzer aes368@cornell.edu 11/23/14 Sensors Perception of the outside world Cameras, DVL, Sonar, Pressure Accelerometers, Gyroscopes, Magnetometers Position vs

More information

Quaternions and Rotations

Quaternions and Rotations CSCI 420 Computer Graphics Lecture 20 and Rotations Rotations Motion Capture [Angel Ch. 3.14] Rotations Very important in computer animation and robotics Joint angles, rigid body orientations, camera parameters

More information

Quaternions and Rotations

Quaternions and Rotations CSCI 480 Computer Graphics Lecture 20 and Rotations April 6, 2011 Jernej Barbic Rotations Motion Capture [Ch. 4.12] University of Southern California http://www-bcf.usc.edu/~jbarbic/cs480-s11/ 1 Rotations

More information

Mobile Robotics. Mathematics, Models, and Methods. HI Cambridge. Alonzo Kelly. Carnegie Mellon University UNIVERSITY PRESS

Mobile Robotics. Mathematics, Models, and Methods. HI Cambridge. Alonzo Kelly. Carnegie Mellon University UNIVERSITY PRESS Mobile Robotics Mathematics, Models, and Methods Alonzo Kelly Carnegie Mellon University HI Cambridge UNIVERSITY PRESS Contents Preface page xiii 1 Introduction 1 1.1 Applications of Mobile Robots 2 1.2

More information

12.1 Quaternions and Rotations

12.1 Quaternions and Rotations Fall 2015 CSCI 420 Computer Graphics 12.1 Quaternions and Rotations Hao Li http://cs420.hao-li.com 1 Rotations Very important in computer animation and robotics Joint angles, rigid body orientations, camera

More information

Lecture Note 3: Rotational Motion

Lecture Note 3: Rotational Motion ECE5463: Introduction to Robotics Lecture Note 3: Rotational Motion Prof. Wei Zhang Department of Electrical and Computer Engineering Ohio State University Columbus, Ohio, USA Spring 2018 Lecture 3 (ECE5463

More information

Centre for Autonomous Systems

Centre for Autonomous Systems Robot Henrik I Centre for Autonomous Systems Kungl Tekniska Högskolan hic@kth.se 27th April 2005 Outline 1 duction 2 Kinematic and Constraints 3 Mobile Robot 4 Mobile Robot 5 Beyond Basic 6 Kinematic 7

More information

MTRX4700 Experimental Robotics

MTRX4700 Experimental Robotics MTRX 4700 : Experimental Robotics Lecture 2 Stefan B. Williams Slide 1 Course Outline Week Date Content Labs Due Dates 1 5 Mar Introduction, history & philosophy of robotics 2 12 Mar Robot kinematics &

More information

Visual Recognition: Image Formation

Visual Recognition: Image Formation Visual Recognition: Image Formation Raquel Urtasun TTI Chicago Jan 5, 2012 Raquel Urtasun (TTI-C) Visual Recognition Jan 5, 2012 1 / 61 Today s lecture... Fundamentals of image formation You should know

More information

Articulated Characters

Articulated Characters Articulated Characters Skeleton A skeleton is a framework of rigid body bones connected by articulated joints Used as an (invisible?) armature to position and orient geometry (usually surface triangles)

More information

Agenda. Rotations. Camera models. Camera calibration. Homographies

Agenda. Rotations. Camera models. Camera calibration. Homographies Agenda Rotations Camera models Camera calibration Homographies D Rotations R Y = Z r r r r r r r r r Y Z Think of as change of basis where ri = r(i,:) are orthonormal basis vectors r rotated coordinate

More information

Transformation. Jane Li Assistant Professor Mechanical Engineering & Robotics Engineering

Transformation. Jane Li Assistant Professor Mechanical Engineering & Robotics Engineering RBE 550 MOTION PLANNING BASED ON DR. DMITRY BERENSON S RBE 550 Transformation Jane Li Assistant Professor Mechanical Engineering & Robotics Engineering http://users.wpi.edu/~zli11 Announcement Project

More information

A novel approach to motion tracking with wearable sensors based on Probabilistic Graphical Models

A novel approach to motion tracking with wearable sensors based on Probabilistic Graphical Models A novel approach to motion tracking with wearable sensors based on Probabilistic Graphical Models Emanuele Ruffaldi Lorenzo Peppoloni Alessandro Filippeschi Carlo Alberto Avizzano 2014 IEEE International

More information

Computer Animation Fundamentals. Animation Methods Keyframing Interpolation Kinematics Inverse Kinematics

Computer Animation Fundamentals. Animation Methods Keyframing Interpolation Kinematics Inverse Kinematics Computer Animation Fundamentals Animation Methods Keyframing Interpolation Kinematics Inverse Kinematics Lecture 21 6.837 Fall 2001 Conventional Animation Draw each frame of the animation great control

More information

Autonomous Navigation for Flying Robots

Autonomous Navigation for Flying Robots Computer Vision Group Prof. Daniel Cremers Autonomous Navigation for Flying Robots Lecture 3.1: 3D Geometry Jürgen Sturm Technische Universität München Points in 3D 3D point Augmented vector Homogeneous

More information

Stable Vision-Aided Navigation for Large-Area Augmented Reality

Stable Vision-Aided Navigation for Large-Area Augmented Reality Stable Vision-Aided Navigation for Large-Area Augmented Reality Taragay Oskiper, Han-Pang Chiu, Zhiwei Zhu Supun Samarasekera, Rakesh Teddy Kumar Vision and Robotics Laboratory SRI-International Sarnoff,

More information

Motivation. Parametric Curves (later Surfaces) Outline. Tangents, Normals, Binormals. Arclength. Advanced Computer Graphics (Fall 2010)

Motivation. Parametric Curves (later Surfaces) Outline. Tangents, Normals, Binormals. Arclength. Advanced Computer Graphics (Fall 2010) Advanced Computer Graphics (Fall 2010) CS 283, Lecture 19: Basic Geometric Concepts and Rotations Ravi Ramamoorthi http://inst.eecs.berkeley.edu/~cs283/fa10 Motivation Moving from rendering to simulation,

More information

Image processing and features

Image processing and features Image processing and features Gabriele Bleser gabriele.bleser@dfki.de Thanks to Harald Wuest, Folker Wientapper and Marc Pollefeys Introduction Previous lectures: geometry Pose estimation Epipolar geometry

More information

CMPUT 412 Motion Control Wheeled robots. Csaba Szepesvári University of Alberta

CMPUT 412 Motion Control Wheeled robots. Csaba Szepesvári University of Alberta CMPUT 412 Motion Control Wheeled robots Csaba Szepesvári University of Alberta 1 Motion Control (wheeled robots) Requirements Kinematic/dynamic model of the robot Model of the interaction between the wheel

More information

Animation. CS 4620 Lecture 32. Cornell CS4620 Fall Kavita Bala

Animation. CS 4620 Lecture 32. Cornell CS4620 Fall Kavita Bala Animation CS 4620 Lecture 32 Cornell CS4620 Fall 2015 1 What is animation? Modeling = specifying shape using all the tools we ve seen: hierarchies, meshes, curved surfaces Animation = specifying shape

More information

Animations. Hakan Bilen University of Edinburgh. Computer Graphics Fall Some slides are courtesy of Steve Marschner and Kavita Bala

Animations. Hakan Bilen University of Edinburgh. Computer Graphics Fall Some slides are courtesy of Steve Marschner and Kavita Bala Animations Hakan Bilen University of Edinburgh Computer Graphics Fall 2017 Some slides are courtesy of Steve Marschner and Kavita Bala Animation Artistic process What are animators trying to do? What tools

More information

Animation. Keyframe animation. CS4620/5620: Lecture 30. Rigid motion: the simplest deformation. Controlling shape for animation

Animation. Keyframe animation. CS4620/5620: Lecture 30. Rigid motion: the simplest deformation. Controlling shape for animation Keyframe animation CS4620/5620: Lecture 30 Animation Keyframing is the technique used for pose-to-pose animation User creates key poses just enough to indicate what the motion is supposed to be Interpolate

More information

Animation Lecture 10 Slide Fall 2003

Animation Lecture 10 Slide Fall 2003 Animation Lecture 10 Slide 1 6.837 Fall 2003 Conventional Animation Draw each frame of the animation great control tedious Reduce burden with cel animation layer keyframe inbetween cel panoramas (Disney

More information

CALCULATING TRANSFORMATIONS OF KINEMATIC CHAINS USING HOMOGENEOUS COORDINATES

CALCULATING TRANSFORMATIONS OF KINEMATIC CHAINS USING HOMOGENEOUS COORDINATES CALCULATING TRANSFORMATIONS OF KINEMATIC CHAINS USING HOMOGENEOUS COORDINATES YINGYING REN Abstract. In this paper, the applications of homogeneous coordinates are discussed to obtain an efficient model

More information

Visualizing Quaternions

Visualizing Quaternions Visualizing Quaternions Andrew J. Hanson Computer Science Department Indiana University Siggraph 1 Tutorial 1 GRAND PLAN I: Fundamentals of Quaternions II: Visualizing Quaternion Geometry III: Quaternion

More information

POME A mobile camera system for accurate indoor pose

POME A mobile camera system for accurate indoor pose POME A mobile camera system for accurate indoor pose Paul Montgomery & Andreas Winter November 2 2016 2010. All rights reserved. 1 ICT Intelligent Construction Tools A 50-50 joint venture between Trimble

More information

E80. Experimental Engineering. Lecture 9 Inertial Measurement

E80. Experimental Engineering. Lecture 9 Inertial Measurement Lecture 9 Inertial Measurement http://www.volker-doormann.org/physics.htm Feb. 19, 2013 Christopher M. Clark Where is the rocket? Outline Sensors People Accelerometers Gyroscopes Representations State

More information

Fusion of unsynchronized optical tracker and inertial sensor in EKF framework for in-car Augmented Reality delay reduction

Fusion of unsynchronized optical tracker and inertial sensor in EKF framework for in-car Augmented Reality delay reduction Fusion of unsynchronized optical tracker and inertial sensor in EKF framework for in-car Augmented Reality delay reduction Jason Rambach 1, Alain Pagani 1, Sebastian Lampe 2, Ruben Reiser 1, Manthan Pancholi

More information

An idea which can be used once is a trick. If it can be used more than once it becomes a method

An idea which can be used once is a trick. If it can be used more than once it becomes a method An idea which can be used once is a trick. If it can be used more than once it becomes a method - George Polya and Gabor Szego University of Texas at Arlington Rigid Body Transformations & Generalized

More information

METR 4202: Advanced Control & Robotics

METR 4202: Advanced Control & Robotics Position & Orientation & State t home with Homogenous Transformations METR 4202: dvanced Control & Robotics Drs Surya Singh, Paul Pounds, and Hanna Kurniawati Lecture # 2 July 30, 2012 metr4202@itee.uq.edu.au

More information

Augmented Reality, Advanced SLAM, Applications

Augmented Reality, Advanced SLAM, Applications Augmented Reality, Advanced SLAM, Applications Prof. Didier Stricker & Dr. Alain Pagani alain.pagani@dfki.de Lecture 3D Computer Vision AR, SLAM, Applications 1 Introduction Previous lectures: Basics (camera,

More information

FAB verses tradition camera-based motion capture systems

FAB verses tradition camera-based motion capture systems FAB verses tradition camera-based motion capture systems The advent of micromachined inertial sensors, such as rate gyroscopes and accelerometers, has made new navigation and tracking technologies possible.

More information

Massachusetts Institute of Technology Department of Computer Science and Electrical Engineering 6.801/6.866 Machine Vision QUIZ II

Massachusetts Institute of Technology Department of Computer Science and Electrical Engineering 6.801/6.866 Machine Vision QUIZ II Massachusetts Institute of Technology Department of Computer Science and Electrical Engineering 6.801/6.866 Machine Vision QUIZ II Handed out: 001 Nov. 30th Due on: 001 Dec. 10th Problem 1: (a (b Interior

More information

Today. Today. Introduction. Matrices. Matrices. Computergrafik. Transformations & matrices Introduction Matrices

Today. Today. Introduction. Matrices. Matrices. Computergrafik. Transformations & matrices Introduction Matrices Computergrafik Matthias Zwicker Universität Bern Herbst 2008 Today Transformations & matrices Introduction Matrices Homogeneous Affine transformations Concatenating transformations Change of Common coordinate

More information

Quaternions and Rotations

Quaternions and Rotations CSCI 520 Computer Animation and Simulation Quaternions and Rotations Jernej Barbic University of Southern California 1 Rotations Very important in computer animation and robotics Joint angles, rigid body

More information

Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview

Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview Arun Das 05/09/2017 Arun Das Waterloo Autonomous Vehicles Lab Introduction What s in a name? Arun Das Waterloo Autonomous

More information

CS354 Computer Graphics Rotations and Quaternions

CS354 Computer Graphics Rotations and Quaternions Slide Credit: Don Fussell CS354 Computer Graphics Rotations and Quaternions Qixing Huang April 4th 2018 Orientation Position and Orientation The position of an object can be represented as a translation

More information

Unit 2: Locomotion Kinematics of Wheeled Robots: Part 3

Unit 2: Locomotion Kinematics of Wheeled Robots: Part 3 Unit 2: Locomotion Kinematics of Wheeled Robots: Part 3 Computer Science 4766/6778 Department of Computer Science Memorial University of Newfoundland January 28, 2014 COMP 4766/6778 (MUN) Kinematics of

More information

Robots are built to accomplish complex and difficult tasks that require highly non-linear motions.

Robots are built to accomplish complex and difficult tasks that require highly non-linear motions. Path and Trajectory specification Robots are built to accomplish complex and difficult tasks that require highly non-linear motions. Specifying the desired motion to achieve a specified goal is often a

More information

Outline. ETN-FPI Training School on Plenoptic Sensing

Outline. ETN-FPI Training School on Plenoptic Sensing Outline Introduction Part I: Basics of Mathematical Optimization Linear Least Squares Nonlinear Optimization Part II: Basics of Computer Vision Camera Model Multi-Camera Model Multi-Camera Calibration

More information

Autonomous Navigation for Flying Robots

Autonomous Navigation for Flying Robots Computer Vision Group Prof. Daniel Cremers Autonomous Navigation for Flying Robots Lecture 3.2: Sensors Jürgen Sturm Technische Universität München Sensors IMUs (inertial measurement units) Accelerometers

More information

Analysis of Euler Angles in a Simple Two-Axis Gimbals Set

Analysis of Euler Angles in a Simple Two-Axis Gimbals Set Vol:5, No:9, 2 Analysis of Euler Angles in a Simple Two-Axis Gimbals Set Ma Myint Myint Aye International Science Index, Mechanical and Mechatronics Engineering Vol:5, No:9, 2 waset.org/publication/358

More information

(1) and s k ωk. p k vk q

(1) and s k ωk. p k vk q Sensing and Perception: Localization and positioning Isaac Sog Project Assignment: GNSS aided INS In this project assignment you will wor with a type of navigation system referred to as a global navigation

More information

Jacobian: Velocities and Static Forces 1/4

Jacobian: Velocities and Static Forces 1/4 Jacobian: Velocities and Static Forces /4 Models of Robot Manipulation - EE 54 - Department of Electrical Engineering - University of Washington Kinematics Relations - Joint & Cartesian Spaces A robot

More information

CS612 - Algorithms in Bioinformatics

CS612 - Algorithms in Bioinformatics Fall 2017 Structural Manipulation November 22, 2017 Rapid Structural Analysis Methods Emergence of large structural databases which do not allow manual (visual) analysis and require efficient 3-D search

More information

CS231A Midterm Review. Friday 5/6/2016

CS231A Midterm Review. Friday 5/6/2016 CS231A Midterm Review Friday 5/6/2016 Outline General Logistics Camera Models Non-perspective cameras Calibration Single View Metrology Epipolar Geometry Structure from Motion Active Stereo and Volumetric

More information

Laser sensors. Transmitter. Receiver. Basilio Bona ROBOTICA 03CFIOR

Laser sensors. Transmitter. Receiver. Basilio Bona ROBOTICA 03CFIOR Mobile & Service Robotics Sensors for Robotics 3 Laser sensors Rays are transmitted and received coaxially The target is illuminated by collimated rays The receiver measures the time of flight (back and

More information

Introduction to Robotics

Introduction to Robotics Introduction to Robotics Ph.D. Antonio Marin-Hernandez Artificial Intelligence Department Universidad Veracruzana Sebastian Camacho # 5 Xalapa, Veracruz Robotics Action and Perception LAAS-CNRS 7, av du

More information

CS283: Robotics Fall 2016: Sensors

CS283: Robotics Fall 2016: Sensors CS283: Robotics Fall 2016: Sensors Sören Schwertfeger / 师泽仁 ShanghaiTech University Robotics ShanghaiTech University - SIST - 23.09.2016 2 REVIEW TRANSFORMS Robotics ShanghaiTech University - SIST - 23.09.2016

More information

Motion Control (wheeled robots)

Motion Control (wheeled robots) Motion Control (wheeled robots) Requirements for Motion Control Kinematic / dynamic model of the robot Model of the interaction between the wheel and the ground Definition of required motion -> speed control,

More information

Structure from Motion. Prof. Marco Marcon

Structure from Motion. Prof. Marco Marcon Structure from Motion Prof. Marco Marcon Summing-up 2 Stereo is the most powerful clue for determining the structure of a scene Another important clue is the relative motion between the scene and (mono)

More information

ME 597: AUTONOMOUS MOBILE ROBOTICS SECTION 2 COORDINATE TRANSFORMS. Prof. Steven Waslander

ME 597: AUTONOMOUS MOBILE ROBOTICS SECTION 2 COORDINATE TRANSFORMS. Prof. Steven Waslander ME 597: AUTONOMOUS MOILE ROOTICS SECTION 2 COORDINATE TRANSFORMS Prof. Steven Waslander OUTLINE Coordinate Frames and Transforms Rotation Matrices Euler Angles Quaternions Homogeneous Transforms 2 COORDINATE

More information

3D Model Acquisition by Tracking 2D Wireframes

3D Model Acquisition by Tracking 2D Wireframes 3D Model Acquisition by Tracking 2D Wireframes M. Brown, T. Drummond and R. Cipolla {96mab twd20 cipolla}@eng.cam.ac.uk Department of Engineering University of Cambridge Cambridge CB2 1PZ, UK Abstract

More information

Practical Robotics (PRAC)

Practical Robotics (PRAC) Practical Robotics (PRAC) A Mobile Robot Navigation System (1) - Sensor and Kinematic Modelling Nick Pears University of York, Department of Computer Science December 17, 2014 nep (UoY CS) PRAC Practical

More information

Rotations (and other transformations) Rotation as rotation matrix. Storage. Apply to vector matrix vector multiply (15 flops)

Rotations (and other transformations) Rotation as rotation matrix. Storage. Apply to vector matrix vector multiply (15 flops) Cornell University CS 569: Interactive Computer Graphics Rotations (and other transformations) Lecture 4 2008 Steve Marschner 1 Rotation as rotation matrix 9 floats orthogonal and unit length columns and

More information

Image Formation. Antonino Furnari. Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania

Image Formation. Antonino Furnari. Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania Image Formation Antonino Furnari Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania furnari@dmi.unict.it 18/03/2014 Outline Introduction; Geometric Primitives

More information

Chapter 4 Dynamics. Part Constrained Kinematics and Dynamics. Mobile Robotics - Prof Alonzo Kelly, CMU RI

Chapter 4 Dynamics. Part Constrained Kinematics and Dynamics. Mobile Robotics - Prof Alonzo Kelly, CMU RI Chapter 4 Dynamics Part 2 4.3 Constrained Kinematics and Dynamics 1 Outline 4.3 Constrained Kinematics and Dynamics 4.3.1 Constraints of Disallowed Direction 4.3.2 Constraints of Rolling without Slipping

More information

Lecture 22 of 41. Animation 2 of 3: Rotations, Quaternions Dynamics & Kinematics

Lecture 22 of 41. Animation 2 of 3: Rotations, Quaternions Dynamics & Kinematics Animation 2 of 3: Rotations, Quaternions Dynamics & Kinematics William H. Hsu Department of Computing and Information Sciences, KSU KSOL course pages: http://bit.ly/hgvxlh / http://bit.ly/evizre Public

More information

Lecture 22 of 41. Animation 2 of 3: Rotations, Quaternions Dynamics & Kinematics

Lecture 22 of 41. Animation 2 of 3: Rotations, Quaternions Dynamics & Kinematics Animation 2 of 3: Rotations, Quaternions Dynamics & Kinematics William H. Hsu Department of Computing and Information Sciences, KSU KSOL course pages: http://bit.ly/hgvxlh / http://bit.ly/evizre Public

More information

Inverse Kinematics. Given a desired position (p) & orientation (R) of the end-effector

Inverse Kinematics. Given a desired position (p) & orientation (R) of the end-effector Inverse Kinematics Given a desired position (p) & orientation (R) of the end-effector q ( q, q, q ) 1 2 n Find the joint variables which can bring the robot the desired configuration z y x 1 The Inverse

More information

Kinematics, Kinematics Chains CS 685

Kinematics, Kinematics Chains CS 685 Kinematics, Kinematics Chains CS 685 Previously Representation of rigid body motion Two different interpretations - as transformations between different coord. frames - as operators acting on a rigid body

More information

Computer Vision. Coordinates. Prof. Flávio Cardeal DECOM / CEFET- MG.

Computer Vision. Coordinates. Prof. Flávio Cardeal DECOM / CEFET- MG. Computer Vision Coordinates Prof. Flávio Cardeal DECOM / CEFET- MG cardeal@decom.cefetmg.br Abstract This lecture discusses world coordinates and homogeneous coordinates, as well as provides an overview

More information

10/11/07 1. Motion Control (wheeled robots) Representing Robot Position ( ) ( ) [ ] T

10/11/07 1. Motion Control (wheeled robots) Representing Robot Position ( ) ( ) [ ] T 3 3 Motion Control (wheeled robots) Introduction: Mobile Robot Kinematics Requirements for Motion Control Kinematic / dynamic model of the robot Model of the interaction between the wheel and the ground

More information

Linear algebra deals with matrixes: two-dimensional arrays of values. Here s a matrix: [ x + 5y + 7z 9x + 3y + 11z

Linear algebra deals with matrixes: two-dimensional arrays of values. Here s a matrix: [ x + 5y + 7z 9x + 3y + 11z Basic Linear Algebra Linear algebra deals with matrixes: two-dimensional arrays of values. Here s a matrix: [ 1 5 ] 7 9 3 11 Often matrices are used to describe in a simpler way a series of linear equations.

More information

Chapter 4: Kinematics of Rigid Bodies

Chapter 4: Kinematics of Rigid Bodies Chapter 4: Kinematics of Rigid Bodies Advanced Dynamics Lecturer: Hossein Nejat Fall 2016 A rigid body is defined to be a collection of particles whose distance of separation is invariant. In this circumstance,

More information

Inertial Measurement Units I!

Inertial Measurement Units I! ! Inertial Measurement Units I! Gordon Wetzstein! Stanford University! EE 267 Virtual Reality! Lecture 9! stanford.edu/class/ee267/!! Lecture Overview! coordinate systems (world, body/sensor, inertial,

More information

Camera Projection Models We will introduce different camera projection models that relate the location of an image point to the coordinates of the

Camera Projection Models We will introduce different camera projection models that relate the location of an image point to the coordinates of the Camera Projection Models We will introduce different camera projection models that relate the location of an image point to the coordinates of the corresponding 3D points. The projection models include:

More information

CSE 252B: Computer Vision II

CSE 252B: Computer Vision II CSE 252B: Computer Vision II Lecturer: Serge Belongie Scribe: Sameer Agarwal LECTURE 1 Image Formation 1.1. The geometry of image formation We begin by considering the process of image formation when a

More information

Performance Study of Quaternion and Matrix Based Orientation for Camera Calibration

Performance Study of Quaternion and Matrix Based Orientation for Camera Calibration Performance Study of Quaternion and Matrix Based Orientation for Camera Calibration Rigoberto Juarez-Salazar 1, Carlos Robledo-Sánchez 2, Fermín Guerrero-Sánchez 2, J. Jacobo Oliveros-Oliveros 2, C. Meneses-Fabian

More information

Data-driven Approaches to Simulation (Motion Capture)

Data-driven Approaches to Simulation (Motion Capture) 1 Data-driven Approaches to Simulation (Motion Capture) Ting-Chun Sun tingchun.sun@usc.edu Preface The lecture slides [1] are made by Jessica Hodgins [2], who is a professor in Computer Science Department

More information

Quaternions & Rotation in 3D Space

Quaternions & Rotation in 3D Space Quaternions & Rotation in 3D Space 1 Overview Quaternions: definition Quaternion properties Quaternions and rotation matrices Quaternion-rotation matrices relationship Spherical linear interpolation Concluding

More information

Visualizing Quaternions

Visualizing Quaternions Visualizing Quaternions Andrew J. Hanson Computer Science Department Indiana University Siggraph 25 Tutorial OUTLINE I: (45min) Twisting Belts, Rolling Balls, and Locking Gimbals: Explaining Rotation Sequences

More information

Hand-Eye Calibration from Image Derivatives

Hand-Eye Calibration from Image Derivatives Hand-Eye Calibration from Image Derivatives Abstract In this paper it is shown how to perform hand-eye calibration using only the normal flow field and knowledge about the motion of the hand. The proposed

More information

VINet: Visual-Inertial Odometry as a Sequence-to-Sequence Learning Problem

VINet: Visual-Inertial Odometry as a Sequence-to-Sequence Learning Problem VINet: Visual-Inertial Odometry as a Sequence-to-Sequence Learning Problem Presented by: Justin Gorgen Yen-ting Chen Hao-en Sung Haifeng Huang University of California, San Diego May 23, 2017 Original

More information

Homogeneous Coordinates. Lecture18: Camera Models. Representation of Line and Point in 2D. Cross Product. Overall scaling is NOT important.

Homogeneous Coordinates. Lecture18: Camera Models. Representation of Line and Point in 2D. Cross Product. Overall scaling is NOT important. Homogeneous Coordinates Overall scaling is NOT important. CSED44:Introduction to Computer Vision (207F) Lecture8: Camera Models Bohyung Han CSE, POSTECH bhhan@postech.ac.kr (",, ) ()", ), )) ) 0 It is

More information

EE565:Mobile Robotics Lecture 2

EE565:Mobile Robotics Lecture 2 EE565:Mobile Robotics Lecture 2 Welcome Dr. Ing. Ahmad Kamal Nasir Organization Lab Course Lab grading policy (40%) Attendance = 10 % In-Lab tasks = 30 % Lab assignment + viva = 60 % Make a group Either

More information

3D Transformations. CS 4620 Lecture 10. Cornell CS4620 Fall 2014 Lecture Steve Marschner (with previous instructors James/Bala)

3D Transformations. CS 4620 Lecture 10. Cornell CS4620 Fall 2014 Lecture Steve Marschner (with previous instructors James/Bala) 3D Transformations CS 4620 Lecture 10 1 Translation 2 Scaling 3 Rotation about z axis 4 Rotation about x axis 5 Rotation about y axis 6 Properties of Matrices Translations: linear part is the identity

More information

θ x Week Date Lecture (M: 2:05p-3:50, 50-N202) 1 23-Jul Introduction + Representing Position & Orientation & State 2 30-Jul

θ x Week Date Lecture (M: 2:05p-3:50, 50-N202) 1 23-Jul Introduction + Representing Position & Orientation & State 2 30-Jul θ x 2018 School of Information Technology and Electrical Engineering at the University of Queensland Lecture Schedule Week Date Lecture (M: 2:05p-3:50, 50-N202) 1 23-Jul Introduction + Representing Position

More information

MEAM 620: HW 1. Sachin Chitta Assigned: January 10, 2007 Due: January 22, January 10, 2007

MEAM 620: HW 1. Sachin Chitta Assigned: January 10, 2007 Due: January 22, January 10, 2007 MEAM 620: HW 1 Sachin Chitta (sachinc@grasp.upenn.edu) Assigned: January 10, 2007 Due: January 22, 2006 January 10, 2007 1: MATLAB Programming assignment Using MATLAB, write the following functions: 1.

More information

Vision Review: Image Formation. Course web page:

Vision Review: Image Formation. Course web page: Vision Review: Image Formation Course web page: www.cis.udel.edu/~cer/arv September 10, 2002 Announcements Lecture on Thursday will be about Matlab; next Tuesday will be Image Processing The dates some

More information

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 253

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 253 Index 3D reconstruction, 123 5+1-point algorithm, 274 5-point algorithm, 260 7-point algorithm, 255 8-point algorithm, 253 affine point, 43 affine transformation, 55 affine transformation group, 55 affine

More information

a a= a a =a a 1 =1 Division turned out to be equivalent to multiplication: a b= a b =a 1 b

a a= a a =a a 1 =1 Division turned out to be equivalent to multiplication: a b= a b =a 1 b MATH 245 Extra Effort ( points) My assistant read through my first draft, got half a page in, and skipped to the end. So I will save you the flipping. Here is the assignment. Do just one of them. All the

More information

Kinematics. Kinematics analyzes the geometry of a manipulator, robot or machine motion. The essential concept is a position.

Kinematics. Kinematics analyzes the geometry of a manipulator, robot or machine motion. The essential concept is a position. Kinematics Kinematics analyzes the geometry of a manipulator, robot or machine motion. The essential concept is a position. 1/31 Statics deals with the forces and moments which are aplied on the mechanism

More information

Why animate humans? Why is this hard? Aspects of the Problem. These lectures. Animation Apreciation 101

Why animate humans? Why is this hard? Aspects of the Problem. These lectures. Animation Apreciation 101 Animation by Example Lecture 1: Introduction, Human Representation Michael Gleicher University of Wisconsin- Madison www.cs.wisc.edu/~gleicher www.cs.wisc.edu/graphics Why animate humans? Movies Television

More information

Introduction to Homogeneous coordinates

Introduction to Homogeneous coordinates Last class we considered smooth translations and rotations of the camera coordinate system and the resulting motions of points in the image projection plane. These two transformations were expressed mathematically

More information

3D Transformations. CS 4620 Lecture Kavita Bala w/ prior instructor Steve Marschner. Cornell CS4620 Fall 2015 Lecture 11

3D Transformations. CS 4620 Lecture Kavita Bala w/ prior instructor Steve Marschner. Cornell CS4620 Fall 2015 Lecture 11 3D Transformations CS 4620 Lecture 11 1 Announcements A2 due tomorrow Demos on Monday Please sign up for a slot Post on piazza 2 Translation 3 Scaling 4 Rotation about z axis 5 Rotation about x axis 6

More information

Introduction to Computer Vision

Introduction to Computer Vision Introduction to Computer Vision Michael J. Black Nov 2009 Perspective projection and affine motion Goals Today Perspective projection 3D motion Wed Projects Friday Regularization and robust statistics

More information

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing Visual servoing vision allows a robotic system to obtain geometrical and qualitative information on the surrounding environment high level control motion planning (look-and-move visual grasping) low level

More information

Model-based Visual Tracking:

Model-based Visual Tracking: Technische Universität München Model-based Visual Tracking: the OpenTL framework Giorgio Panin Technische Universität München Institut für Informatik Lehrstuhl für Echtzeitsysteme und Robotik (Prof. Alois

More information

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 263

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 263 Index 3D reconstruction, 125 5+1-point algorithm, 284 5-point algorithm, 270 7-point algorithm, 265 8-point algorithm, 263 affine point, 45 affine transformation, 57 affine transformation group, 57 affine

More information

CS 775: Advanced Computer Graphics. Lecture 3 : Kinematics

CS 775: Advanced Computer Graphics. Lecture 3 : Kinematics CS 775: Advanced Computer Graphics Lecture 3 : Kinematics Traditional Cell Animation, hand drawn, 2D Lead Animator for keyframes http://animation.about.com/od/flashanimationtutorials/ss/flash31detanim2.htm

More information

CALIBRATION BETWEEN DEPTH AND COLOR SENSORS FOR COMMODITY DEPTH CAMERAS. Cha Zhang and Zhengyou Zhang

CALIBRATION BETWEEN DEPTH AND COLOR SENSORS FOR COMMODITY DEPTH CAMERAS. Cha Zhang and Zhengyou Zhang CALIBRATION BETWEEN DEPTH AND COLOR SENSORS FOR COMMODITY DEPTH CAMERAS Cha Zhang and Zhengyou Zhang Communication and Collaboration Systems Group, Microsoft Research {chazhang, zhang}@microsoft.com ABSTRACT

More information

Table of Contents. Chapter 1. Modeling and Identification of Serial Robots... 1 Wisama KHALIL and Etienne DOMBRE

Table of Contents. Chapter 1. Modeling and Identification of Serial Robots... 1 Wisama KHALIL and Etienne DOMBRE Chapter 1. Modeling and Identification of Serial Robots.... 1 Wisama KHALIL and Etienne DOMBRE 1.1. Introduction... 1 1.2. Geometric modeling... 2 1.2.1. Geometric description... 2 1.2.2. Direct geometric

More information