Practical Robotics (PRAC)

Size: px
Start display at page:

Download "Practical Robotics (PRAC)"

Transcription

1 Practical Robotics (PRAC) A Mobile Robot Navigation System (1) - Sensor and Kinematic Modelling Nick Pears University of York, Department of Computer Science December 17, 2014 nep (UoY CS) PRAC Practical Robotics December 17, / 31

2 Practical Robotics (PRAC): Module outline - first half, Nick Pears and Alan Millard (labs) Lecture 1: A Mobile Robot Navigation System (1) - Sensor and Kinematic Modelling (Nick Pears, week 2) Lab 1a (week 2): Calibrating and characterising IR range sensors. Lab 1b (week 3): Calibrating and characterising the odometry system. Lecture 2: A Mobile Robot Navigation System (2) - Steering Control and Integrated Navigation System (Nick Pears, week 3) Lab 2a (week 4): A steering control system for wall following. Lab 2b (week 5): An integrated robot navigation system (time permitting!) nep (UoY CS) PRAC Practical Robotics December 17, / 31

3 Practical Robotics (PRAC): Module Outline - second half, Alan Millard and Adrian Bors Lecture 3 (Alan Millard, week 4): ROS and SLAM. Lab 3a (week 6): The Robot Operating System (ROS) Lab 3b (week 7): Swarm PI robot lab A Lab 3c (week 8): Swarm PI robot lab B Lecture 4 (Adrian Bors, week 5): Computer Vision (CV) for Mobile Robots (Optical Flow). Lab 4a (week 9): CV lab A Lab 4b (week 10): CV lab B nep (UoY CS) PRAC Practical Robotics December 17, / 31

4 Outline of this lecture The first part of this lecture will overview modelling in general and then discuss the kinematic modelling on the Pololu m3pi robot. 1 An overview of modelling in robotics. 2 Kinematic modelling of the robot. The second part of this lecture will look at each of the two main sensor systems that we will use 1 Calibrating, characterising and using the IR distance sensors in the navigation system (extereoceptive sensors). 2 Calibrating, characterising and using the odometry sensors in the navigation system (proprioceptive sensors). nep (UoY CS) PRAC Practical Robotics December 17, / 31

5 Lecture 1, part 1: What is robot modelling? 1 Kinematics modelling: mathematical representation of (i) the geometric layout of sensors and actuators, and the overall physical robot structure connecting these together and (ii) movement of the robot across all possible degrees of freedom. 2 Dynamics modelling: mathematical representation of how driving forces and torques in robot s actuators result in robot motion; again, over all possible degrees of freedom. 3 Sensor modelling: mathematical representation of the operation and characteristics of the robot s sensors. 4 Environment modelling: mathematical representation of the environment that the robot(s) operate in. 5 Multi-agent modelling: mathematical representation of other robots and other dynamic agents (eg. humans, pets). nep (UoY CS) PRAC Practical Robotics December 17, / 31

6 Why do we model? (1) If we model everything (robot, environment, other agents), we have a robot simulation. In terms of research: cheaper, easier, unsupervised, and (with enough compute power) possible to run faster than real-time, which is important for certain types of highly iterative algorithms (eg genetic algorithms). Beware of the gap between simulation and reality, due to unmodelled effects and inaccuracies in modelling: the reality gap. Modelling of the robot s kinematics and dynamics allows us to design control systems for smooth, efficient, accurate robot motion and actions (real or simulated). Often, direct manual measurement of model parameters to sufficient accuracy is difficult, and hence some form of model calibration procedures are required. nep (UoY CS) PRAC Practical Robotics December 17, / 31

7 Why do we model? (2) Sensor modelling allows us to characterise the behaviour of a sensor system, including operating range and measurement uncertainties due to various types of noise. Again, sensor parameters usually need to be determined by a calibration procedure. If calibration is possible through the normal operation of the robot, it is often called self calibration (eg cross referencing across several sensor systems). One aspect of autonomy is for a robot to learn its own models, at least parameters, sometimes even the overall model structure. nep (UoY CS) PRAC Practical Robotics December 17, / 31

8 Why do we model? (3) Environment modelling is essential for the autonomy of mobile platforms, as witnessed by the mapping aspect of SLAM algorithms (Simultaneous Localisation and Mapping). For a robot to behave intelligently in a space shared with multiple agents (other robots and humans), the robot must have a means of predicting the effect of its own behaviour within a dynamic environment, which may rapidly change over time, therefore multi-agent modelling is necessary. Look-ahead simulation can be used to implement short-term tactical behaviours. nep (UoY CS) PRAC Practical Robotics December 17, / 31

9 Kinematic modelling and steering control used in this study 1 Kinematic modelling of a differential wheeled robot (tank-like steering). 2 Controller that employs a potential field around a demand path allows recovery from large errors. Figure : The Pololu m3pi robot nep (UoY CS) PRAC Practical Robotics December 17, / 31

10 The robot frame and kinematic model parameters Differential drive is very simple to implement and three wheels do not require a suspension system. Kinematic model parameters: let a be the radius of the drive wheels and b be the wheelbase. i.e the distance between the wheels. Note also that there are potentially 4x3 sensor position parameters (discussed later). Infra red distance sensor x R R R R ( xs, ys, θs ) sensor coordinates in robot centred frame ω L y R θ R ω R a, drive wheel radius b, wheelbase Figure : The robot frame and kinematic parameters nep (UoY CS) PRAC Practical Robotics December 17, / 31

11 The robot frame and kinematic model parameters 1 Origin of the robot s coordinate system can be anywhere, natural choice is the midpoint of the drive wheels - why? 2 Note that the superscript R indicates a robot centred coorinate frame. 3 In this navigation system, there are three different coordinate frames: R - the robot centred frame (one frame). L - a local frame describing robot position relative to the start of some straight line path segment (n seg frames for n seg path segments). G - a global frame in which to describe path segments, targets (walls), and accumulated position via odometry (one frame). 4 Make sure you use the correct frame and know how to switch between frames using a rigid Euclidean transformation (rotation and 2D translation, reviewed later). nep (UoY CS) PRAC Practical Robotics December 17, / 31

12 Kinematic modelling of a differential drive 1 Let the angular velocities of the left and right wheels be ω L,R 2 Signs of angular velocities are defined for forward motion. 3 Looking from outside the robot, right wheel is positive clockwise, left is positive anticlockwise. 4 How do ω L,R relate to the robot s (tangential) speed, v, and turning radius r = 1 κ? (nb: both v and κ are signed scalars.) v θ r Figure : Robot on a constant radius, moves through angle θ in time t nep (UoY CS) PRAC Practical Robotics December 17, / 31

13 Kinematic modelling of a differential drive (r b 2 ) θ = aω L t (1) (r + b 2 ) θ = aω R t (2) Dividing the first equation (blue line, previous fig) by the second equation (red line, previous fig) gives Rearranging this equation (do it yourself later) 2r b 2r + b = ω L ω R (3) r = b(ω R + ω L ) 2(ω R ω L ) = b ω mean ω diff (4) nep (UoY CS) PRAC Practical Robotics December 17, / 31

14 Kinematic modelling of a differential drive Observation: the turning radius is independent of the drive wheel radius, a, as expected (?) Beware of th singularity (divide by zero) for a straight path (infinite radius), maybe it is better to computer turning curvature? κ = 1 r = 1 b ω diff ω mean (5) Note that κ = 0 for the same angular velocities, i.e. when ω diff = 0. However, now we have a singularity when turning on the spot, ω L = ω R hence ω mean = 0. Practical solution: compute value of ω mean and for values lower than some minimum, switch to turning radius representation. nep (UoY CS) PRAC Practical Robotics December 17, / 31

15 Kinematic modelling of a differential drive 1 We note that v = θr is linear in r and that the robot s origin is the midpoint between the two drive wheels. 2 This implies that the tangential speed, v, of the robot s origin along its curved path must be the average of the two tangential wheel speeds, as follows: v = v R + v L 2 = a(ω R + ω L ) 2 = aω mean (6) Thus the transformation (ω L, ω R ) (v, κ) from angular wheel speeds to tangential speed and turning curvature is v = a(ω R + ω L ) 2 = aω mean, κ = 2(ω R ω L ) b(ω R + ω L ) = ω diff bω mean. (7) nep (UoY CS) PRAC Practical Robotics December 17, / 31

16 Kinematic modelling of a differential drive 1 However, in our robot system, we want to specify the (signed scalars) robot speed (v) and steering curvature (κ), using software routines for velocity and steering control respectively. 2 Thus we need to invert the relation on the previous slide, in order to get the required mapping (v, κ) (ω R, ω L ) 3 To make v and κ independent of each other, we need to increase the speed of the right wheel by ω and decrease the left wheel speed by the same amount, relative to the mean angular wheel speed defined by the desired velocity, v. nep (UoY CS) PRAC Practical Robotics December 17, / 31

17 Kinematic modelling of a differential drive It is straightforward to show the relation below (prove it for yourself) ω = b (κv) (8) 2a hence the required relation (v, κ) (ω L, ω R ) is given by ω R = v κb (1 + a 2 ) (9) ω L = v κb (1 a 2 ) (10) For high positive curvatures, this may result in a negative value for ω L and the wheel rotation direction would need to be reversed. Similarly, for high negative curvatures, the right wheel rotation direction may need reversing. nep (UoY CS) PRAC Practical Robotics December 17, / 31

18 Lecture 1, part 2: Sensors used in this study 1 Infra-red (IR) distance sensors: extereoceptive not always available due to limited operating range errors fairly predictable and do not grow over time 2 Optical odometry: proprioceptive always available, errors much less predictable, not independent (x, y dependent on θ error) and grow with time 3 Note the complementary nature of these two sensor systems : they can be used to implement a predict-correct (odometry prediction, IR correction) localisation system. 4 Both systems require modelling : parameters to be found via direct measurement or calibration. nep (UoY CS) PRAC Practical Robotics December 17, / 31

19 The IR distance sensors that we will use Range sensor (10-80cm), with PSD (position sensitive detector), IRED (infrared emitting diode) and signal processing circuit. Reflectivity of the target object and environmental temperature don t strongly affect the measurement, due to the employed geometric triangulation method. Three connections, 5V voltage supply (30mA consumed), ground and analogue signal. Valid signal around 60mS after power up. Sample-and-hold arrangement with around 40mS hold interval, hence 25Hz is the approximate maximum sampling rate. nep (UoY CS) PRAC Practical Robotics December 17, / 31

20 IR distance sensor characteristics Output voltage is roughly in proportion to the position of the imaged IR spot on the PSD. This PSD position is roughly proportional to the inverse of the target s distance (a property of triangulation geometry). nep (UoY CS) PRAC Practical Robotics December 17, / 31

21 Triangulation geometry, analyse using similar triangles, z = fb p (11) target object illuminated spot depth, z IR emitter triangulation baseline, B lens, focal length f PSD with spot focussed at position, p p nep (UoY CS) PRAC Practical Robotics December 17, / 31

22 Triangulation geometry Image position, p and sensor range z are inversely proportional. For a small change in p (eg, due to sensor noise), what is the change in z? z p = fb p 2 = z2 fb (12) If random noise level on p was constant over all distances (z), measurement repeatability would drop as square of distance, just due to triangulation geometry. However, the signal drops as square of distance which, for a PSD, causes random noise on p to also increase with the square of distance. Combining both the geometric effect in (1) with the photometric effect in (2) results a ranging RMS error that increases with the fourth power of distance! nep (UoY CS) PRAC Practical Robotics December 17, / 31

23 IR sensor calibration How do we convert a raw voltage reading (via ADC) to a distance measurement? Choosing a method and finding that method s values or parameters is called sensor calibration. Possible methods: 1 Look up table voltage - distance with linear interpolation (piecewise linear sensor model). Look up table voltage - reciprocal distance with linear interpolation (more accurate?) 2 Fit a parametric function using least-squares model fitting Can use either the characteristic in (1) or (2) above using, for example, a polynomial (what order? linear, quadratic?) nep (UoY CS) PRAC Practical Robotics December 17, / 31

24 Local pose estimation with IR sensors 1 For local pose estimation, measure distance and orientation relative to target (wall). 2 This may be used to correct any errors in global position estimation (from odometry). 3 Need the position and orientation of the IR sensors in the robot s coordinate frame, (x R s, y R s, θ R s ) i for i = Assuming good sensor orientation alignment, and accurate symmetrical sensor distribution we may reduce from 12 to 2 parameters (α, β), see next slide. 5 In the practical labs, decide for yourself whether or not this regular rectangle simplifying assumption is acceptable in terms of accuracy in local pose estimation. nep (UoY CS) PRAC Practical Robotics December 17, / 31

25 Local pose estimation with IR sensors y L t d LF x R R R R ( xs, ys, θs ) sensor coordinates in robot centred frame y L t y L y R θ R θ L β α y L Note that y L is positive here θ L is negative here d LR Infra red distance sensor y L θ L x L local (path) frame where y L t θ L = tan 1 ( dlr d LF 2β ), y L = y L t d LR + α cos θ L (13) is the distance between the demand path and the target (wall). nep (UoY CS) PRAC Practical Robotics December 17, / 31

26 Odometry using optical shaft encoders 1 For every revolution of a drive wheel, n digital pulses are generated. 2 There are two square wave signal 90 degrees out of phase - phase indicates wheel s rotational direction (may not need this - why?). 3 How do we determine the robot s global position update based on the incremental change in drive wheel rotation over time t? 4 Suppose the changes in pulse counts over t are w L,R, then comparing to Eqn. 7, r = 1 κ = b( w R + w L ) 2( w R w L ) (14) θ = 2πa nb ( w R w L ) (15) nep (UoY CS) PRAC Practical Robotics December 17, / 31

27 Odometry using optical shaft encoders Odometry software function should consider two cases 1 When the (signed) returned count from each wheel is different. 2 When the (signed) return count from each wheel is the same, straight line, no rotation, hence r =. The first case is described overleaf, the second (simpler case) is left as an exercise. nep (UoY CS) PRAC Practical Robotics December 17, / 31

28 Odometry - position updates in the robot s own frame [ x R, y R ] = [r sin θ, r(1 cos θ)]; (16) r θ rsin θ x R y R r(1 cos θ ) nep (UoY CS) PRAC Practical Robotics December 17, / 31

29 Odometry - resolved into a global frame The coordinate changes in the robot s own frame need to be mapped into a global frame, y G G θ G x Global frame Resolve the red components yourself and check with eqns. θ r(1 cos θ ) r rsin θ θ G rsin θ rsin θ cos θ G sin θ G nep (UoY CS) PRAC Practical Robotics December 17, / 31

30 Odometry equations - in a global frame x G y G θ G i+1 x G y G θ G i i+1 x G x R = y + y (17) θ G θ i x G cos θ G sin θ G 0 r sin θ = y G + sin θ G cos θ G 0 r(1 cos θ) (18) θ G θ nep (UoY CS) PRAC Practical Robotics December 17, / 31

31 Closing slide This lecture will help you with the first two practical classes: Lab 1a (week 2): Calibrating, characterising and using the IR range sensors. Lab 1b (week 3): Calibrating, characterising and using the odometry sensors. Any questions? nep (UoY CS) PRAC Practical Robotics December 17, / 31

Encoder applications. I Most common use case: Combination with motors

Encoder applications. I Most common use case: Combination with motors 3.5 Rotation / Motion - Encoder applications 64-424 Intelligent Robotics Encoder applications I Most common use case: Combination with motors I Used to measure relative rotation angle, rotational direction

More information

CS283: Robotics Fall 2016: Sensors

CS283: Robotics Fall 2016: Sensors CS283: Robotics Fall 2016: Sensors Sören Schwertfeger / 师泽仁 ShanghaiTech University Robotics ShanghaiTech University - SIST - 23.09.2016 2 REVIEW TRANSFORMS Robotics ShanghaiTech University - SIST - 23.09.2016

More information

Localization, Where am I?

Localization, Where am I? 5.1 Localization, Where am I?? position Position Update (Estimation?) Encoder Prediction of Position (e.g. odometry) YES matched observations Map data base predicted position Matching Odometry, Dead Reckoning

More information

Chapter 4 Dynamics. Part Constrained Kinematics and Dynamics. Mobile Robotics - Prof Alonzo Kelly, CMU RI

Chapter 4 Dynamics. Part Constrained Kinematics and Dynamics. Mobile Robotics - Prof Alonzo Kelly, CMU RI Chapter 4 Dynamics Part 2 4.3 Constrained Kinematics and Dynamics 1 Outline 4.3 Constrained Kinematics and Dynamics 4.3.1 Constraints of Disallowed Direction 4.3.2 Constraints of Rolling without Slipping

More information

10/11/07 1. Motion Control (wheeled robots) Representing Robot Position ( ) ( ) [ ] T

10/11/07 1. Motion Control (wheeled robots) Representing Robot Position ( ) ( ) [ ] T 3 3 Motion Control (wheeled robots) Introduction: Mobile Robot Kinematics Requirements for Motion Control Kinematic / dynamic model of the robot Model of the interaction between the wheel and the ground

More information

Centre for Autonomous Systems

Centre for Autonomous Systems Robot Henrik I Centre for Autonomous Systems Kungl Tekniska Högskolan hic@kth.se 27th April 2005 Outline 1 duction 2 Kinematic and Constraints 3 Mobile Robot 4 Mobile Robot 5 Beyond Basic 6 Kinematic 7

More information

Laser sensors. Transmitter. Receiver. Basilio Bona ROBOTICA 03CFIOR

Laser sensors. Transmitter. Receiver. Basilio Bona ROBOTICA 03CFIOR Mobile & Service Robotics Sensors for Robotics 3 Laser sensors Rays are transmitted and received coaxially The target is illuminated by collimated rays The receiver measures the time of flight (back and

More information

A Simple Introduction to Omni Roller Robots (3rd April 2015)

A Simple Introduction to Omni Roller Robots (3rd April 2015) A Simple Introduction to Omni Roller Robots (3rd April 2015) Omni wheels have rollers all the way round the tread so they can slip laterally as well as drive in the direction of a regular wheel. The three-wheeled

More information

Motion Control (wheeled robots)

Motion Control (wheeled robots) Motion Control (wheeled robots) Requirements for Motion Control Kinematic / dynamic model of the robot Model of the interaction between the wheel and the ground Definition of required motion -> speed control,

More information

EE565:Mobile Robotics Lecture 2

EE565:Mobile Robotics Lecture 2 EE565:Mobile Robotics Lecture 2 Welcome Dr. Ing. Ahmad Kamal Nasir Organization Lab Course Lab grading policy (40%) Attendance = 10 % In-Lab tasks = 30 % Lab assignment + viva = 60 % Make a group Either

More information

Robotics. Lecture 5: Monte Carlo Localisation. See course website for up to date information.

Robotics. Lecture 5: Monte Carlo Localisation. See course website  for up to date information. Robotics Lecture 5: Monte Carlo Localisation See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information. Andrew Davison Department of Computing Imperial College London Review:

More information

Robotics (Kinematics) Winter 1393 Bonab University

Robotics (Kinematics) Winter 1393 Bonab University Robotics () Winter 1393 Bonab University : most basic study of how mechanical systems behave Introduction Need to understand the mechanical behavior for: Design Control Both: Manipulators, Mobile Robots

More information

Introduction to Robotics

Introduction to Robotics Introduction to Robotics Ph.D. Antonio Marin-Hernandez Artificial Intelligence Department Universidad Veracruzana Sebastian Camacho # 5 Xalapa, Veracruz Robotics Action and Perception LAAS-CNRS 7, av du

More information

Introduction to Mobile Robotics

Introduction to Mobile Robotics Introduction to Mobile Robotics Olivier Aycard Associate Professor University of Grenoble Laboratoire d Informatique de Grenoble http://membres-liglab.imag.fr/aycard 1/29 Some examples of mobile robots

More information

CMPUT 412 Motion Control Wheeled robots. Csaba Szepesvári University of Alberta

CMPUT 412 Motion Control Wheeled robots. Csaba Szepesvári University of Alberta CMPUT 412 Motion Control Wheeled robots Csaba Szepesvári University of Alberta 1 Motion Control (wheeled robots) Requirements Kinematic/dynamic model of the robot Model of the interaction between the wheel

More information

Motion Capture & Simulation

Motion Capture & Simulation Motion Capture & Simulation Motion Capture Character Reconstructions Joint Angles Need 3 points to compute a rigid body coordinate frame 1 st point gives 3D translation, 2 nd point gives 2 angles, 3 rd

More information

MTRX4700 Experimental Robotics

MTRX4700 Experimental Robotics MTRX 4700 : Experimental Robotics Lecture 2 Stefan B. Williams Slide 1 Course Outline Week Date Content Labs Due Dates 1 5 Mar Introduction, history & philosophy of robotics 2 12 Mar Robot kinematics &

More information

1 Differential Drive Kinematics

1 Differential Drive Kinematics CS W4733 NOTES - Differential Drive Robots Note: these notes were compiled from Dudek and Jenkin, Computational Principles of Mobile Robotics. 1 Differential Drive Kinematics Many mobile robots use a drive

More information

Sensor Modalities. Sensor modality: Different modalities:

Sensor Modalities. Sensor modality: Different modalities: Sensor Modalities Sensor modality: Sensors which measure same form of energy and process it in similar ways Modality refers to the raw input used by the sensors Different modalities: Sound Pressure Temperature

More information

Range Sensors (time of flight) (1)

Range Sensors (time of flight) (1) Range Sensors (time of flight) (1) Large range distance measurement -> called range sensors Range information: key element for localization and environment modeling Ultrasonic sensors, infra-red sensors

More information

Linear algebra deals with matrixes: two-dimensional arrays of values. Here s a matrix: [ x + 5y + 7z 9x + 3y + 11z

Linear algebra deals with matrixes: two-dimensional arrays of values. Here s a matrix: [ x + 5y + 7z 9x + 3y + 11z Basic Linear Algebra Linear algebra deals with matrixes: two-dimensional arrays of values. Here s a matrix: [ 1 5 ] 7 9 3 11 Often matrices are used to describe in a simpler way a series of linear equations.

More information

Introduction to Mobile Robotics

Introduction to Mobile Robotics Introduction to Mobile Robotics Olivier Aycard Associate Professor University of Grenoble Laboratoire d Informatique de Grenoble http://membres-liglab.imag.fr/aycard olivier. 1/22 What is a robot? Robot

More information

CHAPTER 3 MATHEMATICAL MODEL

CHAPTER 3 MATHEMATICAL MODEL 38 CHAPTER 3 MATHEMATICAL MODEL 3.1 KINEMATIC MODEL 3.1.1 Introduction The kinematic model of a mobile robot, represented by a set of equations, allows estimation of the robot s evolution on its trajectory,

More information

10/5/09 1. d = 2. Range Sensors (time of flight) (2) Ultrasonic Sensor (time of flight, sound) (1) Ultrasonic Sensor (time of flight, sound) (2) 4.1.

10/5/09 1. d = 2. Range Sensors (time of flight) (2) Ultrasonic Sensor (time of flight, sound) (1) Ultrasonic Sensor (time of flight, sound) (2) 4.1. Range Sensors (time of flight) (1) Range Sensors (time of flight) (2) arge range distance measurement -> called range sensors Range information: key element for localization and environment modeling Ultrasonic

More information

Unit 2: Locomotion Kinematics of Wheeled Robots: Part 3

Unit 2: Locomotion Kinematics of Wheeled Robots: Part 3 Unit 2: Locomotion Kinematics of Wheeled Robots: Part 3 Computer Science 4766/6778 Department of Computer Science Memorial University of Newfoundland January 28, 2014 COMP 4766/6778 (MUN) Kinematics of

More information

Kinematics, Kinematics Chains CS 685

Kinematics, Kinematics Chains CS 685 Kinematics, Kinematics Chains CS 685 Previously Representation of rigid body motion Two different interpretations - as transformations between different coord. frames - as operators acting on a rigid body

More information

MTRX4700: Experimental Robotics

MTRX4700: Experimental Robotics Stefan B. Williams April, 2013 MTR4700: Experimental Robotics Assignment 3 Note: This assignment contributes 10% towards your final mark. This assignment is due on Friday, May 10 th during Week 9 before

More information

ME 597/747 Autonomous Mobile Robots. Mid Term Exam. Duration: 2 hour Total Marks: 100

ME 597/747 Autonomous Mobile Robots. Mid Term Exam. Duration: 2 hour Total Marks: 100 ME 597/747 Autonomous Mobile Robots Mid Term Exam Duration: 2 hour Total Marks: 100 Instructions: Read the exam carefully before starting. Equations are at the back, but they are NOT necessarily valid

More information

Fundamental problems in mobile robotics

Fundamental problems in mobile robotics ROBOTICS 01PEEQW Basilio Bona DAUIN Politecnico di Torino Mobile & Service Robotics Kinematics Fundamental problems in mobile robotics Locomotion: how the robot moves in the environment Perception: how

More information

Mobile Robotics. Mathematics, Models, and Methods. HI Cambridge. Alonzo Kelly. Carnegie Mellon University UNIVERSITY PRESS

Mobile Robotics. Mathematics, Models, and Methods. HI Cambridge. Alonzo Kelly. Carnegie Mellon University UNIVERSITY PRESS Mobile Robotics Mathematics, Models, and Methods Alonzo Kelly Carnegie Mellon University HI Cambridge UNIVERSITY PRESS Contents Preface page xiii 1 Introduction 1 1.1 Applications of Mobile Robots 2 1.2

More information

Motion Models (cont) 1 3/15/2018

Motion Models (cont) 1 3/15/2018 Motion Models (cont) 1 3/15/018 Computing the Density to compute,, and use the appropriate probability density function; i.e., for zeromean Gaussian noise: 3/15/018 Sampling from the Velocity Motion Model

More information

METR 4202: Advanced Control & Robotics

METR 4202: Advanced Control & Robotics Position & Orientation & State t home with Homogenous Transformations METR 4202: dvanced Control & Robotics Drs Surya Singh, Paul Pounds, and Hanna Kurniawati Lecture # 2 July 30, 2012 metr4202@itee.uq.edu.au

More information

Exam in DD2426 Robotics and Autonomous Systems

Exam in DD2426 Robotics and Autonomous Systems Exam in DD2426 Robotics and Autonomous Systems Lecturer: Patric Jensfelt KTH, March 16, 2010, 9-12 No aids are allowed on the exam, i.e. no notes, no books, no calculators, etc. You need a minimum of 20

More information

Waypoint Navigation with Position and Heading Control using Complex Vector Fields for an Ackermann Steering Autonomous Vehicle

Waypoint Navigation with Position and Heading Control using Complex Vector Fields for an Ackermann Steering Autonomous Vehicle Waypoint Navigation with Position and Heading Control using Complex Vector Fields for an Ackermann Steering Autonomous Vehicle Tommie J. Liddy and Tien-Fu Lu School of Mechanical Engineering; The University

More information

WORKSPACE AGILITY FOR ROBOTIC ARM Karna Patel

WORKSPACE AGILITY FOR ROBOTIC ARM Karna Patel ISSN 30-9135 1 International Journal of Advance Research, IJOAR.org Volume 4, Issue 1, January 016, Online: ISSN 30-9135 WORKSPACE AGILITY FOR ROBOTIC ARM Karna Patel Karna Patel is currently pursuing

More information

Localization and Map Building

Localization and Map Building Localization and Map Building Noise and aliasing; odometric position estimation To localize or not to localize Belief representation Map representation Probabilistic map-based localization Other examples

More information

Jacobians. 6.1 Linearized Kinematics. Y: = k2( e6)

Jacobians. 6.1 Linearized Kinematics. Y: = k2( e6) Jacobians 6.1 Linearized Kinematics In previous chapters we have seen how kinematics relates the joint angles to the position and orientation of the robot's endeffector. This means that, for a serial robot,

More information

Introduction to Information Science and Technology (IST) Part IV: Intelligent Machines and Robotics Planning

Introduction to Information Science and Technology (IST) Part IV: Intelligent Machines and Robotics Planning Introduction to Information Science and Technology (IST) Part IV: Intelligent Machines and Robotics Planning Sören Schwertfeger / 师泽仁 ShanghaiTech University ShanghaiTech University - SIST - 10.05.2017

More information

Mobile Robot Kinematics

Mobile Robot Kinematics Mobile Robot Kinematics Dr. Kurtuluş Erinç Akdoğan kurtuluserinc@cankaya.edu.tr INTRODUCTION Kinematics is the most basic study of how mechanical systems behave required to design to control Manipulator

More information

Computer Vision I - Algorithms and Applications: Multi-View 3D reconstruction

Computer Vision I - Algorithms and Applications: Multi-View 3D reconstruction Computer Vision I - Algorithms and Applications: Multi-View 3D reconstruction Carsten Rother 09/12/2013 Computer Vision I: Multi-View 3D reconstruction Roadmap this lecture Computer Vision I: Multi-View

More information

CSE 490R P1 - Localization using Particle Filters Due date: Sun, Jan 28-11:59 PM

CSE 490R P1 - Localization using Particle Filters Due date: Sun, Jan 28-11:59 PM CSE 490R P1 - Localization using Particle Filters Due date: Sun, Jan 28-11:59 PM 1 Introduction In this assignment you will implement a particle filter to localize your car within a known map. This will

More information

Robotics and Autonomous Systems

Robotics and Autonomous Systems Robotics and Autonomous Systems Lecture 6: Perception/Odometry Terry Payne Department of Computer Science University of Liverpool 1 / 47 Today We ll talk about perception and motor control. 2 / 47 Perception

More information

Robotics and Autonomous Systems

Robotics and Autonomous Systems Robotics and Autonomous Systems Lecture 6: Perception/Odometry Simon Parsons Department of Computer Science University of Liverpool 1 / 47 Today We ll talk about perception and motor control. 2 / 47 Perception

More information

Essential Physics I. Lecture 13:

Essential Physics I. Lecture 13: Essential Physics I E I Lecture 13: 11-07-16 Reminders No lecture: Monday 18th July (holiday) Essay due: Monday 25th July, 4:30 pm 2 weeks!! Exam: Monday 1st August, 4:30 pm Announcements 250 word essay

More information

Mechanisms. Updated: 18Apr16 v7

Mechanisms. Updated: 18Apr16 v7 Mechanisms Updated: 8Apr6 v7 Mechanism Converts input motion or force into a desired output with four combinations of input and output motion Rotational to Oscillating Rotational to Rotational Rotational

More information

Kinematics of Wheeled Robots

Kinematics of Wheeled Robots CSE 390/MEAM 40 Kinematics of Wheeled Robots Professor Vijay Kumar Department of Mechanical Engineering and Applied Mechanics University of Pennsylvania September 16, 006 1 Introduction In this chapter,

More information

Omni Stereo Vision of Cooperative Mobile Robots

Omni Stereo Vision of Cooperative Mobile Robots Omni Stereo Vision of Cooperative Mobile Robots Zhigang Zhu*, Jizhong Xiao** *Department of Computer Science **Department of Electrical Engineering The City College of the City University of New York (CUNY)

More information

Graphics and Interaction Transformation geometry and homogeneous coordinates

Graphics and Interaction Transformation geometry and homogeneous coordinates 433-324 Graphics and Interaction Transformation geometry and homogeneous coordinates Department of Computer Science and Software Engineering The Lecture outline Introduction Vectors and matrices Translation

More information

Session #5 2D Mechanisms: Mobility, Kinematic Analysis & Synthesis

Session #5 2D Mechanisms: Mobility, Kinematic Analysis & Synthesis Session #5 2D Mechanisms: Mobility, Kinematic Analysis & Synthesis Courtesy of Design Simulation Technologies, Inc. Used with permission. Dan Frey Today s Agenda Collect assignment #2 Begin mechanisms

More information

COMP30019 Graphics and Interaction Transformation geometry and homogeneous coordinates

COMP30019 Graphics and Interaction Transformation geometry and homogeneous coordinates COMP30019 Graphics and Interaction Transformation geometry and homogeneous coordinates Department of Computer Science and Software Engineering The Lecture outline Introduction Vectors and matrices Translation

More information

CSE 252B: Computer Vision II

CSE 252B: Computer Vision II CSE 252B: Computer Vision II Lecturer: Serge Belongie Scribe: Sameer Agarwal LECTURE 1 Image Formation 1.1. The geometry of image formation We begin by considering the process of image formation when a

More information

Lecture «Robot Dynamics»: Multi-body Kinematics

Lecture «Robot Dynamics»: Multi-body Kinematics Lecture «Robot Dynamics»: Multi-body Kinematics 151-0851-00 V lecture: CAB G11 Tuesday 10:15 12:00, every week exercise: HG E1.2 Wednesday 8:15 10:00, according to schedule (about every 2nd week) Marco

More information

Lecture «Robot Dynamics»: Kinematics 3

Lecture «Robot Dynamics»: Kinematics 3 Lecture «Robot Dynamics»: Kinematics 3 151-0851-00 V lecture: CAB G11 Tuesday 10:15 12:00, every week exercise: HG E1.2 Wednesday 8:15 10:00, according to schedule (about every 2nd week) Marco Hutter,

More information

7 3-Sep Localization and Navigation (GPS, INS, & SLAM) 8 10-Sep State-space modelling & Controller Design 9 17-Sep Vision-based control

7 3-Sep Localization and Navigation (GPS, INS, & SLAM) 8 10-Sep State-space modelling & Controller Design 9 17-Sep Vision-based control RoboticsCourseWare Contributor 2012 School of Information Technology and Electrical Engineering at the University of Queensland Schedule Week Date Lecture (M: 12-1:30, 43-102) 1 23-Jul Introduction Representing

More information

Humanoid Robotics. Least Squares. Maren Bennewitz

Humanoid Robotics. Least Squares. Maren Bennewitz Humanoid Robotics Least Squares Maren Bennewitz Goal of This Lecture Introduction into least squares Use it yourself for odometry calibration, later in the lecture: camera and whole-body self-calibration

More information

ROBOTICS AND AUTONOMOUS SYSTEMS

ROBOTICS AND AUTONOMOUS SYSTEMS ROBOTICS AND AUTONOMOUS SYSTEMS Simon Parsons Department of Computer Science University of Liverpool LECTURE 6 PERCEPTION/ODOMETRY comp329-2013-parsons-lect06 2/43 Today We ll talk about perception and

More information

Lecture «Robot Dynamics»: Kinematics 3

Lecture «Robot Dynamics»: Kinematics 3 Lecture «Robot Dynamics»: Kinematics 3 151-0851-00 V lecture: CAB G11 Tuesday 10:15 12:00, every week exercise: HG E1.2 Wednesday 8:15 10:00, according to schedule (about every 2nd week) office hour: LEE

More information

Time Optimal Trajectories for Bounded Velocity Differential Drive Robots

Time Optimal Trajectories for Bounded Velocity Differential Drive Robots Time Optimal Trajectories for Bounded Velocity Differential Drive Robots Devin J. Balkcom Matthew T. Mason Robotics Institute and Computer Science Department Carnegie Mellon University Pittsburgh PA 53

More information

C / 35. C18 Computer Vision. David Murray. dwm/courses/4cv.

C / 35. C18 Computer Vision. David Murray.   dwm/courses/4cv. C18 2015 1 / 35 C18 Computer Vision David Murray david.murray@eng.ox.ac.uk www.robots.ox.ac.uk/ dwm/courses/4cv Michaelmas 2015 C18 2015 2 / 35 Computer Vision: This time... 1. Introduction; imaging geometry;

More information

Autonomous Mobile Robot Design

Autonomous Mobile Robot Design Autonomous Mobile Robot Design Topic: EKF-based SLAM Dr. Kostas Alexis (CSE) These slides have partially relied on the course of C. Stachniss, Robot Mapping - WS 2013/14 Autonomous Robot Challenges Where

More information

3. The three points (2, 4, 1), (1, 2, 2) and (5, 2, 2) determine a plane. Which of the following points is in that plane?

3. The three points (2, 4, 1), (1, 2, 2) and (5, 2, 2) determine a plane. Which of the following points is in that plane? Math 4 Practice Problems for Midterm. A unit vector that is perpendicular to both V =, 3, and W = 4,, is (a) V W (b) V W (c) 5 6 V W (d) 3 6 V W (e) 7 6 V W. In three dimensions, the graph of the equation

More information

INTRODUCTION REFLECTION AND REFRACTION AT BOUNDARIES. Introduction. Reflection and refraction at boundaries. Reflection at a single surface

INTRODUCTION REFLECTION AND REFRACTION AT BOUNDARIES. Introduction. Reflection and refraction at boundaries. Reflection at a single surface Chapter 8 GEOMETRICAL OPTICS Introduction Reflection and refraction at boundaries. Reflection at a single surface Refraction at a single boundary Dispersion Summary INTRODUCTION It has been shown that

More information

Differential Geometry: Circle Patterns (Part 1) [Discrete Conformal Mappinngs via Circle Patterns. Kharevych, Springborn and Schröder]

Differential Geometry: Circle Patterns (Part 1) [Discrete Conformal Mappinngs via Circle Patterns. Kharevych, Springborn and Schröder] Differential Geometry: Circle Patterns (Part 1) [Discrete Conformal Mappinngs via Circle Patterns. Kharevych, Springborn and Schröder] Preliminaries Recall: Given a smooth function f:r R, the function

More information

Animation. CS 4620 Lecture 32. Cornell CS4620 Fall Kavita Bala

Animation. CS 4620 Lecture 32. Cornell CS4620 Fall Kavita Bala Animation CS 4620 Lecture 32 Cornell CS4620 Fall 2015 1 What is animation? Modeling = specifying shape using all the tools we ve seen: hierarchies, meshes, curved surfaces Animation = specifying shape

More information

TEAMS National Competition Middle School Version Photometry Solution Manual 25 Questions

TEAMS National Competition Middle School Version Photometry Solution Manual 25 Questions TEAMS National Competition Middle School Version Photometry Solution Manual 25 Questions Page 1 of 14 Photometry Questions 1. When an upright object is placed between the focal point of a lens and a converging

More information

Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview

Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview Arun Das 05/09/2017 Arun Das Waterloo Autonomous Vehicles Lab Introduction What s in a name? Arun Das Waterloo Autonomous

More information

Chapter 11. Parametric Equations And Polar Coordinates

Chapter 11. Parametric Equations And Polar Coordinates Instructor: Prof. Dr. Ayman H. Sakka Chapter 11 Parametric Equations And Polar Coordinates In this chapter we study new ways to define curves in the plane, give geometric definitions of parabolas, ellipses,

More information

Inverse Kinematics. Given a desired position (p) & orientation (R) of the end-effector

Inverse Kinematics. Given a desired position (p) & orientation (R) of the end-effector Inverse Kinematics Given a desired position (p) & orientation (R) of the end-effector q ( q, q, q ) 1 2 n Find the joint variables which can bring the robot the desired configuration z y x 1 The Inverse

More information

Parameterization. Michael S. Floater. November 10, 2011

Parameterization. Michael S. Floater. November 10, 2011 Parameterization Michael S. Floater November 10, 2011 Triangular meshes are often used to represent surfaces, at least initially, one reason being that meshes are relatively easy to generate from point

More information

Data Association for SLAM

Data Association for SLAM CALIFORNIA INSTITUTE OF TECHNOLOGY ME/CS 132a, Winter 2011 Lab #2 Due: Mar 10th, 2011 Part I Data Association for SLAM 1 Introduction For this part, you will experiment with a simulation of an EKF SLAM

More information

Vision Review: Image Formation. Course web page:

Vision Review: Image Formation. Course web page: Vision Review: Image Formation Course web page: www.cis.udel.edu/~cer/arv September 10, 2002 Announcements Lecture on Thursday will be about Matlab; next Tuesday will be Image Processing The dates some

More information

MEAM 520. Mobile Robots

MEAM 520. Mobile Robots MEAM 520 Mobile Robots Katherine J. Kuchenbecker, Ph.D. General Robotics, Automation, Sensing, and Perception Lab (GRASP) MEAM Department, SEAS, Universit of Pennslvania Lecture 22: December 6, 2012 T

More information

Kinematics of Closed Chains

Kinematics of Closed Chains Chapter 7 Kinematics of Closed Chains Any kinematic chain that contains one or more loops is called a closed chain. Several examples of closed chains were encountered in Chapter 2, from the planar four-bar

More information

Homework Assignment /645 Fall Instructions and Score Sheet (hand in with answers)

Homework Assignment /645 Fall Instructions and Score Sheet (hand in with answers) Homework Assignment 4 600.445/645 Fall 2018 Instructions and Score Sheet (hand in with answers Name Email Other contact information (optional Signature (required I have followed the rules in completing

More information

Cinematica dei Robot Mobili su Ruote. Corso di Robotica Prof. Davide Brugali Università degli Studi di Bergamo

Cinematica dei Robot Mobili su Ruote. Corso di Robotica Prof. Davide Brugali Università degli Studi di Bergamo Cinematica dei Robot Mobili su Ruote Corso di Robotica Prof. Davide Brugali Università degli Studi di Bergamo Riferimenti bibliografici Roland SIEGWART, Illah R. NOURBAKHSH Introduction to Autonomous Mobile

More information

MEM380 Applied Autonomous Robots Winter Robot Kinematics

MEM380 Applied Autonomous Robots Winter Robot Kinematics MEM38 Applied Autonomous obots Winter obot Kinematics Coordinate Transformations Motivation Ultimatel, we are interested in the motion of the robot with respect to a global or inertial navigation frame

More information

0. Introduction: What is Computer Graphics? 1. Basics of scan conversion (line drawing) 2. Representing 2D curves

0. Introduction: What is Computer Graphics? 1. Basics of scan conversion (line drawing) 2. Representing 2D curves CSC 418/2504: Computer Graphics Course web site (includes course information sheet): http://www.dgp.toronto.edu/~elf Instructor: Eugene Fiume Office: BA 5266 Phone: 416 978 5472 (not a reliable way) Email:

More information

Robotic Perception and Action: Vehicle SLAM Assignment

Robotic Perception and Action: Vehicle SLAM Assignment Robotic Perception and Action: Vehicle SLAM Assignment Mariolino De Cecco Mariolino De Cecco, Mattia Tavernini 1 CONTENTS Vehicle SLAM Assignment Contents Assignment Scenario 3 Odometry Localization...........................................

More information

Physics Tutorial 2: Numerical Integration Methods

Physics Tutorial 2: Numerical Integration Methods Physics Tutorial 2: Numerical Integration Methods Summary The concept of numerically solving differential equations is explained, and applied to real time gaming simulation. Some objects are moved in a

More information

Planar Robot Kinematics

Planar Robot Kinematics V. Kumar lanar Robot Kinematics The mathematical modeling of spatial linkages is quite involved. t is useful to start with planar robots because the kinematics of planar mechanisms is generally much simpler

More information

Offline Simultaneous Localization and Mapping (SLAM) using Miniature Robots

Offline Simultaneous Localization and Mapping (SLAM) using Miniature Robots Offline Simultaneous Localization and Mapping (SLAM) using Miniature Robots Objectives SLAM approaches SLAM for ALICE EKF for Navigation Mapping and Network Modeling Test results Philipp Schaer and Adrian

More information

Parallel Lines Investigation

Parallel Lines Investigation Year 9 - The Maths Knowledge Autumn 1 (x, y) Along the corridor, up the stairs (3,1) x = 3 Gradient (-5,-2) (0,0) y-intercept Vertical lines are always x = y = 6 Horizontal lines are always y = Parallel

More information

Review of Trigonometry

Review of Trigonometry Worksheet 8 Properties of Trigonometric Functions Section Review of Trigonometry This section reviews some of the material covered in Worksheets 8, and The reader should be familiar with the trig ratios,

More information

EXAM SOLUTIONS. Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006,

EXAM SOLUTIONS. Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006, School of Computer Science and Communication, KTH Danica Kragic EXAM SOLUTIONS Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006, 14.00 19.00 Grade table 0-25 U 26-35 3 36-45

More information

1 Introduction. Control 1:

1 Introduction. Control 1: 1 1 Introduction Hierarchical Control for Mobile Robots Motive Autonomy Control Level We build controllers because: Real hardware ultimately responds to forces, energy, power, etc. whereas we are concerned

More information

AUTONOMOUS PLANETARY ROVER CONTROL USING INVERSE SIMULATION

AUTONOMOUS PLANETARY ROVER CONTROL USING INVERSE SIMULATION AUTONOMOUS PLANETARY ROVER CONTROL USING INVERSE SIMULATION Kevin Worrall (1), Douglas Thomson (1), Euan McGookin (1), Thaleia Flessa (1) (1)University of Glasgow, Glasgow, G12 8QQ, UK, Email: kevin.worrall@glasgow.ac.uk

More information

Computer Vision. Coordinates. Prof. Flávio Cardeal DECOM / CEFET- MG.

Computer Vision. Coordinates. Prof. Flávio Cardeal DECOM / CEFET- MG. Computer Vision Coordinates Prof. Flávio Cardeal DECOM / CEFET- MG cardeal@decom.cefetmg.br Abstract This lecture discusses world coordinates and homogeneous coordinates, as well as provides an overview

More information

Srikumar Ramalingam. Review. 3D Reconstruction. Pose Estimation Revisited. School of Computing University of Utah

Srikumar Ramalingam. Review. 3D Reconstruction. Pose Estimation Revisited. School of Computing University of Utah School of Computing University of Utah Presentation Outline 1 2 3 Forward Projection (Reminder) u v 1 KR ( I t ) X m Y m Z m 1 Backward Projection (Reminder) Q K 1 q Q K 1 u v 1 What is pose estimation?

More information

Lab1: Use of Word and Excel

Lab1: Use of Word and Excel Dr. Fritz Wilhelm; physics 230 Lab1: Use of Word and Excel Page 1 of 9 Lab partners: Download this page onto your computer. Also download the template file which you can use whenever you start your lab

More information

Robot. A thesis presented to. the faculty of. In partial fulfillment. of the requirements for the degree. Master of Science. Zachary J.

Robot. A thesis presented to. the faculty of. In partial fulfillment. of the requirements for the degree. Master of Science. Zachary J. Uncertainty Analysis throughout the Workspace of a Macro/Micro Cable Suspended Robot A thesis presented to the faculty of the Russ College of Engineering and Technology of Ohio University In partial fulfillment

More information

TRIGONOMETRIC FUNCTIONS

TRIGONOMETRIC FUNCTIONS Chapter TRIGONOMETRIC FUNCTIONS.1 Introduction A mathematician knows how to solve a problem, he can not solve it. MILNE The word trigonometry is derived from the Greek words trigon and metron and it means

More information

Localization, Mapping and Exploration with Multiple Robots. Dr. Daisy Tang

Localization, Mapping and Exploration with Multiple Robots. Dr. Daisy Tang Localization, Mapping and Exploration with Multiple Robots Dr. Daisy Tang Two Presentations A real-time algorithm for mobile robot mapping with applications to multi-robot and 3D mapping, by Thrun, Burgard

More information

Math for Geometric Optics

Math for Geometric Optics Algebra skills Math for Geometric Optics general rules some common types of equations units problems with several variables (substitution) Geometry basics Trigonometry Pythagorean theorem definitions,

More information

Basics of Localization, Mapping and SLAM. Jari Saarinen Aalto University Department of Automation and systems Technology

Basics of Localization, Mapping and SLAM. Jari Saarinen Aalto University Department of Automation and systems Technology Basics of Localization, Mapping and SLAM Jari Saarinen Aalto University Department of Automation and systems Technology Content Introduction to Problem (s) Localization A few basic equations Dead Reckoning

More information

LIGHT: Two-slit Interference

LIGHT: Two-slit Interference LIGHT: Two-slit Interference Objective: To study interference of light waves and verify the wave nature of light. Apparatus: Two red lasers (wavelength, λ = 633 nm); two orange lasers (λ = 612 nm); two

More information

CS 130 Final. Fall 2015

CS 130 Final. Fall 2015 CS 130 Final Fall 2015 Name Student ID Signature You may not ask any questions during the test. If you believe that there is something wrong with a question, write down what you think the question is trying

More information

MCE/EEC 647/747: Robot Dynamics and Control. Lecture 3: Forward and Inverse Kinematics

MCE/EEC 647/747: Robot Dynamics and Control. Lecture 3: Forward and Inverse Kinematics MCE/EEC 647/747: Robot Dynamics and Control Lecture 3: Forward and Inverse Kinematics Denavit-Hartenberg Convention Reading: SHV Chapter 3 Mechanical Engineering Hanz Richter, PhD MCE503 p.1/12 Aims of

More information

TEAMS National Competition High School Version Photometry Solution Manual 25 Questions

TEAMS National Competition High School Version Photometry Solution Manual 25 Questions TEAMS National Competition High School Version Photometry Solution Manual 25 Questions Page 1 of 15 Photometry Questions 1. When an upright object is placed between the focal point of a lens and a converging

More information

Physics 11 Chapter 18: Ray Optics

Physics 11 Chapter 18: Ray Optics Physics 11 Chapter 18: Ray Optics "... Everything can be taken from a man but one thing; the last of the human freedoms to choose one s attitude in any given set of circumstances, to choose one s own way.

More information

PARAMETRIC EQUATIONS AND POLAR COORDINATES

PARAMETRIC EQUATIONS AND POLAR COORDINATES 10 PARAMETRIC EQUATIONS AND POLAR COORDINATES PARAMETRIC EQUATIONS & POLAR COORDINATES A coordinate system represents a point in the plane by an ordered pair of numbers called coordinates. PARAMETRIC EQUATIONS

More information