Practical Robotics (PRAC)
|
|
- Priscilla Jordan
- 5 years ago
- Views:
Transcription
1 Practical Robotics (PRAC) A Mobile Robot Navigation System (1) - Sensor and Kinematic Modelling Nick Pears University of York, Department of Computer Science December 17, 2014 nep (UoY CS) PRAC Practical Robotics December 17, / 31
2 Practical Robotics (PRAC): Module outline - first half, Nick Pears and Alan Millard (labs) Lecture 1: A Mobile Robot Navigation System (1) - Sensor and Kinematic Modelling (Nick Pears, week 2) Lab 1a (week 2): Calibrating and characterising IR range sensors. Lab 1b (week 3): Calibrating and characterising the odometry system. Lecture 2: A Mobile Robot Navigation System (2) - Steering Control and Integrated Navigation System (Nick Pears, week 3) Lab 2a (week 4): A steering control system for wall following. Lab 2b (week 5): An integrated robot navigation system (time permitting!) nep (UoY CS) PRAC Practical Robotics December 17, / 31
3 Practical Robotics (PRAC): Module Outline - second half, Alan Millard and Adrian Bors Lecture 3 (Alan Millard, week 4): ROS and SLAM. Lab 3a (week 6): The Robot Operating System (ROS) Lab 3b (week 7): Swarm PI robot lab A Lab 3c (week 8): Swarm PI robot lab B Lecture 4 (Adrian Bors, week 5): Computer Vision (CV) for Mobile Robots (Optical Flow). Lab 4a (week 9): CV lab A Lab 4b (week 10): CV lab B nep (UoY CS) PRAC Practical Robotics December 17, / 31
4 Outline of this lecture The first part of this lecture will overview modelling in general and then discuss the kinematic modelling on the Pololu m3pi robot. 1 An overview of modelling in robotics. 2 Kinematic modelling of the robot. The second part of this lecture will look at each of the two main sensor systems that we will use 1 Calibrating, characterising and using the IR distance sensors in the navigation system (extereoceptive sensors). 2 Calibrating, characterising and using the odometry sensors in the navigation system (proprioceptive sensors). nep (UoY CS) PRAC Practical Robotics December 17, / 31
5 Lecture 1, part 1: What is robot modelling? 1 Kinematics modelling: mathematical representation of (i) the geometric layout of sensors and actuators, and the overall physical robot structure connecting these together and (ii) movement of the robot across all possible degrees of freedom. 2 Dynamics modelling: mathematical representation of how driving forces and torques in robot s actuators result in robot motion; again, over all possible degrees of freedom. 3 Sensor modelling: mathematical representation of the operation and characteristics of the robot s sensors. 4 Environment modelling: mathematical representation of the environment that the robot(s) operate in. 5 Multi-agent modelling: mathematical representation of other robots and other dynamic agents (eg. humans, pets). nep (UoY CS) PRAC Practical Robotics December 17, / 31
6 Why do we model? (1) If we model everything (robot, environment, other agents), we have a robot simulation. In terms of research: cheaper, easier, unsupervised, and (with enough compute power) possible to run faster than real-time, which is important for certain types of highly iterative algorithms (eg genetic algorithms). Beware of the gap between simulation and reality, due to unmodelled effects and inaccuracies in modelling: the reality gap. Modelling of the robot s kinematics and dynamics allows us to design control systems for smooth, efficient, accurate robot motion and actions (real or simulated). Often, direct manual measurement of model parameters to sufficient accuracy is difficult, and hence some form of model calibration procedures are required. nep (UoY CS) PRAC Practical Robotics December 17, / 31
7 Why do we model? (2) Sensor modelling allows us to characterise the behaviour of a sensor system, including operating range and measurement uncertainties due to various types of noise. Again, sensor parameters usually need to be determined by a calibration procedure. If calibration is possible through the normal operation of the robot, it is often called self calibration (eg cross referencing across several sensor systems). One aspect of autonomy is for a robot to learn its own models, at least parameters, sometimes even the overall model structure. nep (UoY CS) PRAC Practical Robotics December 17, / 31
8 Why do we model? (3) Environment modelling is essential for the autonomy of mobile platforms, as witnessed by the mapping aspect of SLAM algorithms (Simultaneous Localisation and Mapping). For a robot to behave intelligently in a space shared with multiple agents (other robots and humans), the robot must have a means of predicting the effect of its own behaviour within a dynamic environment, which may rapidly change over time, therefore multi-agent modelling is necessary. Look-ahead simulation can be used to implement short-term tactical behaviours. nep (UoY CS) PRAC Practical Robotics December 17, / 31
9 Kinematic modelling and steering control used in this study 1 Kinematic modelling of a differential wheeled robot (tank-like steering). 2 Controller that employs a potential field around a demand path allows recovery from large errors. Figure : The Pololu m3pi robot nep (UoY CS) PRAC Practical Robotics December 17, / 31
10 The robot frame and kinematic model parameters Differential drive is very simple to implement and three wheels do not require a suspension system. Kinematic model parameters: let a be the radius of the drive wheels and b be the wheelbase. i.e the distance between the wheels. Note also that there are potentially 4x3 sensor position parameters (discussed later). Infra red distance sensor x R R R R ( xs, ys, θs ) sensor coordinates in robot centred frame ω L y R θ R ω R a, drive wheel radius b, wheelbase Figure : The robot frame and kinematic parameters nep (UoY CS) PRAC Practical Robotics December 17, / 31
11 The robot frame and kinematic model parameters 1 Origin of the robot s coordinate system can be anywhere, natural choice is the midpoint of the drive wheels - why? 2 Note that the superscript R indicates a robot centred coorinate frame. 3 In this navigation system, there are three different coordinate frames: R - the robot centred frame (one frame). L - a local frame describing robot position relative to the start of some straight line path segment (n seg frames for n seg path segments). G - a global frame in which to describe path segments, targets (walls), and accumulated position via odometry (one frame). 4 Make sure you use the correct frame and know how to switch between frames using a rigid Euclidean transformation (rotation and 2D translation, reviewed later). nep (UoY CS) PRAC Practical Robotics December 17, / 31
12 Kinematic modelling of a differential drive 1 Let the angular velocities of the left and right wheels be ω L,R 2 Signs of angular velocities are defined for forward motion. 3 Looking from outside the robot, right wheel is positive clockwise, left is positive anticlockwise. 4 How do ω L,R relate to the robot s (tangential) speed, v, and turning radius r = 1 κ? (nb: both v and κ are signed scalars.) v θ r Figure : Robot on a constant radius, moves through angle θ in time t nep (UoY CS) PRAC Practical Robotics December 17, / 31
13 Kinematic modelling of a differential drive (r b 2 ) θ = aω L t (1) (r + b 2 ) θ = aω R t (2) Dividing the first equation (blue line, previous fig) by the second equation (red line, previous fig) gives Rearranging this equation (do it yourself later) 2r b 2r + b = ω L ω R (3) r = b(ω R + ω L ) 2(ω R ω L ) = b ω mean ω diff (4) nep (UoY CS) PRAC Practical Robotics December 17, / 31
14 Kinematic modelling of a differential drive Observation: the turning radius is independent of the drive wheel radius, a, as expected (?) Beware of th singularity (divide by zero) for a straight path (infinite radius), maybe it is better to computer turning curvature? κ = 1 r = 1 b ω diff ω mean (5) Note that κ = 0 for the same angular velocities, i.e. when ω diff = 0. However, now we have a singularity when turning on the spot, ω L = ω R hence ω mean = 0. Practical solution: compute value of ω mean and for values lower than some minimum, switch to turning radius representation. nep (UoY CS) PRAC Practical Robotics December 17, / 31
15 Kinematic modelling of a differential drive 1 We note that v = θr is linear in r and that the robot s origin is the midpoint between the two drive wheels. 2 This implies that the tangential speed, v, of the robot s origin along its curved path must be the average of the two tangential wheel speeds, as follows: v = v R + v L 2 = a(ω R + ω L ) 2 = aω mean (6) Thus the transformation (ω L, ω R ) (v, κ) from angular wheel speeds to tangential speed and turning curvature is v = a(ω R + ω L ) 2 = aω mean, κ = 2(ω R ω L ) b(ω R + ω L ) = ω diff bω mean. (7) nep (UoY CS) PRAC Practical Robotics December 17, / 31
16 Kinematic modelling of a differential drive 1 However, in our robot system, we want to specify the (signed scalars) robot speed (v) and steering curvature (κ), using software routines for velocity and steering control respectively. 2 Thus we need to invert the relation on the previous slide, in order to get the required mapping (v, κ) (ω R, ω L ) 3 To make v and κ independent of each other, we need to increase the speed of the right wheel by ω and decrease the left wheel speed by the same amount, relative to the mean angular wheel speed defined by the desired velocity, v. nep (UoY CS) PRAC Practical Robotics December 17, / 31
17 Kinematic modelling of a differential drive It is straightforward to show the relation below (prove it for yourself) ω = b (κv) (8) 2a hence the required relation (v, κ) (ω L, ω R ) is given by ω R = v κb (1 + a 2 ) (9) ω L = v κb (1 a 2 ) (10) For high positive curvatures, this may result in a negative value for ω L and the wheel rotation direction would need to be reversed. Similarly, for high negative curvatures, the right wheel rotation direction may need reversing. nep (UoY CS) PRAC Practical Robotics December 17, / 31
18 Lecture 1, part 2: Sensors used in this study 1 Infra-red (IR) distance sensors: extereoceptive not always available due to limited operating range errors fairly predictable and do not grow over time 2 Optical odometry: proprioceptive always available, errors much less predictable, not independent (x, y dependent on θ error) and grow with time 3 Note the complementary nature of these two sensor systems : they can be used to implement a predict-correct (odometry prediction, IR correction) localisation system. 4 Both systems require modelling : parameters to be found via direct measurement or calibration. nep (UoY CS) PRAC Practical Robotics December 17, / 31
19 The IR distance sensors that we will use Range sensor (10-80cm), with PSD (position sensitive detector), IRED (infrared emitting diode) and signal processing circuit. Reflectivity of the target object and environmental temperature don t strongly affect the measurement, due to the employed geometric triangulation method. Three connections, 5V voltage supply (30mA consumed), ground and analogue signal. Valid signal around 60mS after power up. Sample-and-hold arrangement with around 40mS hold interval, hence 25Hz is the approximate maximum sampling rate. nep (UoY CS) PRAC Practical Robotics December 17, / 31
20 IR distance sensor characteristics Output voltage is roughly in proportion to the position of the imaged IR spot on the PSD. This PSD position is roughly proportional to the inverse of the target s distance (a property of triangulation geometry). nep (UoY CS) PRAC Practical Robotics December 17, / 31
21 Triangulation geometry, analyse using similar triangles, z = fb p (11) target object illuminated spot depth, z IR emitter triangulation baseline, B lens, focal length f PSD with spot focussed at position, p p nep (UoY CS) PRAC Practical Robotics December 17, / 31
22 Triangulation geometry Image position, p and sensor range z are inversely proportional. For a small change in p (eg, due to sensor noise), what is the change in z? z p = fb p 2 = z2 fb (12) If random noise level on p was constant over all distances (z), measurement repeatability would drop as square of distance, just due to triangulation geometry. However, the signal drops as square of distance which, for a PSD, causes random noise on p to also increase with the square of distance. Combining both the geometric effect in (1) with the photometric effect in (2) results a ranging RMS error that increases with the fourth power of distance! nep (UoY CS) PRAC Practical Robotics December 17, / 31
23 IR sensor calibration How do we convert a raw voltage reading (via ADC) to a distance measurement? Choosing a method and finding that method s values or parameters is called sensor calibration. Possible methods: 1 Look up table voltage - distance with linear interpolation (piecewise linear sensor model). Look up table voltage - reciprocal distance with linear interpolation (more accurate?) 2 Fit a parametric function using least-squares model fitting Can use either the characteristic in (1) or (2) above using, for example, a polynomial (what order? linear, quadratic?) nep (UoY CS) PRAC Practical Robotics December 17, / 31
24 Local pose estimation with IR sensors 1 For local pose estimation, measure distance and orientation relative to target (wall). 2 This may be used to correct any errors in global position estimation (from odometry). 3 Need the position and orientation of the IR sensors in the robot s coordinate frame, (x R s, y R s, θ R s ) i for i = Assuming good sensor orientation alignment, and accurate symmetrical sensor distribution we may reduce from 12 to 2 parameters (α, β), see next slide. 5 In the practical labs, decide for yourself whether or not this regular rectangle simplifying assumption is acceptable in terms of accuracy in local pose estimation. nep (UoY CS) PRAC Practical Robotics December 17, / 31
25 Local pose estimation with IR sensors y L t d LF x R R R R ( xs, ys, θs ) sensor coordinates in robot centred frame y L t y L y R θ R θ L β α y L Note that y L is positive here θ L is negative here d LR Infra red distance sensor y L θ L x L local (path) frame where y L t θ L = tan 1 ( dlr d LF 2β ), y L = y L t d LR + α cos θ L (13) is the distance between the demand path and the target (wall). nep (UoY CS) PRAC Practical Robotics December 17, / 31
26 Odometry using optical shaft encoders 1 For every revolution of a drive wheel, n digital pulses are generated. 2 There are two square wave signal 90 degrees out of phase - phase indicates wheel s rotational direction (may not need this - why?). 3 How do we determine the robot s global position update based on the incremental change in drive wheel rotation over time t? 4 Suppose the changes in pulse counts over t are w L,R, then comparing to Eqn. 7, r = 1 κ = b( w R + w L ) 2( w R w L ) (14) θ = 2πa nb ( w R w L ) (15) nep (UoY CS) PRAC Practical Robotics December 17, / 31
27 Odometry using optical shaft encoders Odometry software function should consider two cases 1 When the (signed) returned count from each wheel is different. 2 When the (signed) return count from each wheel is the same, straight line, no rotation, hence r =. The first case is described overleaf, the second (simpler case) is left as an exercise. nep (UoY CS) PRAC Practical Robotics December 17, / 31
28 Odometry - position updates in the robot s own frame [ x R, y R ] = [r sin θ, r(1 cos θ)]; (16) r θ rsin θ x R y R r(1 cos θ ) nep (UoY CS) PRAC Practical Robotics December 17, / 31
29 Odometry - resolved into a global frame The coordinate changes in the robot s own frame need to be mapped into a global frame, y G G θ G x Global frame Resolve the red components yourself and check with eqns. θ r(1 cos θ ) r rsin θ θ G rsin θ rsin θ cos θ G sin θ G nep (UoY CS) PRAC Practical Robotics December 17, / 31
30 Odometry equations - in a global frame x G y G θ G i+1 x G y G θ G i i+1 x G x R = y + y (17) θ G θ i x G cos θ G sin θ G 0 r sin θ = y G + sin θ G cos θ G 0 r(1 cos θ) (18) θ G θ nep (UoY CS) PRAC Practical Robotics December 17, / 31
31 Closing slide This lecture will help you with the first two practical classes: Lab 1a (week 2): Calibrating, characterising and using the IR range sensors. Lab 1b (week 3): Calibrating, characterising and using the odometry sensors. Any questions? nep (UoY CS) PRAC Practical Robotics December 17, / 31
Encoder applications. I Most common use case: Combination with motors
3.5 Rotation / Motion - Encoder applications 64-424 Intelligent Robotics Encoder applications I Most common use case: Combination with motors I Used to measure relative rotation angle, rotational direction
More informationCS283: Robotics Fall 2016: Sensors
CS283: Robotics Fall 2016: Sensors Sören Schwertfeger / 师泽仁 ShanghaiTech University Robotics ShanghaiTech University - SIST - 23.09.2016 2 REVIEW TRANSFORMS Robotics ShanghaiTech University - SIST - 23.09.2016
More informationLocalization, Where am I?
5.1 Localization, Where am I?? position Position Update (Estimation?) Encoder Prediction of Position (e.g. odometry) YES matched observations Map data base predicted position Matching Odometry, Dead Reckoning
More informationChapter 4 Dynamics. Part Constrained Kinematics and Dynamics. Mobile Robotics - Prof Alonzo Kelly, CMU RI
Chapter 4 Dynamics Part 2 4.3 Constrained Kinematics and Dynamics 1 Outline 4.3 Constrained Kinematics and Dynamics 4.3.1 Constraints of Disallowed Direction 4.3.2 Constraints of Rolling without Slipping
More information10/11/07 1. Motion Control (wheeled robots) Representing Robot Position ( ) ( ) [ ] T
3 3 Motion Control (wheeled robots) Introduction: Mobile Robot Kinematics Requirements for Motion Control Kinematic / dynamic model of the robot Model of the interaction between the wheel and the ground
More informationCentre for Autonomous Systems
Robot Henrik I Centre for Autonomous Systems Kungl Tekniska Högskolan hic@kth.se 27th April 2005 Outline 1 duction 2 Kinematic and Constraints 3 Mobile Robot 4 Mobile Robot 5 Beyond Basic 6 Kinematic 7
More informationLaser sensors. Transmitter. Receiver. Basilio Bona ROBOTICA 03CFIOR
Mobile & Service Robotics Sensors for Robotics 3 Laser sensors Rays are transmitted and received coaxially The target is illuminated by collimated rays The receiver measures the time of flight (back and
More informationA Simple Introduction to Omni Roller Robots (3rd April 2015)
A Simple Introduction to Omni Roller Robots (3rd April 2015) Omni wheels have rollers all the way round the tread so they can slip laterally as well as drive in the direction of a regular wheel. The three-wheeled
More informationMotion Control (wheeled robots)
Motion Control (wheeled robots) Requirements for Motion Control Kinematic / dynamic model of the robot Model of the interaction between the wheel and the ground Definition of required motion -> speed control,
More informationEE565:Mobile Robotics Lecture 2
EE565:Mobile Robotics Lecture 2 Welcome Dr. Ing. Ahmad Kamal Nasir Organization Lab Course Lab grading policy (40%) Attendance = 10 % In-Lab tasks = 30 % Lab assignment + viva = 60 % Make a group Either
More informationRobotics. Lecture 5: Monte Carlo Localisation. See course website for up to date information.
Robotics Lecture 5: Monte Carlo Localisation See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information. Andrew Davison Department of Computing Imperial College London Review:
More informationRobotics (Kinematics) Winter 1393 Bonab University
Robotics () Winter 1393 Bonab University : most basic study of how mechanical systems behave Introduction Need to understand the mechanical behavior for: Design Control Both: Manipulators, Mobile Robots
More informationIntroduction to Robotics
Introduction to Robotics Ph.D. Antonio Marin-Hernandez Artificial Intelligence Department Universidad Veracruzana Sebastian Camacho # 5 Xalapa, Veracruz Robotics Action and Perception LAAS-CNRS 7, av du
More informationIntroduction to Mobile Robotics
Introduction to Mobile Robotics Olivier Aycard Associate Professor University of Grenoble Laboratoire d Informatique de Grenoble http://membres-liglab.imag.fr/aycard 1/29 Some examples of mobile robots
More informationCMPUT 412 Motion Control Wheeled robots. Csaba Szepesvári University of Alberta
CMPUT 412 Motion Control Wheeled robots Csaba Szepesvári University of Alberta 1 Motion Control (wheeled robots) Requirements Kinematic/dynamic model of the robot Model of the interaction between the wheel
More informationMotion Capture & Simulation
Motion Capture & Simulation Motion Capture Character Reconstructions Joint Angles Need 3 points to compute a rigid body coordinate frame 1 st point gives 3D translation, 2 nd point gives 2 angles, 3 rd
More informationMTRX4700 Experimental Robotics
MTRX 4700 : Experimental Robotics Lecture 2 Stefan B. Williams Slide 1 Course Outline Week Date Content Labs Due Dates 1 5 Mar Introduction, history & philosophy of robotics 2 12 Mar Robot kinematics &
More information1 Differential Drive Kinematics
CS W4733 NOTES - Differential Drive Robots Note: these notes were compiled from Dudek and Jenkin, Computational Principles of Mobile Robotics. 1 Differential Drive Kinematics Many mobile robots use a drive
More informationSensor Modalities. Sensor modality: Different modalities:
Sensor Modalities Sensor modality: Sensors which measure same form of energy and process it in similar ways Modality refers to the raw input used by the sensors Different modalities: Sound Pressure Temperature
More informationRange Sensors (time of flight) (1)
Range Sensors (time of flight) (1) Large range distance measurement -> called range sensors Range information: key element for localization and environment modeling Ultrasonic sensors, infra-red sensors
More informationLinear algebra deals with matrixes: two-dimensional arrays of values. Here s a matrix: [ x + 5y + 7z 9x + 3y + 11z
Basic Linear Algebra Linear algebra deals with matrixes: two-dimensional arrays of values. Here s a matrix: [ 1 5 ] 7 9 3 11 Often matrices are used to describe in a simpler way a series of linear equations.
More informationIntroduction to Mobile Robotics
Introduction to Mobile Robotics Olivier Aycard Associate Professor University of Grenoble Laboratoire d Informatique de Grenoble http://membres-liglab.imag.fr/aycard olivier. 1/22 What is a robot? Robot
More informationCHAPTER 3 MATHEMATICAL MODEL
38 CHAPTER 3 MATHEMATICAL MODEL 3.1 KINEMATIC MODEL 3.1.1 Introduction The kinematic model of a mobile robot, represented by a set of equations, allows estimation of the robot s evolution on its trajectory,
More information10/5/09 1. d = 2. Range Sensors (time of flight) (2) Ultrasonic Sensor (time of flight, sound) (1) Ultrasonic Sensor (time of flight, sound) (2) 4.1.
Range Sensors (time of flight) (1) Range Sensors (time of flight) (2) arge range distance measurement -> called range sensors Range information: key element for localization and environment modeling Ultrasonic
More informationUnit 2: Locomotion Kinematics of Wheeled Robots: Part 3
Unit 2: Locomotion Kinematics of Wheeled Robots: Part 3 Computer Science 4766/6778 Department of Computer Science Memorial University of Newfoundland January 28, 2014 COMP 4766/6778 (MUN) Kinematics of
More informationKinematics, Kinematics Chains CS 685
Kinematics, Kinematics Chains CS 685 Previously Representation of rigid body motion Two different interpretations - as transformations between different coord. frames - as operators acting on a rigid body
More informationMTRX4700: Experimental Robotics
Stefan B. Williams April, 2013 MTR4700: Experimental Robotics Assignment 3 Note: This assignment contributes 10% towards your final mark. This assignment is due on Friday, May 10 th during Week 9 before
More informationME 597/747 Autonomous Mobile Robots. Mid Term Exam. Duration: 2 hour Total Marks: 100
ME 597/747 Autonomous Mobile Robots Mid Term Exam Duration: 2 hour Total Marks: 100 Instructions: Read the exam carefully before starting. Equations are at the back, but they are NOT necessarily valid
More informationFundamental problems in mobile robotics
ROBOTICS 01PEEQW Basilio Bona DAUIN Politecnico di Torino Mobile & Service Robotics Kinematics Fundamental problems in mobile robotics Locomotion: how the robot moves in the environment Perception: how
More informationMobile Robotics. Mathematics, Models, and Methods. HI Cambridge. Alonzo Kelly. Carnegie Mellon University UNIVERSITY PRESS
Mobile Robotics Mathematics, Models, and Methods Alonzo Kelly Carnegie Mellon University HI Cambridge UNIVERSITY PRESS Contents Preface page xiii 1 Introduction 1 1.1 Applications of Mobile Robots 2 1.2
More informationMotion Models (cont) 1 3/15/2018
Motion Models (cont) 1 3/15/018 Computing the Density to compute,, and use the appropriate probability density function; i.e., for zeromean Gaussian noise: 3/15/018 Sampling from the Velocity Motion Model
More informationMETR 4202: Advanced Control & Robotics
Position & Orientation & State t home with Homogenous Transformations METR 4202: dvanced Control & Robotics Drs Surya Singh, Paul Pounds, and Hanna Kurniawati Lecture # 2 July 30, 2012 metr4202@itee.uq.edu.au
More informationExam in DD2426 Robotics and Autonomous Systems
Exam in DD2426 Robotics and Autonomous Systems Lecturer: Patric Jensfelt KTH, March 16, 2010, 9-12 No aids are allowed on the exam, i.e. no notes, no books, no calculators, etc. You need a minimum of 20
More informationWaypoint Navigation with Position and Heading Control using Complex Vector Fields for an Ackermann Steering Autonomous Vehicle
Waypoint Navigation with Position and Heading Control using Complex Vector Fields for an Ackermann Steering Autonomous Vehicle Tommie J. Liddy and Tien-Fu Lu School of Mechanical Engineering; The University
More informationWORKSPACE AGILITY FOR ROBOTIC ARM Karna Patel
ISSN 30-9135 1 International Journal of Advance Research, IJOAR.org Volume 4, Issue 1, January 016, Online: ISSN 30-9135 WORKSPACE AGILITY FOR ROBOTIC ARM Karna Patel Karna Patel is currently pursuing
More informationLocalization and Map Building
Localization and Map Building Noise and aliasing; odometric position estimation To localize or not to localize Belief representation Map representation Probabilistic map-based localization Other examples
More informationJacobians. 6.1 Linearized Kinematics. Y: = k2( e6)
Jacobians 6.1 Linearized Kinematics In previous chapters we have seen how kinematics relates the joint angles to the position and orientation of the robot's endeffector. This means that, for a serial robot,
More informationIntroduction to Information Science and Technology (IST) Part IV: Intelligent Machines and Robotics Planning
Introduction to Information Science and Technology (IST) Part IV: Intelligent Machines and Robotics Planning Sören Schwertfeger / 师泽仁 ShanghaiTech University ShanghaiTech University - SIST - 10.05.2017
More informationMobile Robot Kinematics
Mobile Robot Kinematics Dr. Kurtuluş Erinç Akdoğan kurtuluserinc@cankaya.edu.tr INTRODUCTION Kinematics is the most basic study of how mechanical systems behave required to design to control Manipulator
More informationComputer Vision I - Algorithms and Applications: Multi-View 3D reconstruction
Computer Vision I - Algorithms and Applications: Multi-View 3D reconstruction Carsten Rother 09/12/2013 Computer Vision I: Multi-View 3D reconstruction Roadmap this lecture Computer Vision I: Multi-View
More informationCSE 490R P1 - Localization using Particle Filters Due date: Sun, Jan 28-11:59 PM
CSE 490R P1 - Localization using Particle Filters Due date: Sun, Jan 28-11:59 PM 1 Introduction In this assignment you will implement a particle filter to localize your car within a known map. This will
More informationRobotics and Autonomous Systems
Robotics and Autonomous Systems Lecture 6: Perception/Odometry Terry Payne Department of Computer Science University of Liverpool 1 / 47 Today We ll talk about perception and motor control. 2 / 47 Perception
More informationRobotics and Autonomous Systems
Robotics and Autonomous Systems Lecture 6: Perception/Odometry Simon Parsons Department of Computer Science University of Liverpool 1 / 47 Today We ll talk about perception and motor control. 2 / 47 Perception
More informationEssential Physics I. Lecture 13:
Essential Physics I E I Lecture 13: 11-07-16 Reminders No lecture: Monday 18th July (holiday) Essay due: Monday 25th July, 4:30 pm 2 weeks!! Exam: Monday 1st August, 4:30 pm Announcements 250 word essay
More informationMechanisms. Updated: 18Apr16 v7
Mechanisms Updated: 8Apr6 v7 Mechanism Converts input motion or force into a desired output with four combinations of input and output motion Rotational to Oscillating Rotational to Rotational Rotational
More informationKinematics of Wheeled Robots
CSE 390/MEAM 40 Kinematics of Wheeled Robots Professor Vijay Kumar Department of Mechanical Engineering and Applied Mechanics University of Pennsylvania September 16, 006 1 Introduction In this chapter,
More informationOmni Stereo Vision of Cooperative Mobile Robots
Omni Stereo Vision of Cooperative Mobile Robots Zhigang Zhu*, Jizhong Xiao** *Department of Computer Science **Department of Electrical Engineering The City College of the City University of New York (CUNY)
More informationGraphics and Interaction Transformation geometry and homogeneous coordinates
433-324 Graphics and Interaction Transformation geometry and homogeneous coordinates Department of Computer Science and Software Engineering The Lecture outline Introduction Vectors and matrices Translation
More informationSession #5 2D Mechanisms: Mobility, Kinematic Analysis & Synthesis
Session #5 2D Mechanisms: Mobility, Kinematic Analysis & Synthesis Courtesy of Design Simulation Technologies, Inc. Used with permission. Dan Frey Today s Agenda Collect assignment #2 Begin mechanisms
More informationCOMP30019 Graphics and Interaction Transformation geometry and homogeneous coordinates
COMP30019 Graphics and Interaction Transformation geometry and homogeneous coordinates Department of Computer Science and Software Engineering The Lecture outline Introduction Vectors and matrices Translation
More informationCSE 252B: Computer Vision II
CSE 252B: Computer Vision II Lecturer: Serge Belongie Scribe: Sameer Agarwal LECTURE 1 Image Formation 1.1. The geometry of image formation We begin by considering the process of image formation when a
More informationLecture «Robot Dynamics»: Multi-body Kinematics
Lecture «Robot Dynamics»: Multi-body Kinematics 151-0851-00 V lecture: CAB G11 Tuesday 10:15 12:00, every week exercise: HG E1.2 Wednesday 8:15 10:00, according to schedule (about every 2nd week) Marco
More informationLecture «Robot Dynamics»: Kinematics 3
Lecture «Robot Dynamics»: Kinematics 3 151-0851-00 V lecture: CAB G11 Tuesday 10:15 12:00, every week exercise: HG E1.2 Wednesday 8:15 10:00, according to schedule (about every 2nd week) Marco Hutter,
More information7 3-Sep Localization and Navigation (GPS, INS, & SLAM) 8 10-Sep State-space modelling & Controller Design 9 17-Sep Vision-based control
RoboticsCourseWare Contributor 2012 School of Information Technology and Electrical Engineering at the University of Queensland Schedule Week Date Lecture (M: 12-1:30, 43-102) 1 23-Jul Introduction Representing
More informationHumanoid Robotics. Least Squares. Maren Bennewitz
Humanoid Robotics Least Squares Maren Bennewitz Goal of This Lecture Introduction into least squares Use it yourself for odometry calibration, later in the lecture: camera and whole-body self-calibration
More informationROBOTICS AND AUTONOMOUS SYSTEMS
ROBOTICS AND AUTONOMOUS SYSTEMS Simon Parsons Department of Computer Science University of Liverpool LECTURE 6 PERCEPTION/ODOMETRY comp329-2013-parsons-lect06 2/43 Today We ll talk about perception and
More informationLecture «Robot Dynamics»: Kinematics 3
Lecture «Robot Dynamics»: Kinematics 3 151-0851-00 V lecture: CAB G11 Tuesday 10:15 12:00, every week exercise: HG E1.2 Wednesday 8:15 10:00, according to schedule (about every 2nd week) office hour: LEE
More informationTime Optimal Trajectories for Bounded Velocity Differential Drive Robots
Time Optimal Trajectories for Bounded Velocity Differential Drive Robots Devin J. Balkcom Matthew T. Mason Robotics Institute and Computer Science Department Carnegie Mellon University Pittsburgh PA 53
More informationC / 35. C18 Computer Vision. David Murray. dwm/courses/4cv.
C18 2015 1 / 35 C18 Computer Vision David Murray david.murray@eng.ox.ac.uk www.robots.ox.ac.uk/ dwm/courses/4cv Michaelmas 2015 C18 2015 2 / 35 Computer Vision: This time... 1. Introduction; imaging geometry;
More informationAutonomous Mobile Robot Design
Autonomous Mobile Robot Design Topic: EKF-based SLAM Dr. Kostas Alexis (CSE) These slides have partially relied on the course of C. Stachniss, Robot Mapping - WS 2013/14 Autonomous Robot Challenges Where
More information3. The three points (2, 4, 1), (1, 2, 2) and (5, 2, 2) determine a plane. Which of the following points is in that plane?
Math 4 Practice Problems for Midterm. A unit vector that is perpendicular to both V =, 3, and W = 4,, is (a) V W (b) V W (c) 5 6 V W (d) 3 6 V W (e) 7 6 V W. In three dimensions, the graph of the equation
More informationINTRODUCTION REFLECTION AND REFRACTION AT BOUNDARIES. Introduction. Reflection and refraction at boundaries. Reflection at a single surface
Chapter 8 GEOMETRICAL OPTICS Introduction Reflection and refraction at boundaries. Reflection at a single surface Refraction at a single boundary Dispersion Summary INTRODUCTION It has been shown that
More informationDifferential Geometry: Circle Patterns (Part 1) [Discrete Conformal Mappinngs via Circle Patterns. Kharevych, Springborn and Schröder]
Differential Geometry: Circle Patterns (Part 1) [Discrete Conformal Mappinngs via Circle Patterns. Kharevych, Springborn and Schröder] Preliminaries Recall: Given a smooth function f:r R, the function
More informationAnimation. CS 4620 Lecture 32. Cornell CS4620 Fall Kavita Bala
Animation CS 4620 Lecture 32 Cornell CS4620 Fall 2015 1 What is animation? Modeling = specifying shape using all the tools we ve seen: hierarchies, meshes, curved surfaces Animation = specifying shape
More informationTEAMS National Competition Middle School Version Photometry Solution Manual 25 Questions
TEAMS National Competition Middle School Version Photometry Solution Manual 25 Questions Page 1 of 14 Photometry Questions 1. When an upright object is placed between the focal point of a lens and a converging
More informationNonlinear State Estimation for Robotics and Computer Vision Applications: An Overview
Nonlinear State Estimation for Robotics and Computer Vision Applications: An Overview Arun Das 05/09/2017 Arun Das Waterloo Autonomous Vehicles Lab Introduction What s in a name? Arun Das Waterloo Autonomous
More informationChapter 11. Parametric Equations And Polar Coordinates
Instructor: Prof. Dr. Ayman H. Sakka Chapter 11 Parametric Equations And Polar Coordinates In this chapter we study new ways to define curves in the plane, give geometric definitions of parabolas, ellipses,
More informationInverse Kinematics. Given a desired position (p) & orientation (R) of the end-effector
Inverse Kinematics Given a desired position (p) & orientation (R) of the end-effector q ( q, q, q ) 1 2 n Find the joint variables which can bring the robot the desired configuration z y x 1 The Inverse
More informationParameterization. Michael S. Floater. November 10, 2011
Parameterization Michael S. Floater November 10, 2011 Triangular meshes are often used to represent surfaces, at least initially, one reason being that meshes are relatively easy to generate from point
More informationData Association for SLAM
CALIFORNIA INSTITUTE OF TECHNOLOGY ME/CS 132a, Winter 2011 Lab #2 Due: Mar 10th, 2011 Part I Data Association for SLAM 1 Introduction For this part, you will experiment with a simulation of an EKF SLAM
More informationVision Review: Image Formation. Course web page:
Vision Review: Image Formation Course web page: www.cis.udel.edu/~cer/arv September 10, 2002 Announcements Lecture on Thursday will be about Matlab; next Tuesday will be Image Processing The dates some
More informationMEAM 520. Mobile Robots
MEAM 520 Mobile Robots Katherine J. Kuchenbecker, Ph.D. General Robotics, Automation, Sensing, and Perception Lab (GRASP) MEAM Department, SEAS, Universit of Pennslvania Lecture 22: December 6, 2012 T
More informationKinematics of Closed Chains
Chapter 7 Kinematics of Closed Chains Any kinematic chain that contains one or more loops is called a closed chain. Several examples of closed chains were encountered in Chapter 2, from the planar four-bar
More informationHomework Assignment /645 Fall Instructions and Score Sheet (hand in with answers)
Homework Assignment 4 600.445/645 Fall 2018 Instructions and Score Sheet (hand in with answers Name Email Other contact information (optional Signature (required I have followed the rules in completing
More informationCinematica dei Robot Mobili su Ruote. Corso di Robotica Prof. Davide Brugali Università degli Studi di Bergamo
Cinematica dei Robot Mobili su Ruote Corso di Robotica Prof. Davide Brugali Università degli Studi di Bergamo Riferimenti bibliografici Roland SIEGWART, Illah R. NOURBAKHSH Introduction to Autonomous Mobile
More informationMEM380 Applied Autonomous Robots Winter Robot Kinematics
MEM38 Applied Autonomous obots Winter obot Kinematics Coordinate Transformations Motivation Ultimatel, we are interested in the motion of the robot with respect to a global or inertial navigation frame
More information0. Introduction: What is Computer Graphics? 1. Basics of scan conversion (line drawing) 2. Representing 2D curves
CSC 418/2504: Computer Graphics Course web site (includes course information sheet): http://www.dgp.toronto.edu/~elf Instructor: Eugene Fiume Office: BA 5266 Phone: 416 978 5472 (not a reliable way) Email:
More informationRobotic Perception and Action: Vehicle SLAM Assignment
Robotic Perception and Action: Vehicle SLAM Assignment Mariolino De Cecco Mariolino De Cecco, Mattia Tavernini 1 CONTENTS Vehicle SLAM Assignment Contents Assignment Scenario 3 Odometry Localization...........................................
More informationPhysics Tutorial 2: Numerical Integration Methods
Physics Tutorial 2: Numerical Integration Methods Summary The concept of numerically solving differential equations is explained, and applied to real time gaming simulation. Some objects are moved in a
More informationPlanar Robot Kinematics
V. Kumar lanar Robot Kinematics The mathematical modeling of spatial linkages is quite involved. t is useful to start with planar robots because the kinematics of planar mechanisms is generally much simpler
More informationOffline Simultaneous Localization and Mapping (SLAM) using Miniature Robots
Offline Simultaneous Localization and Mapping (SLAM) using Miniature Robots Objectives SLAM approaches SLAM for ALICE EKF for Navigation Mapping and Network Modeling Test results Philipp Schaer and Adrian
More informationParallel Lines Investigation
Year 9 - The Maths Knowledge Autumn 1 (x, y) Along the corridor, up the stairs (3,1) x = 3 Gradient (-5,-2) (0,0) y-intercept Vertical lines are always x = y = 6 Horizontal lines are always y = Parallel
More informationReview of Trigonometry
Worksheet 8 Properties of Trigonometric Functions Section Review of Trigonometry This section reviews some of the material covered in Worksheets 8, and The reader should be familiar with the trig ratios,
More informationEXAM SOLUTIONS. Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006,
School of Computer Science and Communication, KTH Danica Kragic EXAM SOLUTIONS Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006, 14.00 19.00 Grade table 0-25 U 26-35 3 36-45
More information1 Introduction. Control 1:
1 1 Introduction Hierarchical Control for Mobile Robots Motive Autonomy Control Level We build controllers because: Real hardware ultimately responds to forces, energy, power, etc. whereas we are concerned
More informationAUTONOMOUS PLANETARY ROVER CONTROL USING INVERSE SIMULATION
AUTONOMOUS PLANETARY ROVER CONTROL USING INVERSE SIMULATION Kevin Worrall (1), Douglas Thomson (1), Euan McGookin (1), Thaleia Flessa (1) (1)University of Glasgow, Glasgow, G12 8QQ, UK, Email: kevin.worrall@glasgow.ac.uk
More informationComputer Vision. Coordinates. Prof. Flávio Cardeal DECOM / CEFET- MG.
Computer Vision Coordinates Prof. Flávio Cardeal DECOM / CEFET- MG cardeal@decom.cefetmg.br Abstract This lecture discusses world coordinates and homogeneous coordinates, as well as provides an overview
More informationSrikumar Ramalingam. Review. 3D Reconstruction. Pose Estimation Revisited. School of Computing University of Utah
School of Computing University of Utah Presentation Outline 1 2 3 Forward Projection (Reminder) u v 1 KR ( I t ) X m Y m Z m 1 Backward Projection (Reminder) Q K 1 q Q K 1 u v 1 What is pose estimation?
More informationLab1: Use of Word and Excel
Dr. Fritz Wilhelm; physics 230 Lab1: Use of Word and Excel Page 1 of 9 Lab partners: Download this page onto your computer. Also download the template file which you can use whenever you start your lab
More informationRobot. A thesis presented to. the faculty of. In partial fulfillment. of the requirements for the degree. Master of Science. Zachary J.
Uncertainty Analysis throughout the Workspace of a Macro/Micro Cable Suspended Robot A thesis presented to the faculty of the Russ College of Engineering and Technology of Ohio University In partial fulfillment
More informationTRIGONOMETRIC FUNCTIONS
Chapter TRIGONOMETRIC FUNCTIONS.1 Introduction A mathematician knows how to solve a problem, he can not solve it. MILNE The word trigonometry is derived from the Greek words trigon and metron and it means
More informationLocalization, Mapping and Exploration with Multiple Robots. Dr. Daisy Tang
Localization, Mapping and Exploration with Multiple Robots Dr. Daisy Tang Two Presentations A real-time algorithm for mobile robot mapping with applications to multi-robot and 3D mapping, by Thrun, Burgard
More informationMath for Geometric Optics
Algebra skills Math for Geometric Optics general rules some common types of equations units problems with several variables (substitution) Geometry basics Trigonometry Pythagorean theorem definitions,
More informationBasics of Localization, Mapping and SLAM. Jari Saarinen Aalto University Department of Automation and systems Technology
Basics of Localization, Mapping and SLAM Jari Saarinen Aalto University Department of Automation and systems Technology Content Introduction to Problem (s) Localization A few basic equations Dead Reckoning
More informationLIGHT: Two-slit Interference
LIGHT: Two-slit Interference Objective: To study interference of light waves and verify the wave nature of light. Apparatus: Two red lasers (wavelength, λ = 633 nm); two orange lasers (λ = 612 nm); two
More informationCS 130 Final. Fall 2015
CS 130 Final Fall 2015 Name Student ID Signature You may not ask any questions during the test. If you believe that there is something wrong with a question, write down what you think the question is trying
More informationMCE/EEC 647/747: Robot Dynamics and Control. Lecture 3: Forward and Inverse Kinematics
MCE/EEC 647/747: Robot Dynamics and Control Lecture 3: Forward and Inverse Kinematics Denavit-Hartenberg Convention Reading: SHV Chapter 3 Mechanical Engineering Hanz Richter, PhD MCE503 p.1/12 Aims of
More informationTEAMS National Competition High School Version Photometry Solution Manual 25 Questions
TEAMS National Competition High School Version Photometry Solution Manual 25 Questions Page 1 of 15 Photometry Questions 1. When an upright object is placed between the focal point of a lens and a converging
More informationPhysics 11 Chapter 18: Ray Optics
Physics 11 Chapter 18: Ray Optics "... Everything can be taken from a man but one thing; the last of the human freedoms to choose one s attitude in any given set of circumstances, to choose one s own way.
More informationPARAMETRIC EQUATIONS AND POLAR COORDINATES
10 PARAMETRIC EQUATIONS AND POLAR COORDINATES PARAMETRIC EQUATIONS & POLAR COORDINATES A coordinate system represents a point in the plane by an ordered pair of numbers called coordinates. PARAMETRIC EQUATIONS
More information