Action Recognition Through an Action Generation Mechanism

Size: px
Start display at page:

Download "Action Recognition Through an Action Generation Mechanism"

Transcription

1 Action Recognition Through an Action Generation Mechanism Barış Akgün Doruk Tunaoğlu Kovan Research Lab. Computer Engineering Department, Middle East Technical University, Ankara / Turkey Erol Şahin erol@ceng.metu.edu.tr Abstract In this paper, we studied how actions of others can be recognized by the very same mechanisms that generate similar actions. We used action generation mechanisms to simulate actions and compare resulting trajectories with the observed one as the action unfolds. Specifically, we modified Dynamic Movement Primitives (DMP) for online action recognition and used them as our action generation mechanism. A human demonstrator applied three different actions on two different objects. Recordings of these actions were used to train DMPs and test our approach. We showed that our system is capable of online action recognition within approximately the first one-third of the observed action with a success rate of over %90. Online capabilities of the recognition system were also tested in an interactive game with the robot. 1. Introduction Action recognition, the understanding of what a human (or another robot) is doing, is an essential competence for a robot to interact and communicate with humans. Nature might have already provided an elegant way for action recognition through mirror neurons. These neurons, discovered in the area F5 (and some other areas) of the premotor cortex of macaque monkeys, fire both during the generation and the observation of goal-directed actions (Rizzolatti et al., 1996, Gallese et al., 1996). Grasping is a prominent example of these actions. Grasping related mirror neurons of a macaque monkey are active both when the monkey grasps a piece of food and when a human or another monkey does the same thing. In this context, goal-oriented means that the action is directed towards an object being the food in the previous example. These neurons do not respond when the observed agent tries to grasp the air where there is no object. The dual response characteristics is attributed to action recognition hypothesis of mirror neurons; action generation and recognition shares the very same neural circuitry and that motor system is activated during action recognition. This paper proposes an online action recognition scheme using an action generation mechanism taking its inspiration from mirror neurons. 2. Related Work There have been various modelling efforts and studies related to integrated action generationrecognition approaches after the discovery of mirror neurons. Some computational models are developed to explain and/or mimic the mirror neuron system and some other models to directly be used in robotics (action generation etc.). From the robotics side, these models mostly deal with action generation, action learning and imitation, mentioning recognition as a side study. A review regarding the computational models of imitation learning is given in (Schaal et al., 2003a). In addition, (Oztop et al., 2006) presents a review about imitation methods and their relations to mirror neurons. This literature survey will present computational models that could be used as integrated action generation-recognition mechanisms. Recurrent Neural Network with Parametric Biases (RNNPB) is a modified Jordan-Type recurrent neural network with some special neurons in its input layer that can act both as input and output, according to context. These neurons constitute the parametric bias (PB) vector (Tani, 2003). RNNPBs are used to learn, generate and recognize actions (Tani et al., 2004). PB vectors are calculated per action but weights of the network are shared across different actions. For generation, learned PB vectors are input to the network according to the desired action. However, goal setting is not possible. Recognition mode is similar to learning mode; PB vectors are calculated from the observed action. Then the

2 calculated PB vector is compared with the existing vectors for recognition. PB vectors are used as output in the learning and recognition mode and used as input in the generation mode. Mental State Inference (MSI) (Oztop et al., 2005) model is a motor control model augmented with an action recognition scheme. Action generation mode incorporates a forward model, to compensate for sensory processing delays. This forward model is exploited for action recognition. Actor is assumed to make a certain action and a mental simulation of that action is carried out by the observer, using the forward model. The simulated and observed action trajectories are compared and according to the error, assumed action is updated. Modular Selection and Identification for Control (MOSAIC) (Wolpert and Kawato, 1998, Haruno et al., 2001) is an architecture for adaptive motor control. Given a desired trajectory, this architecture creates necessary motor commands to achieve it. The controller-predictor pairs compete and cooperate to handle a given control task in action generation mode. The contribution of a pair to the total control is proportional to its predictor s performance. Predictors performances are represented with responsibility signals. For action recognition, an observed trajectory is fed to the system as the desired trajectory and the responsibility signals are calculated which are then compared with the ones in the observer s memory. Hierarchical Attentive Multiple Models for Execution and Recognition (HAMMER) architecture is first used for imitation in (Demiris and Hayes, 2002) and recognition in (Demiris and Khadhouri, 2006). It consists of multiple forward and inverse models, as in MOSAIC. However, only a single pair handles the control task which is determined by its forward model s prediction performance, called the confidence value. For recognition, observed trajectory is given to the architecture and the confidence values are calculated. The action which has the highest confidence value is the one recognized. Dynamic Movement Primitives (DMP) (Ijspeert et al., 2001, Schaal et al., 2003b) are nonlinear dynamic systems which are used for imitation learning, action generation and recognition. The imitation learning consists of estimating parameters of a function approximator and these parameters are stored to be used during action generation. For action generation, goal and corresponding learned parameters are fed to the system. Recognition with DMPs are done by treating the observed action as a new one to be learned (Ijspeert et al., 2003). Then the calculated parameters are compared with the ones in observer s repertoire. There are some drawbacks of the aforementioned architectures for action recognition. RNNPBs are limited to offline recognition since PB vectors are calculated iteratively and all of the action needs to be observed. DMPs also need to observe the whole action to calculate its parameters and use them for recognition. Depending on implementation, MO- SAIC can suffer from a similar problem. MOSAIC and MSI methods use instantaneous error to update responsibility signals and action estimates respectively which is prone to noise. MOSAIC, HAMMER and MSI architectures do not provide a way for imitation learning and recognition is done on the assumption that the observer s and the actor s actions are similar. Goal-setting is not available for the RN- NPB approach, which limits action recognition on a single goal. 3. System Architecture In this section, we first describe the DMP framework and then present how we extended DMPs for online action recognition. 3.1 Dynamical Movement Primitives A DMP generates a trajectory x(t) by defining a differential equation for ẍ(t). There are more than one DMP formulation. The one that we are interested is; ẍ = K(g x) Dẋ + K f(s, w), (1) where K is the spring constant, D is the damping coefficient, g is the goal position, s is the phase variable, w is the parameters of the function approximator f and x, ẋ and ẍ are respectively the position, velocity and acceleration in the task space. The phase variable s is defined to decay from 1 to 0 in time as: s 0 = 1, (2) ṡ = αs, (3) = s = e α(t t0), (4) where α is calculated by forcing s = s f at the end of the demonstrated behavior. Equation 1 defines a DMP for a one dimensional system. For a multi-dimensional system, a separate DMP is learned for each dimension, bound via the common phase variable s. Note that DMP s can be defined in the joint space (with joint angles being the variable) as well as in the Cartesian space. We can analyze the behavior of the system in two parts: The canonical part (K(g x) Dẋ) which is inspired by mass-spring-damper equations and the non-linear part (K f(s, w)) which perturbs the massspring-damper system. The canonical part has a global asymptotic stable equilibrium point which is {x, ẋ} = {g, 0}. In other words, given any initial condition the system is guaranteed to converge to g with zero velocity as

3 time goes to. Moreover, if we ensure D = 2 K then the system will be critically damped and it will converge to the equilibrium point as quickly as possible without any oscillation. For a multi-dimensional system, the canonical part generates a simple linear trajectory between the starting point and the goal point in the task space. The non-linear part perturbs the canonical part in order to generate complex trajectories. For this, the perturbing function f(s, w) learns to approximate the additional input to the dynamics generated by the canonical part necessary to follow a trajectory, through its parameters w which are estimated from demonstration. 3.2 Learning by Demonstration For our case, learning by demonstration (LbD) refers to teaching a robot an action by demonstrating that action (as opposed to programming it). The presented recognition scheme requires that the actor and the observer have similar action generation mechanisms and LbD is used to ensure this. For DMPs, LbD is done by calculating the parameters w from a demonstrated action. A demonstrated action has different trajectories for each of its dimension. Given trajectories, x(t), for each of the dimensions, w of each dimension can be calculated as follows: First, the phase variable s(t) is calculated from equation 4 where α is calculated by forcing a final value (s f ) to the phase variable. Then for each of the dimensions, f(t) is calculated as: f(t) = ẍ(t) K (g x(t)) + Dẋ(t). (5) K Since s(t) and f(t) are calculated, and since s(t) is a monotonically decreasing function between 1 and 0, f(s) is also known for s [0 1]. Finally w for each dimension is estimated by applying linear regression using radial basis functions where s values are treated as inputs and the corresponding f(s) values are treated as targets. 3.3 Extending DMPs for Online Recognition There are some major problems of DMPs for online recognition, mostly because of the non-linear part: 1. The lack of an online recognition scheme. 2. The use of a phase variable that runs in an openloop fashion (i.e., not related to x or ẋ). 3. A consequence of the second problem: α in equations 3 and 4 has to be calculated from the end of the motion which prevents the recognition to be done in online fashion. 4. Another consequence of the second problem: s depends on starting time of an action (Equation 4) which may not be observed or estimated by an observer since it is an internal parameter of the actor. In order to circumvent these problems, we replaced f(s, w) with f( z, w) in equation 1, where vector z consists of variables that are calculated from the current state only. We chose to represent our state in Cartesian space instead of joint space. The reasoning behind is that it is easier to observe the end-effector of the actor in Cartesian space than the angular positions of his joints. We used object-centered Cartesian positions and velocities as the variables of z. The objectcentered representation implicitly sets g as zero for all dimensions. Such a choice of hand-object relations is inspired from (Oztop and Arbib, 2002, Oztop et al., 2005). Using hand-object relations for generation and recognition greatly reduces actorobserver differences and allows seamless use of generation mechanisms in recognition. As stated before, our hand-object relations are defined as relative position and velocity of the end-effector with respect to the object. Learning the non-linear part means learning a mapping from z to f where elements of f are the f value for each DOF. Given N recorded trajectories x i (t) for an action, each trajectory is translated to an object-centered reference frame. Then, x i (t) and x i (t) are calculated through simple differentiation. f i (t) is calculated using equation 5, and z i (t) is constructed by concatenating x i (t) and x i (t). Finally, a feed-forward neural network is trained by feeding all z values and the corresponding f values as inputs and target values respectively. 3.4 Recognition The aim of the presented approach is to recognize the observed action, and the object being acted upon, in an online manner. For recognition, initial state observations are given to action generation systems as initial conditions and future trajectories are simulated. These are then compared with the observed trajectory by calculating the cumulative error; err i (t c ) = t c t=t 0 x o (t) x i (t), (6) where err i is the i th behavior s cumulative error, t 0 is the observed initial time, t c is the current time, x o is the observed position, x i is the i th behavior s simulated position. Then, to have a quantitative measure of similarity

4 Figure 1: Recognition architecture flow diagram. Motion capture system tracks possible objects and the end-effector from which state variables are calculated. When recognition starts, the observed state is used as initial values to the learned behaviors (not shown). The learned behaviors (i.e., generation systems) then simulate future trajectories. As the action unfolds, observed and simulated trajectories are compared and responsibility signals are calculated. From these signals, recognition decision is made according to threshold. between actions, we define recognition signals as: rs i (t c ) = j e erri(tc) e errj(tc). (7) Recognition signals could be interpreted as the likelihood of the observed motion to be the corresponding simulated action. These are similar to the responsibility signals in (Haruno et al., 2001). However, instantaneous error is used in responsibility signal calculation, which is prone to noise and the authors also point this out to be problematic. During the recognition phase, if the ratio of the highest recognition signal and the second highest gets above a certain threshold (Υ), the recognition decision is made and the behavior corresponding to the highest signal is chosen. In case there are multiple objects in the environment, behaviors are simulated for each object, by calculating hand-object relations accordingly. If we have m behaviors and n objects in the scene, then m n behaviors are simulated and compared. This also allows to recognize which object is being acted upon. We do not learn a separate network for the same behavior on different objects. We use the same neural network for learning and generating the trajectory of one behavior on all objects; only changing inputs according to the object. The flow-chart of the overall recognition architecture is depicted in figure Experimental Framework 4.1 Setup There are two objects, a motion capture system, an actor (human) and an observer (robot) in our setup. The actor and the observer face each other and the objects are hanging in between as shown in figure 2. The motion capture system is used for tracking the centers of the objects and the end-effector of the ac-

5 tor. We used three reaching behaviors: R(X): Reaching object X from right D(X): Reaching object X directly L(X): Reaching object X from left which can be applied to one of the two objects: LO: Left object (wrt. the actor) RO: Right object (wrt. the actor) The use of multiple behaviors on multiple objects aimed to show that the proposed method can learn and recognize different behaviors for different goals. These behaviors and objects are depicted in figure 3. The actor can apply one of the behaviors on one of the objects. Thus, the recognition decision has to be made among six choices. Note that we do not use a different neural network for all behavior-object pairs but use a different one for each behavior. The inputs given to the same network changes between objects since hand-object relations are different. Figure 3: Depiction of behaviors and approximate geometric information from top view of the setup hand trajectories and the object centers are recorded with the motion capture system. The beginning and the end of the recordings are truncated automatically to capture the start and the end of the actual motion, using a velocity threshold (0.05m/s). Samples from the recorded trajectories are shown in figure 4. 60% of the processed recordings are used for training and the rest are left for testing. For different threshold (Υ) values, we have recorded decision times for each testing sample and measured the success rates. We have selected K = 10 and D = 2 10 empirically. The simulate-observe-compare loop of the approach is run with the same frequency as the motion capture system, 30Hz. Figure 2: Experimental setup We used the VisualeyezTM VZ motion capture system (Phoenix Technologies Incorporated). This device measures the 3D position of active markers, which are attached to the points of interest. In our setup, the markers are attached to the centers of the objects and to the right wrist (the end-effector) of the actor as shown in figure 2. The wrist is chosen to minimize occlusion during tracking. The frequency of the sensor is set to 30Hz per marker. 4.2 Experiments The actor performed 50 repetitions of each behavior on each object (a total of 300 repetitions) and the 1 Figure 4: A subset of recorded human trajectories 5. Results and Discussions We plotted the time evolution of recognition signals for nine different recordings in figure 5. In all plots, recognition signals start from the same initial value, 1 6, corresponding to 6 possible choices (3 actions 2 object). As the action unfolds recognition signal of the corresponding action goes to one while suppressing others. Although there may be a confusion in the beginning parts of an action (e.g. top right plot), our system recognizes it correctly as it unfolds. According to the chosen threshold value Υ, there

6 Figure 5: The time evolution of recognition signals. X-axis is time in seconds and Y-axis is recognition signal magnitude. Figure 6: Recognition rate and decision time (percentage completed) vs. threshold value Υ. Percentage completed is defined as decision time divided by action completion time. is a trade-off between the decision time and the success rate. To optimize Υ, we recorded success rates and mean decision times (for correct decisions) for different Υ values which are shown in figure 6. We chose Υ as 1.9 to obtain 90% recognition rate. On the average, the system makes a decision when 33% of the observed action is completed with this chosen Υ value. Figure 7 plots the histogram of the decision times for this Υ value. Table 1 shows the confusion matrix for the chosen Υ value. Cases where the object acted upon is not correctly decided is low: 25% of wrong decisions and 2.5% of all decisions. Since we want to make recognition before the action is completed, there are Figure 7: Distribution of decision times (percentage completed) for Υ = 1.9 confusions between reaching right of the left object and reaching left of the right object (See figure 3). Also, there is a confusion between reaching directly and reaching from left for the left object. These are expected since human motion has noise and variance between repetitions and a demonstrator may not give the initial curvature expected from the action every time (see figure 4). 6. Robot Demonstration To demonstrate that our system works in an online fashion, we tested it in an interactive game with the humanoid robot icub. The robot s head and eyes have a total of 6 DOF and arms have 7 DOF each, excluding the hands (Metta et al., 2007). It also has

7 Figure 8: Demonstrations with the robot: Each row shows a different demonstration. The first column shows the starting point of the actions. The second column shows the point where the system recognizes the action (indicated by the eye-blink). The third column is the point where the demonstrator finishes his action and the last column is the point where the robot finishes its action. Table 1: Confusion matrix for Υ = 1.9 (Recognition rate = 90%) Object LO RO Object Behavior R D L R D L R LO D L R RO D A facial expressions implemented by leds and moving eyelids. For our demonstration, robot was seated in front of the demonstrator, on the other side of the objects, as shown in figure 2. The interactive game is as follows: The actor applies one of the actions to one of the objects. The robot immediately raises its eyebrows and blinks when it recognizes the action and reacts by turning his head to the predicted object. The robot also makes a hand-coded counter action which is defined as bringing its hand on the opposite side of the object. The reason for using hand-coded actions is that action generation is out of the scope of this paper. DMPs have already been shown to have good generation characteristics. Snapshots from a captured video of the game can be seen in figure 8. This figures show that our action recognition method can be used for online interaction with robots. 7. Conclusion We demonstrated an online recognition architecture which can recognize an action before it is completed. Our recognition architecture is based on a generation system where we have modified DMPs to overcome some of their shortcomings. Our modifications remove the dependence of DMPs on the initial position and time (i.e., internal parameters). Also, we have changed recognition approach in (Ijspeert et al., 2003) to allow for online recognition. We have defined recognition signals to have a quantitative way for measuring similar actions. The temporal profile of our recognition signals are similar to mirror neuron responses, although more experiments and comparison should be done to claim functional similarity. Although seemingly similar, our work differs from HAMMER and MOSAIC in major aspects. We are working in trajectory space and our controller and trajectory generator are decoupled; i.e., we do not need to calculate motor commands for recognition. Our architecture has the ability to learn actions by demonstration, which is absent in HAMMER and MOSAIC. Both in HAMMER and MOSAIC, it is assumed that the actor and the observer have sim-

8 ilar internal models, which cannot be guaranteed. HAMMER is more suited for recognizing straight line trajectories. Although a solution is proposed in (Demiris and Simmons, 2006) by using via-points, everything needs to be calculated offline. Our recognition metric, recognition signals, is robust against noise since it uses the cumulative error unlike MO- SAIC and MSI. Acknowledgements This research was supported by European Commission under the ROSSI project(fp ) and by TÜBİTAK under the project 109E033. Barış Akgün acknowledges the support of the TÜBİTAK 2228 graduate student fellowship and Doruk Tunaoğlu acknowledges the support of the TÜBİTAK 2210 graduate student fellowship. References Demiris, J. and Hayes, G. (2002). Imitation as a dual-route process featuring predictive and learning components; a biologically-plausible computational model. In Dautenhahn, K. and Nehaniv, C., Imitation in animals and artifacts. MIT Press. Demiris, Y. and Khadhouri, B. (2006). Hierarchical attentive multiple models for execution and recognition of actions. Robotics and Autonomous Systems, 54: Demiris, Y. and Simmons, G. (2006). Perceiving the unusual: Temporal properties of hierarchical motor representations for action perception. Neural Networks, 19: Gallese, V., Fadiga, L., Fogassi, L., and Rizzolatti, G. (1996). Action recognition in the premotor cortex. Brain, 119(2): Haruno, M., Wolpert, D. M., and Kawato, M. M. (2001). Mosaic model for sensorimotor learning and control. Neural Computation, 13(10): Ijspeert, A., Nakanishi, J., and Schaal, S. (2001). Trajectory formation for imitation with nonlinear dynamical systems. Proceedings 2001 IEEE/RSJ International Conference on Intelligent Robots and Systems. Expanding the Societal Role of Robotics in the the Next Millennium (Cat. No.01CH37180), pages Ijspeert, A. J., Nakanishi, J., and Schaal, S. (2003). Learning attractor landscapes for learning motor primitives. In Becker, S., Thrun, S., and Obermayer, K., (Eds.), Advances in Neural Information Processing Systems, volume 15, pages MIT-Press. Metta, G., Sandini, G., Vernon, D., Natale, L., and Nori, F. (2007). The icub cognitive humanoid robot: An open-system research platform for enactive cognition. In 50 Years of Artificial Intelligence, pages Springer Berlin / Heidelberg. Oztop, E. and Arbib, M. A. (2002). Schema design and implementation of the grasp-related mirror neuron system. Biological Cybernetics, 87: Oztop, E., Kawato, M., and Arbib, M. (2006). Mirror neurons and imitation: A computationally guided review. Neural Networks, 19: Oztop, E., Wolpert, D., and Kawato, M. (2005). Mental state inference using visual control parameters. Cognitive Brain Research, 22: Rizzolatti, G., Fadiga, L., Gallese, V., and Fogassi, L. (1996). Premotor cortex and the recognition of motor actions. Cognitive brain research, 3: Schaal, S., Ijspeert, A., and Billard, A. (2003a). Computational approaches to motor learning by imitation. Philosophical Transactions: Biological Sciences, 358: Schaal, S., Peters, J., Nakanishi, J., and Ijspeert, A. (2003b). Learning movement primitives. In International Symposium on Robotics Research. Springer. Tani, J. (2003). Learning to generate articulated behavior through the bottom-up and the topdown interaction processes. Neural Networks, 16: Tani, J., Ito, M., and Sugita, Y. (2004). Selforganization of distributedly represented multiple behavior schemata in a mirror system : reviews of robot experiments using RNNPB. Neural Networks, 17: Wolpert, D. and Kawato, M. (1998). Multiple paired forward and inverse models for motor control. Neural Networks, 11:

Learning and Generalization of Motor Skills by Learning from Demonstration

Learning and Generalization of Motor Skills by Learning from Demonstration Learning and Generalization of Motor Skills by Learning from Demonstration Peter Pastor, Heiko Hoffmann, Tamim Asfour, and Stefan Schaal Abstract We provide a general approach for learning robotic motor

More information

Learning to bounce a ball with a robotic arm

Learning to bounce a ball with a robotic arm Eric Wolter TU Darmstadt Thorsten Baark TU Darmstadt Abstract Bouncing a ball is a fun and challenging task for humans. It requires fine and complex motor controls and thus is an interesting problem for

More information

Distance-based Kernels for Dynamical Movement Primitives

Distance-based Kernels for Dynamical Movement Primitives Distance-based Kernels for Dynamical Movement Primitives Diego ESCUDERO-RODRIGO a,1, René ALQUEZAR a, a Institut de Robotica i Informatica Industrial, CSIC-UPC, Spain Abstract. In the Anchoring Problem

More information

Perceptual Perspective Taking and Action Recognition

Perceptual Perspective Taking and Action Recognition Perceptual Perspective Taking and Action Recognition Matthew Johnson; Yiannis Demiris Department of Electrical and Electronic Engineering Imperial College London Exhibition Road, London, SW7 2BT {matthew.johnson,

More information

Simulating Mirror Neurons

Simulating Mirror Neurons Simulating Mirror Neurons Kate Derosier and Jackie Kay Abstract Mirror neurons are motor cortical neurons found in primates and thought to be important for imitative learning. Multiple attempts have been

More information

Movement Learning and Control for Robots in Interaction

Movement Learning and Control for Robots in Interaction Movement Learning and Control for Robots in Interaction Dr.-Ing. Michael Gienger Honda Research Institute Europe Carl-Legien-Strasse 30 63073 Offenbach am Main Seminar Talk Machine Learning Lab, University

More information

Teaching a robot to perform a basketball shot using EM-based reinforcement learning methods

Teaching a robot to perform a basketball shot using EM-based reinforcement learning methods Teaching a robot to perform a basketball shot using EM-based reinforcement learning methods Tobias Michels TU Darmstadt Aaron Hochländer TU Darmstadt Abstract In this paper we experiment with reinforcement

More information

Robot learning for ball bouncing

Robot learning for ball bouncing Robot learning for ball bouncing Denny Dittmar Denny.Dittmar@stud.tu-darmstadt.de Bernhard Koch Bernhard.Koch@stud.tu-darmstadt.de Abstract For robots automatically learning to solve a given task is still

More information

Movement reproduction and obstacle avoidance with dynamic movement primitives and potential fields

Movement reproduction and obstacle avoidance with dynamic movement primitives and potential fields 8 8 th IEEE-RAS International Conference on Humanoid Robots December 1 ~ 3, 8 / Daejeon, Korea Movement reproduction and obstacle avoidance with dynamic movement primitives and potential fields Dae-Hyung

More information

Motion Recognition and Generation for Humanoid based on Visual-Somatic Field Mapping

Motion Recognition and Generation for Humanoid based on Visual-Somatic Field Mapping Motion Recognition and Generation for Humanoid based on Visual-Somatic Field Mapping 1 Masaki Ogino, 1 Shigeo Matsuyama, 1 Jun ichiro Ooga, and 1, Minoru Asada 1 Dept. of Adaptive Machine Systems, HANDAI

More information

A Modular Software Framework for Eye-Hand Coordination in Humanoid Robots

A Modular Software Framework for Eye-Hand Coordination in Humanoid Robots A Modular Software Framework for Eye-Hand Coordination in Humanoid Robots Jurgen Leitner, Simon Harding, Alexander Forster and Peter Corke Presentation: Hana Fusman Introduction/ Overview The goal of their

More information

Monocular Virtual Trajectory Estimation with Dynamical Primitives

Monocular Virtual Trajectory Estimation with Dynamical Primitives Monocular Virtual Trajectory Estimation with Dynamical Primitives Odest Chadwicke Jenkins, Germán González, Matthew M. Loper Department of Computer Science Brown University Providence, RI 02912-1910 cjenkins,gerg,matt@cs.brown.edu

More information

Multiple Object Manipulation: is structural modularity necessary? A study of the MOSAIC and CARMA models

Multiple Object Manipulation: is structural modularity necessary? A study of the MOSAIC and CARMA models Multiple Object Manipulation: is structural modularity necessary? A study of the MOSAIC and CARMA models Stéphane Lallée (stephane.lallee@inserm.fr) INSERM U846, Stem-Cell and Brain Research Institute

More information

Movement Imitation with Nonlinear Dynamical Systems in Humanoid Robots

Movement Imitation with Nonlinear Dynamical Systems in Humanoid Robots Movement Imitation with Nonlinear Dynamical Systems in Humanoid Robots Auke Jan Ijspeert & Jun Nakanishi & Stefan Schaal Computational Learning and Motor Control Laboratory University of Southern California,

More information

Applying the Episodic Natural Actor-Critic Architecture to Motor Primitive Learning

Applying the Episodic Natural Actor-Critic Architecture to Motor Primitive Learning Applying the Episodic Natural Actor-Critic Architecture to Motor Primitive Learning Jan Peters 1, Stefan Schaal 1 University of Southern California, Los Angeles CA 90089, USA Abstract. In this paper, we

More information

Inverse KKT Motion Optimization: A Newton Method to Efficiently Extract Task Spaces and Cost Parameters from Demonstrations

Inverse KKT Motion Optimization: A Newton Method to Efficiently Extract Task Spaces and Cost Parameters from Demonstrations Inverse KKT Motion Optimization: A Newton Method to Efficiently Extract Task Spaces and Cost Parameters from Demonstrations Peter Englert Machine Learning and Robotics Lab Universität Stuttgart Germany

More information

Active Sensing based Dynamical Object Feature Extraction

Active Sensing based Dynamical Object Feature Extraction 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems Acropolis Convention Center Nice, France, Sept, 22-26, 2008 Active Sensing based Dynamical Object Feature Extraction Shun Nishide,

More information

A NOUVELLE MOTION STATE-FEEDBACK CONTROL SCHEME FOR RIGID ROBOTIC MANIPULATORS

A NOUVELLE MOTION STATE-FEEDBACK CONTROL SCHEME FOR RIGID ROBOTIC MANIPULATORS A NOUVELLE MOTION STATE-FEEDBACK CONTROL SCHEME FOR RIGID ROBOTIC MANIPULATORS Ahmad Manasra, 135037@ppu.edu.ps Department of Mechanical Engineering, Palestine Polytechnic University, Hebron, Palestine

More information

Cecilia Laschi The BioRobotics Institute Scuola Superiore Sant Anna, Pisa

Cecilia Laschi The BioRobotics Institute Scuola Superiore Sant Anna, Pisa University of Pisa Master of Science in Computer Science Course of Robotics (ROB) A.Y. 2016/17 cecilia.laschi@santannapisa.it http://didawiki.cli.di.unipi.it/doku.php/magistraleinformatica/rob/start Robot

More information

Extended Hopfield Network for Sequence Learning: Application to Gesture Recognition

Extended Hopfield Network for Sequence Learning: Application to Gesture Recognition Extended Hopfield Network for Sequence Learning: Application to Gesture Recognition André Maurer, Micha Hersch and Aude G. Billard Ecole Polytechnique Fédérale de Lausanne (EPFL) Swiss Federal Institute

More information

Parametric Primitives for Motor Representation and Control

Parametric Primitives for Motor Representation and Control Proceedings, IEEE International Conference on Robotics and Automation (ICRA-2002) volume 1, pages 863-868, Washington DC, May 11-15, 2002 Parametric Primitives for Motor Representation and Control R. Amit

More information

The Organization of Cortex-Ganglia-Thalamus to Generate Movements From Motor Primitives: a Model for Developmental Robotics

The Organization of Cortex-Ganglia-Thalamus to Generate Movements From Motor Primitives: a Model for Developmental Robotics The Organization of Cortex-Ganglia-Thalamus to Generate Movements From Motor Primitives: a Model for Developmental Robotics Alessio Mauro Franchi 1, Danilo Attuario 2, and Giuseppina Gini 3 Department

More information

The Fly & Anti-Fly Missile

The Fly & Anti-Fly Missile The Fly & Anti-Fly Missile Rick Tilley Florida State University (USA) rt05c@my.fsu.edu Abstract Linear Regression with Gradient Descent are used in many machine learning applications. The algorithms are

More information

Learning of a ball-in-a-cup playing robot

Learning of a ball-in-a-cup playing robot Learning of a ball-in-a-cup playing robot Bojan Nemec, Matej Zorko, Leon Žlajpah Robotics Laboratory, Jožef Stefan Institute Jamova 39, 1001 Ljubljana, Slovenia E-mail: bojannemec@ijssi Abstract In the

More information

IAV PREPRINTS 5th IFAC/EURON Symposium on Intelligent Autonomous Vehicles Instituto Superior Técnico, Lisboa, Portugal July 5-7, 2004

IAV PREPRINTS 5th IFAC/EURON Symposium on Intelligent Autonomous Vehicles Instituto Superior Técnico, Lisboa, Portugal July 5-7, 2004 IAV24 - PREPRINTS 5th IFAC/EURON Symposium on Intelligent Autonomous Vehicles Instituto Superior Técnico, Lisboa, Portugal July 5-7, 24 VISUOMOTOR MAPS FOR ROBOT TASK LEARNING THROUGH IMITATION Raquel

More information

Motor control learning and modular control architectures. Francesco Nori

Motor control learning and modular control architectures. Francesco Nori Motor control learning and modular control architectures Francesco Nori Italian Institute of Technology, Genova, ITALY Robotics Brain and Cognitive Sciences Department, (former) member of LIRA-Lab Giorgio

More information

and dispersion was used to demonstrate higher-level group behaviors including flocking, foraging/collection, and herding (Matarić 1995). We also demon

and dispersion was used to demonstrate higher-level group behaviors including flocking, foraging/collection, and herding (Matarić 1995). We also demon Primitives and Behavior-Based Architectures for Interactive Entertainment Odest Chadwicke Jenkins and Maja J Matarić Computer Science Department University of Southern California 941 West 37th Place, Mailcode

More information

Instant Prediction for Reactive Motions with Planning

Instant Prediction for Reactive Motions with Planning The 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems October 11-15, 2009 St. Louis, USA Instant Prediction for Reactive Motions with Planning Hisashi Sugiura, Herbert Janßen, and

More information

Kinesthetic Teaching via Fast Marching Square

Kinesthetic Teaching via Fast Marching Square Kinesthetic Teaching via Fast Marching Square Javier V. Gómez, David Álvarez, Santiago Garrido and Luis Moreno Abstract This paper presents a novel robotic learning technique based on Fast Marching Square

More information

ROBOTICS 01PEEQW. Basilio Bona DAUIN Politecnico di Torino

ROBOTICS 01PEEQW. Basilio Bona DAUIN Politecnico di Torino ROBOTICS 01PEEQW Basilio Bona DAUIN Politecnico di Torino Control Part 4 Other control strategies These slides are devoted to two advanced control approaches, namely Operational space control Interaction

More information

Learning Multiple Models of Non-Linear Dynamics for Control under Varying Contexts

Learning Multiple Models of Non-Linear Dynamics for Control under Varying Contexts Learning Multiple Models of Non-Linear Dynamics for Control under Varying Contexts Georgios Petkos, Marc Toussaint, and Sethu Vijayakumar Institute of Perception, Action and Behaviour, School of Informatics

More information

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing Visual servoing vision allows a robotic system to obtain geometrical and qualitative information on the surrounding environment high level control motion planning (look-and-move visual grasping) low level

More information

Automatic Control Industrial robotics

Automatic Control Industrial robotics Automatic Control Industrial robotics Prof. Luca Bascetta (luca.bascetta@polimi.it) Politecnico di Milano Dipartimento di Elettronica, Informazione e Bioingegneria Prof. Luca Bascetta Industrial robots

More information

Handwriting Trajectory Movements Controlled by a Bêta-Elliptic Model

Handwriting Trajectory Movements Controlled by a Bêta-Elliptic Model Handwriting Trajectory Movements Controlled by a Bêta-Elliptic Model Hala BEZINE, Adel M. ALIMI, and Nabil DERBEL Research Group on Intelligent Machines (REGIM) Laboratory of Intelligent Control and Optimization

More information

Design and Development of Cartesian Robot for Machining with Error Compensation and Chatter Reduction

Design and Development of Cartesian Robot for Machining with Error Compensation and Chatter Reduction International Journal of Engineering Research and Technology. ISSN 0974-3154 Volume 6, Number 4 (2013), pp. 449-454 International Research Publication House http://www.irphouse.com Design and Development

More information

Motion Capture & Simulation

Motion Capture & Simulation Motion Capture & Simulation Motion Capture Character Reconstructions Joint Angles Need 3 points to compute a rigid body coordinate frame 1 st point gives 3D translation, 2 nd point gives 2 angles, 3 rd

More information

A study of human locomotor trajectories

A study of human locomotor trajectories A study of human locomotor trajectories Phạm Quang Cường December, 29 Thèse réalisée sous la direction de P r Alain Berthoz, au Laboratoire de Physiologie de la Perception et de l Action Collège de France,

More information

Predicting Object Dynamics from Visual Images through Active Sensing Experiences

Predicting Object Dynamics from Visual Images through Active Sensing Experiences 2007 IEEE International Conference on Robotics and Automation Roma, Italy, 10-14 April 2007 ThC3.5 Predicting Object Dynamics from Visual Images through Active Sensing Experiences Shun Nishide, Tetsuya

More information

Exploiting Vestibular Output during Learning Results in Naturally Curved Reaching Trajectories

Exploiting Vestibular Output during Learning Results in Naturally Curved Reaching Trajectories Exploiting Vestibular Output during Learning Results in Naturally Curved Reaching Trajectories Ganghua Sun and Brian Scassellati Computer Science Department, Yale University New Haven, Connecticut 06511,

More information

Simulation in Computer Graphics. Deformable Objects. Matthias Teschner. Computer Science Department University of Freiburg

Simulation in Computer Graphics. Deformable Objects. Matthias Teschner. Computer Science Department University of Freiburg Simulation in Computer Graphics Deformable Objects Matthias Teschner Computer Science Department University of Freiburg Outline introduction forces performance collision handling visualization University

More information

Control 2. Keypoints: Given desired behaviour, determine control signals Inverse models:

Control 2. Keypoints: Given desired behaviour, determine control signals Inverse models: Control 2 Keypoints: Given desired behaviour, determine control signals Inverse models: Inverting the forward model for simple linear dynamic system Problems for more complex systems Open loop control:

More information

This is an author produced version of Definition and composition of motor primitives using latent force models and hidden Markov models.

This is an author produced version of Definition and composition of motor primitives using latent force models and hidden Markov models. This is an author produced version of Definition and composition of motor primitives using latent force models and hidden Markov models. White Rose Research Online URL for this paper: http://eprints.whiterose.ac.uk/116580/

More information

Learning Motor Behaviors: Past & Present Work

Learning Motor Behaviors: Past & Present Work Stefan Schaal Computer Science & Neuroscience University of Southern California, Los Angeles & ATR Computational Neuroscience Laboratory Kyoto, Japan sschaal@usc.edu http://www-clmc.usc.edu Learning Motor

More information

Nao Devils Dortmund. Team Description Paper for RoboCup Matthias Hofmann, Ingmar Schwarz, and Oliver Urbann

Nao Devils Dortmund. Team Description Paper for RoboCup Matthias Hofmann, Ingmar Schwarz, and Oliver Urbann Nao Devils Dortmund Team Description Paper for RoboCup 2017 Matthias Hofmann, Ingmar Schwarz, and Oliver Urbann Robotics Research Institute Section Information Technology TU Dortmund University 44221 Dortmund,

More information

Deriving Action and Behavior Primitives from Human Motion Data

Deriving Action and Behavior Primitives from Human Motion Data In Proceedings of 2002 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS-2002), Lausanne, Switzerland, Sept. 30 - Oct. 4, 2002, pp. 2551-2556 Deriving Action and Behavior Primitives

More information

Learning Task-Space Tracking Control with Kernels

Learning Task-Space Tracking Control with Kernels Learning Task-Space Tracking Control with Kernels Duy Nguyen-Tuong 1, Jan Peters 2 Max Planck Institute for Intelligent Systems, Spemannstraße 38, 72076 Tübingen 1,2 Universität Darmstadt, Intelligent

More information

Human Motion Reconstruction by Direct Control of Marker Trajectories

Human Motion Reconstruction by Direct Control of Marker Trajectories Human Motion Reconstruction by Direct Control of Marker Trajectories Emel Demircan, Luis Sentis, Vincent De Sapio and Oussama Khatib Artificial Intelligence Laboratory, Stanford University, Stanford, CA

More information

Incremental learning of full body motions via adaptive Factorial Hidden Markov Models

Incremental learning of full body motions via adaptive Factorial Hidden Markov Models Incremental learning of full body motions via adaptive Factorial Hidden Markov Models Dana Kulić, Wataru Takano and Yoshihiko Nakamura Department of Mechano-Informatics, University of Tokyo 7-3- Hongo,

More information

Analyzing and Segmenting Finger Gestures in Meaningful Phases

Analyzing and Segmenting Finger Gestures in Meaningful Phases 2014 11th International Conference on Computer Graphics, Imaging and Visualization Analyzing and Segmenting Finger Gestures in Meaningful Phases Christos Mousas Paul Newbury Dept. of Informatics University

More information

A Segmentation Algorithm for the Comparison of Human Limb Trajectories Marc Pomplun 1 and Maja J. Matarić 2 1 Department of Computer Science, York Uni

A Segmentation Algorithm for the Comparison of Human Limb Trajectories Marc Pomplun 1 and Maja J. Matarić 2 1 Department of Computer Science, York Uni Title: A Segmentation Algorithm for the Comparison of Human Limb Trajectories Authors: Marc Pomplun and Maja J. Matarić Running head: Segmentation Algorithm Corresponding author: Marc Pomplun Department

More information

Algorithm for Arm Position Reconstruction from Optical Motion Capture Data with Noisy or Missing Data

Algorithm for Arm Position Reconstruction from Optical Motion Capture Data with Noisy or Missing Data Algorithm for Arm Position Reconstruction from Optical Motion Capture Data with Noisy or Missing Data Meredith Moore Neuroscience and Computer Science Drake University Des Moines, IA, 50311 Meredith.Moore@drake.edu

More information

Modeling Tool-Body Assimilation using Second-order Recurrent Neural Network

Modeling Tool-Body Assimilation using Second-order Recurrent Neural Network The 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems October -5, 2009 St. Louis, USA Modeling Tool-Body Assimilation using Second-order Recurrent Neural Network Shun Nishide, Tatsuhiro

More information

Globally Stabilized 3L Curve Fitting

Globally Stabilized 3L Curve Fitting Globally Stabilized 3L Curve Fitting Turker Sahin and Mustafa Unel Department of Computer Engineering, Gebze Institute of Technology Cayirova Campus 44 Gebze/Kocaeli Turkey {htsahin,munel}@bilmuh.gyte.edu.tr

More information

A New Algorithm for Measuring and Optimizing the Manipulability Index

A New Algorithm for Measuring and Optimizing the Manipulability Index DOI 10.1007/s10846-009-9388-9 A New Algorithm for Measuring and Optimizing the Manipulability Index Ayssam Yehia Elkady Mohammed Mohammed Tarek Sobh Received: 16 September 2009 / Accepted: 27 October 2009

More information

Locally Weighted Learning

Locally Weighted Learning Locally Weighted Learning Peter Englert Department of Computer Science TU Darmstadt englert.peter@gmx.de Abstract Locally Weighted Learning is a class of function approximation techniques, where a prediction

More information

Open and Closed-Loop Task Space Trajectory Control of Redundant Robots Using Learned Models

Open and Closed-Loop Task Space Trajectory Control of Redundant Robots Using Learned Models Open and Closed-Loop Task Space Trajectory Control of Redundant Robots Using Learned Models Bruno Damas Lorenzo Jamone José Santos-Victor Abstract This paper presents a comparison of open-loop and closed-loop

More information

Towards Autonomous Robotic Valve Turning

Towards Autonomous Robotic Valve Turning BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 12, No 3 Sofia 2012 Print ISSN: 1311-9702; Online ISSN: 1314-4081 DOI: 10.2478/cait-2012-0018 Towards Autonomous Robotic Valve

More information

PRACTICAL SESSION 4: FORWARD DYNAMICS. Arturo Gil Aparicio.

PRACTICAL SESSION 4: FORWARD DYNAMICS. Arturo Gil Aparicio. PRACTICAL SESSION 4: FORWARD DYNAMICS Arturo Gil Aparicio arturo.gil@umh.es OBJECTIVES After this practical session, the student should be able to: Simulate the movement of a simple mechanism using the

More information

Indian Institute of Technology Kanpur. Visuomotor Learning Using Image Manifolds: ST-GK Problem

Indian Institute of Technology Kanpur. Visuomotor Learning Using Image Manifolds: ST-GK Problem Indian Institute of Technology Kanpur Introduction to Cognitive Science SE367A Visuomotor Learning Using Image Manifolds: ST-GK Problem Author: Anurag Misra Department of Computer Science and Engineering

More information

Representation and Generalization of Bi-manual Skills from Kinesthetic Teaching

Representation and Generalization of Bi-manual Skills from Kinesthetic Teaching Representation and Generalization of Bi-manual Skills from Kinesthetic Teaching René Felix Reinhart, Andre Lemme and Jochen Jakob Steil Research Institute for Cognition and Robotics (CoR-Lab) & Faculty

More information

Using Artificial Neural Networks for Prediction Of Dynamic Human Motion

Using Artificial Neural Networks for Prediction Of Dynamic Human Motion ABSTRACT Using Artificial Neural Networks for Prediction Of Dynamic Human Motion Researchers in robotics and other human-related fields have been studying human motion behaviors to understand and mimic

More information

Robot Vision without Calibration

Robot Vision without Calibration XIV Imeko World Congress. Tampere, 6/97 Robot Vision without Calibration Volker Graefe Institute of Measurement Science Universität der Bw München 85577 Neubiberg, Germany Phone: +49 89 6004-3590, -3587;

More information

Fitting (LMedS, RANSAC)

Fitting (LMedS, RANSAC) Fitting (LMedS, RANSAC) Thursday, 23/03/2017 Antonis Argyros e-mail: argyros@csd.uoc.gr LMedS and RANSAC What if we have very many outliers? 2 1 Least Median of Squares ri : Residuals Least Squares n 2

More information

10/25/2018. Robotics and automation. Dr. Ibrahim Al-Naimi. Chapter two. Introduction To Robot Manipulators

10/25/2018. Robotics and automation. Dr. Ibrahim Al-Naimi. Chapter two. Introduction To Robot Manipulators Robotics and automation Dr. Ibrahim Al-Naimi Chapter two Introduction To Robot Manipulators 1 Robotic Industrial Manipulators A robot manipulator is an electronically controlled mechanism, consisting of

More information

Online movement adaptation based on previous sensor experiences

Online movement adaptation based on previous sensor experiences 211 IEEE/RSJ International Conference on Intelligent Robots and Systems September 25-3, 211. San Francisco, CA, USA Online movement adaptation based on previous sensor experiences Peter Pastor, Ludovic

More information

Tracking Human Motion and Actions for Interactive Robots

Tracking Human Motion and Actions for Interactive Robots Tracking Human Motion and Actions for Interactive Robots Odest Chadwicke Jenkins Dept. of Computer Science Brown University Providence, RI, USA cjenkins@cs.brown.edu German González Computer Vision Lab

More information

Automated Modularization of Human Motion into Actions and Behaviors

Automated Modularization of Human Motion into Actions and Behaviors USC Center for Robotics and Embedded Systems Technical Report CRES-02-002 Automated Modularization of Human Motion into Actions and Behaviors Odest Chadwicke Jenkins Robotics Research Laboratory Department

More information

Deep Learning. Architecture Design for. Sargur N. Srihari

Deep Learning. Architecture Design for. Sargur N. Srihari Architecture Design for Deep Learning Sargur N. srihari@cedar.buffalo.edu 1 Topics Overview 1. Example: Learning XOR 2. Gradient-Based Learning 3. Hidden Units 4. Architecture Design 5. Backpropagation

More information

6. NEURAL NETWORK BASED PATH PLANNING ALGORITHM 6.1 INTRODUCTION

6. NEURAL NETWORK BASED PATH PLANNING ALGORITHM 6.1 INTRODUCTION 6 NEURAL NETWORK BASED PATH PLANNING ALGORITHM 61 INTRODUCTION In previous chapters path planning algorithms such as trigonometry based path planning algorithm and direction based path planning algorithm

More information

Automated Parameterization of the Joint Space Dynamics of a Robotic Arm. Josh Petersen

Automated Parameterization of the Joint Space Dynamics of a Robotic Arm. Josh Petersen Automated Parameterization of the Joint Space Dynamics of a Robotic Arm Josh Petersen Introduction The goal of my project was to use machine learning to fully automate the parameterization of the joint

More information

David Galdeano. LIRMM-UM2, Montpellier, France. Members of CST: Philippe Fraisse, Ahmed Chemori, Sébatien Krut and André Crosnier

David Galdeano. LIRMM-UM2, Montpellier, France. Members of CST: Philippe Fraisse, Ahmed Chemori, Sébatien Krut and André Crosnier David Galdeano LIRMM-UM2, Montpellier, France Members of CST: Philippe Fraisse, Ahmed Chemori, Sébatien Krut and André Crosnier Montpellier, Thursday September 27, 2012 Outline of the presentation Context

More information

MCE/EEC 647/747: Robot Dynamics and Control. Lecture 1: Introduction

MCE/EEC 647/747: Robot Dynamics and Control. Lecture 1: Introduction MCE/EEC 647/747: Robot Dynamics and Control Lecture 1: Introduction Reading: SHV Chapter 1 Robotics and Automation Handbook, Chapter 1 Assigned readings from several articles. Cleveland State University

More information

13. Learning Ballistic Movementsof a Robot Arm 212

13. Learning Ballistic Movementsof a Robot Arm 212 13. Learning Ballistic Movementsof a Robot Arm 212 13. LEARNING BALLISTIC MOVEMENTS OF A ROBOT ARM 13.1 Problem and Model Approach After a sufficiently long training phase, the network described in the

More information

Learning Inverse Dynamics: a Comparison

Learning Inverse Dynamics: a Comparison Learning Inverse Dynamics: a Comparison Duy Nguyen-Tuong, Jan Peters, Matthias Seeger, Bernhard Schölkopf Max Planck Institute for Biological Cybernetics Spemannstraße 38, 72076 Tübingen - Germany Abstract.

More information

Simulation. x i. x i+1. degrees of freedom equations of motion. Newtonian laws gravity. ground contact forces

Simulation. x i. x i+1. degrees of freedom equations of motion. Newtonian laws gravity. ground contact forces Dynamic Controllers Simulation x i Newtonian laws gravity ground contact forces x i+1. x degrees of freedom equations of motion Simulation + Control x i Newtonian laws gravity ground contact forces internal

More information

COMP 175 COMPUTER GRAPHICS. Lecture 10: Animation. COMP 175: Computer Graphics March 12, Erik Anderson 08 Animation

COMP 175 COMPUTER GRAPHICS. Lecture 10: Animation. COMP 175: Computer Graphics March 12, Erik Anderson 08 Animation Lecture 10: Animation COMP 175: Computer Graphics March 12, 2018 1/37 Recap on Camera and the GL Matrix Stack } Go over the GL Matrix Stack 2/37 Topics in Animation } Physics (dynamics, simulation, mechanics)

More information

Learning algorithms for physical systems: challenges and solutions

Learning algorithms for physical systems: challenges and solutions Learning algorithms for physical systems: challenges and solutions Ion Matei Palo Alto Research Center 2018 PARC 1 All Rights Reserved System analytics: how things are done Use of models (physics) to inform

More information

Reinforcement Learning for Appearance Based Visual Servoing in Robotic Manipulation

Reinforcement Learning for Appearance Based Visual Servoing in Robotic Manipulation Reinforcement Learning for Appearance Based Visual Servoing in Robotic Manipulation UMAR KHAN, LIAQUAT ALI KHAN, S. ZAHID HUSSAIN Department of Mechatronics Engineering AIR University E-9, Islamabad PAKISTAN

More information

Image Compression: An Artificial Neural Network Approach

Image Compression: An Artificial Neural Network Approach Image Compression: An Artificial Neural Network Approach Anjana B 1, Mrs Shreeja R 2 1 Department of Computer Science and Engineering, Calicut University, Kuttippuram 2 Department of Computer Science and

More information

Optimization of a two-link Robotic Manipulator

Optimization of a two-link Robotic Manipulator Optimization of a two-link Robotic Manipulator Zachary Renwick, Yalım Yıldırım April 22, 2016 Abstract Although robots are used in many processes in research and industry, they are generally not customized

More information

1724. Mobile manipulators collision-free trajectory planning with regard to end-effector vibrations elimination

1724. Mobile manipulators collision-free trajectory planning with regard to end-effector vibrations elimination 1724. Mobile manipulators collision-free trajectory planning with regard to end-effector vibrations elimination Iwona Pajak 1, Grzegorz Pajak 2 University of Zielona Gora, Faculty of Mechanical Engineering,

More information

Realtime Avoidance of Fast Moving Objects: A Dynamical System-based Approach

Realtime Avoidance of Fast Moving Objects: A Dynamical System-based Approach Realtime Avoidance of Fast Moving Objects: A Dynamical System-based Approach S. Mohammad Khansari-Zadeh and Aude Billard Ecole Polytechnique Federale de Lausanne, LASA Laboratory {mohammad.khansari,aude.billard}@epfl.ch

More information

Activity recognition with echo state networks using 3D body joints and objects category

Activity recognition with echo state networks using 3D body joints and objects category Activity recognition with echo state networks using 3D body joints and objects category Luiza Mici, Xavier Hinaut and Stefan Wermter University of Hamburg - Department of Informatics, Vogt-Koelln-Strasse

More information

Skill. Robot/ Controller

Skill. Robot/ Controller Skill Acquisition from Human Demonstration Using a Hidden Markov Model G. E. Hovland, P. Sikka and B. J. McCarragher Department of Engineering Faculty of Engineering and Information Technology The Australian

More information

Representability of Human Motions by Factorial Hidden Markov Models

Representability of Human Motions by Factorial Hidden Markov Models Proceedings of the 2007 IEEE/RSJ International Conference on Intelligent Robots and Systems San Diego, CA, USA, Oct 29 - Nov 2, 2007 WeC10.1 Representability of Human Motions by Factorial Hidden Markov

More information

Animation. Computer Graphics COMP 770 (236) Spring Instructor: Brandon Lloyd 4/23/07 1

Animation. Computer Graphics COMP 770 (236) Spring Instructor: Brandon Lloyd 4/23/07 1 Animation Computer Graphics COMP 770 (236) Spring 2007 Instructor: Brandon Lloyd 4/23/07 1 Today s Topics Interpolation Forward and inverse kinematics Rigid body simulation Fluids Particle systems Behavioral

More information

Motion Planning for Dynamic Knotting of a Flexible Rope with a High-speed Robot Arm

Motion Planning for Dynamic Knotting of a Flexible Rope with a High-speed Robot Arm The 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems October 18-22, 2010, Taipei, Taiwan Motion Planning for Dynamic Knotting of a Flexible Rope with a High-speed Robot Arm Yuji

More information

Written exams of Robotics 2

Written exams of Robotics 2 Written exams of Robotics 2 http://www.diag.uniroma1.it/~deluca/rob2_en.html All materials are in English, unless indicated (oldies are in Year Date (mm.dd) Number of exercises Topics 2018 07.11 4 Inertia

More information

Facial Expression Simulator Control System using OpenGL Graphical Simulator Indra Adji Sulistijono, Rachman Ardyansyah Keywords 1.

Facial Expression Simulator Control System using OpenGL Graphical Simulator Indra Adji Sulistijono, Rachman Ardyansyah Keywords 1. 6789012345678901234567890123456789012345 2 Facial Expression Simulator Control System using OpenGL Graphical Simulator 1 2 Indra Adji Sulistijono, Rachman Ardyansyah Dept. of Mechatronics Eng., Electronics

More information

Applying Neural Network Architecture for Inverse Kinematics Problem in Robotics

Applying Neural Network Architecture for Inverse Kinematics Problem in Robotics J. Software Engineering & Applications, 2010, 3: 230-239 doi:10.4236/jsea.2010.33028 Published Online March 2010 (http://www.scirp.org/journal/jsea) Applying Neural Network Architecture for Inverse Kinematics

More information

Inverse Kinematics. Given a desired position (p) & orientation (R) of the end-effector

Inverse Kinematics. Given a desired position (p) & orientation (R) of the end-effector Inverse Kinematics Given a desired position (p) & orientation (R) of the end-effector q ( q, q, q ) 1 2 n Find the joint variables which can bring the robot the desired configuration z y x 1 The Inverse

More information

Topics in AI (CPSC 532L): Multimodal Learning with Vision, Language and Sound. Lecture 12: Deep Reinforcement Learning

Topics in AI (CPSC 532L): Multimodal Learning with Vision, Language and Sound. Lecture 12: Deep Reinforcement Learning Topics in AI (CPSC 532L): Multimodal Learning with Vision, Language and Sound Lecture 12: Deep Reinforcement Learning Types of Learning Supervised training Learning from the teacher Training data includes

More information

Incremental Learning of Robot Dynamics using Random Features

Incremental Learning of Robot Dynamics using Random Features Incremental Learning of Robot Dynamics using Random Features Arjan Gijsberts, Giorgio Metta Cognitive Humanoids Laboratory Dept. of Robotics, Brain and Cognitive Sciences Italian Institute of Technology

More information

Elastic Bands: Connecting Path Planning and Control

Elastic Bands: Connecting Path Planning and Control Elastic Bands: Connecting Path Planning and Control Sean Quinlan and Oussama Khatib Robotics Laboratory Computer Science Department Stanford University Abstract Elastic bands are proposed as the basis

More information

Learning Deformations of Human Arm Movement to Adapt to Environmental Constraints

Learning Deformations of Human Arm Movement to Adapt to Environmental Constraints Learning Deformations of Human Arm Movement to Adapt to Environmental Constraints Stephan Al-Zubi and Gerald Sommer Cognitive Systems, Christian Albrechts University, Kiel, Germany Abstract. We propose

More information

Graph-based High Level Motion Segmentation using Normalized Cuts

Graph-based High Level Motion Segmentation using Normalized Cuts Graph-based High Level Motion Segmentation using Normalized Cuts Sungju Yun, Anjin Park and Keechul Jung Abstract Motion capture devices have been utilized in producing several contents, such as movies

More information

SCREW-BASED RELATIVE JACOBIAN FOR MANIPULATORS COOPERATING IN A TASK

SCREW-BASED RELATIVE JACOBIAN FOR MANIPULATORS COOPERATING IN A TASK ABCM Symposium Series in Mechatronics - Vol. 3 - pp.276-285 Copyright c 2008 by ABCM SCREW-BASED RELATIVE JACOBIAN FOR MANIPULATORS COOPERATING IN A TASK Luiz Ribeiro, ribeiro@ime.eb.br Raul Guenther,

More information

NEURAL NETWORK VISUALIZATION

NEURAL NETWORK VISUALIZATION Neural Network Visualization 465 NEURAL NETWORK VISUALIZATION Jakub Wejchert Gerald Tesauro IB M Research T.J. Watson Research Center Yorktown Heights NY 10598 ABSTRACT We have developed graphics to visualize

More information

Introduction to ANSYS DesignXplorer

Introduction to ANSYS DesignXplorer Lecture 4 14. 5 Release Introduction to ANSYS DesignXplorer 1 2013 ANSYS, Inc. September 27, 2013 s are functions of different nature where the output parameters are described in terms of the input parameters

More information

Task analysis based on observing hands and objects by vision

Task analysis based on observing hands and objects by vision Task analysis based on observing hands and objects by vision Yoshihiro SATO Keni Bernardin Hiroshi KIMURA Katsushi IKEUCHI Univ. of Electro-Communications Univ. of Karlsruhe Univ. of Tokyo Abstract In

More information