Learning algorithms for physical systems: challenges and solutions

Similar documents
Autonomous Mobile Robot Design

Written exams of Robotics 2

Mobile Robotics. Mathematics, Models, and Methods. HI Cambridge. Alonzo Kelly. Carnegie Mellon University UNIVERSITY PRESS

Lecture VI: Constraints and Controllers

Lecture VI: Constraints and Controllers. Parts Based on Erin Catto s Box2D Tutorial

Inverse KKT Motion Optimization: A Newton Method to Efficiently Extract Task Spaces and Cost Parameters from Demonstrations

Module 1 Lecture Notes 2. Optimization Problem and Model Formulation

COURSE WEBPAGE. Peter Orbanz Applied Data Mining

Developing Algorithms for Robotics and Autonomous Systems

Mass-Spring Systems. Last Time?

Lesson 1: Introduction to Pro/MECHANICA Motion

Prof. Fanny Ficuciello Robotics for Bioengineering Trajectory planning

Table of contents for: Waves and Mean Flows by Oliver Bühler Cambridge University Press 2009 Monographs on Mechanics. Contents.

Planning and Control: Markov Decision Processes

Acknowledgements. Prof. Dan Negrut Prof. Darryl Thelen Prof. Michael Zinn. SBEL Colleagues: Hammad Mazar, Toby Heyn, Manoj Kumar

Turning an Automated System into an Autonomous system using Model-Based Design Autonomous Tech Conference 2018

Animation. Computer Graphics COMP 770 (236) Spring Instructor: Brandon Lloyd 4/23/07 1

ROBOTICS 01PEEQW. Basilio Bona DAUIN Politecnico di Torino

Introduction to Control Systems Design

Mathematics in Orbit

Manipulator trajectory planning

Goals: Course Unit: Describing Moving Objects Different Ways of Representing Functions Vector-valued Functions, or Parametric Curves

Can we quantify the hardness of learning manipulation? Kris Hauser Department of Electrical and Computer Engineering Duke University

Simulation in Computer Graphics. Introduction. Matthias Teschner. Computer Science Department University of Freiburg

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing

Neural Networks for Obstacle Avoidance

Chapter 3: Kinematics Locomotion. Ross Hatton and Howie Choset

1. Introduction 1 2. Mathematical Representation of Robots

Apprenticeship Learning for Reinforcement Learning. with application to RC helicopter flight Ritwik Anand, Nick Haliday, Audrey Huang

Path Planning. Marcello Restelli. Dipartimento di Elettronica e Informazione Politecnico di Milano tel:

Data Mining Chapter 8: Search and Optimization Methods Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University

A simple example. Assume we want to find the change in the rotation angles to get the end effector to G. Effect of changing s

ECE276B: Planning & Learning in Robotics Lecture 5: Configuration Space

Character Animation 1

ADAPTIVE TILE CODING METHODS FOR THE GENERALIZATION OF VALUE FUNCTIONS IN THE RL STATE SPACE A THESIS SUBMITTED TO THE FACULTY OF THE GRADUATE SCHOOL

10/11/07 1. Motion Control (wheeled robots) Representing Robot Position ( ) ( ) [ ] T

Optimization of a two-link Robotic Manipulator

Balancing Control of Two Wheeled Mobile Robot Based on Decoupling Controller

Model learning for robot control: a survey

Applying Supervised Learning

COPYRIGHTED MATERIAL INTRODUCTION CHAPTER 1

Lecture «Robot Dynamics»: Kinematic Control

CSIS Introduction to Pattern Classification. CSIS Introduction to Pattern Classification 1

Skill. Robot/ Controller

DEEP LEARNING REVIEW. Yann LeCun, Yoshua Bengio & Geoffrey Hinton Nature Presented by Divya Chitimalla

Final Exam Practice Fall Semester, 2012

Accelerating Reinforcement Learning in Engineering Systems

Applying Neural Network Architecture for Inverse Kinematics Problem in Robotics

Lecture 3: Motion Planning 2

Example 24 Spring-back

Parallel Robots. Mechanics and Control H AMID D. TAG HI RAD. CRC Press. Taylor & Francis Group. Taylor & Francis Croup, Boca Raton London NewYoric

Machine Learning Applications for Data Center Optimization

To Do. History of Computer Animation. These Lectures. 2D and 3D Animation. Computer Animation. Foundations of Computer Graphics (Spring 2010)

Vinnie Saini Cloud Solution Architect Big Data & AI

M. Sc. (Artificial Intelligence and Machine Learning)

Reinforcement Learning for Appearance Based Visual Servoing in Robotic Manipulation

Introduction to Artificial Intelligence

W4. Perception & Situation Awareness & Decision making

Trajectory planning in Cartesian space

ANALYTICAL MODEL OF THE CUTTING PROCESS WITH SCISSORS-ROBOT FOR HAPTIC SIMULATION

Inverse Kinematics (part 1) CSE169: Computer Animation Instructor: Steve Rotenberg UCSD, Winter 2018

ME964 High Performance Computing for Engineering Applications

Numerical Methods for (Time-Dependent) HJ PDEs

Motion Planning of Extreme Locomotion Maneuvers Using Multi-Contact Dynamics and Numerical Integration

Last Time? Inverse Kinematics. Today. Keyframing. Physically-Based Animation. Procedural Animation

Overview. Animation is a big topic We will concentrate on character animation as is used in many games today. humans, animals, monsters, robots, etc.

Designing a Pick and Place Robotics Application Using MATLAB and Simulink

Homogenization and numerical Upscaling. Unsaturated flow and two-phase flow

Topics in AI (CPSC 532L): Multimodal Learning with Vision, Language and Sound. Lecture 12: Deep Reinforcement Learning

Elastic Bands: Connecting Path Planning and Control

Template Matching Rigid Motion

INTRODUCTION CHAPTER 1

EEG Imaginary Body Kinematics Regression. Justin Kilmarx, David Saffo, and Lucien Ng

3D Face and Hand Tracking for American Sign Language Recognition

Using RecurDyn. Contents

Matching. Compare region of image to region of image. Today, simplest kind of matching. Intensities similar.

Introduction. Linear because it requires linear functions. Programming as synonymous of planning.

Character Animation. Presented by: Pam Chow

Non-holonomic Planning

Variational Integration Methods for Simulating and Designing Systems with Simultaneous Impact. Todd Murphey Northwestern University June 27, 2010

Supervised Learning with Neural Networks. We now look at how an agent might learn to solve a general problem by seeing examples.

ÉCOLE POLYTECHNIQUE DE MONTRÉAL

Research on time optimal trajectory planning of 7-DOF manipulator based on genetic algorithm

Deep Learning. Roland Olsson

IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 10, NO. 6, NOVEMBER Inverting Feedforward Neural Networks Using Linear and Nonlinear Programming

ECS289: Scalable Machine Learning

Fast marching methods

Machine Learning in Biology

Classification. 1 o Semestre 2007/2008

Control of industrial robots. Kinematic redundancy

Coupled Simulations Using different Tools

Introduction to Optimization Problems and Methods

Introduction to C omputational F luid Dynamics. D. Murrin

Lecture 20: Neural Networks for NLP. Zubin Pahuja

CS770/870 Spring 2017 Animation Basics

CS770/870 Spring 2017 Animation Basics

CGT 581 G Fluids. Overview. Some terms. Some terms

Problem characteristics. Dynamic Optimization. Examples. Policies vs. Trajectories. Planning using dynamic optimization. Dynamic Optimization Issues

MCE/EEC 647/747: Robot Dynamics and Control. Lecture 1: Introduction

Face Detection and Tracking Control with Omni Car

Transcription:

Learning algorithms for physical systems: challenges and solutions Ion Matei Palo Alto Research Center 2018 PARC 1 All Rights Reserved

System analytics: how things are done Use of models (physics) to inform AI algorithms Use of AI algorithms to inform about the physics 2

What are we trying to avoid Solution: ADD MORE INTELIGENCE www.youtube.com 3

None Path forward Automation Reconfigure Reoptimize Reactive maintenance Fail and fix No automation Preventative maintenance Scheduled Low automation Conditionbased maintenance Predictive More automation Selfadaptive assets Proactively reconfigures Highly autonomous Selfcoordinated assets Recognizes key objectives Completely autonomous Awareness None Sensing 4

Two Systematic Approaches to Increased Automation Simulink, MATLAB, Modelica, Simscape, Easy5 1 2 3 4 5 Close Acceleration Constant velocity Braking Open 1 9 8 7 6 Close Ripples Braking Constant velocity Purely data-driven Acceleration Purely behavioral Model-based 5

Data driven methods Pros: A plethora of statistical models (regression models, decision trees, neural networks) More and more efficient algorithms for training 1 2 3 4 5 Easy access to documentation and Close Acceleration Constant velocity Braking Open training platforms 1 9 8 7 6 Close Ripples Braking Constant velocity Purely data-driven Acceleration Do not require very specialized training 6

Data driven methods Cons: 1 2 3 4 5 Close Acceleration Constant velocity Braking 1 9 8 7 6 Close Ripples Braking Constant velocity Purely data-driven Acceleration Open May require a lot of data which is not always easy to obtain (assets do NOT fail very often) May take a long time to train Loss of explainability (systems are seen as black boxes nothing is known about the internal structure and behavior) 7

Data driven methods Training: 1 2 3 4 5 Close Acceleration Constant velocity Braking Open 1 9 8 7 6 Close Ripples Braking Constant velocity Purely data-driven Acceleration 8

Data driven methods enablers Hardware Technology GPU Cheap storage NVIDIA Big data www.blogs.gartner.com 9

Data-driven methods success stories Object detection and tracking www.towardsdatascience.com Speech recognition and translations www.medium.com Text generation http://colah.github.io 10

Data-driven methods success stories, but Judea Pearl, 2011 ACM Turing award All the impressive achievements of deep learning amount to just curve fitting. 11

Model-based methods Often in Simulink, MATLAB, Modelica, Simscape, Easy5 Pros: Vast history of model-based results (reasoning, control, diagnosis, prognosis) Purely behavioral Model-based Well established modeling languages and tools (Matlab, Modelica, Simulink, OpenModelica) Support for both causal (input-output maps) and acausal (physics-based) models Explainability (can connect to particular components and physical processes) Require less data 12

Two Systematic Approaches to System Analytics Often in Simulink, MATLAB, Modelica, Simscape, Easy5 Cons: Work well for specific classes of problems (e.g. linear systems) Purely behavioral Model-based Does not always scale to complex systems Modeling complex systems can be expensive Requires deep expertise Models are not very accurate due to simplifying assumptions 13

Model-based methods Often in Simulink, MATLAB, Modelica, Simscape, Easy5 Hybrid 1 2 3 4 5 Close Acceleration Constant velocity Braking Open 1 9 8 7 6 Close Ripples Braking Constant velocity Purely data-driven Acceleration Purely behavioral How to benefit the from the combining the two approaches? 15

MODELS MACHINE LEARNING MAKE USE OF SYSTEM PROPERTIES (REGULARITIES) THAT ARE INFORMED BY (PARTIAL) MODELS 16

System regularities Inverted pendulum control Pendulum stabilization when starting from two opposing angles The force needed to stabilize the pendulum when starting from a left angle can be used when starting from an opposing right angle by changing its sign Animation produced using a Modelica model 2018 PARC All Rights Reserved 17

System regularities Inverted pendulum control: symmetric motion and control 2018 PARC All Rights Reserved 18

System regularities Robotics: symmetries in complex robot motion Animation produced using a Modelica model 19

System regularities Quadrotor motion planning Using rotational symmetries we can generate a multitude of additional feasible trajectories from one initial trajectory Matlab generated animation 20

System regularities Geometric symmetries Transformations that take a system trajectory and produce another system trajectory Can have discrete symmetries: Rotations at a fixed angle Parametrized symmetries Lie symmetries Can check for typical symmetries (scaling, translations, rotations) Can be learned from data 21

Leveraging system regularities for policy learning Reinforcement learning: Area of machine learning and control that tells us what actions (controls) should we take when interacting with an environment to maximize some reward. Suitable for incomplete, or model free case (or when other methods fail) Requires large amount of data (experience) to learn the best policy 22

Leveraging system regularities for policy learning Symmetry enhanced reinforcement learning: Use geometric symmetries to augment the training data set used for policy learning Using a discrete symmetry can double the size of the training data set More symmetries more data Policy Policy update Policy* Policy update Experience Symmetry map Experience* Policy** 23

Leveraging system regularities for policy learning Inverted pendulum example: State-space discretization: 7 points for the positions, 2 points for the velocity, 8 points for the angle, 2 points for the angular velocity Action space: Training with symmetry use Training without symmetry use Smaller number of failures (faster convergence) Higher rewards Less uncertainty Standard reinforcement algorithm Symmetry enhanced reinforcement algorithm 24

MACHINE LEARNING MODELS MAKE USE OF MACHINE LEARNING TOOLS (ALGORITHMS) TO LEARN (PHYSICAL) MODELS 25

CAUSAL VS. ACAUSAL MODELS CAUSAL MODELS Output is completely determined by the current and past inputs, states and outputs Typical in machine learning, control and signal processing ACAUSAL MODELS There is no clear notion of input and outputs Behavior described by constraints between variables 26

LEARNING PHYSICAL MODELS IN PARTIALLY KNOWN SYSTEMS Learn a model that fits the data What representation? How can I make sure it makes sense? 27

LEARNING PHYSICAL MODELS IN PARTIALLY KNOWN SYSTEMS What representation? Parametrized constraint equation: How can I make sure it makes sense? Impose a priori feasibility constraints on parameters (component does not generate energy, e.g., passive) 28

LEARNING PHYSICAL MODELS IN PARTIALLY KNOWN SYSTEMS What representation? Parametrized constraint equation: How can I make sure it makes sense? Joint learning of best parameter and their feasibility space (optimization problem with unknown constraints) 29

LEARNING PHYSICAL MODELS IN PARTIALLY KNOWN SYSTEMS Strategy for joint learning of parameters and their constraints: explore-exploit feasibility space 30

LEARNING PHYSICAL MODELS IN PARTIALLY KNOWN SYSTEMS How do I learn the feasibility space? Train a classifier Use the model simulator to label points Unfeasible hyperplane Simulation fails Optimization step Update separation between feasible and unfeasible sets Feasible hyperplane Current point 31

LEARNING PHYSICAL MODELS IN PARTIALLY KNOWN SYSTEMS What you would expect to learn? At least a local separating hyperplane Powell Simplex Quadratic programming 32

LEARNING PHYSICAL MODELS: OTHERS IDEAS Build neural network like representations Neural network cell: Linear part + activation function (nonlinear) Acausal neural network cell Nonlinearities in the damper and springs 33

LEARNING PHYSICAL MODELS: OTHERS IDEAS Build neural network like representations Need to add boundary conditions Much more difficult to train forward propagation needs to simulate a dynamical system backward propagation require computation of gradients 34

LEARNING PHYSICAL MODELS: OTHERS IDEAS Discovering Hamiltonians, Lagrangians and other laws of geometric and momentum conservations: Symmetries and invariants underlie almost all physical law Use of genetic algorithms (the main challenge is avoiding trivialities) Hamiltonian M. Schmidt, Hod Lipson, Distilling Free-Form Natural Laws from Experimental Data, Science Magazine 35

None Where we are today Automation Reconfigure Reoptimize Reactive maintenance Fail and fix No automation Preventative maintenance Scheduled Low automation Conditionbased maintenance Predictive More automation current focus Selfadaptive assets Proactively reconfigures Highly autonomous Selfcoordinated assets Recognizes key objectives Completely autonomous Awareness None Sensing 37

What s next Seamless integration between model-based and data-driven methods Explainable AI (what are we learning?) Assured AI (what can we guarantee?) Design use cases 38