Application of Deep Learning Techniques in Satellite Telemetry Analysis.

Similar documents
Evaluation of Machine Learning Algorithms for Satellite Operations Support

Code Mania Artificial Intelligence: a. Module - 1: Introduction to Artificial intelligence and Python:

Machine Learning 13. week

Keras: Handwritten Digit Recognition using MNIST Dataset

Recurrent Neural Network (RNN) Industrial AI Lab.

Machine Learning With Python. Bin Chen Nov. 7, 2017 Research Computing Center

Deep Learning with Tensorflow AlexNet

Tutorial on Keras CAP ADVANCED COMPUTER VISION SPRING 2018 KISHAN S ATHREY

An Exploration of Computer Vision Techniques for Bird Species Classification

Keras: Handwritten Digit Recognition using MNIST Dataset

Characterization and Benchmarking of Deep Learning. Natalia Vassilieva, PhD Sr. Research Manager

Deep Learning in Visual Recognition. Thanks Da Zhang for the slides

LSTM and its variants for visual recognition. Xiaodan Liang Sun Yat-sen University

International Journal of Computer Engineering and Applications, Volume XII, Special Issue, September 18,

SEMANTIC COMPUTING. Lecture 8: Introduction to Deep Learning. TU Dresden, 7 December Dagmar Gromann International Center For Computational Logic

MoonRiver: Deep Neural Network in C++

NVIDIA DLI HANDS-ON TRAINING COURSE CATALOG

A Quick Guide on Training a neural network using Keras.

INTRODUCTION TO DEEP LEARNING

Deep Learning. Volker Tresp Summer 2014

On the Effectiveness of Neural Networks Classifying the MNIST Dataset

Disguised Face Identification (DFI) with Facial KeyPoints using Spatial Fusion Convolutional Network. Nathan Sun CIS601

Dynamic Routing Between Capsules

DEEP LEARNING REVIEW. Yann LeCun, Yoshua Bengio & Geoffrey Hinton Nature Presented by Divya Chitimalla

Improving the way neural networks learn Srikumar Ramalingam School of Computing University of Utah

Deep Learning. Visualizing and Understanding Convolutional Networks. Christopher Funk. Pennsylvania State University.

CIS581: Computer Vision and Computational Photography Project 4, Part B: Convolutional Neural Networks (CNNs) Due: Dec.11, 2017 at 11:59 pm

NVIDIA DEEP LEARNING INSTITUTE

Object Detection Lecture Introduction to deep learning (CNN) Idar Dyrdal

Study of Residual Networks for Image Recognition

Recurrent Neural Networks and Transfer Learning for Action Recognition

Sequence Modeling: Recurrent and Recursive Nets. By Pyry Takala 14 Oct 2015

All You Want To Know About CNNs. Yukun Zhu

Convolutional Neural Networks

! References: ! Computer eyesight gets a lot more accurate, NY Times. ! Stanford CS 231n. ! Christopher Olah s blog. ! Take ECS 174!

Free Learning OpenCV 3 Computer Vision With Python - Second Edition Ebooks Online

Machine Learning. Deep Learning. Eric Xing (and Pengtao Xie) , Fall Lecture 8, October 6, Eric CMU,

Deep Learning for Computer Vision II

Usable while performant: the challenges building. Soumith Chintala

CS 523: Multimedia Systems

Neural Networks for unsupervised learning From Principal Components Analysis to Autoencoders to semantic hashing

Using Machine Learning for Classification of Cancer Cells

Asynchronous Parallel Learning for Neural Networks and Structured Models with Dense Features

Regionlet Object Detector with Hand-crafted and CNN Feature

PLT: Inception (cuz there are so many layers)

Index. Umberto Michelucci 2018 U. Michelucci, Applied Deep Learning,

ImageNet Classification with Deep Convolutional Neural Networks

Classifying Depositional Environments in Satellite Images

Facial Keypoint Detection

CSC 578 Neural Networks and Deep Learning

Machine Learning. MGS Lecture 3: Deep Learning

An Introduction to Deep Learning with RapidMiner. Philipp Schlunder - RapidMiner Research

SEMANTIC COMPUTING. Lecture 9: Deep Learning: Recurrent Neural Networks (RNNs) TU Dresden, 21 December 2018

Project 3 Q&A. Jonathan Krause

Back propagation Algorithm:

Rotation Invariance Neural Network

LSTM: An Image Classification Model Based on Fashion-MNIST Dataset

Deep neural networks

ConvolutionalNN's... ConvNet's... deep learnig

Deep Learning Benchmarks Mumtaz Vauhkonen, Quaizar Vohra, Saurabh Madaan Collaboration with Adam Coates, Stanford Unviersity

Neural Network Neurons

Predicting resonant properties of plasmonic structures by deep learning

Convolution Neural Network for Traditional Chinese Calligraphy Recognition

Multi-Glance Attention Models For Image Classification

DEEP LEARNING IN PYTHON. Introduction to deep learning

5th Vienna Deep Learning Meetup

Natural Language Processing CS 6320 Lecture 6 Neural Language Models. Instructor: Sanda Harabagiu

Deep Learning Frameworks. COSC 7336: Advanced Natural Language Processing Fall 2017

Convolutional Neural Networks

Event: PASS SQL Saturday - DC 2018 Presenter: Jon Tupitza, CTO Architect

CNN optimization. Rassadin A

Introduction to Deep Learning in Signal Processing & Communications with MATLAB

Deep Learning. Deep Learning provided breakthrough results in speech recognition and image classification. Why?

Deep Learning with R. Francesca Lazzeri Data Scientist II - Microsoft, AI Research

Deep learning for dense per-pixel prediction. Chunhua Shen The University of Adelaide, Australia

ROB 537: Learning-Based Control

Deep Learning for Remote Sensing

Adversarial Examples and Adversarial Training. Ian Goodfellow, Staff Research Scientist, Google Brain CS 231n, Stanford University,

Lecture 2 Notes. Outline. Neural Networks. The Big Idea. Architecture. Instructors: Parth Shah, Riju Pahwa

Convolutional Neural Networks. Computer Vision Jia-Bin Huang, Virginia Tech

Data Mining. Neural Networks

A Deep Learning Approach to Vehicle Speed Estimation

Defense Data Generation in Distributed Deep Learning System Se-Yoon Oh / ADD-IDAR

CS489/698: Intro to ML

CAP 6412 Advanced Computer Vision

Lecture 17: Neural Networks and Deep Learning. Instructor: Saravanan Thirumuruganathan

Practical Deep Learning

Homework 01 : Deep learning Tutorial

Counting Passenger Vehicles from Satellite Imagery

COMP 551 Applied Machine Learning Lecture 16: Deep Learning

CIS680: Vision & Learning Assignment 2.b: RPN, Faster R-CNN and Mask R-CNN Due: Nov. 21, 2018 at 11:59 pm

How to Develop Encoder-Decoder LSTMs

Accelerating Convolutional Neural Nets. Yunming Zhang

Su et al. Shape Descriptors - III

NVIDIA FOR DEEP LEARNING. Bill Veenhuis

Deep Face Recognition. Nathan Sun

Spatial Localization and Detection. Lecture 8-1

Neural Network and Deep Learning. Donglin Zeng, Department of Biostatistics, University of North Carolina

Lecture: Deep Convolutional Neural Networks

Introduction to Deep Learning for Facial Understanding Part III: Regional CNNs

Transcription:

Application of Deep Learning Techniques in Satellite Telemetry Analysis. Greg Adamski, Member of Technical Staff L3 Technologies Telemetry and RF Products Julian Spencer Jones, Spacecraft Engineer Telenor Satellite AS 2018 by L3 Technologies and Telenor Satellite AS. Published by The Aerospace Corporation with permission.

Agenda Artificial Neural Networks (ANNs) and Deep Learning Classic machine learning vs. ANNs Selected Neural Network architectures and their applications Fully Connected (Dense) Convolutional Neural Networks (CNNs) Recurrent Neural Networks (RNNS and LSTMs) Predicting thermal parameter values based on historical record using LSTMs network

Artificial Neural Networks and Deep Learning What they are not: Brand new toy for Data Scientists Invented in 1940s, so it s been around for a while. Recent years have brought tremendous advancements in architectures, training approaches as well as computing power allowing to train large networks with a lot of data Magic bullet to solve all the world problems (yet) Some areas are much better researched than others: computer vision, natural language processing, playing computer games (old Atari arcade, Chess, Go) A foolproof way for your boss to get you a new video card (well, maybe) Training ANNs tends to be very computationally expensive. Using modern video cards can speed up the process by at least an order of magnitude (state of the art. CPU vs decent GPU)

Artificial Neural Networks and Deep Learning What they are a network of connected functional units sometimes called neurons where outputs from one layer are fed into another layer A data modeling tool that can approximate any function [see https://en.wikipedia.org/wiki/universal_approximation _theorem] A powerful machine learning tool A tool that automatically detects numerous input features without the need for complex feature engineering Carefully engineered features as inputs can in some situations improve performance Hidden Input Output Simple ANN diagram

Fully Connected Network Each neuron in layer n receives input from all neurons in layer n 1 Typically the activation of the neuron is calculated by calculating a weighed sum of the inputs and applying a non linearity to the result Very useful for classification and regression problems Does not deal well with context data: E.g. for face recognition, looks at individual pixels, but has no ability to look at a broader context (like body parts etc.) For best performance, the inputs should be well engineered features Frequently used in tandem with CNNs and RNNs for final classification/regression

Convolutional Neural Networks (CNNs) Each convolution layer applies a convolution filter to the input vector: 6 3 1 4 4 2 4 2 1 2 3 5 4 1 2 9 1 1 0 0 9 4 5 6 6 6 3 5 8 Convolution filters are learned during the training process This allows the network to learn to recognize features within the input without explicitly being told to.

Convolutional Neural Networks (CNNs) For example for image data it can detect edges, gradients (early layers) or complex elements such as noses or books in deeper layers (images) [1],[2] For telemetry data, 1d convolutions may return 1d time series features relevant to the problem domain [3] These features can then be used to train a fully connected network to perform the task

CNNs applications Image recognition Given a list of labeled images, define which class the image belongs to. For example for satellite imagery data the network can say if the picture contains trees, crops, cars, buildings etc. Object detection Parse a satellite image to mark where each car, building or ship are (both labels and bounding boxes) Image processing Advanced scaling image augmentation and processing techniques (super resolution, style transfer) which could enhance the quality of satellite imagery

Recurrent Neural Netwoks (RNNs) Output of a layer is re fed back as input into that layer Neuron retains some state between executions Allows the neural network to learn the context Popular in natural language processing Gaining popularity in time series data analysis LSTM Long Short Term Memory a variant of RNN input activation activation

Problem Statement Given a historical record of thermal performance on a Telenor satellite, predict the future behavior Allow the algorithm to discover data trends on its own without manual feature engineering Allows the algorithm to be re used on various measurands More specifically: Based on a number of <n> length sequences, build a model that predicts sample n+1 Important: values are aggregated over 20 minute periods to remove noisce and scaled to zero mean, unit variance

Architecture 3 layer network with LSTM layers responsible for keeping the context and single normal neuron providing value calculation 100 samples of historical values used for input 101 st sample used for output Mean square error loss function Layer 1 Layer 2 Layer 3 input LST M LST M Dens e tanh prediction

Training process and results After 4 training epochs some patterns emerge Blue context data Green actual values Orange predicted values Note: predictions are performed one sample at a time with a fixed window. With each step, oldest sample is purged while the newest prediction is used as ground truth

Training process and results 2 After 8 epochs, the recurring pattern is captured

Training process and results 3 After 64 epochs a more precise trend is captured reasonably well Spikes within each period are modeled in the first two prediction cycles

Results summary Input data had a very easy shape to capture, more research and analysis needed for other data types. Our research shows that for a GEO satellite, parameters show a very strong 24 hour periodicity [as found by dr. Alan Higgings] Event longer term predictions seem to capture the longer term trend reasonably well. This is despite the fact that after 100 predicted samples, all predictions are based on past predictions Training is expensive. When CPU is used the process may take minutes per epoch. With a GPU with appropriate tuning each epoch takes only a few seconds. Each prediction takes << 1 second, however longer term predictions are not immediate

Results applicability and future work The main area of application is to create a dynamic limit/anomaly detection framework. Further work is needed to tune the input algorithms for other data types. Initial results are promising though as the model can predict over 100 samples into the future with high accuracy. Training models on larger sets of parameters can be expensive Need to better quantify prediction performance when running on a CPU vs. a GPU Need to better determine the applicability of deep learning models vs. more traditional approaches

Helpful libraries Keras Provides a high level API for building neural networks. Uses other libraries such as Tensorflow or Theano for heavy lifting Tensorflow Neural network/numerical processing library from google Numpy Numerical processing for python Matplotlib Invaluable tool for result visualization

References [1] Alex Krizhevsky et al, ImageNet Classification with Deep ConvolutionalNeural Networks, https://www.cs.toronto.edu/~kriz/imagenet_classification_with_deep_co nvolutional.pdf [2] Jason Yosinski et al. Understanding Neural Networks Through Deep Visualization, https://arxiv.org/pdf/1506.06579.pdf [3] Cardiologist Level Arrhythmia Detection with Convolutional Neural Networks, Pranav Rajpurkar et al., https://arxiv.org/pdf/1707.01836.pdf

Learning resources Deep learning, parts 1 and 2, Rachel Thomas, Jeremy Howard, http://www.fast.ai/ Coursera.org deep learning specialization, Andrew Ng, https://www.deeplearning.ai/