Data Intensive processing with irods and the middleware CiGri for the Whisper project Xavier Briand

Similar documents
Automating Real-time Seismic Analysis

Introduction to Grid Computing

irods for Data Management and Archiving UGM 2018 Masilamani Subramanyam

The VERCE Architecture for Data-Intensive Seismology

Chapter 4:- Introduction to Grid and its Evolution. Prepared By:- NITIN PANDYA Assistant Professor SVBIT.

Cluster Computing. Resource and Job Management for HPC 16/08/2010 SC-CAMP. ( SC-CAMP) Cluster Computing 16/08/ / 50

THE ATLAS DISTRIBUTED DATA MANAGEMENT SYSTEM & DATABASES

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Evolution of the ATLAS PanDA Workload Management System for Exascale Computational Science

ICAT Job Portal. a generic job submission system built on a scientific data catalog. IWSG 2013 ETH, Zurich, Switzerland 3-5 June 2013

ODC and future EIDA/ EPOS-S plans within EUDAT2020. Luca Trani and the EIDA Team Acknowledgements to SURFsara and the B2SAFE team

Managing Petabytes of data with irods. Jean-Yves Nief CC-IN2P3 France

Configuring the Oracle Network Environment. Copyright 2009, Oracle. All rights reserved.

Swedish National Storage Infrastructure for Academic Research with irods

A web portal to analyze and distribute cosmology data

Rapid Processing of Synthetic Seismograms using Windows Azure Cloud

SEDA: An Architecture for Well-Conditioned, Scalable Internet Services

SCA19 APRP. Update Andrew Howard - Co-Chair APAN APRP Working Group. nci.org.au

Efficient Scheduling of Scientific Workflows using Hot Metadata in a Multisite Cloud

irods usage at CC-IN2P3 Jean-Yves Nief

Lecture 1: January 23

USE CASES IN SEISMOLOGY. Alberto Michelini INGV

Introduction to Distributed HTC and overlay systems

N. Marusov, I. Semenov

I/O at the Center for Information Services and High Performance Computing

The Canadian CyberSKA Project

Outline. Definition of a Distributed System Goals of a Distributed System Types of Distributed Systems

Grid Scheduling Architectures with Globus

an Object-Based File System for Large-Scale Federated IT Infrastructures

A Simple Mass Storage System for the SRB Data Grid

Science-as-a-Service

Analytics Platform for ATLAS Computing Services

The IBM MobileFirst Platform

High Throughput WAN Data Transfer with Hadoop-based Storage

Genomics on Cisco Metacloud + SwiftStack

Lecture 1: January 22

Lecture 9: MIMD Architectures

OAR batch scheduler and scheduling on Grid'5000

Dell EMC CIFS-ECS Tool

RZG Visualisation Infrastructure

Cloud Computing 2. CSCI 4850/5850 High-Performance Computing Spring 2018

Storage Virtualization. Eric Yen Academia Sinica Grid Computing Centre (ASGC) Taiwan

Grid Computing. MCSN - N. Tonellotto - Distributed Enabling Platforms

MOHA: Many-Task Computing Framework on Hadoop

Lecture 9: MIMD Architectures

Grid Architectural Models

The ATLAS EventIndex: Full chain deployment and first operation

The Why and How of HPC-Cloud Hybrids with OpenStack

Data Access and Analysis with Distributed, Federated Data Servers in climateprediction.net

Cloud Computing and Hadoop Distributed File System. UCSB CS170, Spring 2018

Application of Virtualization Technologies & CernVM. Benedikt Hegner CERN

Jobs Resource Utilization as a Metric for Clusters Comparison and Optimization. Slurm User Group Meeting 9-10 October, 2012

Improve Web Application Performance with Zend Platform

Design of Distributed Data Mining Applications on the KNOWLEDGE GRID

FuncX: A Function Serving Platform for HPC. Ryan Chard 28 Jan 2019

An Introduction to GPFS

Visualization and clusters: collaboration and integration issues. Philip NERI Integrated Solutions Director

Interconnect EGEE and CNGRID e-infrastructures

Policy Based Distributed Data Management Systems

Software composition for scientific workflows

XtreemFS a case for object-based storage in Grid data management. Jan Stender, Zuse Institute Berlin

Next-generation IT Platforms Delivering New Value through Accumulation and Utilization of Big Data

TensorFlow: A System for Learning-Scale Machine Learning. Google Brain

irods usage at CC-IN2P3: a long history

Multi-threaded, discrete event simulation of distributed computing systems

A Container On a Virtual Machine On an HPC? Presentation to HPC Advisory Council. Perth, July 31-Aug 01, 2017

ΠΙΝΑΚΑΣ ΠΛΑΝΟΥ ΕΚΠΑΙΔΕΥΣΗΣ

On-demand provisioning of HEP compute resources on cloud sites and shared HPC centers

Implementing a Genomic Data Management System using irods at Bayer HealthCare

Bigtable. A Distributed Storage System for Structured Data. Presenter: Yunming Zhang Conglong Li. Saturday, September 21, 13

What's New in SUSE LINUX Enterprise Server 9

MetaData Management Control of Distributed Digital Objects using irods. Venkata Raviteja Vutukuri

EMC Isilon. Cisco UCS Director Support for EMC Isilon

Data storage services at KEK/CRC -- status and plan

PROCESSING THE GAIA DATA IN CNES: THE GREAT ADVENTURE INTO HADOOP WORLD

Application Layer in Science Data Systems:

Cycle Sharing Systems

Creating a Recommender System. An Elasticsearch & Apache Spark approach

China Big Data and HPC Initiatives Overview. Xuanhua Shi

Cloud Computing. Summary

REDCap Importing and Exporting (302)

Scientific data management

An exceedingly high-level overview of ambient noise processing with Spark and Hadoop

Report. Middleware Proxy: A Request-Driven Messaging Broker For High Volume Data Distribution

MONTE CARLO SIMULATION FOR RADIOTHERAPY IN A DISTRIBUTED COMPUTING ENVIRONMENT

Leveraging Software-Defined Storage to Meet Today and Tomorrow s Infrastructure Demands

SONAS Best Practices and options for CIFS Scalability

Distributed Scheduling for the Sombrero Single Address Space Distributed Operating System

SUPPORTING EFFICIENT EXECUTION OF MANY-TASK APPLICATIONS WITH EVEREST

Parallel Storage Systems for Large-Scale Machines

OpenIAM Identity and Access Manager Technical Architecture Overview

Architecting Applications to Scale in the Cloud

Hadoop File System S L I D E S M O D I F I E D F R O M P R E S E N T A T I O N B Y B. R A M A M U R T H Y 11/15/2017

ASDF Definition. Release Lion Krischer, James Smith, Jeroen Tromp

PBS PROFESSIONAL VS. MICROSOFT HPC PACK

Analyzing the Performance of IWAVE on a Cluster using HPCToolkit

Case Study: CyberSKA - A Collaborative Platform for Data Intensive Radio Astronomy

ASYNCHRONOUS SHADERS WHITE PAPER 0

Simplifying Collaboration in the Cloud

Overview of ATLAS PanDA Workload Management

Transcription:

and the middleware CiGri for the Whisper project

Use Case of Data-Intensive processing with irods Collaboration between: IT part of Whisper: Sofware development, computation () Platform Ciment: IT infrastructure of the University of Grenoble High Performance Computing HPC Data-Grid: grid manager CiGRi and irods storage ()

Whisper: a European seismological project Main Goal: Detect sligth changes of properties in the solid Earth Kind of Datamining of seismograms A Motivation: Variation of wave speed before eruption (La Reunion 2000)

Main constraint for Whisper: Massive data processing Data provenance: Noise Continuously recorded by seismic stations worldwilde. Europe: Alpes China: LMS USA: USArray (mobile Network) Japan Networks (NIED)... The computations produce even more data... How much data are we talking about? Japanese Network: 20 TB per year Typical processing: 8 TB in 5 days Another: 'read' 3 TB and 'write' 1 TB in 6 hours Many processings have to be tested... 200 TB managed at the same time Main operations: Raw data processing and computation of the correlations

Other constraints for Whisper: Time: Organize a scientific workflow Duration PHD and Posdoc Available IT Infrastrucure Ease of access for researchers Implementing a generic model, technical support Specific tools: Organize distributed data Access to computational ressources Need a Data Grid environment:

Plan 1- Presentation of the IT infrastructure Ciment: Grid computing with CiGri and irdos 2- Presentation of the Whisper Code: Software for data intensive processing 3- Results and feedback: Whisper Use Case irods experience in the context of Ciment

IT infrastructure of the Ciment platform Platform: HPC centre of the University of Grenoble Partial pooling of computing resources 10 computing clusters, 6600 cores + GPUS Local data grid environement: Resources in a local grid of supercomputers Distributed storage storage 700 TB Centralized controller with total observation: monitor computation Grid manager CiGri (OAR project) Short parametric jobs Best effort mode Improves resource usage Whisper case: most perfectly (aka embarrassingly) parrallel but data intensive processing

IRODS infrastructure 3 sites close to supercomputers: An unique zone with the icat server Dozen of Irods nodes Heterogenous WAN connexions Each site has its own 10Gbe local network switch Site specification: Each site has its own irods resource group Store results of computation randomly at the same site Use of LAN and not WAN

IRODS infrastructure More precisely: Site A and B have 10Gbe WAN connexion Site C has only 1Gbe WAN Automatic staging for site C Automatic replication of resoures from sites A and B to site C when data get from site C

IRODS infrastructure Focus on storage nodes: Two RAID arrays 24 TB 48 TB Irods Nodes with Debian GNU/Linux (Easy synchronisation of the system admin with Kanif) Ciment provides a web interface to check resource status Now 700 TB, constant evolution Irods offers great scalability

Cigri middleware Cigri achieves access to all 6600 cores of 10 clusters within Ciment. Perfectly parallel jobs on idle processors Principles Each cluster uses the ressource manager OAR Cigri communicates with the cluster through OAR's RESTful API Cigri submits jobs without exhausting local queues Best effort mode Jobs on idle processors of every computing nodes. Improves computational ressources Jobs have lowest priority but automatic resubmission (CiGri works also in normal mode)

Cigri middleware Run a set of jobs: a campaign User defines the campaign parameters in a JSON file For each cluster (or all clusters): - Accepted clusters, ressources needed - Maximum duration - Maximum number of jobs - Location of the code - Prologue, epilogue And irods? Code and input are retrieved from irods with i-commands (or API) No direct connexion between irods and CiGri But completely complementary through job scripts.

Cigri middleware Technologies, functionalities RestFul API, Ruby, around a PostgreSQL database. Apache with SSL authentication Statistics: execution rate, resubmission rate Email notification of failed jobs Retrieve the jobs' standard output and standard error from CiGri RestFul API. Authentication, security A centralized LDAP infrastructure for Ciment Password method with synchronization with LDAP authentication. Transparent for users: automatic initialization in home directory (.irods/.irodsenv and.irods/.irodsa files)

Whisper Code General design Grouped into 3 parts: - Seismograms (arrays) processing - Computation of correlations (couple of seismograms) - Analysis of the correlations (differents methods) Python language, libraries Scipy and Obspy (seismology) Development driven by performance, evolution and required support.

Whisper Code Raw data (seismograms) processings A flexible way of specifying a pipeline of processing User specifies: - A directory - A set of seismic stations - A set of dates - A sequence of treatments Signal treatment Traces Space O (N) Users can enter their own code into the processing Code: - Scan a directory - Extract all pieces of seismograms - Organize a specific architecture of files of daily seismograms

Whisper Code Computation of the correlations Operation with 2 seismograms that provides a new virtual seismogram Get 'virtual' new observational data Architecture of files that corresponds to all the couple of seismograms. Quadratic space complexity (linear for seismogram processing) Need to store seismograms processing Quadratic space complexity can be critical Optimization

Whisper Code Focus on optimization of the computation of correlations Numerical optimization: - Algorithm uses the fast fourier transform - Pre-calculating 'Good' combinations of small prime numbers - 40% improvement for good cases Main optimization: try to catch the cache - Test the GC behavior in order to follow the cache heuristics - Don't use the Garbage Collector directly! - Write code towards 'optimal' use of the cache.. (and find good unfolding)

Whisper Code Last part of Code: Analysis of correlations Beamforming Further methods - Beamforming - Doublet - Inversion Corr Space O ( N2 ) Analysys of the correlations ( Signal Processing ) Doublet Dts C3 Space O ( N3 ) New filter the signal, computation of the variation velocity of waves (Note C3 is cubic in Space: distribution for quadratic space complexity) Japan NIED

Results: Organization of a computation Either local computation with a dedicated bay or grid computation - Retrieve a dataset - Anticipate computation time - Evaluate required storage capacity - Assess users' computer skills The specification must evolve - New requirements from researchers - Not generic enough - Very complex to evaluate Adaptation to the grid environment: development time?

Results: Whisper computation with Ciment data grid First step: convert data into seismic standard formats Can be a very data intensive processing Minimize concurrency on Irods IT code for randomly spreading files on resources Find a 'good' distribution model Fine tuning distribution of computation and transfer

Results: Seismogram processing on the grid: Ouput of the Japanese Network over one year Seismograms processing: simplest case for data grid process Convert 9 TB into 20 TB of seismic standard format Modeling the distribution: subsets of dates and subsets of stations Adaptation for irods Add python modules to retrieve data on irods: Provide encapsulation of the iget and iput: number of try, waiting time, 'else' command

Results: Adaptation for Cigri: Define a file of parameters: each line corresponds to a job Campaign with 1280 jobs 160 sublists of dates and 8 sublists of stations 'traces160_0_8_1 160 0 8 1' Sublist of dates of index 0 and sublist of stations of index 1 Define the paramter of the campaign (jdl file JSON) - Campaign named 'test_processing' - Run the 'start.bash' with 'param.txt' - Use clusters 'c1' and 'c2' - Best effort mode - Campaign associated to 'whisper' - Maximum number vof jobs - Retrieve 'start.bash' script from irods...

Results: Correlation computation on the grid: Ouput of the Japanese Network over one year Add similar modules to retrieve data on Irods Two types of processes: with one sublist and with two disjoint sublists Transfer is proportional to the distribution for the stations Maximize the distribution of dates Minimize the distribution of stations Fine tuning the size of files transferred between irods and computing nodes: - Correlations into dictionary - Size file between 100 MB and 500 MB - Appropriate for irods infrastructure

Results: One year of the Japan Network on the grid Processing all seismograms now takes half a day (down from 4 months) Computation of 350 Million correlations for the Japanese Network (down from 2 years... ) - Between 9h and 20h (It depends on stacking, overlap) - 'iget' command correponds to 11 TB - 'iput' command correspons to 3.3 TB - Best-effort can increase also transfer... Big change in opportunities to test processing Many scientific results About variation of velocity change for the tohoku earthquake in Japan: in Science, June 2014

Results: irods experience in the context of Ciment Opportunity to test and improve Ciment Infrastructure for Big Data Add new functionalities: - Limit the number of jobs avoid overload irods Infrastructure - Limit the number of simultaneous connexions - Develop wrapper for i-commands: secure_iget For irods: unique namespace over distributed sites Only the meta Catalog can block the system (asynchronous) Simply add a server as an irods resource and define access policy for each project Challenge with small files (< 32 MB) Completly overload 1Gbe link between 2 sites for conversion processing during automatic staging.

Thanks so much! http://whisper.obs.ujf-grenoble.fr https://ciment.ujf-grenoble.fr http://cigri.imag.fr http://irods.org http://oar.imag.fr kanif http://taktuk.gforge.inria.fr/kanif http://www.liglab.fr