Interconnect EGEE and CNGRID e-infrastructures

Similar documents
g-eclipse A Framework for Accessing Grid Infrastructures Nicholas Loulloudes Trainer, University of Cyprus (loulloudes.n_at_cs.ucy.ac.

glite Grid Services Overview

30 Nov Dec Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy

Introduction to Grid Infrastructures

Multiple Broker Support by Grid Portals* Extended Abstract

Garuda : The National Grid Computing Initiative Of India. Natraj A.C, CDAC Knowledge Park, Bangalore.

Cross-Domain Middlewares Interoperability for Distributed Aircraft Design Optimization

On the employment of LCG GRID middleware

EGEE and Interoperation

Grid Scheduling Architectures with Globus

FREE SCIENTIFIC COMPUTING

Grid technologies, solutions and concepts in the synchrotron Elettra

Implementing GRID interoperability

Grid Infrastructure For Collaborative High Performance Scientific Computing

Gergely Sipos MTA SZTAKI

Grid Interoperation and Regional Collaboration

Deploying virtualisation in a production grid

Grid services. Enabling Grids for E-sciencE. Dusan Vudragovic Scientific Computing Laboratory Institute of Physics Belgrade, Serbia

Introduction to Grid Computing

Data Grid Infrastructure for YBJ-ARGO Cosmic-Ray Project

UNICORE Globus: Interoperability of Grid Infrastructures

Grid Computing Middleware. Definitions & functions Middleware components Globus glite

SEDA: An Architecture for Well-Conditioned, Scalable Internet Services

ARC integration for CMS

I Tier-3 di CMS-Italia: stato e prospettive. Hassen Riahi Claudio Grandi Workshop CCR GRID 2011

The University of Oxford campus grid, expansion and integrating new partners. Dr. David Wallom Technical Manager

SLCS and VASH Service Interoperability of Shibboleth and glite

Moving e-infrastructure into a new era the FP7 challenge

Interoperable job submission and management with GridSAM, JMEA, and UNICORE

Chapter 2 Introduction to the WS-PGRADE/gUSE Science Gateway Framework

The Grid. Processing the Data from the World s Largest Scientific Machine II Brazilian LHC Computing Workshop

How to build Scientific Gateways with Vine Toolkit and Liferay/GridSphere framework

A Practical Approach for a Workflow Management System

The INFN Tier1. 1. INFN-CNAF, Italy

Grid Computing. Olivier Dadoun LAL, Orsay. Introduction & Parachute method. Socle 2006 Clermont-Ferrand Orsay)

The glite middleware. Ariel Garcia KIT

SZDG, ecom4com technology, EDGeS-EDGI in large P. Kacsuk MTA SZTAKI

High Throughput WAN Data Transfer with Hadoop-based Storage

Prototypes of a Computational Grid for the Planck Satellite

Grids and Security. Ian Neilson Grid Deployment Group CERN. TF-CSIRT London 27 Jan

Andrea Sciabà CERN, Switzerland

Advanced Job Submission on the Grid

Monitoring the Usage of the ZEUS Analysis Grid

Scalable Computing: Practice and Experience Volume 10, Number 4, pp

Workflow, Planning and Performance Information, information, information Dr Andrew Stephen M c Gough

MOHA: Many-Task Computing Framework on Hadoop

Pan-European Grid einfrastructure for LHC Experiments at CERN - SCL's Activities in EGEE

The Grid: Processing the Data from the World s Largest Scientific Machine

Bob Jones. EGEE and glite are registered trademarks. egee EGEE-III INFSO-RI

JAMILA: A Usable Batch Job Management System to Coordinate Heterogeneous Clusters and Diverse Applications over Grid or Cloud Infrastructure

The EPIKH, GILDA and GISELA Projects

CAS 703 Software Design

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

A Composable Service-Oriented Architecture for Middleware-Independent and Interoperable Grid Job Management

Enabling Grids for E-sciencE. EGEE security pitch. Olle Mulmo. EGEE Chief Security Architect KTH, Sweden. INFSO-RI

Eclipse Technology Project: g-eclipse

Regional e-infrastructures

g-eclipse - An Integrated Framework to Access and Maintain Grid Resources

EUROPEAN MIDDLEWARE INITIATIVE

Access the power of Grid with Eclipse

Software composition for scientific workflows

WMS Application Program Interface: How to integrate them in your code

MyProxy Server Installation

ISTITUTO NAZIONALE DI FISICA NUCLEARE

The LHC Computing Grid

Setup Desktop Grids and Bridges. Tutorial. Robert Lovas, MTA SZTAKI

Heterogeneous Grid Computing: Issues and Early Benchmarks

G-ECLIPSE: A MIDDLEWARE-INDEPENDENT FRAMEWORK TO ACCESS AND MAINTAIN GRID RESOURCES

Workload Management. Stefano Lacaprara. CMS Physics Week, FNAL, 12/16 April Department of Physics INFN and University of Padova

AGATA Analysis on the GRID

Easy Access to Grid Infrastructures

International Collaboration to Extend and Advance Grid Education. glite WMS Workload Management System

High Performance Computing Course Notes Grid Computing I

Problemi di schedulazione distribuita su Grid

The LHC Computing Grid

Gatlet - a Grid Portal Framework

( PROPOSAL ) THE AGATA GRID COMPUTING MODEL FOR DATA MANAGEMENT AND DATA PROCESSING. version 0.6. July 2010 Revised January 2011

Figure 1: VRengine (left rack)

Computing grids, a tool for international collaboration and against digital divide Guy Wormser Director of CNRS Institut des Grilles (CNRS, France)

Chapter 4:- Introduction to Grid and its Evolution. Prepared By:- NITIN PANDYA Assistant Professor SVBIT.

A Login Shell interface for INFN-GRID

Architecture of the WMS

Improved 3G Bridge scalability to support desktop grid executions

HEP Grid Activities in China

CHIPP Phoenix Cluster Inauguration

SDS: A Scalable Data Services System in Data Grid

Ivane Javakhishvili Tbilisi State University High Energy Physics Institute HEPI TSU

DESY. Andreas Gellrich DESY DESY,

LCG-2 and glite Architecture and components

Travelling securely on the Grid to the origin of the Universe

GRID COMPUTING APPLIED TO OFF-LINE AGATA DATA PROCESSING. 2nd EGAN School, December 2012, GSI Darmstadt, Germany

The glite middleware. Presented by John White EGEE-II JRA1 Dep. Manager On behalf of JRA1 Enabling Grids for E-sciencE

Tier2 Centre in Prague

The EU DataGrid Testbed

ACET s e-research Activities

Application of Virtualization Technologies & CernVM. Benedikt Hegner CERN

Overview of HEP software & LCG from the openlab perspective

Leveraging the InCommon Federation to access the NSF TeraGrid

Outline 18/12/2014. Accessing GROMACS on a Science Gateway. GROMACS in a nutshell. GROMACS users in India. GROMACS on GARUDA

Beob Kyun KIM, Christophe BONNAUD {kyun, NSDC / KISTI

Transcription:

Interconnect EGEE and CNGRID e-infrastructures Giuseppe Andronico Interoperability and Interoperation between Europe, India and Asia Workshop Barcelona - Spain, June 2 2007 FP6 2004 Infrastructures 6-SSA-026634

Agenda The EUChinaGRID Project EGEE & CNGRID grid e-infrastructure glite-gos Interoperability & GW approach Gateway design principles Batch job level interoperability process Gilda Testbed in INFN Catania Data Transfer for small data volumes Security issues Future Works 2/20

EUChinaGRID Project Project Banner: Interconnection and Interoperability of Grids between Europe and China The EUChinaGRID Project aims to support the interoperability of the Grid infrastructures in Europe and China for the benefit of escience applications. To foster the creation of a intercontinental escience community: Training people Supporting existing and new applications To support interoperable infrastructure for grid operations between Europe (EGEE) and China (CNGRID) 3/20

EGEE The Enabling Grids for E-sciencE (EGEE) project draws together over 1000 scientist and engineers from more than 90 institutions in 32 countries. The EGEE grid infrastructure is used at present by applications from many scientific disciplines, as high energy physics, life sciences, geology, computational chemistry and so on. It is a general purpose infrastructure. The EGEE grid is composed by over 36000 CPU in addition to 5 Petabytes storage. 30000 concurrent jobs run on the average. glite is the middleware used in EGEE. It provides a set of service (SOA architecture) that allow users to make use of the huge resources of the EGEE infrastructure in a easy way, hiding the complexity. 4/20

CNGRID 5/20 China National Grid (CNGrid) was developed under the support of China s National Hightech R&D Program. CNGrid consists of 8 nodes across the country, including 2 major nodes, Computer Network & Information Center of CAS in Beijing and the Shangai Supercomputer Center in Shangai. CNGrid forms an open grid environment for supporting applications in the field of research, resource and environment, service and manufacture. The aggregated computing capacity of CNGrid exceeds 18 TFlops. The total storage capacity is over 200 TB. The grid middleware, CNGrid GOS, enables the operation of CNGrid and grid applications by supporting resource sharing and cooperative works, which makes CNGrid a testbed for scientific research, technology development, and application demostration.

glite-gos Interoperability & GW approach glite and GOS middlewares adopt very different architectures and for this reason the Gateway approach to make them interoperable seemed to be the best practicable one. The role of the Gateway is to translate requests coming from one middleware into requests compatible with the other. Request such as job submission, job control, output retrieval, and data management have to be taken in account. A method for sharing the knowledge between the 2 middlewares about the status of resource available in their respective infrastructure is the basis for a trivial interoperation prototyping (Information System). 6/20

Role of Gateway Gateway is just a logical component, it can be treat as facade of underlying grid infrastructure: Interface conversion Function mapping etc Gateway should support the following features: Transparent to end users of different grid infrastructures Easy to extend Support massive concurrency and high throughput Support standalone deployment or integrated underlying grid middleware etc Different instances of the gateway could be configured to connect the infrastructure running GOS and glite (to guarantee the scalability of the system) 7/20

Gateway design principle 8/20 Gateway design depends heavily on SEDA model and IoC model The reasons to choose SEDA model and IoC model as follows: SEDA model SEDA is short for Staged Event Driven Architecture Firstly proposed by Matt Welsh, David Culler, and Eric Brewer, from University of California, Berkeley in their paper SEDA: An Architecture for Well-Conditioned, Scalable Internet Services Support massive concurrency, high throughput Simplify the construction of well-conditioned Internet services In our design, we firstly divide the whole process into several independent basic stages and then compose basic stages into different pipelines for different purpose such as glite-to-gos batch job forwarding and so on IoC model IoC is short for Inversion of Control Provide loose couple among different modules and easy to reuse basic modules Assemble new module easily and quickly In our design, HiveMind 1.1 is used as IoC container which is released under LGPL license

Core components of Gateway Core components of gateway including: Pipelines constructed for different purposes Consists of different basic processing stages Used for different purposes such as forwarding batch job from GOS to glite and vice verse Scheduler Execute processing stages at fixed rate One to one mapping between Pipeline and scheduler Thread pool Improves performance One to one mapping between Thread pool and scheduler Different processing stages in same pipeline usually in charge of different concrete functions such as StageIn, StageOut and so on Detailed description of gateway can be found in the next slide 9/20

Detailed description of gateway components Extended GOS forward batch job to gateway scheduler Thread Pool idle threads pool used in schedule GOS batch job batch job WMProxy GOS batch job Pipeline for GOS to glite Pipeline for glite to GOS batch job Extended LCG-CE Extended LCG-CE forward batch job to gateway Thread Pool Scheduler execute stage in pipeline using idle thread from thread pool scheduler 1 Different colors in pipeline standard for different stage which is in charge of concrete function such as data stagein, data stageout and so on 2 Different pipeline have different thread pool and scheduler 10/20

Batch job level interoperability process To implement batch job level interoperability, we do the following work: The gateway needs to develop a converter between the Job Description Languages used in glite (JDL) and GOS (JSDL) The common Web Service interface can be used to submit jobs Extend Job Manager in both glite and GOS Extend glite LCG-CE Job Manager Framework providing a special Job Manager for CNGrid which will forward the batch job to CNGrid for execution Provide Broker plugin for GOS Job Manager framework (a glite based RMS System which can interact with the glite WMProxy module) Sandbox mode data transfer A fast approach for cross-domain security scenario 11/20

Gilda Testbed in INFN Catania Portal or WS-Client WN RSL CE CE CE JSDL portal.ct.infn.it JSDL Gateway component PipeLine4GLite PipeLine4GOS RSL JDL WMS/RB RSL glite-rb2.ct.infn.it JSDL WMProxy gos.ct.infn.it JSDL Extended LCG-CE RSL RSL glite-ui OpenPBS glite-gos.ct.infn.it JDL Command Line 12/20

Batch job level interoperability process JSDL GOS Node portal.ct.infn.it GOS Node gos.ct.infn.it Forwarding JSDL PipeLineForGLite PipeLineForGOS Submit JDL Submit JSDL PipeLineForGOS: Submit 1 Data Transfer Extended JSDL 2 Submit job to GOS LCG-CE Extended LCG-CE: 1 Extend Dispatch Globus batch JobManager job used in LCG-CE 2 Convert JDL to JSDL Gateway Component 3 Submit batch job to PipeLineForGOS glite-gos.ct.infn.it WMProxy PipeLineForGLite: 1 Convert JSDL to JDL glite-rb2.ct.infn.it 2 Data Transfer 3 Submit job to WMProxy WMS glite-ui Submit JDL Submit JDL 13/20

Data Transfer for small data volumes Data transfer operations between CNGrid and EGEE uses sandbox mode All the data transfer will pass the gateway Suitable for small scale data transfer scenario Gateway act as data transfer center and have two different roles at the same time: GridFTP client Gateway upload/download necessary data to/from glite WMS FTP server GOS upload/download data to/from gateway component 14/20

Role of gateway component in Data Transfer FTP Client GridFTP Server FTP Protocol Gateway Component GridFTP Protocol FTP protocol is widely used in CNGrid environment, so we used ftp protocol to transfer data between GOS node and gateway GridFTP Client FTP Server GridFTP protocol is widely used in glite environment, so we used ftp protocol to transfer data between glite RB and gateway Role of gateway in data transfer scenario 15/20

Data Transfer: New development SRM interface deployed on GOS from maintainers Simplify data management interoperability issues A client is available to browse the GOS SE 16/20

Security issues Security module used now is just a fast approach for cross-domain security scenario Predefine some users in GOS and glite for interoperability purpose, just a static approach (Identity Mapping Service) Requests from GOS are submitted to glite use predefined voms proxy Requests from glite are submitted to GOS use predefined the username User Management module is designed to keep mapping relationships Security token service Used to keep, distribute, exchange and verify security token between GOS and glite and provide dynamical approach At present MyProxy Server is setup to store temporary security token, but we plan to replace it with newly developed security token service 17/20

Role of security token service Security token Service WS-Trust Client security token still obey the way that GOS used for service invocation Gateway still obey the way that glite used for service invocation GOS glite Role of security token service in cross-domain security scenario 18/20

Future Works Cross-domain Security Security token service based token distribution in crossdomain scenario: More generic solution for cross-domain security token distribution Comply with WS-Trust specification Large scale data share and transfer between different grid infrastructures At present we make full use of sandbox mode to support data transfer between GOS and glite and all the data that transfer between GOS and glite will pass the gateway, the problems lie in: Do unnecessary data transfer between the two middlewares. Cause problems in the case of large scale data transfer (Sandbox volume is less than 20 MB). Adopt SRM Interface in GOS environment to allow large data volumes interoperability. 19/20

Reference materials SEDA: An Architecture for Well-Conditioned, Scalable Internet Services, Matt Welsh, David Culler, and Eric Brewer, University of California, Berkeley Distributed Computation, technical report of London e- Science Centre (Le-Sc), Imperial College, London State-of-the-art of Interoperability in EUChinaGrid Project, EUChinaGrid Project Annual Work Report HiveMind, www.apache.org 20/20

Thanks for your attentions Any Questions? 21/20