SZDG, ecom4com technology, EDGeS-EDGI in large P. Kacsuk MTA SZTAKI

Similar documents
Setup Desktop Grids and Bridges. Tutorial. Robert Lovas, MTA SZTAKI

BOINC extensions in the SZTAKI DesktopGrid system

Improved 3G Bridge scalability to support desktop grid executions

Workflow applications on EGI with WS-PGRADE. Peter Kacsuk and Zoltan Farkas MTA SZTAKI

Dario Ferrer. 27 September CPC Centre for Parallel Computing University of Westminster. Challenges, problems and solutions for

Gergely Sipos MTA SZTAKI

g-eclipse A Framework for Accessing Grid Infrastructures Nicholas Loulloudes Trainer, University of Cyprus (loulloudes.n_at_cs.ucy.ac.

Multiple Broker Support by Grid Portals* Extended Abstract

EGEE and Interoperation

Chapter 2 Introduction to the WS-PGRADE/gUSE Science Gateway Framework

PoS(ISGC 2011 & OGF 31)119

Overview about the SCI-BUS Project (Communities and sustainability)

The LHC Computing Grid

Deploying virtualisation in a production grid

glite Grid Services Overview

Towards a Security Model to Bridge Internet Desktop Grids and Service Grids

Virtualization for Desktop Grid Clients

30 Nov Dec Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy

EDGI Project Infrastructure Benchmarking

Grid services. Enabling Grids for E-sciencE. Dusan Vudragovic Scientific Computing Laboratory Institute of Physics Belgrade, Serbia

Integration of the guse/ws-pgrade and InSilicoLab portals with DIRAC

Evolution of the ATLAS PanDA Workload Management System for Exascale Computational Science

FREE SCIENTIFIC COMPUTING

Grid Scheduling Architectures with Globus

Table of contents. Friday 23 September

ARC NOX AND THE ROADMAP TO THE UNIFIED EUROPEAN MIDDLEWARE

The glite middleware. Ariel Garcia KIT

Easy Access to Grid Infrastructures

AMGA metadata catalogue system

First European Globus Community Forum Meeting

WestminsterResearch

ARC integration for CMS

Interconnect EGEE and CNGRID e-infrastructures

IDGF International Desktop Grid Federation

Advanced School in High Performance and GRID Computing November Introduction to Grid computing.

Bob Jones. EGEE and glite are registered trademarks. egee EGEE-III INFSO-RI

Access the power of Grid with Eclipse

Introduction of NAREGI-PSE implementation of ACS and Replication feature

Grid technologies, solutions and concepts in the synchrotron Elettra

Integration of Cloud and Grid Middleware at DGRZR

Failover procedure for Grid core services

AutoPyFactory: A Scalable Flexible Pilot Factory Implementation

Grid Infrastructure For Collaborative High Performance Scientific Computing

AAI in EGI Current status

Grid Mashups. Gluing grids together with Condor and BOINC

Cycle Sharing Systems

European Grid Infrastructure

Grid and Cloud Activities in KISTI

Virtualization. A very short summary by Owen Synge

The University of Oxford campus grid, expansion and integrating new partners. Dr. David Wallom Technical Manager

Pan-European Grid einfrastructure for LHC Experiments at CERN - SCL's Activities in EGEE

Grid Computing Middleware. Definitions & functions Middleware components Globus glite

Security issues in hierarchically connected BOINC systems

ATLAS NorduGrid related activities

How to build Scientific Gateways with Vine Toolkit and Liferay/GridSphere framework

The glite middleware. Presented by John White EGEE-II JRA1 Dep. Manager On behalf of JRA1 Enabling Grids for E-sciencE

Developing Microsoft Azure Solutions (70-532) Syllabus

On the employment of LCG GRID middleware

ELFms industrialisation plans

GRIDS INTRODUCTION TO GRID INFRASTRUCTURES. Fabrizio Gagliardi

StratusLab Cloud Distribution Installation. Charles Loomis (CNRS/LAL) 3 July 2014

Scientific Workflows and Cloud Computing. Gideon Juve USC Information Sciences Institute

Presentation Title. Grid Computing Project Officer / Research Assistance. InfoComm Development Center (idec) & Department of Communication

PARALLEL PROGRAM EXECUTION SUPPORT IN THE JGRID SYSTEM

THEBES: THE GRID MIDDLEWARE PROJECT Project Overview, Status Report and Roadmap

Ivane Javakhishvili Tbilisi State University High Energy Physics Institute HEPI TSU

Geant4 on Azure using Docker containers

Hungarian Supercomputing Grid 1

The ATLAS Software Installation System v2 Alessandro De Salvo Mayuko Kataoka, Arturo Sanchez Pineda,Yuri Smirnov CHEP 2015

What s new in HTCondor? What s coming? HTCondor Week 2018 Madison, WI -- May 22, 2018

Opal: Simple Web Services Wrappers for Scientific Applications

Parallel Computing in EGI

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

ALICE Grid Activities in US

Azure Development Course

Grids and Security. Ian Neilson Grid Deployment Group CERN. TF-CSIRT London 27 Jan

Java Development and Grid Computing with the Globus Toolkit Version 3

L3.4. Data Management Techniques. Frederic Desprez Benjamin Isnard Johan Montagnat

ACET s e-research Activities

NAREGI PSE with ACS. S.Kawata 1, H.Usami 2, M.Yamada 3, Y.Miyahara 3, Y.Hayase 4, S.Hwang 2, K.Miura 2. Utsunomiya University 2

Operating the Distributed NDGF Tier-1

CERN: LSF and HTCondor Batch Services

Molecular dynamics simulations in the MolDynGrid Virtual Laboratory by means of ARC between Grid and Cloud

The LGI Pilot job portal. EGI Technical Forum 20 September 2011 Jan Just Keijser Willem van Engen Mark Somers

Work Queue + Python. A Framework For Scalable Scientific Ensemble Applications

The OpenNebula Virtual Infrastructure Engine

Fault tolerance based on the Publishsubscribe Paradigm for the BonjourGrid Middleware

Developing Microsoft Azure Solutions (70-532) Syllabus

ARC-XWCH bridge: Running ARC jobs on the XtremWeb-CH volunteer

Cloud interoperability and elasticity with COMPSs

Developing Microsoft Azure Solutions

Ganga The Job Submission Tool. WeiLong Ueng

BUILDING SCIENCE GATEWAYS BY UTILIZING THE GENERIC WS-PGRADE/GUSE WORKFLOW SYSTEM

An Eclipse-based Environment for Programming and Using Service-Oriented Grid

Europeana DSI 2 Access to Digital Resources of European Heritage

XRAY Grid TO BE OR NOT TO BE?

The PanDA System in the ATLAS Experiment

EMI Deployment Planning. C. Aiftimiei D. Dongiovanni INFN

MATLAB. Senior Application Engineer The MathWorks Korea The MathWorks, Inc. 2

The EPIKH, GILDA and GISELA Projects

Transcription:

SZDG, ecom4com technology, EDGeS-EDGI in large P. Kacsuk MTA SZTAKI The EDGI/EDGeS projects receive(d) Community research funding 1

Outline of the talk SZTAKI Desktop Grid (SZDG) SZDG technology: ecom4com EDGeS EDGI 2

SZTAKI Desktop Grid global version Main objectives: Demonstrate the power of the Desktop Grid concept Support Hungarian scientific applications Introducing DG technology in Hungary Three steps for everybody to try and use the technology: 1. Donate one PC to test the client site 2. Port application to the DG server and register PCs for that application 3. Set up a DG server for the community (univ., company, city, etc.) SZTAKI helps in steps 2 and 3 as explained in detail at http://www.desktopgrid.hu/ 3

4

5

Lessons learnt from SZDG project BOINC is excellent to create volunteer DG projects Reliable, stable, robust and scalable technology Drawback Difficult to port applications to BOINC There is no user interface through which endusers could easily launch applications with their parameters 6

Question: Can we use BOINC to build institutional DGs? Yes, but in this case we have to solve the following problems: Abandon the credit system Enable Fast deployment Fast application porting Easy usage by end-users Possible extension with service grids desktop grids clouds 7

ecom4com SZTAKI developed a new technology based on BOINC and SZTAKI Desktop Grid experiences. The new technology is called ecom4com (e- Computing for Communities) The goal of the ecom4com technology is to quickly build and easily use BOINC-based DG systems Easily run PS-nodes of workflows in such DG systems. 8

A real example: CancerGrid workflow N = 30K, M = 100 => about 0.5 year execution time x1 NxM= 3 million x1 xn xn xn NxM x1 Generator job xn xn N=30K xn Generator job NxM= 3 million NxM 9

Accessing Desktop Grids by a Grid portal WS- PGRADE Grid portal University DG Local DG Local DG University Faculty DG Volunteer DG Public DG 10

ecom4com software ecom4com technology provides the following components on top of BOINC: Easy setup: Debian packages for the server For Volunteer Computing: Global For Private/Local Desktop Grids: Local Simplify writing applications: DC-API Support porting legacy applications: GenWrapper WS-PGRADE workflow portal for easily run applications Allow setups peculiar to DGs: clusters as resources, 3G Bridge enables hierarchy, bridging to/from other systems (service grids, clouds) Security enhancements: certificates, virtualisation VMs to easily deploy DG systems with bridges 11

Automatic generation of master and client code SZTAKI developed the DC-API (Distributed Computing API) that can automatically generate WUs for PS jobs at the server of the DG system SZTAKI developed a generic wrapper that can eliminate the boincification of the client code automatically generates the client code without modification of the original code 12

WS- PGRADE (User IF) The ecom4com architecture to support PS legacy applications guse Storage DG Server BOINC Server Components Scheduler BOINC client (WF representation) guse (Workflow enactor) guse Local Submitter guse WS Submitter Author: Peter guse Kacsuk Data server BOINC Task DB DesktopGrid BOINC client 13

WS- PGRADE (User IF) The ecom4com architecture to support PS legacy applications guse Storage DG Server BOINC Server Components Scheduler BOINC client DC-API cli GenWrapper for batch execution Legacy Application (WF representation) guse (Workflow enactor) guse Local Submitter guse WS Submitter Author: Peter guse Kacsuk Data server BOINC Task DB DesktopGrid BOINC client DC-API cli GenWrapper for batch execution Legacy Application 14

WS- PGRADE (User IF) The ecom4com architecture to support PS legacy applications guse Storage DG Server guse DG Submitter Job descr. Job Database (Description of Jobs: Apps, Args, I/O files) BOINC Server Components Scheduler BOINC client DC-API cli GenWrapper for batch execution Legacy Application (WF representation) guse (Workflow enactor) guse Local Submitter guse WS Submitter Author: Peter guse Kacsuk Job descr. Queue Manager Scheduling policy Batch creation DC-API master guse-dg integration WU Data server BOINC Task DB DesktopGrid BOINC client DC-API cli GenWrapper for batch execution Legacy Application 15

DC-API: Writing an application Simple API to hide the grid infrastructure from application developers Usable with minimal set of functions but has additional features that can be used when needed Allows application deployment on different grid infrastructures without source code modification via different backends: Standalone (local) for testing BOINC Condor XtremWeb 16

DC-API application Application Master DC-API BOINC (Public or Local DG) G r i d Condor Pool XtremWeb DC-API DC-API Application Worker Application Worker DC-API 17 Application Worker

Generic Wrapper (GenWrapper) Objective of GenWrapper The features of BOINC wrapper is not enough (e.g. patching config files on client machines, generating extra messages, independent jobs in a WU, etc.) Wanted to be prepared for unknown requirements might be raised by future applications (e.g. Cancergrid) We did not want to extend the BOINC wrapper to make it an XML-based programming language, we choose to BOINCify an existing language -> Bourne shell 18

Wrapping on the DG client GenWrapper Interfaces DC-API on behalf of the application prepares environment for the application unpacks application binaries executes application handles multiple ins/outs BOINC client DC-API cli GenWrapper execute BOINC client DC-API cli Application inputs Legacy Application outputs 19

SZTAKI Desktop Grid local version Main objective: Enable the creation of local DGs for any community Demonstrate how to create such a system Building production Grids requires huge effort and represents a privilege for those organizations where high Grid expertise is available SZTAKI Desktop Grid local version is built on the ecom4com technology Using the local SZDG package Any organization can build a local DG in a day with minimal cost (a strong PC is enough as a server machine) The applications of the local community will be executed by the spare PC cycles of the local community There is no limitation for the applied PCs, all the PCs of the organization can be exploited (heterogeneous Grid) Users of the local SZDG can access the local DG system via WS- PGRADE portal 20

The CancerGrid System (CSG) WS-PGRADE Portal DG jobs Local jobs executing workflows 3G Bridge Job 1 Job 2 Job N Local Resource browsing molecules BOINC server WU 1 WU 2 WU N Portal Storage BOINC client GenWrapper for batch execution WU X Legacy WU Y Application Legacy Application DG Presentation clients from for: all 6 partners th IDGF Tutorial Portal and Desktop Grid server (4 selectable workflow applications) molecule database 21 Molecule database server

Generic Grid-Grid (3G) Bridge BOINC Job source BOINC client 3G Bridge Job Database DC-API Plugin BOINC EGEE ARC Job source EGEE GRAM Job source ARC Job Handler Interface Queue Manager GridHandler Interface EGEE Plugin XtremWe b Plugin EGEE Xtrem Web WS- PGRADE portals Job source WS-PGRADE submitter BES Plugin ARC and Unicor e

The Grid Ecosystem EDGeS Supercomputer based SGs (DEISA) Cluster based service grids (SGs) (EGEE, NorduGrid, etc.) Very expensive, small number of sites and very large number of cores MPI appls Moderately expensive, moderate number of sites and CPUs any appls Desktop grids (DGs) (volunteer DGs home computers, institutional DGs institutional desktops) Inexpensive, very large number of CPUs (~10K 1M) Bag of task appls 23

glite->dg infrastructure glite VO1 glite VO2 glite VOn Jobs EDGeS bridge services WUs 3G Bridge 3G Bridge 3G Bridge EDGeS@home SZDG Local DGs 24

Production glite DG Infrastructure of EDGeS glite based service grid Core Service Resource Bridge XtremWebbased DGs voms lfc AR BOINC-based DGs bdii lb wms myproxy IN2P3 local DG bridge EGEE-DG bridge (ce1) bridge UoW local DGs IN2P3 @home Almere public DGs bridge bridge bridge bridge SZDGr, Fundecyt, @home Almere, Ibercivis public DGs VO1 VO2 25

Desktop Grid types to support SG VOs Volunteer desktop grids: Edges@home City grid: Almeregrid School Grid: Extremadura School Grid Local University Grid: University of Westminster 26

27

Plans for EDGI EDGeS scope only for compute intensive applications for EGEE (glite) EDGI scope for both compute and data intensive applications for EMI/EGI (glite, ARC, Unicore) Extend Desktop Grids with Clouds for QoS 28

Summary ecom4com technology enables to easily build and program local DG systems Easily extend them with other DGs, grids and clouds EDGeS created production infrastructure to extend DG systems with grids Grids with DGs EDGeS creates production infrastructure to extend Grids with DGs DG systems with clouds Based on these new scenarios the DG concept will be much more popular than before (see the case in Europe) We invite you to join us in IDGF and let s collaborate 29

Thank you for your attention Any questions? For more information please visit the EDGeS and EDGI Websites: http://www.edges-grid.eu/ http://edgi-project.eu and/or send e-mail to me: kacsuk@sztaki.hu 30