EUROPEAN MIDDLEWARE INITIATIVE

Size: px
Start display at page:

Download "EUROPEAN MIDDLEWARE INITIATIVE"

Transcription

1 EUROPEAN MIDDLEWARE INITIATIVE MSA2.2 - CONTINUOUS INTEGRATION AND C ERTIFICATION TESTBEDS IN PLACE EC MILESTONE: MS23 Document identifier: EMI_MS23_v1.0.doc Activity: Lead Partner: Document status: Document link: SA2 INFN Final

2 Copyright notice: Copyright (c) Members of the EMI Collaboration See for details on the copyright holders. EMI ( European Middleware Initiative ) is a project partially funded by the European Commission. For more information on the project, its partners and contributors please see This document is released under the Open Access license. You are permitted to copy and distribute verbatim copies of this document containing this copyright notice, but modifying this document is not allowed. You are permitted to copy this document in whole or in part into other documents if you attach the following reference to the copied elements: "Copyright (C) Members of the EMI Collaboration. ". The information contained in this document represents the views of EMI as of the date they are published. EMI does not guarantee that any information contained herein is error-free, or up to date. EMI MAKES NO WARRANTIES, EXPRESS, IMPLIED, OR STATUTORY, BY PUBLISHING THIS DOCUMENT.

3 Delivery Slip Name Partner / Activity Date Signature From Danilo N.Dongiovanni INFN 20/11/2010 Approved by PEB 26/11/2010 Document Log Issue Date Comment Author / Partner 1 08/05/2010 First version available for revision within SA2 Danilo Dongiovanni 2 25/09/2010 Internal Review Alberto Aimar 3 10/08/2010 Added Unicore Resources Danilo Dongiovanni 4 10/28/2010 Added EMI Testbed GGUS Support Unit Danilo Dongiovanni 5 11/02/ /19/2010 Changed document structure: milestone report on first page and technical details in following sections Changed document structure: milestone report on first page. Summary of technical details reported in the documentation url in Annex. Danilo Dongiovanni Danilo Dongiovanni Document Change Record Issue Item Reason for Change 1 2 3

4 MILESTONE REPORT EMI continuous integration and certification testbed has been put in place and it is currently available for Product Team developers to perform integration tests. The infrastructural and operational resources in place consist of: HW/SW resources provided (currently 90 server instances) with: OS installed (one among agreed platforms for EMI, i.e. SL5 x32 or x86_64), network connection and utilities. Server host certificates (required in ARC and glite, containers certificate in UNICORE not related to host). Also a tool for testing certificates generation was made available (ARC). Server level monitoring tool to collect statistics about server availability/reliability. Software Product installed and configured. Full coverage of already available products mentioned in EMI Technical plan was granted. dcache middleware testing instances were provided by DESY partner and integrated in the testbed. A set of configurable information system service instances publishing resources available in the testbed. User Interface accounts provided to PT testers on demand. Operational resources provided: A Virtual Organization for testing has been created: testers.emi-eu.eu EMI Testbed Documentation Instances logbook reporting details on installed software version and configuration. A GGUS Support Unit has been created for support requests handling and tracking Communication channel and task tracking tools for SA2.6 task members activity coordination Technical details and full documentation about the infrastructural and operational resources together with a description of testing scenarios and use cases supported can be found at the following public url: A summary of contents there available are also attached in the Annex of this document.

5 ANNEX: Technical Details on implemented testbed infrastructure 1.1. SUPPORTED TESTING SCENARIOS The testbed model definition moved from EMI Dow and task participants experience and resources and took advantage of discussion within SA2 and with developers representatives (JRA1, SA1, Release and SW Area Manager). The information was collected through surveys or meeting and summarized in minutes or documents available at SA2.6 task homepage 1. A detailed presentation of EMI integration testbed models is object of the DSA2.4 deliverable (ref. SA2.6 task homepage); therefore here we focus on the testing scenarios supported in the testbed infrastructure put in place: 1. EMI Internal Testing scenario A: Integration testing within a minor release (no backward compatibility broken), so that a Release Candidate 2 Service Version (RCSV in the following) can be tested VERSUS other Version Services (PVS in the following). This implies a distributed testbed of production services available for each middleware stack, with possible multiple instances for central services. This could also imply cases of RCSV vs. other RCSV or RCSV vs. (RCSV + PVS): imagine the case of two interacting products in which a common bug is fixed contemporaneously, that would imply RCSV for both of them to be tested together with all other services at production version. Key performance indicators KSA2.1, KSA2.2 will apply to this testbed. 2. EMI Internal Testing scenario B: Integration testing for a major release (where it is allowed to have new features or backward compatibility broken for many services). This implies a testbed of RCSV available for each middleware stack, so basically this means providing hardware with platform installed for Product Teams (PT) to install needed RCSV and allow them for previewing other's PT RCSV. Key performance indicators KSA2.1, KSA2.2 will apply to this testbed DEFAULT USE CASES SUPPORTED Use Case A: developer John needs to test the correct interaction between service X (Release Candidate version) and services Y ( Version), Z (Release Candidate Version). Solution: service X is configured to see resources published in the chosen EMI Testbed central information system instance. Depending on the test performed John may need some configuration effort on services Y or Z, to ensure they can interact with X. John sends a support request to EMI Testbed group (see section 1.5). Use Case B: developer John needs to test the correct interaction between his service X (version X.X.X installed on some instance of his PT) and services Y ( Version), Z (Release Candidate Version). Solution: service X is configured to see resources published in chosen the EMI Testbed central information system instance. He can also setup a new information system merging information from both the mentioned central information system and a local information system publishing some development resources, building a custom testbed view. Notice that service Y and Z will not be configured to see resources out of EMI Testbed. Use Case C: developer John needs to test the correct interaction between his service X (Release Candidate Version) and services Y ( Version), Z (Release Candidate Version) not currently in the testbed, through a User Interface (ex. Job submission from UI involving broker, information system, storage element, compute element). Solution: John requests (see line 3 in this table) an account on one of the User Interfaces provided in the testbed, which is configured to see resources published in the chosen EMI Testbed central information system instance. Depending on the Here we assume that for each service a single Release Candidate version per Release exists.

6 test performed John may need some configuration effort on services Y or Z, to ensure they can interact with X. Moreover John needs service Z to be installed in the Testbed. John sends a support request to EMI Testbed group (see section 1.5) TESTBED INFRASTRUCTURE INVENTORY ARC ARC 3 middleware currently deploys 11 products on Fedora, Debian, RedHat, Ubuntu, Windows and MacOSX platforms. The following services (multiple instances for some services) were made available: Product Name Version Platform Partner Site GIIS service (ARC LDAP- Infosys) Release CentOS5.5 i386 Nagios Release Instant CA Release 0.9 CentOS5.5 i386 CentOS5.5 i386 Classic ARC Grid Monitor Release CentOS5.5 i386 WS-ARC Grid Monitor Release Candidate CentOS5.5 i386 ARC ISIS service (4 Instances) Release Candidate CentOS5.5 i386; SLC5.3/x86; Debian Lenny /x86, NIIF Classic ARC CE Release CE1 type Release 1.1 A-REX Release 1.1 Bartender service (2 instances) Release 1.1 CentOS5.5 i386 CentOS5.5 i386 SLC5.3/x86; Debian Lenny /x8, NIIF Debian Lenny /x86; CentOS5.5/x86_64, NIIF AHash service (2 instances) Release 1.1 Classic ARC clients Release WS-ARC clients Release 1.1 ARC data clients Release Debian Lenny /x86 Ubuntu Hardy /i386 Ubuntu Hardy /i386 Ubuntu Hardy /i386 NIIF 3

7 Librarian service (3 instances) Release 1.1 Echo service Release 1.1 Shepherd service (2 instances) Release 1.1 CentOS5.5/x86_64 ; Debian Lenny /x86 SLC5.3/x86 ; Debian Lenny /x86 Debian Lenny /x86, NIIF, NIIF NIIF glite GLite 4 middleware currently deploys 19 products in Release 3.1 SL4 OS and 17 products in Release 3.2 SL5 (when service implementation changes without affecting API services are counted as one, ex. glitevoms mysql/oracle database implementation) on SL4 and SL5 platforms. The following services (multiple instances for some services) were made available: Product Name Version Platform Partner Site glitewms (3 instances) glite 3.1 ; RC SL4 INFN, CERN dgas ig_hlr Version ig48_sl4 SL4 INFN glite-cream 3.2 Version(LSF); SL5/x86_64;S LC4.8/x86 INFN glite UI (3 instances) 3.1 ; 3.2, RC SLC4.8; SL5/x86_64 INFN, CERN GliteBDII (site; Top) SLC4.8/x86 CERN glite-px SLC4.8/x86 CERN glite-lcgce SLC4.8/x86 CERN glite WN ; ; SL4.8/x86, SL4.8/x86_64; SL5.5/x86_64 CERN glite-lfc_mysql SL4.8/x86_64 CERN glite-voms ; 3.2 SLC4.8/x86; SL5.5/x86_64 CERN glite-fts_oracle SLC4.8/x86 CERN glite-vobox SLC4.8/x86 CERN glite-se_dpm_mysql SL5.5/x86_64 CERN 4

8 glite-dpm_pool SLC4.8/x86 CERN glite-se_dpm_mysql / disk SLC4.8/x86 CERN Nagios ; SLC5.5/x86_6 4 ; SLC4.8/x86 glite (9 instances) , RC SLC5.5/x86_6 4 ; SLC4.8/x86 CERN CESNET STORM INFN grid Release _ig50_sl4 Version SL4 INFN UNICORE UNICORE 5 middleware currently deploys 11 products without particular dependencies making it executable on Linux, Windows and MacOSX platforms, generally deployed on opensuse 11.2 for certification purposes. Product Name Version Platform Partner Site Gateway ; opensuse 11.3 JUELICH Registry ; opensuse 11.3 JUELICH X incl. XNJS ; opensuse 11.3 JUELICH OGSA-BES interfaces ; opensuse 11.3 JUELICH HiLA 2.1 ; opensuse 11.3 JUELICH XUUDB ; opensuse 11.3 JUELICH UVOS ; opensuse 11.3 JUELICH Command line Client (UCC) ; opensuse 11.3 JUELICH dcache Also dcache 6 certification resources were kindly made available for integration testing purposes from dcache EMI partners. In particular resources below can be accessed through CERN gliteui instance. Product Name Version Platform Partner Site

9 BDII SL4 DESY dcache A SL5; SL4 32bit PNFS; SL4 64bit PNFS dcache B Sl5; SL4 32bit Chimera; SL4 64bit Chimera; DESY DESY 1.4. MONITORING AND KPI Key performance indicators KSA2.1 and KSA2.2 reported in table below, imply automatic monitoring solutions for resources able to produce statistics on server's availability and reliability. Each middleware currently has a monitoring solution deployed: ARC (Ganglia, GridMonitor, Nagios), glite (Nagios), UNICORE (Nagios). The Key Performance Indicators. CODE KPI Description Method to Measure Estimated Targets KSA2.1 SA2 Services Reliability % uptime dependent only on the SA2 services themselves (individual KPIs for test beds, repository, etc) Participating sites monitoring tools 99.00% KSA2.2 SA2 Services Availability Total % uptime including the underlying suppliers (individual KPIs for test beds, repository, etc) Participating sites monitoring tools 97.00% Availability and reliability statistics are currently provided just on testbed server instances, not on services given the fact that not all services have availability/reliability metrics defined and tools to measure them. Concerning the adopted tool to monitor instances and produce statistics, all middlewares plan to converge on Nagios 7 solution for two suitable features for our task purposes: Evolution of Nagios into grid monitoring service monitoring 8 Nagios, which is expected to provide metrics for Solution for geographical distribution: a second level Nagios can implement a central instance, republishing and aggregating data coming from local sites Nagios instances. Initially, availability and reliability statistics periodically produced by local sites Nagios instances will be made available in the testbed public documentation center for SA2.6 described in section

10 1.5. DOCUMENTATION For public documentation of EMI Internal Testbed resources the following web page was put in place reporting: Description of supported testing scenarios Role and duties of SA2.6 task and PT contribution Procedures for testbed update requests Testbed Monitoring solutions Procedure to enable testers.emi-eu.eu VO EMI Testbed Coverage of EMI components Testbed Inventory with a list of provided instances specifying: Middleware Suite, Service Deployed, Platform, Server Hostname, Site Location, reference Product Team, Status Logbook Status Logbook field in previous table is a link to a instance-specific web page describing the hardware details of instance, software version installed, configuration information and history of updates. The maintenance of this page is in charge of people performing installation; configuration or updates (can be PT members) USER SUPPORT, COORDINATION AND COMMUNICATION Both coordination and installation/configuration activities concerning testbeds require clear channels of communication and a way to track the effort of people involved. As mentioned in DS2.4 deliverable document, a distributed effort model for testbed setup and maintenance was adopted, with the possible involvement of Product Teams members as support effort for service installation and configuration. To coordinate and track all the distributed effort the following solution was adopted: User Support Requests Handling: An emi-support-testbed support unit has been created in GGUS, for testbed support requests reception. Representatives from all partners contributing to testbed are members of the support unit, and agreed on a 2 working days response time on best effort. The Support Unit will be part of next GGUS release. The adoption of GGUS will give a common framework to handle both requests coming from EMI developers or users and those coming from external users (e.g. the users of large scale testbed involving other projects or partners contribution). Communication: SA2.6: an emi-sa26@eu-emi.eu was created both for task internal communication and for testbed requests reception. Activity Tracking: SA2.6 and Product Teams activities on testbed will be tracked through Savannah 9 tasks. An emi-sa2-testbed Savannah squad has been created to submit requests. Product Team squads have been created to track PT activities on testbeds 1.7. TESTBED UPDATE Resources made available for integration testing described in section 1.2 form a first nucleus of EMI testbed, putting together services currently used for certification from all middlewares converging into EMI. 9

11 EMI testbed evolution is then strictly connected to its actual usage by product teams members performing integration testing. In fact depending on the specific integration test to be performed different coordination, installation or configuration activities can be required. So we expect EMI internal testbed customers (PTs, SW Area Manager, SA1, and JRA1) to submit support requests for the following expected cases, treated as described below: Requests for configuration support of existing services. These requests may include enabling VO/users, making services to talk to each other, custom bdii setup etc. Procedure: Open a GGUS ticket assigned to EMI-Testbeds Support Unit explaining your testing and configuration needs. The request will be then evaluated and tracked into a savannah task on the testbed squad. If needed a PT members of services involved in the test will be contacted and their contribute will be tracked by savannah Requests for new services setup (or particular service RC versions setup): Open a GGUS ticket assigned to EMI-Testbeds Support Unit explaining: Your testing needs and the type and version of services you need to be installed and the PT producing that service Please also specify if you need the service to be included in the permanent EMI testbed. The request will be then evaluated and tracked into a savannah task on the testbed squad. If needed, the involved PT members will be then contacted and their contribute tracked by savannah tasks. Requests for specific testbed (in this category: performance tests, security tests, data management tests, etc.): Open a GGUS ticket assigned to EMI-Testbeds Support Unit explaining: Your testing needs and an estimate of HW and SW requirements for your test PTs involved in the setup and suggestions on possible sites/pt/ngi that may help in the setup. The request will be then evaluated and tracked into a savannah task on the testbed squad. The period of time you expect to have the testbed on for The involved PT members will be then contacted and their contribute tracked by savannah tasks TESTBED ACCESSIBILITY User Interface Service Instances: As default use case we assume testbed users to have direct access just to user interface instances, that is just to grid middleware access point services. To request an account on a EMI Testbed User Interfaces instance, every user with a valid certificate from a trusted Certification Authority, should send a user support request following the procedure and tools described at point 4 in this table. Notice that it is also possible to install the set of clients directly on personal machine (ex. usual use case in UNICORE). Other Service Instances: Root access on other services can be granted on request, depending on the local sites security policy (which generally is also subjected to national laws about traceability of access on servers). If the access is required for debugging or logs exploration purposes, logs sharing solutions will be implemented on demand (publishing of logs on public AFS area, GridFTP downloads, https access) TESTBED RESOURCES DISCOVERY Information Systems Configuration: Each of the middleware has a service for resource discovery and publication (ARC, glite BDII, Unicore Registry). A central information system instance was configured for each middleware publishing the resources in the testbed. Cross-middleware

12 compatibility among existing information system services is in EMI plans, and EMI Testbed will reflect that integration once it will be technically available. Implications for testbed usage: the set of resources visible to the end users (developers) depends on the configuration of their access point (the information system instance configured in the User Interface instance user is logged on). In practice user can build a custom testbed by selecting needed resources from the pool of those published in the central information system or merging them with other resources published on other information system (ex. Product Team internal development testbed).

EUROPEAN MIDDLEWARE INITIATIVE

EUROPEAN MIDDLEWARE INITIATIVE EUROPEAN MIDDLEWARE INITIATIVE DSA2.4 - CONTINUOUS INTEGRATION AND CERTIFICATION TESTBEDS EU DELIVERABLE: D4.4 Document identifier: EMI-DSA2.4-1277550-Integration_Testbeds_v1.0.doc Activity: Lead Partner:

More information

EUROPEAN MIDDLEWARE INITIATIVE

EUROPEAN MIDDLEWARE INITIATIVE EUROPEAN MIDDLEWARE INITIATIVE DSA2.3.1 - PERIODIC QA REPORTS EU DELIVERABLE: D4.3.1 Document identifier: EMI-DSA2.3.1-QAReport-Final.doc Date: 31/07/2010 Activity: Lead Partner: Document status: Document

More information

EUROPEAN MIDDLEWARE INITIATIVE

EUROPEAN MIDDLEWARE INITIATIVE EUROPEAN MIDDLEWARE INITIATIVE DJRA1.6.1 INTEGRATION WORK PLAN AND STATUS REPORT EC DELIVERABLE: D5.6.1 Document identifier: EMI-DJRA1.6.1-1277592- Integration_Work_Plan_v1.0.doc Activity: Lead Partner:

More information

EMI Deployment Planning. C. Aiftimiei D. Dongiovanni INFN

EMI Deployment Planning. C. Aiftimiei D. Dongiovanni INFN EMI Deployment Planning C. Aiftimiei D. Dongiovanni INFN Outline Migrating to EMI: WHY What's new: EMI Overview Products, Platforms, Repos, Dependencies, Support / Release Cycle Migrating to EMI: HOW Admin

More information

EUROPEAN MIDDLEWARE INITIATIVE

EUROPEAN MIDDLEWARE INITIATIVE EUROPEAN MIDDLEWARE INITIATIVE MNA1.3 - TECHNICAL COLL ABORATI ON PROCEDURES WITH EGI AND PRACE ARE EST ABLISHED EU MILESTONE: MS3 Document identifier: EMI_MS3.doc Activity: Lead Partner: Document status:

More information

EGEE and Interoperation

EGEE and Interoperation EGEE and Interoperation Laurence Field CERN-IT-GD ISGC 2008 www.eu-egee.org EGEE and glite are registered trademarks Overview The grid problem definition GLite and EGEE The interoperability problem The

More information

Jozef Cernak, Marek Kocan, Eva Cernakova (P. J. Safarik University in Kosice, Kosice, Slovak Republic)

Jozef Cernak, Marek Kocan, Eva Cernakova (P. J. Safarik University in Kosice, Kosice, Slovak Republic) ARC tools for revision and nightly functional tests Jozef Cernak, Marek Kocan, Eva Cernakova (P. J. Safarik University in Kosice, Kosice, Slovak Republic) Outline Testing strategy in ARC ARC-EMI testing

More information

glite Grid Services Overview

glite Grid Services Overview The EPIKH Project (Exchange Programme to advance e-infrastructure Know-How) glite Grid Services Overview Antonio Calanducci INFN Catania Joint GISELA/EPIKH School for Grid Site Administrators Valparaiso,

More information

The glite middleware. Presented by John White EGEE-II JRA1 Dep. Manager On behalf of JRA1 Enabling Grids for E-sciencE

The glite middleware. Presented by John White EGEE-II JRA1 Dep. Manager On behalf of JRA1 Enabling Grids for E-sciencE The glite middleware Presented by John White EGEE-II JRA1 Dep. Manager On behalf of JRA1 John.White@cern.ch www.eu-egee.org EGEE and glite are registered trademarks Outline glite distributions Software

More information

EUROPEAN MIDDLEWARE INITIATIVE

EUROPEAN MIDDLEWARE INITIATIVE EUROPEAN MIDDLEWARE INITIATIVE MSA1.2.1 - EMI REF ERENCE RELE AS ES EC MILESTONE: MS18 Document identifier: EMI_MS18_v1.0.doc Activity: Lead Partner: Document status: Document link: SA1 INFN Final http://cdsweb.cern.ch/record/1277546?ln=en

More information

EUROPEAN MIDDLEWARE INITIATIVE

EUROPEAN MIDDLEWARE INITIATIVE EUROPEAN MIDDLEWARE INITIATIVE DSA2.2.2 - QA TOOL S DOCUMEN T ATION EU DELIVERABLE: D4.2.2 Document identifier: EMI-DSA2.2.2-1277590- QA_Tools_Documentation-Rev.1-v1.0.doc Activity: Lead Partner: Document

More information

PoS(EGICF12-EMITC2)081

PoS(EGICF12-EMITC2)081 University of Oslo, P.b.1048 Blindern, N-0316 Oslo, Norway E-mail: aleksandr.konstantinov@fys.uio.no Martin Skou Andersen Niels Bohr Institute, Blegdamsvej 17, 2100 København Ø, Denmark E-mail: skou@nbi.ku.dk

More information

ARC NOX AND THE ROADMAP TO THE UNIFIED EUROPEAN MIDDLEWARE

ARC NOX AND THE ROADMAP TO THE UNIFIED EUROPEAN MIDDLEWARE ARC NOX AND THE ROADMAP TO THE UNIFIED EUROPEAN MIDDLEWARE GRID-2010, Dubna, July 2 2010 Oxana Smirnova (on behalf of the NorduGrid Collaboration) Outlook Usage of ARC in NDGF and ATLAS Overview of the

More information

30 Nov Dec Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy

30 Nov Dec Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy Why the Grid? Science is becoming increasingly digital and needs to deal with increasing amounts of

More information

Outline. Infrastructure and operations architecture. Operations. Services Monitoring and management tools

Outline. Infrastructure and operations architecture. Operations. Services Monitoring and management tools EGI-InSPIRE EGI Operations Tiziana Ferrari/EGI.eu EGI Chief Operations Officer 1 Outline Infrastructure and operations architecture Services Monitoring and management tools Operations 2 Installed Capacity

More information

ARC integration for CMS

ARC integration for CMS ARC integration for CMS ARC integration for CMS Erik Edelmann 2, Laurence Field 3, Jaime Frey 4, Michael Grønager 2, Kalle Happonen 1, Daniel Johansson 2, Josva Kleist 2, Jukka Klem 1, Jesper Koivumäki

More information

EMI Componets Installation And Configuration

EMI Componets Installation And Configuration EMI Componets Installation And Configuration Sara Bertocco INFN Padova - GridKA School EMI is partially funded by the European Commission under Grant Agreement RI-261611 Tour in the EMI site EMI site:

More information

On the EGI Operational Level Agreement Framework

On the EGI Operational Level Agreement Framework EGI-InSPIRE On the EGI Operational Level Agreement Framework Tiziana Ferrari, EGI.eu EGI Chief Operations Officer 1 Outline EGI and its ecosystem EGI Service Infrastructure Operational level agreements

More information

Eclipse Technology Project: g-eclipse

Eclipse Technology Project: g-eclipse (Incubation) Document classification: Made available under the Eclipse Public License v1.0. Date: September 11, 2007 Abstract: This document contains the Release Review Documentation for the Eclipse Technology

More information

EGI Operations and Best Practices

EGI Operations and Best Practices EGI-InSPIRE EGI Operations and Best Practices Tiziana Ferrari/EGI.eu CHAIN-REDS Workshop ISGC 2013 CHAIN-REDS workshop, ISGC, March 2013 1 EGI Infrastructure EGI Infrastructure is the composition of multiple

More information

EUROPEAN MIDDLEWARE INITIATIVE

EUROPEAN MIDDLEWARE INITIATIVE EUROPEAN MIDDLEWARE INITIATIVE DSA2.2.3 - QA TOOLS DOCUMENTATION EU DELIVERABLE: D4.2.3 Document identifier: EMI-DSA2.2.3-1277591-QA_Tools_Documentationv1.0.doc Activity: SA2 Lead Partner: CERN Document

More information

Status of KISTI Tier2 Center for ALICE

Status of KISTI Tier2 Center for ALICE APCTP 2009 LHC Physics Workshop at Korea Status of KISTI Tier2 Center for ALICE August 27, 2009 Soonwook Hwang KISTI e-science Division 1 Outline ALICE Computing Model KISTI ALICE Tier2 Center Future Plan

More information

EUROPEAN MIDDLEWARE INITIATIVE

EUROPEAN MIDDLEWARE INITIATIVE EUROPEAN MIDDLEWARE INITIATIVE DJRA1.1.1 - COMPUTE ARE A WORK PLAN AND STATUS REPORT EC DELIVERABLE: D5.1.1 Document identifier: EMI-DJRA1.1.1-1277608- Compute_Area_Work_Plan-v1.0.doc Activity: Lead Partner:

More information

Integration of Cloud and Grid Middleware at DGRZR

Integration of Cloud and Grid Middleware at DGRZR D- of International Symposium on Computing 2010 Stefan Freitag Robotics Research Institute Dortmund University of Technology March 12, 2010 Overview D- 1 D- Resource Center Ruhr 2 Clouds in the German

More information

Monitoring System for the GRID Monte Carlo Mass Production in the H1 Experiment at DESY

Monitoring System for the GRID Monte Carlo Mass Production in the H1 Experiment at DESY Journal of Physics: Conference Series OPEN ACCESS Monitoring System for the GRID Monte Carlo Mass Production in the H1 Experiment at DESY To cite this article: Elena Bystritskaya et al 2014 J. Phys.: Conf.

More information

EUROPEAN MIDDLEWARE INITIATIVE

EUROPEAN MIDDLEWARE INITIATIVE EUROPEAN MIDDLEWARE INITIATIVE VOMS CORE AND WMS SECURITY ASSESSMENT EMI DOCUMENT Document identifier: EMI-DOC-SA2- VOMS_WMS_Security_Assessment_v1.0.doc Activity: Lead Partner: Document status: Document

More information

I Tier-3 di CMS-Italia: stato e prospettive. Hassen Riahi Claudio Grandi Workshop CCR GRID 2011

I Tier-3 di CMS-Italia: stato e prospettive. Hassen Riahi Claudio Grandi Workshop CCR GRID 2011 I Tier-3 di CMS-Italia: stato e prospettive Claudio Grandi Workshop CCR GRID 2011 Outline INFN Perugia Tier-3 R&D Computing centre: activities, storage and batch system CMS services: bottlenecks and workarounds

More information

Provisioning of Grid Middleware for EGI in the framework of EGI InSPIRE

Provisioning of Grid Middleware for EGI in the framework of EGI InSPIRE Ibergrid 2010 Provisioning of Grid Middleware for EGI in the framework of EGI InSPIRE M. David G. Borges, J. Gomes, I. Campos, A. Lopez, P. Orviz, J. López Cacheiro, C. Fernandez and A. Simon LIP, CSIC,

More information

ALICE Grid Activities in US

ALICE Grid Activities in US ALICE Grid Activities in US 1 ALICE-USA Computing Project ALICE-USA Collaboration formed to focus on the ALICE EMCal project Construction, installation, testing and integration participating institutions

More information

Grid Computing Middleware. Definitions & functions Middleware components Globus glite

Grid Computing Middleware. Definitions & functions Middleware components Globus glite Seminar Review 1 Topics Grid Computing Middleware Grid Resource Management Grid Computing Security Applications of SOA and Web Services Semantic Grid Grid & E-Science Grid Economics Cloud Computing 2 Grid

More information

g-eclipse A Framework for Accessing Grid Infrastructures Nicholas Loulloudes Trainer, University of Cyprus (loulloudes.n_at_cs.ucy.ac.

g-eclipse A Framework for Accessing Grid Infrastructures Nicholas Loulloudes Trainer, University of Cyprus (loulloudes.n_at_cs.ucy.ac. g-eclipse A Framework for Accessing Grid Infrastructures Trainer, University of Cyprus (loulloudes.n_at_cs.ucy.ac.cy) EGEE Training the Trainers May 6 th, 2009 Outline Grid Reality The Problem g-eclipse

More information

EUROPEAN MIDDLEWARE INITIATIVE

EUROPEAN MIDDLEWARE INITIATIVE EUROPEAN MIDDLEWARE INITIATIVE DSA2.2.1 - QA TOOLS DOCUMENT ATION EU DELIVERABLE: D4.2.1 Document identifier: EMI-DSA2.2.1-1277589- QA_Tools_Documentation-v1.0.doc Activity: Lead Partner: Document status:

More information

Installation Guide. How to install the Active Security monitoring component for int.eu.grid JRA1

Installation Guide. How to install the Active Security monitoring component for int.eu.grid JRA1 Installation Guide How to install the Active Security monitoring component for int.eu.grid JRA1 Document Filename: Workpackage: Partner(s): Lead Partner: Config ID: Document classification: Installation

More information

Regional SEE-GRID-SCI Training for Site Administrators Institute of Physics Belgrade March 5-6, 2009

Regional SEE-GRID-SCI Training for Site Administrators Institute of Physics Belgrade March 5-6, 2009 SEE-GRID-SCI SEE-GRID-SCI Operations Procedures and Tools www.see-grid-sci.eu Regional SEE-GRID-SCI Training for Site Administrators Institute of Physics Belgrade March 5-6, 2009 Antun Balaz Institute

More information

Middleware-Tests with our Xen-based Testcluster

Middleware-Tests with our Xen-based Testcluster Tier-2 meeting March 3, 2008 1 Introduction Overview of the Testcluster Overview of the installed Software Xen 2 Main Original Usage of the Testcluster Present Activities The Testcluster Future Activities

More information

Agenda: Alberto dimeglio (AdM) Balazs Konya (BK) Helmut Heller (HH) Steve Crouch (SC)

Agenda:   Alberto dimeglio (AdM) Balazs Konya (BK) Helmut Heller (HH) Steve Crouch (SC) 1 st Technical Coordination Board Date: 25 th October 2010 Place: Thon Hotel, Brussels Centre. Agenda: https://www.egi.eu/indico/conferencedisplay.py?confid=149 Participants Alberto dimeglio (AdM) Balazs

More information

Access the power of Grid with Eclipse

Access the power of Grid with Eclipse Access the power of Grid with Eclipse Harald Kornmayer (Forschungszentrum Karlsruhe GmbH) Markus Knauer (Innoopract GmbH) October 11th, 2006, Eclipse Summit, Esslingen 2006 by H. Kornmayer, M. Knauer;

More information

The ARC Information System

The ARC Information System The ARC Information System Overview of a GLUE2 compliant production system Florido Paganelli, Lund University EGI Community Forum 2012, Munich, 26-30 March 2012 Outline Existing solutions ARC key concepts

More information

The EU DataGrid Testbed

The EU DataGrid Testbed The EU DataGrid Testbed The European DataGrid Project Team http://www.eudatagrid.org DataGrid is a project funded by the European Union Grid Tutorial 4/3/2004 n 1 Contents User s Perspective of the Grid

More information

Setup Desktop Grids and Bridges. Tutorial. Robert Lovas, MTA SZTAKI

Setup Desktop Grids and Bridges. Tutorial. Robert Lovas, MTA SZTAKI Setup Desktop Grids and Bridges Tutorial Robert Lovas, MTA SZTAKI Outline of the SZDG installation process 1. Installing the base operating system 2. Basic configuration of the operating system 3. Installing

More information

Interoperating AliEn and ARC for a distributed Tier1 in the Nordic countries.

Interoperating AliEn and ARC for a distributed Tier1 in the Nordic countries. for a distributed Tier1 in the Nordic countries. Philippe Gros Lund University, Div. of Experimental High Energy Physics, Box 118, 22100 Lund, Sweden philippe.gros@hep.lu.se Anders Rhod Gregersen NDGF

More information

Interconnect EGEE and CNGRID e-infrastructures

Interconnect EGEE and CNGRID e-infrastructures Interconnect EGEE and CNGRID e-infrastructures Giuseppe Andronico Interoperability and Interoperation between Europe, India and Asia Workshop Barcelona - Spain, June 2 2007 FP6 2004 Infrastructures 6-SSA-026634

More information

EDGI Project Infrastructure Benchmarking

EDGI Project Infrastructure Benchmarking Masters Degree in Informatics Engineering Internship Final Report EDGI Project Infrastructure Benchmarking Serhiy Boychenko Viktorovich serhiy@student.dei.uc.pt Advisor: Filipe João Boavida de Mendonça

More information

DESY. Andreas Gellrich DESY DESY,

DESY. Andreas Gellrich DESY DESY, Grid @ DESY Andreas Gellrich DESY DESY, Legacy Trivially, computing requirements must always be related to the technical abilities at a certain time Until not long ago: (at least in HEP ) Computing was

More information

European Globus Community Forum The Future of Globus in Europe

European Globus Community Forum The Future of Globus in Europe European Globus Community Forum The Future of Globus in Europe Michael Krieger, RISC Software GmbH Matthias Hofmann, TU Dortmund University Globus usage in Europe Do we need Globus in Europe??? Number

More information

Ref. Ares(2015) /12/2015. D9.1 Project Collaborative Workspace Bénédicte Ferreira, IT

Ref. Ares(2015) /12/2015. D9.1 Project Collaborative Workspace Bénédicte Ferreira, IT Ref. Ares(2015)6013006-22/12/2015 D9.1 Project Collaborative Workspace Bénédicte Ferreira, IT AQUAEXCEL 2020 Deliverable D9.1 Executive Summary Objectives: This deliverable presents the AQUAEXCEL 2020

More information

ISTITUTO NAZIONALE DI FISICA NUCLEARE

ISTITUTO NAZIONALE DI FISICA NUCLEARE ISTITUTO NAZIONALE DI FISICA NUCLEARE Sezione di Perugia INFN/TC-05/10 July 4, 2005 DESIGN, IMPLEMENTATION AND CONFIGURATION OF A GRID SITE WITH A PRIVATE NETWORK ARCHITECTURE Leonello Servoli 1,2!, Mirko

More information

The INFN Tier1. 1. INFN-CNAF, Italy

The INFN Tier1. 1. INFN-CNAF, Italy IV WORKSHOP ITALIANO SULLA FISICA DI ATLAS E CMS BOLOGNA, 23-25/11/2006 The INFN Tier1 L. dell Agnello 1), D. Bonacorsi 1), A. Chierici 1), M. Donatelli 1), A. Italiano 1), G. Lo Re 1), B. Martelli 1),

More information

EGI-InSPIRE. ARC-CE IPv6 TESTBED. Barbara Krašovec, Jure Kranjc ARNES. EGI-InSPIRE RI

EGI-InSPIRE. ARC-CE IPv6 TESTBED. Barbara Krašovec, Jure Kranjc ARNES.   EGI-InSPIRE RI EGI-InSPIRE ARC-CE IPv6 TESTBED Barbara Krašovec, Jure Kranjc ARNES Why ARC-CE over IPv6? - IPv4 exhaustion - On Friday 14 th, RIPE NCC has announced that the last /8 is being distributed from available

More information

Easy Access to Grid Infrastructures

Easy Access to Grid Infrastructures Easy Access to Grid Infrastructures Dr. Harald Kornmayer (NEC Laboratories Europe) On behalf of the g-eclipse consortium WP11 Grid Workshop Grenoble, France 09 th of December 2008 Background in astro particle

More information

Grid Computing. MCSN - N. Tonellotto - Distributed Enabling Platforms

Grid Computing. MCSN - N. Tonellotto - Distributed Enabling Platforms Grid Computing 1 Resource sharing Elements of Grid Computing - Computers, data, storage, sensors, networks, - Sharing always conditional: issues of trust, policy, negotiation, payment, Coordinated problem

More information

Andrea Sciabà CERN, Switzerland

Andrea Sciabà CERN, Switzerland Frascati Physics Series Vol. VVVVVV (xxxx), pp. 000-000 XX Conference Location, Date-start - Date-end, Year THE LHC COMPUTING GRID Andrea Sciabà CERN, Switzerland Abstract The LHC experiments will start

More information

European Grid Infrastructure

European Grid Infrastructure EGI-InSPIRE European Grid Infrastructure A pan-european Research Infrastructure supporting the digital European Research Area Michel Drescher Technical Manager, EGI.eu Michel.Drescher@egi.eu TPDL 2013

More information

How to build Scientific Gateways with Vine Toolkit and Liferay/GridSphere framework

How to build Scientific Gateways with Vine Toolkit and Liferay/GridSphere framework How to build Scientific Gateways with Vine Toolkit and Liferay/GridSphere framework Piotr Dziubecki, Piotr Grabowski, Michał Krysiński, Tomasz Kuczyński, Dawid Szejnfeld, Dominik Tarnawczyk, Gosia Wolniewicz

More information

Monte Carlo Production on the Grid by the H1 Collaboration

Monte Carlo Production on the Grid by the H1 Collaboration Journal of Physics: Conference Series Monte Carlo Production on the Grid by the H1 Collaboration To cite this article: E Bystritskaya et al 2012 J. Phys.: Conf. Ser. 396 032067 Recent citations - Monitoring

More information

The glite middleware. Ariel Garcia KIT

The glite middleware. Ariel Garcia KIT The glite middleware Ariel Garcia KIT Overview Background The glite subsystems overview Security Information system Job management Data management Some (my) answers to your questions and random rumblings

More information

A distributed tier-1. International Conference on Computing in High Energy and Nuclear Physics (CHEP 07) IOP Publishing. c 2008 IOP Publishing Ltd 1

A distributed tier-1. International Conference on Computing in High Energy and Nuclear Physics (CHEP 07) IOP Publishing. c 2008 IOP Publishing Ltd 1 A distributed tier-1 L Fischer 1, M Grønager 1, J Kleist 2 and O Smirnova 3 1 NDGF - Nordic DataGrid Facilty, Kastruplundgade 22(1), DK-2770 Kastrup 2 NDGF and Aalborg University, Department of Computer

More information

ATLAS Tier-3 UniGe

ATLAS Tier-3 UniGe ATLAS Tier-3 cluster @ UniGe Luis March and Yann Meunier (Université de Genève) CHIPP + CSCS GRID: Face To Face meeting CERN, September 1st 2016 Description of ATLAS Tier-3 cluster at UniGe The ATLAS Tier-3

More information

Deploying virtualisation in a production grid

Deploying virtualisation in a production grid Deploying virtualisation in a production grid Stephen Childs Trinity College Dublin & Grid-Ireland TERENA NRENs and Grids workshop 2 nd September 2008 www.eu-egee.org EGEE and glite are registered trademarks

More information

Parallel Computing in EGI

Parallel Computing in EGI Parallel Computing in EGI V. Šipková, M. Dobrucký, and P. Slížik Ústav informatiky, Slovenská akadémia vied 845 07 Bratislava, Dúbravská cesta 9 http://www.ui.sav.sk/ {Viera.Sipkova, Miroslav.Dobrucky,

More information

The LHC Computing Grid

The LHC Computing Grid The LHC Computing Grid Gergely Debreczeni (CERN IT/Grid Deployment Group) The data factory of LHC 40 million collisions in each second After on-line triggers and selections, only 100 3-4 MB/event requires

More information

The European DataGRID Production Testbed

The European DataGRID Production Testbed The European DataGRID Production Testbed Franck Bonnassieux CNRS/UREC ENS-Lyon France DataGrid Network Work Package Manager Franck.Bonnassieux@ens-lyon.fr Presentation outline General DataGrid project

More information

Grids and Security. Ian Neilson Grid Deployment Group CERN. TF-CSIRT London 27 Jan

Grids and Security. Ian Neilson Grid Deployment Group CERN. TF-CSIRT London 27 Jan Grids and Security Ian Neilson Grid Deployment Group CERN TF-CSIRT London 27 Jan 2004-1 TOC Background Grids Grid Projects Some Technical Aspects The three or four A s Some Operational Aspects Security

More information

Managing Scientific Computations in Grid Systems

Managing Scientific Computations in Grid Systems Managing Scientific Computations in Grid Systems Salman Toor Division of Scientific Computing Department of Information Technology Uppsala University November 11, 2008 Motivation Integration architecture

More information

On the employment of LCG GRID middleware

On the employment of LCG GRID middleware On the employment of LCG GRID middleware Luben Boyanov, Plamena Nenkova Abstract: This paper describes the functionalities and operation of the LCG GRID middleware. An overview of the development of GRID

More information

The Legnaro-Padova distributed Tier-2: challenges and results

The Legnaro-Padova distributed Tier-2: challenges and results The Legnaro-Padova distributed Tier-2: challenges and results Simone Badoer a, Massimo Biasotto a,fulviacosta b, Alberto Crescente b, Sergio Fantinel a, Roberto Ferrari b, Michele Gulmini a, Gaetano Maron

More information

E G E E - I I. Document identifier: Date: 10/08/06. Document status: Document link:

E G E E - I I. Document identifier: Date: 10/08/06. Document status: Document link: E G E E - I I A F S P O O L A C C O U N T U S E R S G S S K L O G A N D L C M A P S E X T E N S I O N T O S U P P O R T A F S U S E R S A S E G E E P O O L A C C O U N T U S E R S Document identifier:

More information

Overview of HEP software & LCG from the openlab perspective

Overview of HEP software & LCG from the openlab perspective Overview of HEP software & LCG from the openlab perspective Andreas Unterkircher, CERN openlab February 2005 Andreas Unterkircher 1 Contents 1. Opencluster overview 2. High Energy Physics (HEP) software

More information

Deliverable D8.4 Certificate Transparency Log v2.0 Production Service

Deliverable D8.4 Certificate Transparency Log v2.0 Production Service 16-11-2017 Certificate Transparency Log v2.0 Production Contractual Date: 31-10-2017 Actual Date: 16-11-2017 Grant Agreement No.: 731122 Work Package/Activity: 8/JRA2 Task Item: Task 6 Nature of Deliverable:

More information

Introduction to Grid Technology

Introduction to Grid Technology Introduction to Grid Technology B.Ramamurthy 1 Arthur C Clarke s Laws (two of many) Any sufficiently advanced technology is indistinguishable from magic." "The only way of discovering the limits of the

More information

The Role and Functions of European Grid Infrastructure

The Role and Functions of European Grid Infrastructure The Role and Functions of European Grid Infrastructure Luděk Matyska Masaryk University and CESNET Czech Republic (Ludek.Matyska@cesnet.cz) EGI_DS Project Director What is a Grid? A distributed system

More information

THE WIDE AREA GRID. Architecture

THE WIDE AREA GRID. Architecture THE WIDE AREA GRID Architecture Context The Wide Area Grid concept was discussed during several WGISS meetings The idea was to imagine and experiment an infrastructure that could be used by agencies to

More information

Operation of Site Running StratusLab toolkit v1.0

Operation of Site Running StratusLab toolkit v1.0 Operation of Site Running StratusLab toolkit v1.0 Evangelos Floros, Charles Loomis, Christophe Blanchet, David O Callaghan To cite this version: Evangelos Floros, Charles Loomis, Christophe Blanchet, David

More information

Towards sustainability: An interoperability outline for a Regional ARC based infrastructure in the WLCG and EGEE infrastructures

Towards sustainability: An interoperability outline for a Regional ARC based infrastructure in the WLCG and EGEE infrastructures Journal of Physics: Conference Series Towards sustainability: An interoperability outline for a Regional ARC based infrastructure in the WLCG and EGEE infrastructures To cite this article: L Field et al

More information

Computing activities in Napoli. Dr. Silvio Pardi (INFN-Napoli) Belle II Italian collaboration meeting 21 November 2017 Pisa - Italy

Computing activities in Napoli. Dr. Silvio Pardi (INFN-Napoli) Belle II Italian collaboration meeting 21 November 2017 Pisa - Italy Computing activities in Napoli Dr. Silvio Pardi (INFN-Napoli) Belle II Italian collaboration meeting 21 November 2017 Pisa - Italy Activities in Napoli Grid Services Networking Http/Webdav and Dynamic

More information

Monitoring tools in EGEE

Monitoring tools in EGEE Monitoring tools in EGEE Piotr Nyczyk CERN IT/GD Joint OSG and EGEE Operations Workshop - 3 Abingdon, 27-29 September 2005 www.eu-egee.org Kaleidoscope of monitoring tools Monitoring for operations Covered

More information

Grid Scheduling Architectures with Globus

Grid Scheduling Architectures with Globus Grid Scheduling Architectures with Workshop on Scheduling WS 07 Cetraro, Italy July 28, 2007 Ignacio Martin Llorente Distributed Systems Architecture Group Universidad Complutense de Madrid 1/38 Contents

More information

( PROPOSAL ) THE AGATA GRID COMPUTING MODEL FOR DATA MANAGEMENT AND DATA PROCESSING. version 0.6. July 2010 Revised January 2011

( PROPOSAL ) THE AGATA GRID COMPUTING MODEL FOR DATA MANAGEMENT AND DATA PROCESSING. version 0.6. July 2010 Revised January 2011 ( PROPOSAL ) THE AGATA GRID COMPUTING MODEL FOR DATA MANAGEMENT AND DATA PROCESSING version 0.6 July 2010 Revised January 2011 Mohammed Kaci 1 and Victor Méndez 1 For the AGATA collaboration 1 IFIC Grid

More information

Installation of CMSSW in the Grid DESY Computing Seminar May 17th, 2010 Wolf Behrenhoff, Christoph Wissing

Installation of CMSSW in the Grid DESY Computing Seminar May 17th, 2010 Wolf Behrenhoff, Christoph Wissing Installation of CMSSW in the Grid DESY Computing Seminar May 17th, 2010 Wolf Behrenhoff, Christoph Wissing Wolf Behrenhoff, Christoph Wissing DESY Computing Seminar May 17th, 2010 Page 1 Installation of

More information

Chelonia. a lightweight self-healing distributed storage

Chelonia. a lightweight self-healing distributed storage Chelonia a lightweight self-healing distributed storage Zsombor Nagy (zsombor@niif.hu) Salman Toor (salman.toor@it.uu.se) Jon Kerr Nilsen (j.k.nilsen@fys.uio.no) Motivation How to easily... Create a storage

More information

Scientific data management

Scientific data management Scientific data management Storage and data management components Application database Certificate Certificate Authorised users directory Certificate Certificate Researcher Certificate Policies Information

More information

Ivane Javakhishvili Tbilisi State University High Energy Physics Institute HEPI TSU

Ivane Javakhishvili Tbilisi State University High Energy Physics Institute HEPI TSU Ivane Javakhishvili Tbilisi State University High Energy Physics Institute HEPI TSU Grid cluster at the Institute of High Energy Physics of TSU Authors: Arnold Shakhbatyan Prof. Zurab Modebadze Co-authors:

More information

where the Web was born Experience of Adding New Architectures to the LCG Production Environment

where the Web was born Experience of Adding New Architectures to the LCG Production Environment where the Web was born Experience of Adding New Architectures to the LCG Production Environment Andreas Unterkircher, openlab fellow Sverre Jarp, CTO CERN openlab Industrializing the Grid openlab Workshop

More information

Grid Interoperation and Regional Collaboration

Grid Interoperation and Regional Collaboration Grid Interoperation and Regional Collaboration Eric Yen ASGC Academia Sinica Taiwan 23 Jan. 2006 Dreams of Grid Computing Global collaboration across administrative domains by sharing of people, resources,

More information

Monitoring ARC services with GangliARC

Monitoring ARC services with GangliARC Journal of Physics: Conference Series Monitoring ARC services with GangliARC To cite this article: D Cameron and D Karpenko 2012 J. Phys.: Conf. Ser. 396 032018 View the article online for updates and

More information

R-GMA (Relational Grid Monitoring Architecture) for monitoring applications

R-GMA (Relational Grid Monitoring Architecture) for monitoring applications R-GMA (Relational Grid Monitoring Architecture) for monitoring applications www.eu-egee.org egee EGEE-II INFSO-RI-031688 Acknowledgements Slides are taken/derived from the GILDA team Steve Fisher (RAL,

More information

Operating the Distributed NDGF Tier-1

Operating the Distributed NDGF Tier-1 Operating the Distributed NDGF Tier-1 Michael Grønager Technical Coordinator, NDGF International Symposium on Grid Computing 08 Taipei, April 10th 2008 Talk Outline What is NDGF? Why a distributed Tier-1?

More information

Open Source Software Licence at CERN Recommendations from the OSL Task Force François Fluckiger, Editor 20 April; 2012

Open Source Software Licence at CERN Recommendations from the OSL Task Force François Fluckiger, Editor 20 April; 2012 OSL-2012-01-Short version Open Source Licence - Task force Open Source Software Licence at CERN Recommendations from the OSL Task Force François Fluckiger, Editor 20 April; 2012 Main Volume-Short version

More information

EGI-InSPIRE. EGI Applications Database (TNA3.4) William Vassilis Karageorgos, et al.

EGI-InSPIRE. EGI Applications Database (TNA3.4) William Vassilis Karageorgos, et al. EGI-InSPIRE EGI Applications Database (TNA3.4) William Vassilis Karageorgos, et al. Institute of Accelerating Systems and Applications (IASA) Greek Research & Technology Network (GRNET) April 2011 EGI

More information

Chapter 2 Introduction to the WS-PGRADE/gUSE Science Gateway Framework

Chapter 2 Introduction to the WS-PGRADE/gUSE Science Gateway Framework Chapter 2 Introduction to the WS-PGRADE/gUSE Science Gateway Framework Tibor Gottdank Abstract WS-PGRADE/gUSE is a gateway framework that offers a set of highlevel grid and cloud services by which interoperation

More information

Grid Infrastructure For Collaborative High Performance Scientific Computing

Grid Infrastructure For Collaborative High Performance Scientific Computing Computing For Nation Development, February 08 09, 2008 Bharati Vidyapeeth s Institute of Computer Applications and Management, New Delhi Grid Infrastructure For Collaborative High Performance Scientific

More information

First European Globus Community Forum Meeting

First European Globus Community Forum Meeting First European Globus Community Forum Meeting Florian Zrenner (zrenner@lrz.de) Slides from Dr. Helmut Heller (heller@lrz.de) Leibniz Supercomputing Centre (LRZ), Munich, Germany September 7 th, 2011 1

More information

Edinburgh (ECDF) Update

Edinburgh (ECDF) Update Edinburgh (ECDF) Update Wahid Bhimji On behalf of the ECDF Team HepSysMan,10 th June 2010 Edinburgh Setup Hardware upgrades Progress in last year Current Issues June-10 Hepsysman Wahid Bhimji - ECDF 1

More information

Advanced School in High Performance and GRID Computing November Introduction to Grid computing.

Advanced School in High Performance and GRID Computing November Introduction to Grid computing. 1967-14 Advanced School in High Performance and GRID Computing 3-14 November 2008 Introduction to Grid computing. TAFFONI Giuliano Osservatorio Astronomico di Trieste/INAF Via G.B. Tiepolo 11 34131 Trieste

More information

Architecture Proposal

Architecture Proposal Nordic Testbed for Wide Area Computing and Data Handling NORDUGRID-TECH-1 19/02/2002 Architecture Proposal M.Ellert, A.Konstantinov, B.Kónya, O.Smirnova, A.Wäänänen Introduction The document describes

More information

dcache: challenges and opportunities when growing into new communities Paul Millar on behalf of the dcache team

dcache: challenges and opportunities when growing into new communities Paul Millar on behalf of the dcache team dcache: challenges and opportunities when growing into new Paul Millar communities on behalf of the dcache team EMI is partially funded by the European Commission under Grant Agreement RI-261611 Orientation:

More information

E u r o p e a n G r i d I n i t i a t i v e

E u r o p e a n G r i d I n i t i a t i v e E u r o p e a n G r i d I n i t i a t i v e R E S O U R C E I N F R A S T R U C T U R E P R O V I D ER O P E R A T I O N A L L E V E L A G R E E M E N T Document identifier: EGI-RP-OLA-v0.3clean Date:

More information

Introduction to Grid Infrastructures

Introduction to Grid Infrastructures Introduction to Grid Infrastructures Stefano Cozzini 1 and Alessandro Costantini 2 1 CNR-INFM DEMOCRITOS National Simulation Center, Trieste, Italy 2 Department of Chemistry, Università di Perugia, Perugia,

More information

FREE SCIENTIFIC COMPUTING

FREE SCIENTIFIC COMPUTING Institute of Physics, Belgrade Scientific Computing Laboratory FREE SCIENTIFIC COMPUTING GRID COMPUTING Branimir Acković March 4, 2007 Petnica Science Center Overview 1/2 escience Brief History of UNIX

More information

The Grid Monitor. Usage and installation manual. Oxana Smirnova

The Grid Monitor. Usage and installation manual. Oxana Smirnova NORDUGRID NORDUGRID-MANUAL-5 2/5/2017 The Grid Monitor Usage and installation manual Oxana Smirnova Abstract The LDAP-based ARC Grid Monitor is a Web client tool for the ARC Information System, allowing

More information