A distributed tier-1. International Conference on Computing in High Energy and Nuclear Physics (CHEP 07) IOP Publishing. c 2008 IOP Publishing Ltd 1

Similar documents
Analysis of internal network requirements for the distributed Nordic Tier-1

Operating the Distributed NDGF Tier-1

Interoperating AliEn and ARC for a distributed Tier1 in the Nordic countries.

Lessons Learned in the NorduGrid Federation

Distributing storage of LHC data - in the nordic countries

Future trends in distributed infrastructures the Nordic Tier-1 example

The LHC Computing Grid

Andrea Sciabà CERN, Switzerland

WLCG Transfers Dashboard: a Unified Monitoring Tool for Heterogeneous Data Transfers.

Towards sustainability: An interoperability outline for a Regional ARC based infrastructure in the WLCG and EGEE infrastructures

ARC integration for CMS

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

A Simulation Model for Large Scale Distributed Systems

EGEE and Interoperation

The LHC Computing Grid

GRIDS INTRODUCTION TO GRID INFRASTRUCTURES. Fabrizio Gagliardi

Overview of ATLAS PanDA Workload Management

Scientific data management

Status of KISTI Tier2 Center for ALICE

The evolving role of Tier2s in ATLAS with the new Computing and Data Distribution model

PoS(EGICF12-EMITC2)081

IEPSAS-Kosice: experiences in running LCG site

Travelling securely on the Grid to the origin of the Universe

CernVM-FS beyond LHC computing

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

Large scale commissioning and operational experience with tier-2 to tier-2 data transfer links in CMS

Introduction to Grid Infrastructures

HEP replica management

Computing / The DESY Grid Center

The INFN Tier1. 1. INFN-CNAF, Italy

DIRAC pilot framework and the DIRAC Workload Management System

Grid Computing a new tool for science

The Grid: Processing the Data from the World s Largest Scientific Machine

Monitoring the Usage of the ZEUS Analysis Grid

The SweGrid Accounting System

ADAPTIVE AND DYNAMIC LOAD BALANCING METHODOLOGIES FOR DISTRIBUTED ENVIRONMENT

On the EGI Operational Level Agreement Framework

Outline. Infrastructure and operations architecture. Operations. Services Monitoring and management tools

Grid Scheduling Architectures with Globus

Data preservation for the HERA experiments at DESY using dcache technology

PROJECT FINAL REPORT. Tel: Fax:

Monitoring ARC services with GangliARC

Chapter 3. Design of Grid Scheduler. 3.1 Introduction

Summary of the LHC Computing Review

Reliability Engineering Analysis of ATLAS Data Reprocessing Campaigns

Introduction to Geant4

ISTITUTO NAZIONALE DI FISICA NUCLEARE


Grids and Security. Ian Neilson Grid Deployment Group CERN. TF-CSIRT London 27 Jan

First Experience with LCG. Board of Sponsors 3 rd April 2009

Monitoring System for the GRID Monte Carlo Mass Production in the H1 Experiment at DESY

AGIS: The ATLAS Grid Information System

Grid Computing Activities at KIT

Data Management for the World s Largest Machine

Usage statistics and usage patterns on the NorduGrid: Analyzing the logging information collected on one of the largest production Grids of the world

GStat 2.0: Grid Information System Status Monitoring

New strategies of the LHC experiments to meet the computing requirements of the HL-LHC era

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

glite Grid Services Overview

Security Correlation Server System Deployment and Planning Guide

The glite File Transfer Service

Grid Architectural Models

Grid Computing. MCSN - N. Tonellotto - Distributed Enabling Platforms

The LCG 3D Project. Maria Girone, CERN. The 23rd Open Grid Forum - OGF23 4th June 2008, Barcelona. CERN IT Department CH-1211 Genève 23 Switzerland

PoS(EGICF12-EMITC2)106

ATLAS NorduGrid related activities

An Analysis of Storage Interface Usages at a Large, MultiExperiment Tier 1

THE GLOBUS PROJECT. White Paper. GridFTP. Universal Data Transfer for the Grid

The Grid. Processing the Data from the World s Largest Scientific Machine II Brazilian LHC Computing Workshop

Virtualizing a Batch. University Grid Center

The creation of a Tier-1 Data Center for the ALICE experiment in the UNAM. Lukas Nellen ICN-UNAM

Online data storage service strategy for the CERN computer Centre G. Cancio, D. Duellmann, M. Lamanna, A. Pace CERN, Geneva, Switzerland

National R&E Networks: Engines for innovation in research

ACCI Recommendations on Long Term Cyberinfrastructure Issues: Building Future Development

Evolution of the ATLAS PanDA Workload Management System for Exascale Computational Science

N. Marusov, I. Semenov

Uptime and Proactive Support Services

ARC-XWCH bridge: Running ARC jobs on the XtremWeb-CH volunteer

UW-ATLAS Experiences with Condor

DESY. Andreas Gellrich DESY DESY,

Optimizing Parallel Access to the BaBar Database System Using CORBA Servers

30 Nov Dec Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy

ATLAS operations in the GridKa T1/T2 Cloud

Grid Data Management

Introduction to Grid Computing

The ATLAS Tier-3 in Geneva and the Trigger Development Facility

RUSSIAN DATA INTENSIVE GRID (RDIG): CURRENT STATUS AND PERSPECTIVES TOWARD NATIONAL GRID INITIATIVE

On the employment of LCG GRID middleware

Storage and I/O requirements of the LHC experiments

Grid Computing with Voyager

Access the power of Grid with Eclipse

Batch Services at CERN: Status and Future Evolution

ALICE Grid Activities in US

Grid Security Policy

Experiences with the new ATLAS Distributed Data Management System

Pan-European Grid einfrastructure for LHC Experiments at CERN - SCL's Activities in EGEE

Storage Virtualization. Eric Yen Academia Sinica Grid Computing Centre (ASGC) Taiwan

High Performance Computing Course Notes Grid Computing I

Popularity Prediction Tool for ATLAS Distributed Data Management

CHIPP Phoenix Cluster Inauguration

Transcription:

A distributed tier-1 L Fischer 1, M Grønager 1, J Kleist 2 and O Smirnova 3 1 NDGF - Nordic DataGrid Facilty, Kastruplundgade 22(1), DK-2770 Kastrup 2 NDGF and Aalborg University, Department of Computer Science, Selma Lagerlöfsvej 300, DK 9220 Aalborg SØ 3 NDGF and Div. of Experimental High Energy Physics, Institute of Physics, Lund University, Box 118, 22100 Lund, Sweden E-mail: gronager@ndgf.org, lars@nordu.net, kleist@ndgf.org, oxana.smirnova@ndgf.org Abstract. The Tier-1 facility operated by the Nordic DataGrid Facility (NDGF) differs significantly from other Tier-1s in several aspects: firstly, it is not located at one or a few premises, but instead is distributed throughout the Nordic countries; secondly, it is not under the governance of a single organization but instead is a meta-center built of resources under the control of a number of different national organizations. We present some technical implications of these aspects as well as the high-level design of this distributed Tier-1. The focus will be on computing services, storage and monitoring. 1. Introduction This paper describes the Nordic DataGrid Facility (NDGF) [1], a distributed super-computing site spanning four Nordic countries (Denmark, Norway, Sweden and Finland). The paper addresses two goals: to describe the design and operations of the NDGF super-center, and to describe in greater detail how the services specific to the Nordic Tier-1 are integrated into the distributed computing and storage facility. The Nordic countries are each quite small (0.3 to 9 million people), but together they represent a population of 25 million and rank as one of the richest population groups in the world. In order to participate in major international research projects it is beneficial to pool individual Nordic resources, providing a common administrative and management entry point. The past few decades have seen a significant and continuous growth in data from scientific experiments. Today scientific experiments produce vast amounts of data, and storing, accessing, and processing these data is increasingly becoming a challenge. Yet, doing so is a requirement for scientific discoveries. Scientific computing centers today address the need for assistance in optimizing algorithms and software. With the data explosion arises a similar need for optimizing access to databases, and for handling storage and manipulation of databases stored in several locations. The NDGF meta-center connecting several existing scientific computing facilities addresses the following needs: to enable single sign-on access to multiple computing and storage resources distributed over wide area in general, and to enable a pan-nordic storage system in particular. Thus, NDGF strives to fulfil its primary purpose of creating a joint Nordic production grid facility, leveraging existing scientific computing resources and grid facilities in the four c 2008 IOP Publishing Ltd 1

participating countries. It is important to stress that the purpose of NDGF is to foster operational collaboration and joint policy for existing and future resources owned by national e-infrastructures; NDGF does not establish or own hardware resources themselves. In addition to establishing an infrastructure, NDGF has a number of other tasks: to act as a single point of contact for e-science in the Nordic countries; to maximize visibility and impact of Nordic grid efforts outside the Nordic countries, and to represent the community in international efforts; to host major e-science communities and projects, providing resources for coordination and project management; to develop and support production-quality middleware for specific requirements or projects; to support and guide the development efforts of the NorduGrid community. The main customers of NDGF are Nordic research communities with a need for massive computational and data storage resources in conjunction with collaborative work. One such example is the High Energy Physics (HEP) community, specifically the ALICE [2], ATLAS [3], and, to a lesser extent, CMS [4] groups 1. While the HEP community is the first and probably the largest community to request such services, NDGF is also working with other research communities with similar needs, such as environmental science and bioinformatics. By the virtue of NDGF being a Grid service provider, these communities are treated as Virtual Organizations (VOs). Each VO establishes formal relationships with NDGF. This paper will describe one NDGF use case: the Nordic WLCG [5] Tier-1. The paper first presents the technical setup and the services required for a Tier-1; this is followed by a presentation of how operations are handled in this distributed setup. Section 4 describes how NDGF and the NDGF Tier-1 interact with other international projects, and finally Section 5 summarizes and looks ahead on how the successful creation of a distributed Tier-1 can be used as an inspiration for other projects and communities. 2. Services and technical setup The core services needed to form a WLCG Tier-1 are a computing service, a storage service and an information service. There can be multiple computing and storage services per Tier-1. In NDGF, these services are distributed across the different scientific computing centers. Figure 1 illustrates the service distribution as of September 2007. For such a distributed site spanning multiple organizational domains, core services inevitably have to constitute a grid infrastructure [6] by itself. NDGF allows organizations that contribute resources to have their own policies with respect to resource configuration and setup. This results in a highly heterogeneous set of resources: operating systems, batch schedulers, storage systems, network connectivity, policies are all different and highly non-uniform. For this reason, the grid middleware of choice for NDGF sites is NorduGrid s [7] Advanced Resource Connector (ARC) [8], which provides a portable grid solution with a minimal footprint and yet very high performance. ARC facilitates computing services and parts of data management which are immediately associated with computational tasks, such as data stage-in and stage-out, indexing and caching. This optimizes usage of distributed resources in case of data-intensive tasks typical for NDGF user communities. One of the main tasks of NDGF as an infrastructure is to ensure that ARC-enabled sites are equipped with software packages required by different customers. In order to reach this goal, NDGF works on integrating different VO-specific tools and services with ARC and contributes to the development of the ARC middleware itself. 1 There is currently only one CMS group in Nordic countries 2

International Conference on Computing in High Energy and Nuclear Physics (CHEP 07) IOP Publishing Figure 1. Geographical distribution of NDGF services, September 2007 2.1. Detailed service overview of the Tier-1 as of Q3-2007 The computing services of NDGF need to be exposed to the production systems of the supported VOs, particularly ALICE, ATLAS and CMS2. Every production system needs a specific approach, as described later in this section, in the part dedicated to VO-specific services. In case of all the currently supported VOs, the underlying storage service of NDGF relies on dcache [9] for low-level management of data stored across different nodes. Higher-level services are deployed and adapted to the needs of the respective VOs whenever necessary. Common storage solution for all experiments dcache [9] provides most of the low-level services needed to satisfy the HEP VOs requirements for persistent data storage. From the point of view of external interfaces, it offers a single SRM [10] end-point with support for large amounts of both disk and tape, VOMS [11] based authentication and authorization, a number of different transport protocols etc. This setup is illustrated in Figure 2. However, in order to support the distributed nature of of the Nordic Tier-1 (storage is not located at a single physical site, but distributed around the Nordic countries), NDGF is contributing to dcache with a number of improvements to decrease the overhead incurred by distribution. In a companion paper [12] we describe the technical problems and how do we address them. This setup also has a number of administrative and policy-related consequences. There are many administrative domains involved, rules and regulations from more than one country apply and resource providers are concerned about losing control of their resources to an external entity. 2 NDGF provides only Tier-2 services to CMS, still via ARC middleware 3

Figure 2. NDGF Distributed storage setup, based on dcache. Dealing with these issues requires both technical and administrative solutions. On the technical side, support in dcache for delegating control over resources is planned to be put in place, effectively allowing resource providers to manage their own resources directly via the central dcache components. Administratively, the Service Level Agreements and Memoranda of Understanding between NDGF and resource providers, and between CERN [13] (as the WLCG host organisation and Tier-0) and NDGF will clarify under which rules and regulations operation will take place. Other common services A set of stand-alone services required by WLCG is enabled by NDGF. A standard File Transfer Service (FTS) of glite [14] server is installed to facilitate scheduling of data transfers from the Nordic Tier-1 to associated Tier-2 sites and to other Tier-1s. Furthermore, a Distributed Deployment of Databases (3D) service of WLCG is being installed for those experiments that need this service (ATLAS at the moment). Computing resource usage accounting relies on the SGAS [15] system developed by Swegrid [16], which is not a WLCG standard, but offers the needed functionality and is integrated with ARC. SGAS as it is provides most, if not all, of the features needed by NDGF to do accounting of computational usage, such as: authorization via policy management and policy enforcement; decentralized administration of user accounts; each VO has its own associated bank component; support for NorduGrid ARC; 4

interface to WLCG s APEL accounting service. Accounting of storage utilization is an integral part of dcache, and scripts developed for integration with CERN accounting systems already developed at other Tier-1 sites are expected to be usable by NDGF also. Availability calculations through the Service Availability Monitoring system (SAM) was achieved by developing SAM sensors for ARC. ATLAS-specific services ATLAS production and analysis systems submit computational jobs directly to ARC, which simplifies greatly NDGF task of supporting ATLAS VO. However, in addition to installing the above mentioned services, some integration with the ATLAS Distributed Data Management system (DDM) was needed. Internally, ARC uses Globus RLS [17] for data indexing. ATLAS DDM however did not until recently support RLS, which meant that files produced at NDGF were not easily visible from other sites that did not support RLS. To resolve this issue, output files were also registered in a catalogue supported by DDM; as this lead to inconsistencies, ATLAS DDM is being modified by NDGF to support RLS directly. Similarly, the ATLAS VObox service is modified such that data transferred from outside to the NDGF Tier-1 are also registered in RLS. In [18] we describe this setup in more depth. ALICE-specific services Currently, sites running ALICE computations install either a standard ALICE job submission front-end (VObox), or a modified VObox, which instead of submitting jobs directly to LRMS submits jobs to NorduGrid ARC. Investigations are being carried out on whether or not it will be possible to use a single ALICE VObox to cover all the NDGF sites, which may reduce the burden of maintaining both ARC and ALICE front-ends at each site. Data storage for ALICE is provided by through the standard xrootd support in dcache. CMS-specific services It is yet to be investigated what kinds of specific services (if any) are needed to enable CMS production via ARC. 2.2. Other services NDGF also maintains a number of other services not directly related to the LHC experiments. A VOMS server is being setup to allow easy creation and maintenance of VOs. High-level monitoring of the NDGF sites is done using a separate GIIS for the NDGF sites, together with the ARC web-based grid monitor. Finally, a logging service is set up to allow sites to log information about grid jobs. 3. Operation The operation of the NDGF Tier-1 is similar to a single-site computer center. However, as the Tier-1 spans several organizational units at sites distributed over a large geographical area 3, an internal communications process with frequent electronic exchanges and regular face-to-face meetings is required to ensure coherence of the operations team. 3.1. New e-science Projects Figure 3 illustrates the organization of the Tier-1. The Tier-1 for NDGF is synonymous to the Nordic CERN project, and is treated similar to all other e-science projects hosted by NDGF. The CERN project is represented in NDGF by a CERN Advisory Committee, consisting of grantholders of the resources to be shared. The grant-holders apply for resources to their respective 3 The north-south span of NDGF is about 1300km 5

Figure 3. The operational structure of the CERN VO in NDGF National Resource Allocation entities, and the resources are installed at the e-infrastructure centers. The operation of the resources is directed by the NDGF, using Node Coordinators to ensure that the resources are configured according to the requirements of the user communities. In particular, the NDGF CERN Coordinator is responsible for collecting and aggregating the requirements of the CERN user community, with support from the NDGF team and the CERN Advisory Committee. NDGF also has resources (a dedicated developers team) to conduct middleware development, if it is needed to meet additional infrastructure functionality requirements. 3.2. The NDGF Tier-1 The operation of the NDGF Tier-1 resembles the operation of a single-site Tier-1. However, as the NDGF Tier-1 spans several sites in four countries, the operational structure has additional layers and requires an additional set of Service Agreements. As previously indicated, the NDGF Tier-1 does not possess resources, but rather utilizes the resources of the grant-holders to create a Tier-1. The exception is a few smaller installations, such as monitoring and entry-access nodes that are ran centrally at NDGF premises. Figure 4 outlines the agreement structure put in place to support this: Between NDGF and the national e-infrastructure centers there exist a set of MoUs/SLAs that essentially describe that NDGF is used for resource sharing, and that NDGF will support the centers in doing so via providing the National Node Coordinators. 6

Between the National Resource Allocation entities, the funding authorities, and CERN, the WLCG MoU is signed, similarly to other Tier-1s. This MoU obliges the funding authorities to do their best to provide the required funding level to assure the pledged resources for running the Tier-1 can be acquired. NDGF supports this structure and facilitates the creation and maintenance of the agreements, helping the communities to create specifications for optimal operations and facilitating the agreements with the computing centers. An example of such a specification could be that since a resource will be used in an international project, users other than the grant-holder must be allowed to use the resource, and even that the allowance and administration of the user group will be delegated to an international entity in this case the ATLAS, ALICE and CMS VO administration. Other requirements may be of an operational nature, i.e., procedures for trouble ticket handling, service windows, etc. Figure 4. The agreements involved in running the NDGF Tier-1 The operation of a Tier-1 center is directed by the need for storing and processing data from the LHC experiments at CERN. The first data will flow in early 2008 as smaller bursts, and regular data flows will begin during late 2008. The expected data rate to NDGF will be 50-100 MByte/second for about 100 days of LHC operation per year. The data from CERN consists of raw detector data together with data processed into events based on a current calibration data. As the measuring devices are understood better, new calibration data will be created and hence all raw data will need to be reprocessed. 7

Besides data processing from experiements, the Tier-1 will also process simulation of events to predict the physics behind the observed signal. These calculations will be done at Tier-1s as well as on Tier-2s. Hence, the primary Tier-1 tasks can be divided into: recording raw data from CERN and storing them on tape (critical); recording processed data from CERN and storing them on disk (less critical); providing data to other Tier-1 / Tier-2; reprocessing raw data; event simulation calculations. The recording of raw data from CERN is highly critical as the data buffer at CERN can only store 12 hours of data and thus failing to record data for one hour means doubling the transfers for the next hour. Every piece of data from CERN is stored at CERN and at two Tier-1s, for redundancy. This means that if data is lost at a Tier-1, it can be re-staged from another Tier-1, but at a cost of bandwidth use at both the sending and the receiving Tier-1. For this reason, data safety is important. It should be observed that the properties of the system are such that in the event of data loss, it is easier to recover from several instances of loss of small amounts of data than from a single loss of a large amount of data. To ensure a high degree of service stability, in particular for data recording, a distributed storage setup is deployed. The setup is based on dcache with a small central component located at the network ingress point of the LHC OPN. From here, data is distributed to to the participating sites. This system is described in detail in [12]. If one NDGF site is down, another site will be chosen automatically, thereby maximizing uptime for the overall system. The NDGF distributed meta-center is designed to sustain operations even in the event of loss of an entire center. Providing data to other Tier-1s as well as to associated Tier-2s is done using the same distributed storage system. The overhead of having other Tier-2s connected is mainly related to the usage of the public research network, NORDUnet. Data reprocessing and Monte Carlo simulation are initiated by experiment production managers and are run on resources committed to the NDGF Tier-1. The actual production is conducted by production managers from each VO. The production managers have higher priority on the resources. The share of compute power is 80% for production and 20% for general users. General users should be members of one of the supported CERN VOs (ALICE, ATLAS and later also CMS). Production Managers are typical user representatives from the different VOs embedded in one of the Nordic research groups. 3.3. Operational Procedures The operational procedures in the NDGF infrastructure is illustrated in Figure 5. All problems observed or reported go to either the operator on duty (during office hours) or to the NDGF 24x7 first-line support hosted at the NU-NOC (shared with NORDUnet). Outside contact can be made either via a phone or e-mail, depending on the severity of the problem. In addition, problems can be detected through use of the internal monitoring systems If the problem can be solved immediately by the first line support by following standard procedures for bringing service back to normal, a problem report is sent to the operator on duty for later problem tracing and is added to the problem log. If the problem cannot be solved immediately, or it occurs outside office hours, it is up to the operator on duty to take appropriate action on the problem, e.g., forwarding it to the involved site or application expert or by taking direct action. 8

Figure 5. The operational flow in the NDGF Tier-1 The operator on duty (OoD) responsibility rotates on a weekly basis in the NDGF Node Coordinator team. Every Monday both the previous weeks OoD and the new OoD participate in the WLCG Weekly Operation meeting. In addition, and OoD handover is performed. If a problem cannot be fixed by the Node Coordinator team, it may be escalated to the NDGF Technical Coordinator and from there either through management or to the NDGF software development team. 4. Interaction with other projects One of the main driving forces behind the Nordic Data Grid Facility is the desire to make Nordic computational resources available to a wider international community in a coordinated and uniform manner. Nordic academic researchers are involved in a great many collaborative efforts worldwide, and today such collaborations imply shared utilization of all resources available to the effort, irrespective of geography. The sharing of resources is facilitated by modern high-capacity network infrastructure. Resources shared in this manner include area-specific facilities, instruments, devices, and materials, allowing, e.g., international research groups joint user of shared instruments. Increasingly, also computing power, storage resources and electronic databases are shared this way. Typically, members of an international project each contribute computing and storage resources to common pool for the project. From this pool, all project members are granted a uniform access to all available assets, independently of affiliation. Grid technologies have become the standard way of harnessing heterogeneous collaborative resources into consistent 9

distributed facilities. NDGF serves as the gateway for Nordic contributions to such international efforts for sharing of computing and storage resources. NDGF acts as a single point of entry point to such initiatives, providing the pledged services by consolidating available Nordic resources. NDGF can participate as a partner in international projects, can assist Nordic research communities establish collaborations, and can inject the additional resources required to create synergy. In order for NDGF to serve as gateway it is essential that NDGF take active part in collaborations such as EGEE, EGI and OGF, and at the more technical level contribute to development of middleware products, such as NorduGrid s ARC, dcache, and the Globus [19] toolkit. In EGEE, NDGF is becoming a Core Infrastructure Center (CIC). NDGF resources are monitored through EGEE-developed tools such as e.g. Service Availability Monitoring (SAM). Consequently, all NDGF storage can be integrated into the EGEE European Grid Infrastructure. Currently, NDGF and EGEE are developing interfaces that will make NDGF computing resources visible through a single gateway thus making NDGF appear as a single computing facility in EGEE. Furthermore, NDGF will take part in the EGEE CIC-on-Duty rotation. 5. Conclusion We have in this paper presented the technical, organizational and operational aspects of the NDGF distributed Tier-1. It became operational in spring 2007, with the additions of monitoring and accounting over the summer of 2007. Today it is a fully functional Tier-1 and hence stands as a proof of the idea of a distributed computing and storage service spanning multiple organizational domains. The computing service is based not on a normal batch system, but on a grid. The storage service is accessible through a single point of entry, even though the actual storage pools are still hosted, ran and serviced by the local computing centers. As the individual Nordic computing centers cannot provide service levels as high as needed for a Tier-1, redundancy is a core part of the NDGF design, essentially aiming to leave no single point of failure. Operations and development are also done in a distributed fashion, using manpower available at the local computing centers, yet it does not affect the actual stability and responsiveness of the Tier-1 as seen from the outside. To conclude, we see the design a success and we anticipate that in the future other e-science projects will benefit from this approach. Most notably, this experience must assist in establishing a Core Infrastructure Center in the European Grid Infrastructure of the EGEE project. It will be useful also for other Nordic and European e-science projects that need large amounts of computing resources and aim to raise them through sharing and collaboration between multiple e-infrastructure centers. References [1] The Nordic DataGrid Facility http://www.ndgf.org [2] ALICE experment at CERN http://aliceinfo.cern.ch/ [3] ATLAS experiment at CERN http://cern.ch/atlas [4] CMS experiment at CERN http://cms.cern.ch/ [5] Worldwide LHC Comuting Grid http://cern.ch/lcg [6] Foster I and Kesselman C 1999 The Grid: Blueprint for a New Computing Infrastructure (Morgan Kaufmann) [7] NorduGrid Collaboration http://www.nordugrid.org [8] Ellert M et al. 2007 Future Generation Computer Systems 23 219 240 [9] dcache http://www.dcache.org [10] Storage Resource Management Working Group http://sdm.lbl.gov/srm-wg/ [11] Alfieri R et al. 2005 Future Gener. Comput. Syst. 21 549 558 ISSN 0167-739X [12] Behrmann G, Fuhrmann P and Kleist J 2007 A Distributed Storage System with dcache proceedings of CHEP-07 [13] CERN, European Centre for Nuclear Research http://www.cern.ch/ 10

[14] glite, Lightweight Middleware for Grid Computing http://glite.web.cern.ch/glite/ [15] Sgas http://www.sgas.se [16] Swegrid project http://www.swegrid.se/ [17] Chervenak A L et al. 2004 Performance and Scalability of a Replica Location Service (IEEE Computer Society Press) pp 182 191 ISBN 0-7803-2175-4 [18] Behrmann G, Cameron D, Ellert M, Kleist J and Taga A 2007 ATLAS DDM integration in ARC proceedings of CHEP-07 [19] Foster I and Kesselman C 1997 International Journal of Supercomputer Applications 11 115 128 11