Experiences in testing a Grid service in a production environment
|
|
- Silvia White
- 5 years ago
- Views:
Transcription
1 EWDC th European Workshop on Dependable Computing Toulouse, France, May, 2009 Experiences in testing a Grid service in a production environment Flavia Donno CERN, European Organization for Nuclear Research Flavia.Donno@cern.ch Andrea Domenici DIIEIT, Università di Pisa Andrea.Domenici@iet.unipi.it
2 The Worldwide LHC Computing Grid Grid sites in the world (Courtesy of CERN). The Worldwide LHC Computing Grid (WLCG) is one of the largest Grid infrastructures dedicated to high-performance scientific computation, with more than 200 sites all over the world. F. Donno, A. Domenici EWDC09 2/21
3 Mass Storage Systems A robotic tape library. The Grid uses heterogenous Mass Storage Systems (MSS), based on different technologies and with different capabilities and interfaces. F. Donno, A. Domenici EWDC09 3/21
4 Storage Elements A Storage Element (SE) is a Grid Service that provides: A mass storage system. A GridFTP service to provide data transfer in and out of the SE to and from the Grid. Local POSIX-like input/output calls providing application access to the data on the SE. Authentication, authorization and audit/accounting facilities. F. Donno, A. Domenici EWDC09 4/21
5 The Storage Resource Manager The Storage Resource Manager (SRM) is the common interface of the Storage Elements. The SRM specification defines many service requests: Space management functions allow the client to reserve, allocate, release, and manage storage spaces, their types and lifetimes. Data transfer functions have the purpose of getting files into SRM spaces either from the client s space or from other remote storage systems on the Grid, and to retrieve them. Other function classes are Directory, Permission, and Discovery functions. F. Donno, A. Domenici EWDC09 5/21
6 Testing for SRM compliance The goals of testing are: Validating the SRM interface and protocol specification for adherence to the explicit and implicit user requirements, and against inconsistency, incompleteness, or inefficiency; validating the SRM implementations for compliance with the specification; checking the SRM implementations for performance and reliability. Difficulties arise from: Large and complex set of service requests, informal specification, number of different implementations, and number of sites. F. Donno, A. Domenici EWDC09 6/21
7 A typical SRM request srmreservespace Input parameters: TRetentionPolicyInfo unsigned long string unsigned long int retentionpolicyinfo desiredsizeofguaranteedspace authorizationid desiredsizeoftotalspace desiredlifetimeofreservedspace TTransferParameters transferparameters... more optional parameters Output parameters: TReturnStatus returnstatus string requesttoken... more optional parameters F. Donno, A. Domenici EWDC09 7/21
8 Space properties The retentionpolicyinfo parameter specifies two properties of the requested space: Retention policy, likelyhood of file loss: REPLICA, OUTPUT, CUSTODIAL. Access latency, readiness of file access: ONLINE (e.g., disk), NEARLINE (e.g., tape). A storage class is a combination of retention policy and access latency. In the WLCG, the following storage classes are supported: Tape0Disk1: Replica, Online; Tape1Disk1: Custodial, Online; Tape1Disk0: Custodial, Nearline. F. Donno, A. Domenici EWDC09 8/21
9 A large test space The srmreservespace request has nine input arguments. Some arguments range over a finite set of values. Other arguments range over theoretically infinite sets of values. Equivalence partitioning enables us to reduce the number of values to consider... but we are still left with some test cases. And then we have the other 38 requests! F. Donno, A. Domenici EWDC09 9/21
10 Use-case analysis We may shrink the test space by pruning argument values and combinations that may be ruled out based on the actual operating conditions in the WLCG. The SRM specification is very general and flexible: many negotiations are possible; much leeway for implementation or site dependent defaults; allowance for future requirements. The full power of the SRM is currently not used by the implementations... yet they are SRM-compliant. A careful analysis of usage patterns and implementation constraints enables us to significantly reduce the size of the test space. This requires a close interaction between testers, users, and developers. F. Donno, A. Domenici EWDC09 10/21
11 Reshaping the signature We prune the domain of an argument and eliminate some altogether: retentionpolicyinfo: only a few of the possible values are in use. authorizationid: unused, as in the WLCG credentials are not passed as parameters (certificates are used instead). transferparameters: unused, as site dependent defaults are used. Other parameters (not shown) are similarly ignored. However, we consider the validity or absence of a user certificate as an extra argument. We can then test the request with only five variable arguments, thus reducing the test space size to about 200 cases. F. Donno, A. Domenici EWDC09 11/21
12 Modeling constraints and conditions (1) Cause-effect graphing is used to derive test cases covering constraints and operating conditions, e.g.: Causes: 1 retentionpolicyinfo is not NULL 2 retentionpolicyinfo is supported by server requesttoken is returned [11 and 12 mutually exclusive] 12 spacetoken is returned [12 requires 13] 13 sizeofguaranteedreservedspace and lifetimeofreservedspace are returned Effects: sizeofguaranteedreservedspace = default 95 lifetimeofreservedspace = default 96 transferparameters is ignored F. Donno, A. Domenici EWDC09 12/21
13 Modeling constraints and conditions (2) 1 ~ ~ ~ ~ ~ ~ ~ ~ E 12 R 13 ~ ~ ~ Cause-effect graph for the srmreservespace request F. Donno, A. Domenici EWDC09 13/21
14 Error guessing Error guessing = pragmatic knowledge + formalization. Example: formalization of behavior by state machines led to discover unexpected interactions. SURL_Assigned Busy PutDone [retention = CUSTODIAL] PutDone [retention <> CUSTODIAL] BringOnline PrepareToPut [overwrite] ChangeSpaceForFiles ChangeSpaceForFiles Nearline BringOnline Online ReleaseFiles PrepareToGet ReleaseFiles [retention <> CUSTODIAL] AbortRequest Readable AbortFiles ChangeSpaceForFiles PrepareToGet NearlineOnline AbortFiles AbortRequest Partial state machine for a file. F. Donno, A. Domenici EWDC09 14/21
15 Test case families Five families of test cases have been designed: Availability to check the availability in time of the SRM service end-points. Basic to verify basic functionality of the implemented SRM APIs. Use Cases to check boundary conditions, use cases derived by real usage, function interactions, exceptions, etc. Exhaustion to check extreme values and properties of input and output arguments such as length of filenames, URL format, etc. Stress tests to stress the systems, identify race conditions, study the behavior of the system when critical concurrent operations are performed, etc. F. Donno, A. Domenici EWDC09 15/21
16 The SRM testbed The following SRM implementations are being tested: CASTOR developed at CERN, uses tape libraries with disk servers as front-end caches. SRM 2.2 implementation developed at RAL (UK) (4 Tier-1 sites). dcache developed at DESY (Germany), uses multiple MSS backends, both custom and proprietary. SRM 2.2 implementation developed at FNAL (USA) (7 Tier-1 sites). DPM developed at CERN, a disk-only MSS. SRM 2.2 implementation developed at CERN (6 Tier-2 sites). DRM/BeStMan is the LBNL (USA) disk-based storage system. LBNL has been the first promoter of SRM, and this storage system was the first prototype on which SRM has been tested (1 Tier-2 sites). StoRM developed at CNAF (Italy), uses parallel file systems such as GPFS or PVFS. (4 Tier-2 sites). F. Donno, A. Domenici EWDC09 16/21
17 Test execution and analysis Execution framework based on S2 and shell scripts. invoke SRM requests; make checks on return codes; define complex test actions. Automatic execution and result logging six times a day. Monthly plots for each test family. F. Donno, A. Domenici EWDC09 17/21
18 Pre-production testing Basic Tests Jan Mar CASTORCERN DCACHEFNAL DPMCERN DRMLBNL STORM 0.8 No. of failures/no. of tests /11/06 11/25/06 12/09/06 12/23/06 01/06/07 01/20/07 02/03/07 02/17/07 03/03/07 03/17/07 Date F. Donno, A. Domenici EWDC09 18/21
19 In-production testing Deployment Basic Tests May Feb 2009 No. of failures/no. of tests CASTORALICE CASTORASGC CASTORATLAS CASTORCERN CASTORCMS CASTORCNAF CASTORDTEAM CASTORLHCB CASTORPPS DCACHEBNLPR DCACHEDESY DCACHEFZKPR DCACHEIN2P3PR DCACHENDGFPR DCACHEPICPR DCACHESARAPR DCACHESTRESS DCACHETRIUMFPR DCACHEUCSD DPMCERN DPMLAL DPMNIKHEF DRMLBNL /01/08 07/01/08 08/01/08 09/01/08 10/01/08 11/01/08 12/01/08 01/01/09 02/01/09 Date F. Donno, A. Domenici EWDC09 19/21
20 Conclusions A complex Grid service such as the SRM poses a challenge to testers. Standard testing techniques are fundamental... but cannot be applied mechanically. Testers, users, and developers cannot live on different planets. The development of a (semi)formal model has helped design a few families of tests. The testing campaign itself has motivated the developers to reconsider many of the initial assumptions and decisions, leading to solutions that seem to better satisfy the needs of the users. F. Donno, A. Domenici EWDC09 20/21
21 Thank you Merci F. Donno, A. Domenici EWDC09 21/21
Introduction to SRM. Riccardo Zappi 1
Introduction to SRM Grid Storage Resource Manager Riccardo Zappi 1 1 INFN-CNAF, National Center of INFN (National Institute for Nuclear Physic) for Research and Development into the field of Information
More informationUnderstanding StoRM: from introduction to internals
Understanding StoRM: from introduction to internals 13 November 2007 Outline Storage Resource Manager The StoRM service StoRM components and internals Deployment configuration Authorization and ACLs Conclusions.
More informationI Tier-3 di CMS-Italia: stato e prospettive. Hassen Riahi Claudio Grandi Workshop CCR GRID 2011
I Tier-3 di CMS-Italia: stato e prospettive Claudio Grandi Workshop CCR GRID 2011 Outline INFN Perugia Tier-3 R&D Computing centre: activities, storage and batch system CMS services: bottlenecks and workarounds
More informationglite Grid Services Overview
The EPIKH Project (Exchange Programme to advance e-infrastructure Know-How) glite Grid Services Overview Antonio Calanducci INFN Catania Joint GISELA/EPIKH School for Grid Site Administrators Valparaiso,
More informationand the GridKa mass storage system Jos van Wezel / GridKa
and the GridKa mass storage system / GridKa [Tape TSM] staging server 2 Introduction Grid storage and storage middleware dcache h and TSS TSS internals Conclusion and further work 3 FZK/GridKa The GridKa
More informationPromoting Open Standards for Digital Repository. case study examples and challenges
Promoting Open Standards for Digital Repository Infrastructures: case study examples and challenges Flavia Donno CERN P. Fuhrmann, DESY, E. Ronchieri, INFN-CNAF OGF-Europe Community Outreach Seminar Digital
More informationThe CMS experiment workflows on StoRM based storage at Tier-1 and Tier-2 centers
Journal of Physics: Conference Series The CMS experiment workflows on StoRM based storage at Tier-1 and Tier-2 centers To cite this article: D Bonacorsi et al 2010 J. Phys.: Conf. Ser. 219 072027 View
More informationStorage Resource Manager Version 2.2: design, implementation, and testing experience
Storage Resource Manager Version 2.2: design, implementation, and testing experience Flavia Donno (1) Co-authors: Lana Abadie (1),Paolo Badino (1), Jean-Philippe Baud (1), Ezio Corso (2), Shaun De Witt
More informationIntroduction Data Management Jan Just Keijser Nikhef Grid Tutorial, November 2008
Introduction Data Management Jan Just Keijser Nikhef Grid Tutorial, 13-14 November 2008 Outline Introduction SRM Storage Elements in glite LCG File Catalog (LFC) Information System Grid Tutorial, 13-14
More informationThe INFN Tier1. 1. INFN-CNAF, Italy
IV WORKSHOP ITALIANO SULLA FISICA DI ATLAS E CMS BOLOGNA, 23-25/11/2006 The INFN Tier1 L. dell Agnello 1), D. Bonacorsi 1), A. Chierici 1), M. Donatelli 1), A. Italiano 1), G. Lo Re 1), B. Martelli 1),
More informationScientific data processing at global scale The LHC Computing Grid. fabio hernandez
Scientific data processing at global scale The LHC Computing Grid Chengdu (China), July 5th 2011 Who I am 2 Computing science background Working in the field of computing for high-energy physics since
More informationSPINOSO Vincenzo. Optimization of the job submission and data access in a LHC Tier2
EGI User Forum Vilnius, 11-14 April 2011 SPINOSO Vincenzo Optimization of the job submission and data access in a LHC Tier2 Overview User needs Administration issues INFN Bari farm design and deployment
More informationCernVM-FS beyond LHC computing
CernVM-FS beyond LHC computing C Condurache, I Collier STFC Rutherford Appleton Laboratory, Harwell Oxford, Didcot, OX11 0QX, UK E-mail: catalin.condurache@stfc.ac.uk Abstract. In the last three years
More informationScientific data management
Scientific data management Storage and data management components Application database Certificate Certificate Authorised users directory Certificate Certificate Researcher Certificate Policies Information
More informationStorage Resource Management: Concepts, Functionality, and Interface Specification
Storage Resource Management: Concepts, Functionality, and Interface Specification Main author: Arie Shoshani, LBNL Co-authors: Peter Kunszt, Heinz Stockinger, Kurt Stockinger, Erwin Laure (EDG-WP2) Jean-Philippe
More informationChallenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk
Challenges and Evolution of the LHC Production Grid April 13, 2011 Ian Fisk 1 Evolution Uni x ALICE Remote Access PD2P/ Popularity Tier-2 Tier-2 Uni u Open Lab m Tier-2 Science Uni x Grid Uni z USA Tier-2
More informationLong Term Data Preservation for CDF at INFN-CNAF
Long Term Data Preservation for CDF at INFN-CNAF S. Amerio 1, L. Chiarelli 2, L. dell Agnello 3, D. De Girolamo 3, D. Gregori 3, M. Pezzi 3, A. Prosperini 3, P. Ricci 3, F. Rosso 3, and S. Zani 3 1 University
More informationWorkload Management. Stefano Lacaprara. CMS Physics Week, FNAL, 12/16 April Department of Physics INFN and University of Padova
Workload Management Stefano Lacaprara Department of Physics INFN and University of Padova CMS Physics Week, FNAL, 12/16 April 2005 Outline 1 Workload Management: the CMS way General Architecture Present
More informationThe glite middleware. Presented by John White EGEE-II JRA1 Dep. Manager On behalf of JRA1 Enabling Grids for E-sciencE
The glite middleware Presented by John White EGEE-II JRA1 Dep. Manager On behalf of JRA1 John.White@cern.ch www.eu-egee.org EGEE and glite are registered trademarks Outline glite distributions Software
More informationThe LHC Computing Grid
The LHC Computing Grid Visit of Finnish IT Centre for Science CSC Board Members Finland Tuesday 19 th May 2009 Frédéric Hemmer IT Department Head The LHC and Detectors Outline Computing Challenges Current
More informationAndrea Sciabà CERN, Switzerland
Frascati Physics Series Vol. VVVVVV (xxxx), pp. 000-000 XX Conference Location, Date-start - Date-end, Year THE LHC COMPUTING GRID Andrea Sciabà CERN, Switzerland Abstract The LHC experiments will start
More informationA Simple Mass Storage System for the SRB Data Grid
A Simple Mass Storage System for the SRB Data Grid Michael Wan, Arcot Rajasekar, Reagan Moore, Phil Andrews San Diego Supercomputer Center SDSC/UCSD/NPACI Outline Motivations for implementing a Mass Storage
More informationStorage Resource Manager Interface Specification V2.2 Implementations Experience Report
GFD-E.XXXX Grid Storage Resource Management https://forge.gridforum.org/projects/gsm-wg Editors: A. Sim A. Shoshani F. Donno J. Jensen 5/22/2009 Storage Resource Manager Interface Specification V2.2 Implementations
More informationStorage Resource Manager Interface Specification V2.2 Implementations Experience Report
GFD-E.154 Grid Storage Resource Management https://forge.gridforum.org/projects/gsm-wg Editors: A. Sim A. Shoshani F. Donno J. Jensen 8/18/2009 Storage Resource Manager Interface Specification V2.2 Implementations
More informationGrid Infrastructure For Collaborative High Performance Scientific Computing
Computing For Nation Development, February 08 09, 2008 Bharati Vidyapeeth s Institute of Computer Applications and Management, New Delhi Grid Infrastructure For Collaborative High Performance Scientific
More informationOnline data storage service strategy for the CERN computer Centre G. Cancio, D. Duellmann, M. Lamanna, A. Pace CERN, Geneva, Switzerland
Online data storage service strategy for the CERN computer Centre G. Cancio, D. Duellmann, M. Lamanna, A. Pace CERN, Geneva, Switzerland Abstract. The Data and Storage Services group at CERN is conducting
More informationA short introduction to the Worldwide LHC Computing Grid. Maarten Litmaath (CERN)
A short introduction to the Worldwide LHC Computing Grid Maarten Litmaath (CERN) 10-15 PetaByte/year The LHC challenge Data analysis requires at least ~100k typical PC processor cores Scientists in tens
More informationStorage Virtualization. Eric Yen Academia Sinica Grid Computing Centre (ASGC) Taiwan
Storage Virtualization Eric Yen Academia Sinica Grid Computing Centre (ASGC) Taiwan Storage Virtualization In computer science, storage virtualization uses virtualization to enable better functionality
More informationGrid Computing. MCSN - N. Tonellotto - Distributed Enabling Platforms
Grid Computing 1 Resource sharing Elements of Grid Computing - Computers, data, storage, sensors, networks, - Sharing always conditional: issues of trust, policy, negotiation, payment, Coordinated problem
More informationBenchmarking the ATLAS software through the Kit Validation engine
Benchmarking the ATLAS software through the Kit Validation engine Alessandro De Salvo (1), Franco Brasolin (2) (1) Istituto Nazionale di Fisica Nucleare, Sezione di Roma, (2) Istituto Nazionale di Fisica
More informationLawrence Berkeley National Laboratory Recent Work
Lawrence Berkeley National Laboratory Recent Work Title Storage resource managers: Essential components for the grid Permalink https://escholarship.org/uc/item/3qk459bb Journal Grid Resource Management:
More informationNew strategies of the LHC experiments to meet the computing requirements of the HL-LHC era
to meet the computing requirements of the HL-LHC era NPI AS CR Prague/Rez E-mail: adamova@ujf.cas.cz Maarten Litmaath CERN E-mail: Maarten.Litmaath@cern.ch The performance of the Large Hadron Collider
More informationTHE GLOBUS PROJECT. White Paper. GridFTP. Universal Data Transfer for the Grid
THE GLOBUS PROJECT White Paper GridFTP Universal Data Transfer for the Grid WHITE PAPER GridFTP Universal Data Transfer for the Grid September 5, 2000 Copyright 2000, The University of Chicago and The
More informationA copy can be downloaded for personal non-commercial research or study, without prior permission or charge
Bhimji, W., Bland, J., Clark, P. J., Mouzeli, E. G., Skipsey, S., and Walker, C. J. (11) Tuning grid storage resources for LHC data analysis. In: International Conference on Computing in High Energy and
More informationN. Marusov, I. Semenov
GRID TECHNOLOGY FOR CONTROLLED FUSION: CONCEPTION OF THE UNIFIED CYBERSPACE AND ITER DATA MANAGEMENT N. Marusov, I. Semenov Project Center ITER (ITER Russian Domestic Agency N.Marusov@ITERRF.RU) Challenges
More information150 million sensors deliver data. 40 million times per second
CERN June 2007 View of the ATLAS detector (under construction) 150 million sensors deliver data 40 million times per second ATLAS distributed data management software, Don Quijote 2 (DQ2) ATLAS full trigger
More informationDeployment and Testing of Storage Management software, for CMS experiment
www.eu-egee.org Deployment and Testing of Storage Management software, for CMS experiment G.Donvito INFN Bari EGEE is a project funded by the European Union under contract IST-2003-508833 Introduction
More informationHigh Performance Computing Course Notes Grid Computing I
High Performance Computing Course Notes 2008-2009 2009 Grid Computing I Resource Demands Even as computer power, data storage, and communication continue to improve exponentially, resource capacities are
More informationData Access and Data Management
Data Access and Data Management in grids Jos van Wezel Overview Background [KIT, GridKa] Practice [LHC, glite] Data storage systems [dcache a.o.] Data and meta data Intro KIT = FZK + Univ. of Karlsruhe
More informationDIRAC data management: consistency, integrity and coherence of data
Journal of Physics: Conference Series DIRAC data management: consistency, integrity and coherence of data To cite this article: M Bargiotti and A C Smith 2008 J. Phys.: Conf. Ser. 119 062013 Related content
More informationDistributed Data Management on the Grid. Mario Lassnig
Distributed Data Management on the Grid Mario Lassnig Who am I? Mario Lassnig Computer scientist main field of study was theoretical (algorithm design) working on/with distributed and embedded systems
More informationLarge scale commissioning and operational experience with tier-2 to tier-2 data transfer links in CMS
Journal of Physics: Conference Series Large scale commissioning and operational experience with tier-2 to tier-2 data transfer links in CMS To cite this article: J Letts and N Magini 2011 J. Phys.: Conf.
More informationPoS(ACAT2010)039. First sights on a non-grid end-user analysis model on Grid Infrastructure. Roberto Santinelli. Fabrizio Furano.
First sights on a non-grid end-user analysis model on Grid Infrastructure Roberto Santinelli CERN E-mail: roberto.santinelli@cern.ch Fabrizio Furano CERN E-mail: fabrzio.furano@cern.ch Andrew Maier CERN
More informationParallel Storage Systems for Large-Scale Machines
Parallel Storage Systems for Large-Scale Machines Doctoral Showcase Christos FILIPPIDIS (cfjs@outlook.com) Department of Informatics and Telecommunications, National and Kapodistrian University of Athens
More informationKnowledge Discovery Services and Tools on Grids
Knowledge Discovery Services and Tools on Grids DOMENICO TALIA DEIS University of Calabria ITALY talia@deis.unical.it Symposium ISMIS 2003, Maebashi City, Japan, Oct. 29, 2003 OUTLINE Introduction Grid
More informationA scalable storage element and its usage in HEP
AstroGrid D Meeting at MPE 14 15. November 2006 Garching dcache A scalable storage element and its usage in HEP Martin Radicke Patrick Fuhrmann Introduction to dcache 2 Project overview joint venture between
More informationTesting an Open Source installation and server provisioning tool for the INFN CNAF Tier1 Storage system
Testing an Open Source installation and server provisioning tool for the INFN CNAF Tier1 Storage system M Pezzi 1, M Favaro 1, D Gregori 1, PP Ricci 1, V Sapunenko 1 1 INFN CNAF Viale Berti Pichat 6/2
More informationUW-ATLAS Experiences with Condor
UW-ATLAS Experiences with Condor M.Chen, A. Leung, B.Mellado Sau Lan Wu and N.Xu Paradyn / Condor Week, Madison, 05/01/08 Outline Our first success story with Condor - ATLAS production in 2004~2005. CRONUS
More informationData Storage. Paul Millar dcache
Data Storage Paul Millar dcache Overview Introducing storage How storage is used Challenges and future directions 2 (Magnetic) Hard Disks 3 Tape systems 4 Disk enclosures 5 RAID systems 6 Types of RAID
More informationARC integration for CMS
ARC integration for CMS ARC integration for CMS Erik Edelmann 2, Laurence Field 3, Jaime Frey 4, Michael Grønager 2, Kalle Happonen 1, Daniel Johansson 2, Josva Kleist 2, Jukka Klem 1, Jesper Koivumäki
More informationStorage and I/O requirements of the LHC experiments
Storage and I/O requirements of the LHC experiments Sverre Jarp CERN openlab, IT Dept where the Web was born 22 June 2006 OpenFabrics Workshop, Paris 1 Briefly about CERN 22 June 2006 OpenFabrics Workshop,
More informationStorage Resource Sharing with CASTOR.
Storage Resource Sharing with CASTOR Olof Barring, Benjamin Couturier, Jean-Damien Durand, Emil Knezo, Sebastien Ponce (CERN) Vitali Motyakov (IHEP) ben.couturier@cern.ch 16/4/2004 Storage Resource Sharing
More informationClouds in High Energy Physics
Clouds in High Energy Physics Randall Sobie University of Victoria Randall Sobie IPP/Victoria 1 Overview Clouds are integral part of our HEP computing infrastructure Primarily Infrastructure-as-a-Service
More informationWHEN the Large Hadron Collider (LHC) begins operation
2228 IEEE TRANSACTIONS ON NUCLEAR SCIENCE, VOL. 53, NO. 4, AUGUST 2006 Measurement of the LCG2 and Glite File Catalogue s Performance Craig Munro, Birger Koblitz, Nuno Santos, and Akram Khan Abstract When
More informationData Transfers Between LHC Grid Sites Dorian Kcira
Data Transfers Between LHC Grid Sites Dorian Kcira dkcira@caltech.edu Caltech High Energy Physics Group hep.caltech.edu/cms CERN Site: LHC and the Experiments Large Hadron Collider 27 km circumference
More informationDistributing storage of LHC data - in the nordic countries
Distributing storage of LHC data - in the nordic countries Gerd Behrmann INTEGRATE ASG Lund, May 11th, 2016 Agenda WLCG: A world wide computing grid for the LHC NDGF: The Nordic Tier 1 dcache: Distributed
More informationLHCb Computing Resource usage in 2017
LHCb Computing Resource usage in 2017 LHCb-PUB-2018-002 07/03/2018 LHCb Public Note Issue: First version Revision: 0 Reference: LHCb-PUB-2018-002 Created: 1 st February 2018 Last modified: 12 th April
More informationEXAM Administration of Symantec Enterprise Vault 10.0 for Exchange. Buy Full Product.
Symantec EXAM - 250-310 Administration of Symantec Enterprise Vault 10.0 for Exchange Buy Full Product http://www.examskey.com/250-310.html Examskey Symantec 250-310 exam demo product is here for you to
More informationExperience of Data Grid simulation packages using.
Experience of Data Grid simulation packages using. Nechaevskiy A.V. (SINP MSU), Korenkov V.V. (LIT JINR) Dubna, 2008 Contant Operation of LCG DataGrid Errors of FTS services of the Grid. Primary goals
More informationPhilippe Charpentier PH Department CERN, Geneva
Philippe Charpentier PH Department CERN, Geneva Outline Disclaimer: These lectures are not meant at teaching you how to compute on the Grid! I hope it will give you a flavor on what Grid Computing is about
More informationThe Grid. Processing the Data from the World s Largest Scientific Machine II Brazilian LHC Computing Workshop
The Grid Processing the Data from the World s Largest Scientific Machine II Brazilian LHC Computing Workshop Patricia Méndez Lorenzo (IT-GS/EIS), CERN Abstract The world's largest scientific machine will
More informationCHIPP Phoenix Cluster Inauguration
TheComputing Environment for LHC Data Analysis The LHC Computing Grid CHIPP Phoenix Cluster Inauguration Manno, Switzerland 30 May 2008 Les Robertson IT Department - CERN CH-1211 Genève 23 les.robertson@cern.ch
More informationStatus of KISTI Tier2 Center for ALICE
APCTP 2009 LHC Physics Workshop at Korea Status of KISTI Tier2 Center for ALICE August 27, 2009 Soonwook Hwang KISTI e-science Division 1 Outline ALICE Computing Model KISTI ALICE Tier2 Center Future Plan
More informationan Object-Based File System for Large-Scale Federated IT Infrastructures
an Object-Based File System for Large-Scale Federated IT Infrastructures Jan Stender, Zuse Institute Berlin HPC File Systems: From Cluster To Grid October 3-4, 2007 In this talk... Introduction: Object-based
More informationALICE Grid Activities in US
ALICE Grid Activities in US 1 ALICE-USA Computing Project ALICE-USA Collaboration formed to focus on the ALICE EMCal project Construction, installation, testing and integration participating institutions
More informationGrid Architectural Models
Grid Architectural Models Computational Grids - A computational Grid aggregates the processing power from a distributed collection of systems - This type of Grid is primarily composed of low powered computers
More informationData transfer over the wide area network with a large round trip time
Journal of Physics: Conference Series Data transfer over the wide area network with a large round trip time To cite this article: H Matsunaga et al 1 J. Phys.: Conf. Ser. 219 656 Recent citations - A two
More informationLCG data management at IN2P3 CC FTS SRM dcache HPSS
jeudi 26 avril 2007 LCG data management at IN2P3 CC FTS SRM dcache HPSS Jonathan Schaeffer / Lionel Schwarz dcachemaster@cc.in2p3.fr dcache Joint development by FNAL and DESY Cache disk manager with unique
More informationReport. Middleware Proxy: A Request-Driven Messaging Broker For High Volume Data Distribution
CERN-ACC-2013-0237 Wojciech.Sliwinski@cern.ch Report Middleware Proxy: A Request-Driven Messaging Broker For High Volume Data Distribution W. Sliwinski, I. Yastrebov, A. Dworak CERN, Geneva, Switzerland
More informationCERN and Scientific Computing
CERN and Scientific Computing Massimo Lamanna CERN Information Technology Department Experiment Support Group 1960: 26 GeV proton in the 32 cm CERN hydrogen bubble chamber 1960: IBM 709 at the Geneva airport
More informationPatrick Fuhrmann (DESY)
Patrick Fuhrmann (DESY) EMI Data Area lead (on behalf of many people and slides stolen from all over the place) Credits Alejandro Alvarez Alex Sim Claudio Cacciari Christian Bernardt Christian Loeschen
More informationdcache Introduction Course
GRIDKA SCHOOL 2013 KARLSRUHER INSTITUT FÜR TECHNOLOGIE KARLSRUHE August 29, 2013 dcache Introduction Course Overview Chapters I, II and Ⅴ christoph.anton.mitterer@lmu.de I. Introduction To dcache Slide
More informationComputing activities in Napoli. Dr. Silvio Pardi (INFN-Napoli) Belle II Italian collaboration meeting 21 November 2017 Pisa - Italy
Computing activities in Napoli Dr. Silvio Pardi (INFN-Napoli) Belle II Italian collaboration meeting 21 November 2017 Pisa - Italy Activities in Napoli Grid Services Networking Http/Webdav and Dynamic
More informationFile Access Optimization with the Lustre Filesystem at Florida CMS T2
Journal of Physics: Conference Series PAPER OPEN ACCESS File Access Optimization with the Lustre Filesystem at Florida CMS T2 To cite this article: P. Avery et al 215 J. Phys.: Conf. Ser. 664 4228 View
More informationACCURATE STUDY GUIDES, HIGH PASSING RATE! Question & Answer. Dump Step. provides update free of charge in one year!
DUMP STEP Question & Answer ACCURATE STUDY GUIDES, HIGH PASSING RATE! Dump Step provides update free of charge in one year! http://www.dumpstep.com Exam : 250-310 Title : Administration of Symantec Enterprise
More informationThe Grid: Processing the Data from the World s Largest Scientific Machine
The Grid: Processing the Data from the World s Largest Scientific Machine 10th Topical Seminar On Innovative Particle and Radiation Detectors Siena, 1-5 October 2006 Patricia Méndez Lorenzo (IT-PSS/ED),
More informationForschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. Presented by Manfred Alef Contributions of Jos van Wezel, Andreas Heiss
Site Report Presented by Manfred Alef Contributions of Jos van Wezel, Andreas Heiss Grid Computing Centre Karlsruhe (GridKa) Forschungszentrum Karlsruhe Institute for Scientific Computing Hermann-von-Helmholtz-Platz
More informationBig Computing and the Mitchell Institute for Fundamental Physics and Astronomy. David Toback
Big Computing and the Mitchell Institute for Fundamental Physics and Astronomy Texas A&M Big Data Workshop October 2011 January 2015, Texas A&M University Research Topics Seminar 1 Outline Overview of
More informationConference The Data Challenges of the LHC. Reda Tafirout, TRIUMF
Conference 2017 The Data Challenges of the LHC Reda Tafirout, TRIUMF Outline LHC Science goals, tools and data Worldwide LHC Computing Grid Collaboration & Scale Key challenges Networking ATLAS experiment
More informationDynamic Federations. Seamless aggregation of standard-protocol-based storage endpoints
Dynamic Federations Seamless aggregation of standard-protocol-based storage endpoints Fabrizio Furano Patrick Fuhrmann Paul Millar Daniel Becker Adrien Devresse Oliver Keeble Ricardo Brito da Rocha Alejandro
More informationThe Compact Muon Solenoid Experiment. Conference Report. Mailing address: CMS CERN, CH-1211 GENEVA 23, Switzerland
Available on CMS information server CMS CR -2012/140 The Compact Muon Solenoid Experiment Conference Report Mailing address: CMS CERN, CH-1211 GENEVA 23, Switzerland 13 June 2012 (v2, 19 June 2012) No
More informationTransitioning NCAR MSS to HPSS
Transitioning NCAR MSS to HPSS Oct 29, 2009 Erich Thanhardt Overview Transitioning to HPSS Explain rationale behind the move Introduce current HPSS system in house Present transition plans with timelines
More informationLHCb Distributed Conditions Database
LHCb Distributed Conditions Database Marco Clemencic E-mail: marco.clemencic@cern.ch Abstract. The LHCb Conditions Database project provides the necessary tools to handle non-event time-varying data. The
More informationThe CMS Computing Model
The CMS Computing Model Dorian Kcira California Institute of Technology SuperComputing 2009 November 14-20 2009, Portland, OR CERN s Large Hadron Collider 5000+ Physicists/Engineers 300+ Institutes 70+
More informationManaged Data Storage and Data Access Services for Data Grids
Managed Data Storage and Data Access Services for Data Grids 1 M. Ernst, P. Fuhrmann, T. Mkrtchyan DESY J. Bakken, I. Fisk, T. Perelmutov, D. Petravick Fermilab s defined by the GriPhyN Project lobal scientific
More informationExploring cloud storage for scien3fic research
Exploring cloud storage for scien3fic research Fabio Hernandez fabio@in2p3.fr Lu Wang Lu.Wang@ihep.ac.cn 第十六届全国科学计算与信息化会议暨科研大数据论坛 h"p://indico.ihep.ac.cn/conferencedisplay.py?confid=3138 Dalian, July 8th
More informationATLAS computing activities and developments in the Italian Grid cloud
Journal of Physics: Conference Series ATLAS computing activities and developments in the Italian Grid cloud To cite this article: L Rinaldi et al 2012 J. Phys.: Conf. Ser. 396 042052 View the article online
More informationComputing for LHC in Germany
1 Computing for LHC in Germany Günter Quast Universität Karlsruhe (TH) Meeting with RECFA Berlin, October 5th 2007 WLCG Tier1 & Tier2 Additional resources for data analysis - HGF ''Physics at the Terascale''
More informationDeliverable D8.9 - First release of DM services
GridLab - A Grid Application Toolkit and Testbed Deliverable D8.9 - First release of DM services Author(s): Document Filename: Work package: Partner(s): Lead Partner: Config ID: Document classification:
More informationExperience of the WLCG data management system from the first two years of the LHC data taking
Experience of the WLCG data management system from the first two years of the LHC data taking 1 Nuclear Physics Institute, Czech Academy of Sciences Rez near Prague, CZ 25068, Czech Republic E-mail: adamova@ujf.cas.cz
More informationAGIS: The ATLAS Grid Information System
AGIS: The ATLAS Grid Information System Alexey Anisenkov 1, Sergey Belov 2, Alessandro Di Girolamo 3, Stavro Gayazov 1, Alexei Klimentov 4, Danila Oleynik 2, Alexander Senchenko 1 on behalf of the ATLAS
More informationGRID COMPUTING APPLIED TO OFF-LINE AGATA DATA PROCESSING. 2nd EGAN School, December 2012, GSI Darmstadt, Germany
GRID COMPUTING APPLIED TO OFF-LINE AGATA DATA PROCESSING M. KACI mohammed.kaci@ific.uv.es 2nd EGAN School, 03-07 December 2012, GSI Darmstadt, Germany GRID COMPUTING TECHNOLOGY THE EUROPEAN GRID: HISTORY
More informationMetadaten Workshop 26./27. März 2007 Göttingen. Chimera. a new grid enabled name-space service. Martin Radicke. Tigran Mkrtchyan
Metadaten Workshop 26./27. März Chimera a new grid enabled name-space service What is Chimera? a new namespace provider provides a simulated filesystem with additional metadata fast, scalable and based
More information( PROPOSAL ) THE AGATA GRID COMPUTING MODEL FOR DATA MANAGEMENT AND DATA PROCESSING. version 0.6. July 2010 Revised January 2011
( PROPOSAL ) THE AGATA GRID COMPUTING MODEL FOR DATA MANAGEMENT AND DATA PROCESSING version 0.6 July 2010 Revised January 2011 Mohammed Kaci 1 and Victor Méndez 1 For the AGATA collaboration 1 IFIC Grid
More informationBenchmarking third-party-transfer protocols with the FTS
Benchmarking third-party-transfer protocols with the FTS Rizart Dona CERN Summer Student Programme 2018 Supervised by Dr. Simone Campana & Dr. Oliver Keeble 1.Introduction 1 Worldwide LHC Computing Grid
More informationASSIGNMENT- I Topic: Functional Modeling, System Design, Object Design. Submitted by, Roll Numbers:-49-70
ASSIGNMENT- I Topic: Functional Modeling, System Design, Object Design Submitted by, Roll Numbers:-49-70 Functional Models The functional model specifies the results of a computation without specifying
More informationHPSS Treefrog Summary MARCH 1, 2018
HPSS Treefrog Summary MARCH 1, 2018 Disclaimer Forward looking information including schedules and future software reflect current planning that may change and should not be taken as commitments by IBM
More informationHEP replica management
Primary actor Goal in context Scope Level Stakeholders and interests Precondition Minimal guarantees Success guarantees Trigger Technology and data variations Priority Releases Response time Frequency
More informationGrid Data Management
Grid Data Management Week #4 Hardi Teder hardi@eenet.ee University of Tartu March 6th 2013 Overview Grid Data Management Where the Data comes from? Grid Data Management tools 2/33 Grid foundations 3/33
More informationConstant monitoring of multi-site network connectivity at the Tokyo Tier2 center
Constant monitoring of multi-site network connectivity at the Tokyo Tier2 center, T. Mashimo, N. Matsui, H. Matsunaga, H. Sakamoto, I. Ueda International Center for Elementary Particle Physics, The University
More information