where the Web was born Experience of Adding New Architectures to the LCG Production Environment

Similar documents
Overview of HEP software & LCG from the openlab perspective

10 Gbit/s Challenge inside the Openlab framework

Storage and I/O requirements of the LHC experiments

CERN openlab II. CERN openlab and. Sverre Jarp CERN openlab CTO 16 September 2008

FREE SCIENTIFIC COMPUTING

Andrea Sciabà CERN, Switzerland

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. Presented by Manfred Alef Contributions of Jos van Wezel, Andreas Heiss

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Heterogeneous Grid Computing: Issues and Early Benchmarks

Tier-2 structure in Poland. R. Gokieli Institute for Nuclear Studies, Warsaw M. Witek Institute of Nuclear Physics, Cracow

ISTITUTO NAZIONALE DI FISICA NUCLEARE

The LHC Computing Grid

On the employment of LCG GRID middleware

Status of KISTI Tier2 Center for ALICE

Spanish Tier-2. Francisco Matorras (IFCA) Nicanor Colino (CIEMAT) F. Matorras N.Colino, Spain CMS T2,.6 March 2008"

Edinburgh (ECDF) Update

Operating the Distributed NDGF Tier-1

Austrian Federated WLCG Tier-2

The LHC Computing Grid

The INFN Tier1. 1. INFN-CNAF, Italy

The EU DataGrid Testbed

IEPSAS-Kosice: experiences in running LCG site

Considerations for a grid-based Physics Analysis Facility. Dietrich Liko

DESY. Andreas Gellrich DESY DESY,

Tier2 Centre in Prague

g-eclipse A Framework for Accessing Grid Infrastructures Nicholas Loulloudes Trainer, University of Cyprus (loulloudes.n_at_cs.ucy.ac.

The Grid. Processing the Data from the World s Largest Scientific Machine II Brazilian LHC Computing Workshop

CrossGrid testbed status

HEP Grid Activities in China

Summary of the LHC Computing Review

EGEE and Interoperation

ELFms industrialisation plans

Pan-European Grid einfrastructure for LHC Experiments at CERN - SCL's Activities in EGEE

The LHC Computing Grid. Slides mostly by: Dr Ian Bird LCG Project Leader 18 March 2008

Deploying virtualisation in a production grid

First Experience with LCG. Board of Sponsors 3 rd April 2009

The Grid: Processing the Data from the World s Largest Scientific Machine

Grid Challenges and Experience

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

ATLAS Tier-3 UniGe

Ivane Javakhishvili Tbilisi State University High Energy Physics Institute HEPI TSU

CESGA: General description of the institution. FINIS TERRAE: New HPC Supercomputer for RECETGA: Galicia s Science & Technology Network

Middleware-Tests with our Xen-based Testcluster

Grid Computing at Ljubljana and Nova Gorica

Application of Virtualization Technologies & CernVM. Benedikt Hegner CERN

ATLAS NorduGrid related activities

Virtualizing a Batch. University Grid Center

The European DataGRID Production Testbed

The LCG 3D Project. Maria Girone, CERN. The 23rd Open Grid Forum - OGF23 4th June 2008, Barcelona. CERN IT Department CH-1211 Genève 23 Switzerland

Batch Services at CERN: Status and Future Evolution

CSCS CERN videoconference CFD applications

RUSSIAN DATA INTENSIVE GRID (RDIG): CURRENT STATUS AND PERSPECTIVES TOWARD NATIONAL GRID INITIATIVE

PoS(EGICF12-EMITC2)106

Monte Carlo Production on the Grid by the H1 Collaboration

Service Availability Monitor tests for ATLAS

glite Grid Services Overview

Storage Resource Sharing with CASTOR.

CHIPP Phoenix Cluster Inauguration

Introduction to Grid Infrastructures

System upgrade and future perspective for the operation of Tokyo Tier2 center. T. Nakamura, T. Mashimo, N. Matsui, H. Sakamoto and I.

Constant monitoring of multi-site network connectivity at the Tokyo Tier2 center

Computing for LHC in Germany

The creation of a Tier-1 Data Center for the ALICE experiment in the UNAM. Lukas Nellen ICN-UNAM

Virtualization. A very short summary by Owen Synge

Grid-related related Activities around KISTI Tier2 Center for ALICE

Overview. Grid vision Grid application domains The role of CERN in the Grid research Grid Architecture Standards and related activities Summary

High-Energy Physics Data-Storage Challenges

XRAY Grid TO BE OR NOT TO BE?

Grid Infrastructure For Collaborative High Performance Scientific Computing

High Energy Physics data analysis

ARDA status. Massimo Lamanna / CERN. LHCC referees meeting, 28 June

The glite middleware. Presented by John White EGEE-II JRA1 Dep. Manager On behalf of JRA1 Enabling Grids for E-sciencE

The grid for LHC Data Analysis

Figure 1: cstcdie Grid Site architecture

EMI Deployment Planning. C. Aiftimiei D. Dongiovanni INFN

Grid Computing a new tool for science

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

The ARDA project: Grid analysis prototypes of the LHC experiments

ALHAD G. APTE, BARC 2nd GARUDA PARTNERS MEET ON 15th & 16th SEPT. 2006

Belle & Belle II. Takanori Hara (KEK) 9 June, 2015 DPHEP Collaboration CERN

Features and Future. Frédéric Hemmer - CERN Deputy Head of IT Department. Enabling Grids for E-sciencE. BEGrid seminar Brussels, October 27, 2006

Grid and Cloud Activities in KISTI

AMGA metadata catalogue system

Grid Operation at Tokyo Tier-2 Centre for ATLAS

Philippe Charpentier PH Department CERN, Geneva

Tier-2 DESY Volker Gülzow, Peter Wegner

Service Level Agreement Metrics

Ganga - a job management and optimisation tool. A. Maier CERN

CMS Grid Computing at TAMU Performance, Monitoring and Current Status of the Brazos Cluster

Benoit DELAUNAY Benoit DELAUNAY 1

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

Grid Computing Activities at KIT

CMS Belgian T2. G. Bruno UCL, Louvain, Belgium on behalf of the CMS Belgian T2 community. GridKa T1/2 meeting, Karlsruhe Germany February

Support for multiple virtual organizations in the Romanian LCG Federation

Grid Computing. Olivier Dadoun LAL, Orsay. Introduction & Parachute method. Socle 2006 Clermont-Ferrand Orsay)

Advanced School in High Performance and GRID Computing November Introduction to Grid computing.

WHEN the Large Hadron Collider (LHC) begins operation

Servicing HEP experiments with a complete set of ready integreated and configured common software components

I Tier-3 di CMS-Italia: stato e prospettive. Hassen Riahi Claudio Grandi Workshop CCR GRID 2011

Integration of Cloud and Grid Middleware at DGRZR

Transcription:

where the Web was born Experience of Adding New Architectures to the LCG Production Environment Andreas Unterkircher, openlab fellow Sverre Jarp, CTO CERN openlab Industrializing the Grid openlab Workshop 13 June 2005 June 2005 openlab Workshop 1 Introduction June 2005 openlab Workshop 2 1

Grid @ CERN LCG: LHC Computing Grid the deployment project Will run the 24/7 Grid service EGEE: Enabling Grids for E-Science in Europe Started in April 2004 with 70 partners and 32M EU funding Will provide the next generation middleware for LCG CERN openlab for DataGrid applications Started in 2003 - Funded by Industry and CERN Main project: opencluster (including 100 Itanium nodes) R&D aimed at deployment in LCG June 2005 openlab Workshop 3 Computing for LHC Problem: even with an upgraded computer centre, CERN can only provide a fraction of the necessary resources Europe: ~270 institutes ~4500 users Elsewhere: ~200 institutes ~1600 users Solution: computing centres, which were isolated in the past, will now be connected, uniting the computing resources of particle physicists in the world using GRID technologies! June 2005 openlab Workshop 4 2

LCG-2 As of March 2005: biggest Grid project in the world 130 sites in 31 countries 12 000 processors 10 millions Gigabytes storage June 2005 openlab Workshop 5 Openlab: Tight integration with the LCG testbed 4 * GE connections to the backbone 10GE WAN connection 36 Disk Server (dual P4, IDE disks, ~ 1TB disk space each) 4 *ENTERASYS N7 10 GE Switches 2 * Enterasys new Series 10 GE per node 10 GE per node 2 * 100 IA32 CPU Server (dual 2.4 GHz P4, 1 GB mem.) 10GE 1 GE per node 2 * 50 Itanium Server (dual 1.3/1.5 GHz Itanium2, 2 GB mem) 28 TB, IBM StorageTank 12 Tape Server STK 9940B High Througput Prototype (openlab + LCG prototype) June 2005 openlab Workshop 6 3

Service Challenge 3 20 Itanium nodes: June 2005 openlab Workshop 7 64-bit porting project Itanium / Itanium Processor Family (IPF) / IA-64 June 2005 openlab Workshop 8 4

The 64-bit issue What exactly is meant? Simple: Linux on 32-bit hardware uses ILP32 Int = Long = Pointer (32 bit) Linux on 64-bit hardware uses I32LP64 Int stays 32-bit Pointer = Long (64 bit) As a result: For instance: Any attempt to cast a pointer to Int (and back) Fatal error!!! June 2005 openlab Workshop 9 LCG components Soon: glite, FTS from EGEE EDG middleware: workload, monitoring, information mgmt, resource mgmt,... LCG/HEP specific: LFC, dcache, CASTOR... VDT 1.2 Globus (2.x) Scientific Linux 3 (SL3) External software: MySQL, batch system, perl modules, xerces, tomcat,... June 2005 openlab Workshop 10 5

LCG architecture User Interface Minimal LCG site: Computing Element, Worker Node(s), Storage Element, site BDII, R-GMA producer Resource Broker Further LCG nodes: File Catalog (LFC) MyProxy Server Monitoring Higher level BDIIs June 2005 openlab Workshop 11 Timeline Start manual VDT port Start manual EDG port 1 st CE & WN available 1 st Itanium grid job submitted (EIS testbed) Development of Itanium specific installation method (SmartFrog) LCG-2_4_0 with YAIM support on IPF VDT releases IPF rpms LCG build machine on Itanium Sept 2004 Feb May Dec July 2005 March April Installation at CNIC Installation at HP Bristol Installation at HP Puerto Rico Start HEP porting (SEAL,...) Installation at Poznan Supercomputing Center IPF modifications start to get into CVS June 2005 openlab Workshop 12 6

Original LCG build model Check out from source CVS: EDG software LCG specific code Everything else is external. A build machine automatically does the build after the checkout (uses GNU autotools) June 2005 openlab Workshop 13 Initial status (2003) LCG build machine supported only IA-32 with specific version of Red Hat No binaries available for Itanium/IA-64 Hardly any documentation Installation of LCG only via LCFGng (fully automatic, IA-32 only) manual installation was considered to be extremely difficult (EDG manual) June 2005 openlab Workshop 14 7

Initial strategy Started to port everything on our own One doctoral student & one fellow Stephen Eccles (now: Lancaster University) Andreas Unterkircher (CERN) After 6 months we were able to install a minimal (CE,WN,SE) Itanium LCG site and successfully submit jobs. June 2005 openlab Workshop 15 Initial obstacles (1) VDT has its own (not documented) build procedure. We had to do reverse engineering. It was often difficult to find the original sources of rpms. EDG used sometimes special versions of well known libraries (e.g. Boost). June 2005 openlab Workshop 16 8

Initial obstacles (2) EDG build procedure was hard-coded for IA-32 on RH 7.3. As our changes did not get back into the CVS it was difficult for us to keep track with the latest releases The code had, indeed, several 64-bit issues but the complicated build procedures (EDG as well as VDT) caused us much more trouble. June 2005 openlab Workshop 17 Lessons learnt (1) Initial effort was necessary to get noticed by the community. E.g. when VDT saw that we are serious they started to provide Itanium rpms on their own. Vital: Always get changes back into the CVS on a regular basis. June 2005 openlab Workshop 18 9

Lessons learnt (2) Support for different compilers, OS s and architectures should be considered in the build procedure from the beginning and used for testing on a regular basis. From a first proof of concept to a fully supported official release can take a long time: In our case: ~1 year June 2005 openlab Workshop 19 Lessons learnt (3) Porting LCG to Itanium was a chicken and egg problem : LCG was not considering porting as there is no HEP software for Itanium Physicists did not port to Itanium as there were no such resources in LCG. Thus we also started porting of major HEP software (SEAL, POOL, etc.) Note that ALICE has all its software 64-bit clean! Mainly an issue of initial mindset June 2005 openlab Workshop 20 10

Porting to EM64T/AMD64 (1) Should be much easier as IA64 (64bit) code changes are also valid for these platforms. Exactly the same I32LP64 model First one has to ensure that the basic packages (VDT, external software) are available. Getting modifications back to CVS immediately will be important. June 2005 openlab Workshop 21 Porting to EM64T/AMD64 (2) Build procedures not recognizing the architecture could be again the source of much trouble this must be addressed immediately. Hopefully EGEE/gLite will prove to be better in this respect than EDG We (and others) are providing platforms for testing Finally worth mentioning: Some ports of EDG to other platforms (e.g. PowerPC) are available on the Grid-Ireland homepage. June 2005 openlab Workshop 22 11

Overview of 64-bit porting Phase 1 Completed: ROOT (Data analysis framework) http://root.cern.ch/ Geant4 (Physics simulation framework) http://cern.ch/geant4 CLHEP (C++ Class Library) http://proj-clhep.web.cern.ch/proj-clhep/ CASTOR (CERN Hierarchical Storage Manager) http://cern.ch/castor LCG-2 Grid middleware Originated from EDG (European Data Grid) http://lcg.web.cern.ch/lcg/sites/releases.html Itanium version: http://openlab-mu-internal.web.cern.ch/openlab-muinternal/projects/lcgonia64/lcgonia64.asp June 2005 openlab Workshop 23 64-bit porting (cont d) Next aim: Allow the simulation stack of one of the LHC experiments (LHCb) to work on Itanium Set of external packages (Boost, etc.): OK Base set of CERN packages (Geant4, ROOT, CLHEP): OK HEP/LCG packages (SEAL, POOL, PI): In progress Specific packages from the experiment (Gaudi, Gauss, Ganga): In progress Once this experiment s stack is complete, ATLAS and CMS frameworks should also be within range By the way, Intel, Munich is apparently also working on the ATLAS software June 2005 openlab Workshop 24 12

Virtualization project June 2005 openlab Workshop 25 Virtualization Our history Xen benchmarked with CERN simulation workload on IA-32 Work done by summer student 2004 Project work on IO workloads under Xen Two students Project work on Xen on Itanium One of the two students (Master thesis in this semester) Collaboration with HP Labs Additionally: One openlab fellow is continuing the work on IA32 w/linux Fedora version Aim at IO intensive workloads (ROOT analysis, etc.) Rationale: next generation processors (such as IPF Montecito) will have hardware support for virtualization Question: Will virtualization be one of the underpinnings of future Grid security? June 2005 openlab Workshop 26 13

Conclusions The 64-bit port to Itanium has laid the foundation for: The inclusion of Itanium systems in LCG-2 A new architectural dimension in the Grid Heterogeneity A foundation for porting other 64-bit/Linux systems A multi-platform strategy for Grid middleware development June 2005 openlab Workshop 27 BACKUP June 2005 openlab Workshop 28 14