Australian participation in the world's largest eresearch infrastructure: the Worldwide LHC Computing Grid and the EGEE Program.

Similar documents
Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

RUSSIAN DATA INTENSIVE GRID (RDIG): CURRENT STATUS AND PERSPECTIVES TOWARD NATIONAL GRID INITIATIVE

Travelling securely on the Grid to the origin of the Universe

The LHC Computing Grid

Operating the Distributed NDGF Tier-1

The Grid: Processing the Data from the World s Largest Scientific Machine

The LHC Computing Grid. Slides mostly by: Dr Ian Bird LCG Project Leader 18 March 2008

Computing for LHC in Germany

Pan-European Grid einfrastructure for LHC Experiments at CERN - SCL's Activities in EGEE

EUDET Kickoff. Brief Overview of the Meeting at DESY E.Elsen

Grid Computing a new tool for science

ATLAS Experiment and GCE

The EGEE-III Project Towards Sustainable e-infrastructures

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

IEPSAS-Kosice: experiences in running LCG site

Andrea Sciabà CERN, Switzerland

Network Requirements. September, Introduction

The grid for LHC Data Analysis

Historical Milestones for the ATLAS Experiment

The Grid. Processing the Data from the World s Largest Scientific Machine II Brazilian LHC Computing Workshop

First Experience with LCG. Board of Sponsors 3 rd April 2009

Status of KISTI Tier2 Center for ALICE

CHIPP Phoenix Cluster Inauguration

Grid and Cloud Activities in KISTI

Batch Services at CERN: Status and Future Evolution

The LHC Computing Grid

Storage and I/O requirements of the LHC experiments

Support for multiple virtual organizations in the Romanian LCG Federation

Architectural Design and Supervisor RoI Builder

Distributed e-infrastructures for data intensive science

Towards Network Awareness in LHC Computing

CSCS CERN videoconference CFD applications

AGIS: The ATLAS Grid Information System

ATLAS DAQ, EF, LVL2 and DCS

Computing / The DESY Grid Center

CERN openlab II. CERN openlab and. Sverre Jarp CERN openlab CTO 16 September 2008

From raw data to new fundamental particles: The data management lifecycle at the Large Hadron Collider

Virtualizing a Batch. University Grid Center

CC-IN2P3: A High Performance Data Center for Research

Preparing for High-Luminosity LHC. Bob Jones CERN Bob.Jones <at> cern.ch

FREE SCIENTIFIC COMPUTING

RDMS CMS Computing Activities before the LHC start

30 Nov Dec Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy

Austrian Federated WLCG Tier-2

PoS(EGICF12-EMITC2)106

A short introduction to the Worldwide LHC Computing Grid. Maarten Litmaath (CERN)

The LCG 3D Project. Maria Girone, CERN. The 23rd Open Grid Forum - OGF23 4th June 2008, Barcelona. CERN IT Department CH-1211 Genève 23 Switzerland

Grid Interoperation and Regional Collaboration

APAC: A National Research Infrastructure Program

High Performance Computing from an EU perspective

The LHC computing model and its evolution. Dr Bob Jones CERN

where the Web was born Experience of Adding New Architectures to the LCG Production Environment

glite Grid Services Overview

The creation of a Tier-1 Data Center for the ALICE experiment in the UNAM. Lukas Nellen ICN-UNAM

ATLAS High-Level Trigger, Data Acquisition and Controls

Grid Operation at Tokyo Tier-2 Centre for ATLAS

Grid Computing: dealing with GB/s dataflows

Grid Computing: dealing with GB/s dataflows

Overview. About CERN 2 / 11

Experience of the WLCG data management system from the first two years of the LHC data taking

Data Management for the World s Largest Machine

DESY at the LHC. Klaus Mőnig. On behalf of the ATLAS, CMS and the Grid/Tier2 communities

CERN and Scientific Computing

GridPP - The UK Grid for Particle Physics

Grid services. Enabling Grids for E-sciencE. Dusan Vudragovic Scientific Computing Laboratory Institute of Physics Belgrade, Serbia

Federated data storage system prototype for LHC experiments and data intensive science

The Portuguese National Grid Initiative (INGRID)

ALICE Grid Activities in US

Scientific data management

DESY. Andreas Gellrich DESY DESY,

On the employment of LCG GRID middleware

CLOUDS OF JINR, UNIVERSITY OF SOFIA AND INRNE JOIN TOGETHER

Grid Computing at Ljubljana and Nova Gorica

Europe and its Open Science Cloud: the Italian perspective. Luciano Gaido Plan-E meeting, Poznan, April

International Cooperation in High Energy Physics. Barry Barish Caltech 30-Oct-06

Summary of the LHC Computing Review

Outline. Infrastructure and operations architecture. Operations. Services Monitoring and management tools

CouchDB-based system for data management in a Grid environment Implementation and Experience

Computing grids, a tool for international collaboration and against digital divide Guy Wormser Director of CNRS Institut des Grilles (CNRS, France)

CERN s Business Computing

Opportunities A Realistic Study of Costs Associated

The EU DataGrid Testbed

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

Evolution of the ATLAS PanDA Workload Management System for Exascale Computational Science

Tier-2 structure in Poland. R. Gokieli Institute for Nuclear Studies, Warsaw M. Witek Institute of Nuclear Physics, Cracow

Monte Carlo Production on the Grid by the H1 Collaboration

Constant monitoring of multi-site network connectivity at the Tokyo Tier2 center

about us bandwidth changes everything

Visita delegazione ditte italiane

Department of Physics & Astronomy

System upgrade and future perspective for the operation of Tokyo Tier2 center. T. Nakamura, T. Mashimo, N. Matsui, H. Sakamoto and I.

The European DataGRID Production Testbed

On the EGI Operational Level Agreement Framework

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

Evaluation of the computing resources required for a Nordic research exploitation of the LHC

We are IntechOpen, the world s leading publisher of Open Access books Built by scientists, for scientists. International authors and editors

Lessons Learned in the NorduGrid Federation

Clouds in High Energy Physics

Distributed Computing Framework. A. Tsaregorodtsev, CPPM-IN2P3-CNRS, Marseille

1. Introduction. Outline

Transcription:

Australian participation in the world's largest eresearch infrastructure: the Worldwide LHC Computing Grid and the EGEE Program. 27 June 2007 1 High Energy Particle Physics The aim of High Energy Physics is to: Understand the fundamental laws of nature by studying elementary particles and fields and their interactions 27 June 2007 2 1

Peering Back in Time... Studying matter at high energy densities takes us closer to......understanding the physics of the earliest moments of the universe Colliders shed light on questions of cosmological significance... 27 June 2007 3 Experimental High Energy Physics in Australia Close collaboration between: University of Melbourne University of Sydney University of Wollongong Hold joint Research Grants a single community of researchers The Physicists: Academic Staff, Research Fellows, Postgraduate Students, Engineers and Technicians Members of major international HEP experiments... ATLAS at the Large Hadron Collider, CERN, Geneva, Switzerland Belle at the KEK laboratory, Tsukuba, Japan 27 June 2007 4 2

The Large Hadron Collider, CERN The world's largest scientific instrument... Located near Geneva, Switzerland 27 June 2007 5 The Large Hadron Collider, CERN The LHC will smash protons together with energies in the Tera- electronvolt (TeV) range...the energy frontier 27 June 2007 6 3

The LHC Experiments: Four experiments: ATLAS, CMS, LHCb and ALICE 27 June 2007 7 The ATLAS Experiment To investigate the origins of Mass 27 June 2007 8 4

ATLAS will start taking data in 2008... 27 June 2007 9 ATLAS Collaboration 35 Countries 161 Institutions 1830 Scientific Authors Albany, Alberta, NIKHEF Amsterdam, Ankara, LAPP Annecy, Argonne NL, Arizona, UT Arlington, Athens, NTU Athens, Baku, IFAE Barcelona, Belgrade, Bergen, Berkeley LBL and UC, HU Berlin, Bern, Birmingham, Bologna, Bonn, Boston, Brandeis, Bratislava/SAS Kosice, Brookhaven NL, Buenos Aires, Bucharest, Cambridge, Carleton, Casablanca/Rabat, CERN, Chinese Cluster, Chicago, Clermont-Ferrand, Columbia, NBI Copenhagen, Cosenza, AGH UST Cracow, IFJ PAN Cracow, DESY, Dortmund, TU Dresden, JINR Dubna, Duke, Frascati, Freiburg, Geneva, Genoa, Giessen, Glasgow, LPSC Grenoble, Technion Haifa, Hampton, Harvard, Heidelberg, Hiroshima, Hiroshima IT, Indiana, Innsbruck, Iowa SU, Irvine UC, Istanbul Bogazici, KEK, Kobe, Kyoto, Kyoto UE, Lancaster, UN La Plata, Lecce, Lisbon LIP, Liverpool, Ljubljana, QMW London, RHBNC London, UC London, Lund, UA Madrid, Mainz, Manchester, Mannheim, CPPM Marseille, Massachusetts, MIT, Melbourne, Michigan, Michigan SU, Milano, Minsk NAS, Minsk NCPHEP, Montreal, McGill Montreal, FIAN Moscow, ITEP Moscow, MEPhI Moscow, MSU Moscow, Munich LMU, MPI Munich, Nagasaki IAS, Naples, New Mexico, New York, Nijmegen, BINP Novosibirsk, Ohio SU, Okayama, Oklahoma, Oklahoma SU, Oregon, LAL Orsay, Osaka, Oslo, Oxford, Paris VI and VII, Pavia, Pennsylvania, Pisa, Pittsburgh, CAS Prague, CU Prague, TU Prague, IHEP Protvino, Ritsumeikan, UFRJ Rio de Janeiro, Rochester, Rome I, Rome II, Rome III, Rutherford Appleton Laboratory, DAPNIA Saclay, Santa Cruz UC, Sheffield, Shinshu, Siegen, Simon Fraser Burnaby, SLAC, Southern Methodist Dallas, NPI Petersburg, Stockholm, KTH Stockholm, Stony Brook, Sydney, AS Taipei, Tbilisi, Tel Aviv, Thessaloniki, Tokyo ICEPP, Tokyo MU, Toronto, TRIUMF, Tsukuba, Tufts, Udine, Uppsala, Urbana UI, Valencia, UBC Vancouver, Victoria, Washington, Australian Weizmann High Energy Rehovot, Physics Wisconsin, in WLCG Wuppertal, and Yale, EGEE Yerevan 5

Physics at the Tera-Scale The LHC is a frontier machine operating in an unprecedented energy regime: TeV collider Searching for New Physics beyond the Standard Model Enhancing our understanding of the fundamental structure of matter...and throwing a light on the the nature of the universe in the first moments after the Big Bang ATLAS will look for: The elusive Higgs Boson: the origin of mass Supersymmetry Dark Matter, Dark Energy?? Large Extra Dimensions: Microscopic Black Holes Exotics: Magnetic monopoles,...... the Unexpected Black Holes at ATLAS ATLAS was recently honoured by a visit from Stephen Hawking Simulation of a mini black hole event with M BH ~ 8 TeV in ATLAS 6

Data from an LHC Experiment finding a needle in a haystack...finding a needle in a haystack Each event is really an overlay of many interactions We are looking for tracks from one interaction in this soup... 7

Data Mining Particle physicists wade through vast collections of events, searching for increasingly rare processes In principle, each physicist, or each team, is looking for something different. So, we have thousands of scientists searching through vast collections of data: Widely distributed across the Earth Each looking for a different type of needle in the haystack Chaotic data usage patterns! It is essential to deliver high bandwidth data streams to researchers at their home or regional data analysis facilities - a very competitive business The LHC Computing Challenge Data volume High rate x large number of channels x 4 experiments 15 PetaBytes of new data each year Compute power Event complexity x Nb. events x thousands users 100,000s of today's fastest CPUs Worldwide analysis & funding Computing funding locally in major regions & countries Efficient analysis everywhere Equity of Access Deliver the data to the physicists GRID technology 8

The LHC Computing Model Use distributed resources... Funding, access: Take the data to the physicists! The Worldwide LHC Computing Grid The EGEE and OSG projects are the basis of the Worldwide LHC Computing Grid Project WLCG Inter-operation between Grids is working! 9

The Worldwide LHC Computing Grid 37,000 CPUs and 13 Petabytes across 240 sites Up to 98,000 jobs per day: 13,000 non-lhc jobs Other VOs ATLAS transferring 1000 Terabytes per month 27 June 2007 19 EGEE: Enabling Grids for Escience Flagship grid infrastructure co-funded by the European Commission Now in 2 nd phase 91 partners in 32 countries Objectives: Large scale production quality grid infrastructure for eresearch Attract new resources and users from industry and science Maintain and improve glite grid middleware 27 June 2007 20 10

The EGEE Middleware: glite Security Virtual Organization Management: VOMS MyProxy Data management File catalogue: LFC File transfer service: FTS Storage Element: SE Storage Resource Management: SRM Workload management Work Load Management System: WMS Logging and Bookeeping: LB Computing Element: CE Worker Nodes: WN Information System Berkeley Database Information Index: BDII Relational Grid Monitoring Architecture: RGMA Monitoring & visualization (Griview, Dashboard, etc.)... an infrastructure for eresearch EGEE is a production infrastructure Operations Coordination Centre Management, oversight, coordination Regional operations Centres Core support infrastructure Taipei operates Asia-Pacific ROC ( GGUS ) Grid User Support Coordination, management of user support EGEE Network Operations ( ENOC ) Centre Coordination with NRENs & GEANT2 11

Who uses EGEE? Multitude of applications from a growing number of domains Astrophysics Computational Chemistry Earth Sciences Financial Simulation Fusion Geophysics High Energy Physics Life Sciences Multimedia Material Sciences.. High Energy Physics on EGEE Data management: Demonstrated data transfers at nominal rates:1.6 GB/s through File ( FTS ) Transfer Service 1 GB/s with real workloads 2 large experiments transferred >1 PB/month in 2006 Workload management CMS computing service challenge achieved 50k jobs/day CMS aim this year for 100k jobs/day; ATLAS for 60k Reliability and availability Significant effort to ensure Tier 1 sites meet MoU commitments using site and service monitoring Grid is now the primary source of computing resources for experiments at CERN 12

The ATLAS experiment on the WLCG: Heirarchical model: Tier 0 at CERN: Regional Tier 1 centres Take the data to the physicists! Global ATLAS Computing Resource Requirements Worldwide Computing Resources for ATLAS: 2008 2009 2010 Disk (TB) 30,000 45,000 72,000 Tape (TB) 16,000 29,000 48,000 CPU (Opteron 250) 54,000 88,000 150,000 Distributed across: ( CERN ) Tier 0 10 Tier 1 centres across the world 30 Tier 2 centres Australia has deployed a Tier 2 centre at the University of Melbourne Will become a Tier 2 Federation with partner resources at VPAC, AC3, SAPAC... Supported by APAC National Grid program since 2004 27 June 2007 26 13

Australian ATLAS Computing Service Target Resources for Australian facility: 2007 2008 2009 2010 Disk (TB) 30 170 280 440 CPU (Opteron 250) 40 200 300 480 An average Tier 2/3 to serve Australian ATLAS members Essential to ensure Australian researchers are competitive at the LHC Targets will be met by dedicated facility AND shared resources from partners: VPAC, AC3, SAPAC. Currently operating a Pilot Tier 2 service: Production Tier 2 Computing Service will be commissioned from July 2007: Supported by: ARC, DEST, APAC, Universities of Melbourne and Sydney 27 June 2007 27 Australia's Pilot ATLAS Computing Service Since 2006 we have been operating a Pilot Tier 2 site: For commissioning and participation in Pilot LHC Service 39 CPUs and 8 TB storage Supported by Research Computing Services at the University of ( Knijff Melbourne (Dirk van der Certified since July 2006: ATLAS: 15,561 jobs; 38,000 CPU hours BIOMED (Avian Flu): 565 jobs; 9218 CPU hours Jobs from across WLCG A major achievement: Intensive collaboration: Working closely with EGEE, APAC, AARNet, CERN, Taipei, China, Japan - a truly global collaboration 27 June 2007 28 14

Australia's ATLAS Computing Service Many challenges to overcome: glite Middleware deployment: delivering a reliable service ( Rosa (Marco La International long-haul data transfers: Very challenging International R&E networks Substantial support and expertise from AARNet End-point bottlenecks Regional Operations Centre: ASGC, Taipei hosts Asia-Pacific ROC and Tier 1 Strong collaboration and support Regular meetings 27 June 2007 29 Australia's ATLAS Computing Service We are now operating a reliable production facility in the world's largest eresearch infrastructure ( e-infrastructure (research Now hosting 8 Terabytes of simulated ATLAS data replicated from ATLAS collaborators across the globe Capacity increasing to 35 Terabytes from July 2007 Supporting simulated data production, data analysis and detector calibration Building expertise and capacity: Major capacity upgrade for 2008 We will be ready for first data from the LHC in 2008. 27 June 2007 30 15

So, how do researchers use these facilities? EGEE middleware provides: ( SRM ) storage management protocol File Transfer Service (FTS) for managed file transfer queues File Catalogues (LFC) with file replica management ( AMGA ) Metadata Catalogues for physics-level metadata ATLAS deploys a Distributed Data Model on top of these services Data files are organised into data sets We request subscriptions of data sets to our sites. DDM system manages subscriptions and files are transferred through a queuing service Analysis of data goes to the data set Place the data then send jobs to it Racing to discoveries at the LHC... 27 June 2007 31 Benefits from engagement with EGEE program (... relatively ) Easy to exploit The real hard work has been done ( Glenn Substantial expertise in Aus HEP group (Marco, Major support from EGEE on-tap User and operations support, training events Still new developments on the way Exploring the provision of a glite service on the Australian national grid For high energy physics and other communities International interoperation out of the box Going further? We have been invited to join the EGEE-Asia program Working very closely with EGEE partners in Asia-Pacific Want to know more see Marco La Rosa's demonstration at the VeRSI booth... 27 June 2007 32 16

EGEE & WLCG: A Truly Global Grid 27 June 2007 33 17