ALICE Grid Activities in US

Similar documents
EGEE and Interoperation

DVS, GPFS and External Lustre at NERSC How It s Working on Hopper. Tina Butler, Rei Chi Lee, Gregory Butler 05/25/11 CUG 2011

NERSC Site Update. National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory. Richard Gerber

Introduction to FREE National Resources for Scientific Computing. Dana Brunson. Jeff Pummill

NERSC. National Energy Research Scientific Computing Center

The LHC Computing Grid

Magellan Project. Jeff Broughton NERSC Systems Department Head October 7, 2009

Storage Supporting DOE Science

ARC integration for CMS

SLATE. Services Layer at the Edge. First Meeting of the National Research Platform Montana State University August 7-8, 2017

Tuning I/O Performance for Data Intensive Computing. Nicholas J. Wright. lbl.gov

I Tier-3 di CMS-Italia: stato e prospettive. Hassen Riahi Claudio Grandi Workshop CCR GRID 2011

Introduction, History

Co-existence: Can Big Data and Big Computation Co-exist on the Same Systems?

The glite middleware. Presented by John White EGEE-II JRA1 Dep. Manager On behalf of JRA1 Enabling Grids for E-sciencE

30 Nov Dec Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy

Grid and Cloud Activities in KISTI

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Status of KISTI Tier2 Center for ALICE

Big Data 2015: Sponsor and Participants Research Event ""

The LHC Computing Grid

Towards a Strategy for Data Sciences at UW

Tier2 Centers. Rob Gardner. University of Chicago. LHC Software and Computing Review UC San Diego Feb 7-9, 2006

The University of Oxford campus grid, expansion and integrating new partners. Dr. David Wallom Technical Manager

Grid Scheduling Architectures with Globus

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

Transitioning NCAR MSS to HPSS

Garuda : The National Grid Computing Initiative Of India. Natraj A.C, CDAC Knowledge Park, Bangalore.

Clouds at other sites T2-type computing

Clouds in High Energy Physics

Grid Computing Middleware. Definitions & functions Middleware components Globus glite

Operating the Distributed NDGF Tier-1

CISL Update. 29 April Operations and Services Division

A short introduction to the Worldwide LHC Computing Grid. Maarten Litmaath (CERN)

Early Operational Experience with the Cray X1 at the Oak Ridge National Laboratory Center for Computational Sciences

Batch Services at CERN: Status and Future Evolution

Austrian Federated WLCG Tier-2

From raw data to new fundamental particles: The data management lifecycle at the Large Hadron Collider

The Blue Water s File/Archive System. Data Management Challenges Michelle Butler

Scheduling Computational and Storage Resources on the NRP

Ioan Raicu. Everyone else. More information at: Background? What do you want to get out of this course?

Seeking Supernovae in the Clouds: A Performance Study

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

WLCG Lightweight Sites

Amazon Web Services: Performance Analysis of High Performance Computing Applications on the Amazon Web Services Cloud

Evolution of the ATLAS PanDA Workload Management System for Exascale Computational Science

Data storage services at KEK/CRC -- status and plan

Programmable Information Highway (with no Traffic Jams)

Introduction to Grid Computing

Deploying the TeraGrid PKI

ESnet Energy Sciences Network

ATLAS COMPUTING AT OU

Introducing the HTCondor-CE

Table of Contents OSG Blueprint Introduction What Is the OSG? Goals of the Blueprint Document Overview...

The Software Defined Online Storage System at the GridKa WLCG Tier-1 Center

Pan-European Grid einfrastructure for LHC Experiments at CERN - SCL's Activities in EGEE

Travelling securely on the Grid to the origin of the Universe

The Open Science Grid status and architecture

RESEARCH DATA DEPOT AT PURDUE UNIVERSITY

400G: Deployment at a National Lab

Compact Muon Solenoid: Cyberinfrastructure Solutions. Ken Bloom UNL Cyberinfrastructure Workshop -- August 15, 2005

Flying HTCondor at 100gbps Over the Golden State

irods usage at CC-IN2P3: a long history

What s new in HTCondor? What s coming? HTCondor Week 2018 Madison, WI -- May 22, 2018

Distributing storage of LHC data - in the nordic countries

Engagement With Scientific Facilities

High Throughput WAN Data Transfer with Hadoop-based Storage

First Experience with LCG. Board of Sponsors 3 rd April 2009

Cloud Computing with Nimbus

Enabling a SuperFacility with Software Defined Networking

Ivane Javakhishvili Tbilisi State University High Energy Physics Institute HEPI TSU

VC3. Virtual Clusters for Community Computation. DOE NGNS PI Meeting September 27-28, 2017

WVU RESEARCH COMPUTING INTRODUCTION. Introduction to WVU s Research Computing Services

Andrea Sciabà CERN, Switzerland

VMs at a Tier-1 site. EGEE 09, Sander Klous, Nikhef

DESY. Andreas Gellrich DESY DESY,

Advanced School in High Performance and GRID Computing November Introduction to Grid computing.

Grid Interoperation and Regional Collaboration

g-eclipse A Framework for Accessing Grid Infrastructures Nicholas Loulloudes Trainer, University of Cyprus (loulloudes.n_at_cs.ucy.ac.

Benoit DELAUNAY Benoit DELAUNAY 1

Overview of ATLAS PanDA Workload Management

Production Grids. Outline

Considerations for a grid-based Physics Analysis Facility. Dietrich Liko

High-Energy Physics Data-Storage Challenges

New Directions and BNL

International Climate Network Working Group (ICNWG) Meeting

Realtime Data Analytics at NERSC

Current Progress of Grid Project in KMA

One Pool To Rule Them All The CMS HTCondor/glideinWMS Global Pool. D. Mason for CMS Software & Computing

Outline. Infrastructure and operations architecture. Operations. Services Monitoring and management tools

The Cambridge Bio-Medical-Cloud An OpenStack platform for medical analytics and biomedical research

PROOF-Condor integration for ATLAS

The Grid: Processing the Data from the World s Largest Scientific Machine

Cloud Computing for Science

Support for multiple virtual organizations in the Romanian LCG Federation

Scientific Cluster Deployment and Recovery Using puppet to simplify cluster management

ATLAS Experiment and GCE

where the Web was born Experience of Adding New Architectures to the LCG Production Environment

Grid Computing Activities at KIT

HPC Resources & Training

Transcription:

ALICE Grid Activities in US 1

ALICE-USA Computing Project ALICE-USA Collaboration formed to focus on the ALICE EMCal project Construction, installation, testing and integration participating institutions 3 National Labs: LBNL, LLNL, ORNL 7 Universities: Creighton, Houston, Purdue, Tennessee, Texas, Wayne State, Yale ALICE-USA Computing Project was proposed to build a facility to: Enable US ALICE Scientist to pursue research goals Meet ALICE-USA computing obligation to provide its share of computing resources for ALICE data analysis and simulations Proposal settled on 2 Department of Energy (DOE) labs as primary sites Lawrence Livermore National Lab (LLNL) Lawrence Berkeley National Lab (LBNL) National Energy Research Scientific Computing (NERSC) 2

Sites: LLNL- Livermore Computing LLNL Computing Institutional based High Performance Computing Support local Science & Engineering activities (some classified) Developed Very Cost-Effective Procurement & Operations model Very large purchases of scalable units In-house managed & optimized Redhat OS and other software Security requirements make external collaborations a challenge LLNL Computing actively pursued adding more open computing facilities: Green Linux Computing Cluster Green Data Oasis serves Climate modeling data to research community Grid Activities are often a problem, but the ALICE workflow, which allows for no direct user access, fit the Security plan 3

Sites: LBNL-NERSC DOE Office of Science Flagship High Performance Scientific Computing Facility for Scientific Research Available to DOE Office of Science sponsored research Allocations via competitive awards Computing for Scientific Research Two Cray Systems: Hopper & Franklin: ~200k cores Special Clusters: HENP, Visualization, Analytics NERSC Global File System (GPFS): >1PB Archival Storage (HPSS) : ~40PB Data Transfer, Science Gateways & Grid Services Evaluation Systems: GPU & Cloud Clusters Large number of user support services Programming & benchmarking Workflow & Gateway hosting Visualization Tutorials Physics Math + CS Astrophysics Chemistry Climate Combustion Fusion Lattice Gauge Life Sciences Materials Other 4

Sites LBNL-NERSC/PDSF PDSF: High Energy Nuclear Physics Computing Facility Operated by NERSC since mid-90s for HENP community Supports both production and user analysis Nuclear Science Groups: Tier 1 facility for the STAR experiment Neutrino research: SNO, IceCube, KamLAND Physics/Astrophysics Groups: ATLAS Nearby Supernova Factory, Planck DayaBay First Grid-enabled facility within NERSC: PPDG Open Science Grid HENP Data store NERSC/HPSS Archival Storage STAR near top user group of NERSC/HPSS: ~1PB stored ALICE allocation is comparable to the STAR allocation 5

ALICE-USA Computing Project Timeline Proposal reviewed in 2008 Refined during 2009 Specific & yearly HW targets Operational milestones Grid-Enabled CPU & Storage Stress-test services Update security plans Develop resource reporting tools Integrate NERSC HPSS allocations Independent site reviews Plan Steps: Establish operation Register Both as Tier 2 Facilities Transition NERSC Site to Tier 1 Plan accepted by DOE 02/10 HW procurements begin 04/10 Sites begin operation Summer/Fall 2010 6

Facilities: LLNL/LC Single-use facility Grid-only ALICE Grid Facility Site ALICE::LLNL >~700 job-slots (cores) 13.5 khep06 650TB Disk storage Resources fixed for 3 year life-cycle No user accounts for non-llnl persons 7

Facilities: LBNL NERSC/PDSF Multi-group facility: ALICE, STAR, ATLAS, ALICE Grid Facility: Site ALICE::LBL 300-400 jobs slots (cores) ~4.0 khep06 400 TB disk storage PB-scale tape: NERSC/HPSS Resources grow each year by: ~500 cores ~500 TB disk storage ALICE User facility Local user accounts are allowed Used as a grid job-submission site ALICE Allows direct batch submission use of unused grid cycles STAR 8

Open Science Grid (OSG) US project & consortium: national, distributed computing grid for data-intensive research ~30 Participating Virtual Organization (VO) communities ~80 independent self-managed computing centers OSG provides pre-built software bundles Compute Element packages GRAM-2, GRAM-4 GRAM5 Working to provide CREAM option Storage Element packages dcache, BeStMan, Hadoop, xrootd Client Software Submit host (Condor-G ) WN-client Authorization/Authentication Tools Grid Certificate request & retrieval tools VO-discovery maps, VOMS, GUMS OSG Grid Operations Center (GOC) Site registration and availability Resource discovery repositories Accounting/Job statistic reports trouble ticket management 9

ALICE & OSG Open Science Grid & WLCG US ATLAS & CMS major stakeholders Works with WLCG for VOs Provides usage & capacity reports DOEGrids CA partner: OSG Registration Auth. ALICE joined OSG as a VO NERSC/PDSF was already an OSG site LLNL Job Accounting is reported to OSG OSG Accounts from NERSC & LLNL Full LLNL integration is in the ALICE-USA Computing Project Plan Accounting records for ALICE jobs at both sites are being reported to the WLCG via OSG standard reports 10

OSG Software Deployment Model Software is packaged by VDT OSG Validation Test Bed (VTB) 2-3 small sites (1-3 nodes each) Test software deployment & config OSG Integration Test Bed (ITB) Small clusters (4-8 nodes) operated at larger sites Admins deploy software updates & test functionality VOs test workflows on new/updated software Allows New VOs to test building distributed workflows OSG is working to provide a CREAM-CE option LBNL is a participating OSG-ITB site ALICE will run site at LBNL-ITB as part of that integration work OSG Integration Test bed

OSG Interoperability & Outreach OSG active interoperability efforts EGEE/EGI/WLCG US National Grids (TeraGrid) Campus & Regional Grids non-us National Grid Initiatives National Grids in the Americas OSG members collaborate directly osg-americas@opensciencegrid.org Previous workshops & Grid Schools: Costa Rica, Columbia, & Brasil CLCAR 2010 Panel discussion Support National Grid Initiatives with goal of providing International Grid Infrastructure 12

Other US ALICE Grid Sites & Activities Ohio Supercomputing Center, Glenn Cluster ALICE-Ohio State U. Group ~150 cores 30TB Disk Storage University of Houston Planned return to operation ALICE-US Data Transfer Group ESNet, NERSC, LLNL, OSC, OARNet, CERN, ALICE Previous efforts to optimize targeted networks ESNet-KISTI-BNL NERSC-ANL-ORNL, NERSC data transfer group http://fasterdata.es.net 13