Scientific Computing at SLAC. Chuck Boeheim Asst Director: Scientific Computing and Computing Services

Similar documents
Scientific Computing at SLAC

Computing. Richard P. Mount, SLAC. Director, SLAC Computing Services Assistant Director, Research Division. DOE Review

High-Energy Physics Data-Storage Challenges

CERN openlab II. CERN openlab and. Sverre Jarp CERN openlab CTO 16 September 2008

Storage and I/O requirements of the LHC experiments

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

CSCS CERN videoconference CFD applications

ISTITUTO NAZIONALE DI FISICA NUCLEARE

Storage Resource Sharing with CASTOR.

CERN s Business Computing

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

Overview. About CERN 2 / 11

Insight: that s for NSA Decision making: that s for Google, Facebook. so they find the best way to push out adds and products

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Customer Success Story Los Alamos National Laboratory

Database Services at CERN with Oracle 10g RAC and ASM on Commodity HW

Virtualizing a Batch. University Grid Center

IEPSAS-Kosice: experiences in running LCG site

CC-IN2P3: A High Performance Data Center for Research

Scientific Computing at SLAC. Amber Boehnlein

Representing Range Compensators in the TOPAS Monte Carlo System

Andrea Sciabà CERN, Switzerland

CouchDB-based system for data management in a Grid environment Implementation and Experience

The creation of a Tier-1 Data Center for the ALICE experiment in the UNAM. Lukas Nellen ICN-UNAM

ICN for Cloud Networking. Lotfi Benmohamed Advanced Network Technologies Division NIST Information Technology Laboratory

Designing a Multi-petabyte Database for LSST

Existing Tools in HEP and Particle Astrophysics

Network Support for Data Intensive Science

Intel Solid State Drive Data Center Family for PCIe* in Baidu s Data Center Environment

Reliability Engineering Analysis of ATLAS Data Reprocessing Campaigns

Managing Petabytes of data with irods. Jean-Yves Nief CC-IN2P3 France

BaBar Computing: Technologies and Costs

Storage on the Lunatic Fringe. Thomas M. Ruwart University of Minnesota Digital Technology Center Intelligent Storage Consortium

Center for Advanced Computing Research

Grid Computing at the IIHE

Compact Muon Solenoid: Cyberinfrastructure Solutions. Ken Bloom UNL Cyberinfrastructure Workshop -- August 15, 2005

Big Data Analytics and the LHC

SLAC Participation in ATLAS Su Dong

Big Computing and the Mitchell Institute for Fundamental Physics and Astronomy. David Toback

Harnessing Grid Resources to Enable the Dynamic Analysis of Large Astronomy Datasets

Company. Intellectual Property. Headquartered in the Silicon Valley

Netherlands Institute for Radio Astronomy. May 18th, 2009 Hanno Holties

Beyond Petascale. Roger Haskin Manager, Parallel File Systems IBM Almaden Research Center

Gigabyte Bandwidth Enables Global Co-Laboratories

CANARIE: Providing Essential Digital Infrastructure for Canada

Data Transfers Between LHC Grid Sites Dorian Kcira

where the Web was born Experience of Adding New Architectures to the LCG Production Environment

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

Scaling Internet TV Content Delivery ALEX GUTARIN DIRECTOR OF ENGINEERING, NETFLIX

Storage Optimization with Oracle Database 11g

Data Analysis in Experimental Particle Physics

MOHA: Many-Task Computing Framework on Hadoop

CC-IN2P3 / NCSA Meeting May 27-28th,2015

Sun Lustre Storage System Simplifying and Accelerating Lustre Deployments

The LCG 3D Project. Maria Girone, CERN. The 23rd Open Grid Forum - OGF23 4th June 2008, Barcelona. CERN IT Department CH-1211 Genève 23 Switzerland

New strategies of the LHC experiments to meet the computing requirements of the HL-LHC era

Application-Transparent Checkpoint/Restart for MPI Programs over InfiniBand

Giovanni Lamanna LAPP - Laboratoire d'annecy-le-vieux de Physique des Particules, Université de Savoie, CNRS/IN2P3, Annecy-le-Vieux, France

Scientific Instrumentation using NI Technology

Tackling tomorrow s computing challenges today at CERN. Maria Girone CERN openlab CTO

PARALLEL PROCESSING OF LARGE DATA SETS IN PARTICLE PHYSICS

Data transfer over the wide area network with a large round trip time

VIRTUAL OBSERVATORY TECHNOLOGIES

A New NSF TeraGrid Resource for Data-Intensive Science

Evaluation of the computing resources required for a Nordic research exploitation of the LHC

The CMS Computing Model

Offline Tutorial I. Małgorzata Janik Łukasz Graczykowski. Warsaw University of Technology

e-research Infrastructures for e-science Axel Berg SARA national HPC & e-science support center RAMIRI, June 15, 2011

Green Supercomputing

Unified Management for Virtual Storage

Towards a Strategy for Data Sciences at UW

Summary of Data Management Principles

Data storage services at KEK/CRC -- status and plan

SQL Server 2017 for your Mission Critical applications

irods usage at CC-IN2P3: a long history

Programmable Information Highway (with no Traffic Jams)

Top Trends in DBMS & DW

Data Movement & Storage Using the Data Capacitor Filesystem

<Insert Picture Here> Oracle Storage

Improving Network Infrastructure to Enable Large Scale Scientific Data Flows and Collaboration (Award # ) Klara Jelinkova Joseph Ghobrial

General introduction of version 10.4 and prospects

Introduction to Grid Computing

Some Reflections on Advanced Geocomputations and the Data Deluge

A High-Performance Storage and Ultra- High-Speed File Transfer Solution for Collaborative Life Sciences Research

CC-IN2P3 activity. irods in production: irods developpements in Lyon: SRB to irods migration. Hardware setup. Usage. Prospects.

System upgrade and future perspective for the operation of Tokyo Tier2 center. T. Nakamura, T. Mashimo, N. Matsui, H. Sakamoto and I.

Geant4 v9.5. Kernel III. Makoto Asai (SLAC) Geant4 Tutorial Course

Volunteer Computing at CERN

Introduction to Data Management CSE 344

D0 Grid: CCIN2P3 at Lyon

The INFN Tier1. 1. INFN-CNAF, Italy

Travelling securely on the Grid to the origin of the Universe

Data oriented job submission scheme for the PHENIX user analysis in CCJ

<Insert Picture Here> MySQL Web Reference Architectures Building Massively Scalable Web Infrastructure

Overview: Physically- Based Simulation of Single-Event Effects

Coming Changes in Storage Technology. Be Ready or Be Left Behind

ATLAS Experiment and GCE

Advanced School in High Performance and GRID Computing November Introduction to Grid computing.

Andy Kowalski Ian Bird, Bryan Hess

A scalable storage element and its usage in HEP

Transcription:

Scientific Computing at SLAC Chuck Boeheim Asst Director: Scientific Computing and Computing Services 1

Drivers for SLAC Computing Computing to enable today s dataintensive science clusters, interconnects, networks, mass storage,... Computing research to prepare for tomorrow s challenges massive memory, low latency, petascale databases, detector simulation,... 2 2

SLAC Computing Fabric Client Client Client Client Client Client 1700 dual CPU Linux IP Network (Cisco) XROOTD, Objectivity, NFS, AFS, etc. Disk Server Disk Server Disk Server Disk Server Disk Server Disk Server 275 Servers ~700 TB FibreChannel and SATA RAID arrays IP Network (Cisco) HPSS + SLAC enhancements to ROOT and Objectivity server code Tape Server Tape Server Tape Server Tape Server Tape Server 25 dual CPU 40 STK 9940B 6 STK 9840A 6 STK Powderhorn over 2 PB of data 3 3

Storage Growth File Servers 175 200 225 275 700 525 350 GLAST, KIPAC & Others BaBar 175 Jan-03Jul-03 Jan-04 Jul-04 Jan-05 Jul-05 Jan-06 Jul-06 0 Other 4 4

Computing Growth (and Diversification) Number of Nodes 3,000 2003 2004 2005 2006 Retired SpecInt2000 New Clusters 4,500,000 2,250 3,375,000 1,500 2,250,000 750 1,125,000 0 0 2003 2004 2005 2006 Clusters 5 5

Growth and Diversification Continue shared cluster growth as much as practical Increasing MPI (Parallel) capacity and support (Astro, Accelerator, and more) Grid interfaces and support (Atlas et.al.) Large SMPs (Astro) Visualization 6 6

Research - Petacache 7

PetaCache Goals The PetaCache architecture aims at revolutionizing the query and analysis of scientific databases with complex structure. Generally this applies to feature databases (terabytes petabytes) rather than bulk data (petabytes exabytes) The original motivation comes from HEP Sparse (~random) access to tens of terabytes today, petabytes tomorrow Access by thousands of processors today, tens of thousands tomorrow 8

A Decade s Progress in Access Times 9 9

Latency (1) Ideal Memory Client Application 10

Latency (2) Current reality Client Application Data Server Data-Server-Client OS OS OS File System TCP Stack TCP Stack Dis NIC NIC Network Switches 11

Latency (3) Practical Goal Memory Client Application Data Server Data-Server-Client OS OS OS File System TCP Stack TCP Stack Dis NIC NIC Network Switches 12

Prototype Machine Cisco Switch Clients up to 2000 Nodes, each 2 CPU, 2 GB memory Linux Data-Servers 64 Nodes, each Sun V20z, 2 Opteron CPU, 16 GB memory 1TB total Memory Solaris or Linux (mix and match) Cisco Switches PetaCache MICS + HEP- BaBar Funding Existing HEP- Funded BaBar Systems 13

Object-Serving Software Scala (Xrootd/olbd) (Andy Hanushevsky/SLAC) Optimized for read-only access File-access paradigm (filename, offset, bytecount) Make 1000s of servers transparent to user code Load balancing Self-organizing Automatic staging from tape Failure recovery Allows BaBar to start getting benefit from a new data-access architecture within months without changes to user code The application can ignore the hundreds of separate address spaces in the data-cache memory 14

Proposals from SLAC Science BaBar Host part of the (~30 TB) microdst data Access data via pointer skims GLAST Will require a ~2TB intensely accessed database. Have asked to test concepts on a PetaCache Prototype LSST Database Prototyping Proposed tests using a PetaCache prototype 15

Summary Data-intensive science increasingly requires low-latency access to terabytes or petabytes Memory is one key: Commodity DRAM today (increasing total cost by ~2x) Storage-class memory (whatever that will be) in the future Revolutions in scientific data analysis will be another key Current HEP approaches to data analysis assume that random access is prohibitively expensive As a result, permitting random access brings much-less-thanrevolutionary immediate benefit Use the impressive motive force of a major HEP collaboration with huge data-analysis needs to drive the development of techniques for revolutionary exploitation of an above-threshold machine. 16

Research - Very Large Databases 17

SCCS & Very Large Databases 10-year, unique experience with VLDB designing, building, deploying & managing peta-scale production data sets / database world s largest database, 1.4 PB innovative approaches to managing BaBar s data proposed, designed and developed by SCCS SCCS leverages VLDB experience to assist LSST in solving data-related challenges effort started ~end of 2004 18 18

LSST: Data Related Challenges Large volumes 7 PB/year (image and catalog data) 500 TB/year (database) Most today s very large databases : ~10s TB range High availability petabytes 10s of 1000s of disks daily disk failures Real-time requirement Transient alerts generated in < 60 sec Spatial & temporal aspects Most surveys focus on a single dimension All data made public with minimal delay Wide range of users: professional & amateur astronomers, students, general public 19 19

SCCS Role in LSST Data Management Current effort Architecting data access system Database technology choice, design and prototyping Coordinating 2 LSST teams Data Access Working Group Database Group ~12 members from SLAC, NCSA, U. Washington, Johns Hopkins U., U. Arizona, LLNL, SDSC SCCS leads Data Access and Database related effort 20 20

More Details on VLDB Work by SCCS Prototyping at SCCS to uncover peta-scale limitations, bottlenecks & deficiencies, understand needed improvements current focus on MySQL, discussing / planning DB2 tests with IBM Close collaboration with key MySQL developers to develop tools and techniques for supporting petascale databases work include large scale tests, providing input and feature requests Working closely with world-class database gurus includes face-to-face discussions, frequent email exchanges and co-writing papers 21 21

Growing Visibility of SCCS in Astronomical Community Actively working with astronomical community on developing standards working with 4 large astronomical surveys (LSST, SDSS, PanSTARRS, 2MASS) on spatial / GIS 1 support by RDBMSes providing input to NVO 2 Involved in reviewing computing systems of other astronomical surveys Presented at recognized astronomical conferences SPIE Astronomical Telescopes and Instrumentation American Astronomical Society 1 Geographic Information Systems 2 National Virtual Observatory 22 22

Research - Geant4 23

SLAC Geant4 team Geant4 is a toolkit simulating elementary particles passing through and interacting with matter, and modeling the detector apparatus measuring the passage of elementary particles and recording the energy and dose deposition. Geant4 is developed and maintained by an international collaboration. SLAC is the second largest center next to CERN. 24

SLAC Geant4 Team The SLAC Geant4 team is strongly involved in the development and maintenance of Geant4 toolkit. Makoto Asai Deputy-spokesperson, coordinator of Global Architecture, Run&Event and Detector Response WGs Dennis Wright Coordinators of Hadronics and Documentation WGs Joseph Perl Coordinator of Visualization WG Tatsumi Koi Major developer of neutron physics Jane Tinsley Major developer of event biasing options and visualization 25 25

User support and technology transfer BaBar, the pioneer experiment, adopted Geant4 in 2000. They have simulated more than 10 10 events so far at 20 sites in US and Europe. All experimental programs at SLAC including GLAST, EXO and ILC detector study rely on Geant4. All of four LHC experiments at CERN and most of other experiments in US and in the world have already adopted it. 26 26

User support and technology transfer Use of Geant4 not limited to high energy and nuclear physics. Users in space engineering, medical and industrial applications are rapidly expanding. Geant4 is an excellent showcase of technology transfer from particle physics to other fields. The SLAC Geant4 team supports Geant4 for all experiments at SLAC. It also acts as the support center for all other DOE Labs and other users in US. 27

Current Uses of Geant4 BaBar PEP Beamline Beam Transport ATLAS ILC LCD CMS Astro LHCb Medical GLAST LAT EXO 28 28

BaBar BaBar at SLAC is the pioneer experiment in HEP in use of Geant4 Started in 2000 Simulated 10 10 events so far Produced at 20 sites in North America and Europe Current average production rate 6.1 x 10 7 events/week Now simulating PEP beam line as well (-9m < zip < 9m) Courtesy of D.Wright (SLAC) 29

30

Courtesy of F.Longo (INFN/Trieste) 31

Geant4 simulation of EXO Enriched Xenon Observatory for double beta decay Full detector view With Pb shielding Courtesy of J-L.Vuilleumier (Neuchatel) and M.Breidenbach (SLAC) 32

Geant4 for beam transportation Courtesy of V.D.Elvira (FNAL) 33

ILC For International Linear Collider experiment Full Support of SLAC Geant4 Team Courtesy of N.Graf (SLAC) 34

Smart- INTEGRAL AC Cassin LIS Bepi Colombo GLAS Herschel SWIF Astro-E2 XMM-Newton GAI JWS EUSO ISS Columbus AMS MAX 35

Simulation of Single Event Effects on SRAM in space radiation environments 36

GEANT4 based proton dose calculation in a clinical environment: technical aspects, strategies and challenges Harald Paganetti 37

38

Tutorial courses 39

http://top25.sciencedirect.com/index.php?subject_area_id=21 40

http://www.in-cites.com/hotpapers/2004/november04-eng.html http://www.in-cites.com/hotpapers/2005/jan05-eng.html http://www.in-cites.com/hotpapers/2005/mar05-eng.html http://www.in-cites.com/hotpapers/2005/may05-eng.html http://www.in-cites.com/hotpapers/2005/july05-eng.html 41