Benoit DELAUNAY Benoit DELAUNAY 1

Similar documents
LCG data management at IN2P3 CC FTS SRM dcache HPSS

SAM at CCIN2P3 configuration issues

The INFN Tier1. 1. INFN-CNAF, Italy

irods usage at CC-IN2P3: a long history

and the GridKa mass storage system Jos van Wezel / GridKa

A scalable storage element and its usage in HEP

irods usage at CC-IN2P3 Jean-Yves Nief

CC-IN2P3 activity. irods in production: irods developpements in Lyon: SRB to irods migration. Hardware setup. Usage. Prospects.

D0 Grid: CCIN2P3 at Lyon

Storage Resource Sharing with CASTOR.

IBM IBM Open Systems Storage Solutions Version 4. Download Full Version :

Storage Supporting DOE Science

Data Transfers Between LHC Grid Sites Dorian Kcira

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. Presented by Manfred Alef Contributions of Jos van Wezel, Andreas Heiss

Data storage services at KEK/CRC -- status and plan

Storage and I/O requirements of the LHC experiments

The Software Defined Online Storage System at the GridKa WLCG Tier-1 Center

Austrian Federated WLCG Tier-2

High-density Grid storage system optimization at ASGC. Shu-Ting Liao ASGC Operation team ISGC 2011

CMS Belgian T2. G. Bruno UCL, Louvain, Belgium on behalf of the CMS Belgian T2 community. GridKa T1/2 meeting, Karlsruhe Germany February

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Influence of Distributing a Tier-2 Data Storage on Physics Analysis

Spanish Tier-2. Francisco Matorras (IFCA) Nicanor Colino (CIEMAT) F. Matorras N.Colino, Spain CMS T2,.6 March 2008"

Providing a first class, enterprise-level, backup and archive service for Oxford University

IBM System Storage TS1130 Tape Drive Models E06 and other features enhance performance and capacity

Edinburgh (ECDF) Update

IBM System Storage. Tape Library. A highly scalable, tape solution for System z, IBM Virtualization Engine TS7700 and Open Systems.

I Tier-3 di CMS-Italia: stato e prospettive. Hassen Riahi Claudio Grandi Workshop CCR GRID 2011

Table 9. ASCI Data Storage Requirements

High-Energy Physics Data-Storage Challenges

Parallel File Systems. John White Lawrence Berkeley National Lab

Understanding StoRM: from introduction to internals

IBM ProtecTIER and Netbackup OpenStorage (OST)

irods at TACC: Secure Infrastructure for Open Science Chris Jordan

SAS Grid Architecture Solution Using IBM Hardware

Long Term Data Preservation for CDF at INFN-CNAF

HPSS RAIT. A high performance, resilient, fault-tolerant tape data storage class. 1

GPFS on a Cray XT. Shane Canon Data Systems Group Leader Lawrence Berkeley National Laboratory CUG 2009 Atlanta, GA May 4, 2009

System upgrade and future perspective for the operation of Tokyo Tier2 center. T. Nakamura, T. Mashimo, N. Matsui, H. Sakamoto and I.

Symantec Design of DP Solutions for UNIX using NBU 5.0. Download Full Version :

Constant monitoring of multi-site network connectivity at the Tokyo Tier2 center

Site Report. Stephan Wiesand DESY -DV

Parallel Storage Systems for Large-Scale Machines

Solutions for iseries

ALICE Grid Activities in US

Introduction Data Management Jan Just Keijser Nikhef Grid Tutorial, November 2008

SMART SERVER AND STORAGE SOLUTIONS FOR GROWING BUSINESSES

Scaling a Global File System to the Greatest Possible Extent, Performance, Capacity, and Number of Users

Overcoming Obstacles to Petabyte Archives

HPSS Development Update and Roadmap 2017

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

The Legnaro-Padova distributed Tier-2: challenges and results

Technical Requirements

NCP Computing Infrastructure & T2-PK-NCP Site Update. Saqib Haleem National Centre for Physics (NCP), Pakistan

Compact Muon Solenoid: Cyberinfrastructure Solutions. Ken Bloom UNL Cyberinfrastructure Workshop -- August 15, 2005

IBM System i Model 515 offers new levels of price performance

Storage Evaluations at BNL

Ekran System System Requirements and Performance Numbers

Status of KISTI Tier2 Center for ALICE

Computer Science Section. Computational and Information Systems Laboratory National Center for Atmospheric Research

Server Specifications

Opportunities A Realistic Study of Costs Associated

dcache tape pool performance Niklas Edmundsson HPC2N, Umeå University

Analisi Tier2 e Tier3 Esperienze ai Tier-2 Giacinto Donvito INFN-BARI

Oracle 10g on Solaris to Oracle RAC 11gR2 on Linux Upgrade

Exam : Title : High-End Disk for Open Systems V2. Version : DEMO

Scientific Computing at SLAC

where the Web was born Experience of Adding New Architectures to the LCG Production Environment

Experiences with a custom High Availability solution

A Simple Mass Storage System for the SRB Data Grid

PART-I (B) (TECHNICAL SPECIFICATIONS & COMPLIANCE SHEET) Supply and installation of High Performance Computing System

A data Grid testbed environment in Gigabit WAN with HPSS

DVS, GPFS and External Lustre at NERSC How It s Working on Hopper. Tina Butler, Rei Chi Lee, Gregory Butler 05/25/11 CUG 2011

Data Movement & Tiering with DMF 7

Exam Name: XIV Replication and Migration V1

The LCG 3D Project. Maria Girone, CERN. The 23rd Open Grid Forum - OGF23 4th June 2008, Barcelona. CERN IT Department CH-1211 Genève 23 Switzerland

Metadaten Workshop 26./27. März 2007 Göttingen. Chimera. a new grid enabled name-space service. Martin Radicke. Tigran Mkrtchyan

SPINOSO Vincenzo. Optimization of the job submission and data access in a LHC Tier2

Data oriented job submission scheme for the PHENIX user analysis in CCJ

StorNext Supported Operating Systems and Platforms

Data transfer over the wide area network with a large round trip time

ArcExplorer -- Java Edition 9.0 System Requirements

Netherlands Institute for Radio Astronomy. May 18th, 2009 Hanno Holties

Dimensioning storage and computing clusters for efficient high throughput computing

ISTITUTO NAZIONALE DI FISICA NUCLEARE

Current Status of the Ceph Based Storage Systems at the RACF

Guillimin HPC Users Meeting February 11, McGill University / Calcul Québec / Compute Canada Montréal, QC Canada

UW-ATLAS Experiences with Condor

LCE: Lustre at CEA. Stéphane Thiell CEA/DAM

10 Gbit/s Challenge inside the Openlab framework

NetBackup SAN Client and Fibre Transport Troubleshooting Guide. 2 What are the components of the SAN Client feature?

The General Parallel File System

The creation of a Tier-1 Data Center for the ALICE experiment in the UNAM. Lukas Nellen ICN-UNAM

Grid Computing Activities at KIT

An introduction to GPFS Version 3.3

IBM řešení pro větší efektivitu ve správě dat - Store more with less

<Insert Picture Here> Oracle Storage

Monitoring system for geographically distributed datacenters based on Openstack. Gioacchino Vino

Study of the viability of a Green Storage for the ALICE-T1. Eduardo Murrieta Técnico Académico: ICN - UNAM

T e c h n o l o g y. LaserTAPE: The Future of Storage

Transcription:

Benoit DELAUNAY 20091023 Benoit DELAUNAY 1

CC-IN2P3 provides computing and storage for the 4 LHC experiments and many others (astro particles...) A long history of service sharing between experiments Some dedicated resources for LHC experiments Benoit DELAUNAY 2

USER Frontend storage DCACHE RFIO XROOTD irods SRB GPFS Backend storage HPSS Benoit DELAUNAY 3

6 engineers involved in the management of the storage middleware 3 engineers for dcache Lionel SCHWARZ, Jonathan SCHAEFFER, Yvan CALAS 3 engineers for HPSS Pierre-Emmanuel BRINETTE, Andres GOMEZ, Benoit DELAUNAY Benoit DELAUNAY 4

DCACHE Disk storage Access layer RFIO HPSS Disk Storage Tape Storage Benoit DELAUNAY 5

4 Master Servers Scientific Linux 4 16GB memory 8 CPU cores 79 x4540 disk servers SUN Solaris 10 32TB disk storage 2 Gbps Nework Interface Storage: 2500TB Network: 158Gbps Benoit DELAUNAY 6

1 Master Server IBM AIX 5.3 64GB memory 16 CPU cores 27 tape data movers IBM AIX 5.3 Mixed 2Gbps/10Gbps Bandwidth: 70Gbps Benoit DELAUNAY 12 disk data movers RedHat Enterprise Linux 4 40TB disk storage 10Gbps Nework Interface Storage: 480TB Network: 120Gbps 3 libraries STK SL8500 10,000 slots each 13 x 9840 tape drives (20GB) 36 x T10KA tape drives (500GB) 32 x T10KB tape drives (1TB) Max Capacity 30 PB 7

dcache used as a end user storage system and also for data exchange between LCG sites Used for CMS Tier1 and Tier2 at CC-IN2P3 Only one instance for the 4 LHC experiments Storage pools are dedicated to experiments Benoit DELAUNAY 8

Upgrade from dcache v1.9.0 to v1.9.4 (2009/09/22) Migrate from PNFS to CHIMERA (2009/09/28) Complete shutdown during 3 days First observations show that access to metadata has been improved Many storage servers engaged Benoit DELAUNAY 9

39 buffer pools for a total of 661TB T1 / Local read buffer : 360TB T1 / Input buffer : 68TB T1 / Output transfer buffer : 28TB T2 / 205 TB File size distribution since june 2009 : Benoit DELAUNAY 10

HPSS used as a backend storage for dcache System use not dedicated to LHC experiments, but dedicated storage resources 1 logical instance (subsystem) of HPSS for CMS Means dedicated disks and tapes resources CMS data do not share HPSS disks and tapes with others Benoit DELAUNAY 11

Major software upgrade to the version 6.2.2.2 in june 2009 (complete shutdown during 4 days) Hardware platform has been almost totally replaced A new master server and new data movers New tape drives STK T10KB (1TB/tape) HPSS is now more reliable, powerful and capacitive. Benoit DELAUNAY 12

File size base storage policy 4 Class of Service COS10 : small files 0-64MB COS11 : medium files 64MB-512MB COS12 : large files 512MB-8GB COS14 : XL files 8GB-128GB Larger is the file, more powerful is HPSS! Benoit DELAUNAY 13

80 TB allocated disk 1.3 PB on tapes 986,000 files but, 523,000 never read (53%)! Small files (3%) File size distribution : Medium files (1%) Large files (93%) XL files (2%) Benoit DELAUNAY 14

dcache requests file staging one by one HPSS has a very basic behavior when reading files Handles file read requests one by one in the order they were submitted (FIFO) Could be very unefficient when the file lists is disordered and many files are stored on the same tapes A solution for that, submit ordered file list by tapes to HPSS using T-ReqS (Tape Request Scheduler) Benoit DELAUNAY 15

1. dcache asks for n files one by one to be read on m tapes (m<n) 2. HPSS stages files in disorder from tapes to disks (n tape mounts!) DCACHE RFIO HPSS 3 1 3. dcache gets files from HPSS disks via RFIO one by one 2 Benoit DELAUNAY 16

1. dcache asks for n files to be read on m tapes (m<n) 2. T-ReqS reorders file requests by tapes 3. T-ReqS requests the file staging to HPSS 5 RFIO DCACHE 1 2 T-ReqS 3 HPSS 4. HPSS stages files from tapes to disks (m tape mounts!) 4 5. dcache gets files from HPSS disks via RFIO Benoit DELAUNAY 17

dcache upgrade to v1.9.5 (Golden release) on november 2009 HPSS upgrade to v7.x on june 2010 More disks and more tapes A fourth SL8500 in january 2010 (10,000 tape slots + T10KB drives) => tape storage capacity extended to 40PB Call of tender for new disk servers at the end of 2009 Benoit DELAUNAY 18

A lot of work has been done this past year On the hardware infrastructure On the software (dcache and HPSS) We will be still quite busy the next year! We are confident in the ability of the storage system to cope with the LHC experiment requirements. Benoit DELAUNAY 19

Thank you! Benoit DELAUNAY 20