The Software Defined Online Storage System at the GridKa WLCG Tier-1 Center

Similar documents
Grid Computing Activities at KIT

Resilient FTS3 service at GridKa

Extraordinary HPC file system solutions at KIT

CLOUDS OF JINR, UNIVERSITY OF SOFIA AND INRNE JOIN TOGETHER

Lessons learned from Lustre file system operation

Challenges in making Lustre systems reliable

An ESS implementation in a Tier 1 HPC Centre

DDN. DDN Updates. DataDirect Neworks Japan, Inc Nobu Hashizume. DDN Storage 2018 DDN Storage 1

A ClusterStor update. Torben Kling Petersen, PhD. Principal Architect, HPC

Storage for HPC, HPDA and Machine Learning (ML)

Analyzing the High Performance Parallel I/O on LRZ HPC systems. Sandra Méndez. HPC Group, LRZ. June 23, 2016

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. Presented by Manfred Alef Contributions of Jos van Wezel, Andreas Heiss

Using file systems at HC3

SPINOSO Vincenzo. Optimization of the job submission and data access in a LHC Tier2

DDN. DDN Updates. Data DirectNeworks Japan, Inc Shuichi Ihara. DDN Storage 2017 DDN Storage

I Tier-3 di CMS-Italia: stato e prospettive. Hassen Riahi Claudio Grandi Workshop CCR GRID 2011

Status of KISTI Tier2 Center for ALICE

The National Analysis DESY

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

Andrea Sciabà CERN, Switzerland

Benoit DELAUNAY Benoit DELAUNAY 1

CERN Tape Archive (CTA) :

WLCG Transfers Dashboard: a Unified Monitoring Tool for Heterogeneous Data Transfers.

Cold Storage: The Road to Enterprise Ilya Kuznetsov YADRO

The LCG 3D Project. Maria Girone, CERN. The 23rd Open Grid Forum - OGF23 4th June 2008, Barcelona. CERN IT Department CH-1211 Genève 23 Switzerland

Data Management. Parallel Filesystems. Dr David Henty HPC Training and Support

Current Status of the Ceph Based Storage Systems at the RACF

All-Flash High-Performance SAN/NAS Solutions for Virtualization & OLTP

All-Flash High-Performance SAN/NAS Solutions for Virtualization & OLTP

NST6000 UNIFIED HYBRID STORAGE. Performance, Availability and Scale for Any SAN and NAS Workload in Your Environment

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Design Concepts & Capacity Expansions of QNAP RAID 50/60

New Storage Technologies First Impressions: SanDisk IF150 & Intel Omni-Path. Brian Marshall GPFS UG - SC16 November 13, 2016

Lenovo Database Configuration Guide

CERN European Organization for Nuclear Research, 1211 Geneva, CH

and the GridKa mass storage system Jos van Wezel / GridKa

Assistance in Lustre administration

Cluster Setup and Distributed File System

Spanish Tier-2. Francisco Matorras (IFCA) Nicanor Colino (CIEMAT) F. Matorras N.Colino, Spain CMS T2,.6 March 2008"

All-Flash High-Performance SAN/NAS Solutions for Virtualization & OLTP

The INFN Tier1. 1. INFN-CNAF, Italy

ATLAS operations in the GridKa T1/T2 Cloud

Data storage services at KEK/CRC -- status and plan

High-Performance Lustre with Maximum Data Assurance

Next Generation Computing Architectures for Cloud Scale Applications

InfiniBand Networked Flash Storage

Data Transfers Between LHC Grid Sites Dorian Kcira

Xyratex ClusterStor6000 & OneStor

The Legnaro-Padova distributed Tier-2: challenges and results

LHCb Computing Resources: 2018 requests and preview of 2019 requests

Experiences using a multi-tiered GPFS file system at Mount Sinai. Bhupender Thakur Patricia Kovatch Francesca Tartagliogne Dansha Jiang

IBM Spectrum Scale vs EMC Isilon for IBM Spectrum Protect Workloads

Distributing storage of LHC data - in the nordic countries

Data Analytics and Storage System (DASS) Mixing POSIX and Hadoop Architectures. 13 November 2016

System upgrade and future perspective for the operation of Tokyo Tier2 center. T. Nakamura, T. Mashimo, N. Matsui, H. Sakamoto and I.

Seagate ExaScale HPC Storage

Tier-2 structure in Poland. R. Gokieli Institute for Nuclear Studies, Warsaw M. Witek Institute of Nuclear Physics, Cracow

IBM Spectrum NAS. Easy-to-manage software-defined file storage for the enterprise. Overview. Highlights

Austrian Federated WLCG Tier-2

Dell PowerVault MD Family. Modular storage. The Dell PowerVault MD storage family

Dell DL1300 Appliance Interoperability Guide

RAIDIX Data Storage Solution. Clustered Data Storage Based on the RAIDIX Software and GPFS File System

LHCb Computing Resources: 2019 requests and reassessment of 2018 requests

EMC Integrated Infrastructure for VMware. Business Continuity

HIGH-PERFORMANCE STORAGE FOR DISCOVERY THAT SOARS

ALICE Grid Activities in US

Architecting Storage for Semiconductor Design: Manufacturing Preparation

Data center requirements

FLASHARRAY//M Business and IT Transformation in 3U

VMware vsan Ready Nodes

AFM Use Cases Spectrum Scale User Meeting

White Paper. EonStor GS Family Best Practices Guide. Version: 1.1 Updated: Apr., 2018

The Oracle Database Appliance I/O and Performance Architecture

Reconstruyendo una Nube Privada con la Innovadora Hiper-Convergencia Infraestructura Huawei FusionCube Hiper-Convergente

IBM ProtecTIER and Netbackup OpenStorage (OST)

CSCS HPC storage. Hussein N. Harake

Overview. About the Cisco UCS S3260 System

Copyright 2012, Oracle and/or its affiliates. All rights reserved.

DESY. Andreas Gellrich DESY DESY,

Summary of the LHC Computing Review

Isilon Scale Out NAS. Morten Petersen, Senior Systems Engineer, Isilon Division

Mellanox InfiniBand Solutions Accelerate Oracle s Data Center and Cloud Solutions

Increasing Performance of Existing Oracle RAC up to 10X

High Capacity network storage solutions

Nový IBM Storwize V7000 Unified block-file storage system Simon Podepřel Storage Sales 2011 IBM Corporation

Data Protection for Cisco HyperFlex with Veeam Availability Suite. Solution Overview Cisco Public

LHCb Computing Resource usage in 2017

Disaster recovery of the INFN Tier-1 data center: lesson learned. Luca dell Agnello INFN - CNAF CHEP 2018 conference Sofia, July

Building Self-Healing Mass Storage Arrays. for Large Cluster Systems

IT Certification Exams Provider! Weofferfreeupdateserviceforoneyear! h ps://

Solaris Engineered Systems

Beyond Petascale. Roger Haskin Manager, Parallel File Systems IBM Almaden Research Center

Data Access and Data Management

Sun Lustre Storage System Simplifying and Accelerating Lustre Deployments

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

NCP Computing Infrastructure & T2-PK-NCP Site Update. Saqib Haleem National Centre for Physics (NCP), Pakistan

Intel Enterprise Edition Lustre (IEEL-2.3) [DNE-1 enabled] on Dell MD Storage

Dell PowerVault MD Family. Modular Storage. The Dell PowerVault MD Storage Family

Computing for LHC in Germany

I/O at the Center for Information Services and High Performance Computing

Transcription:

The Software Defined Online Storage System at the GridKa WLCG Tier-1 Center CHEP 2018, Sofia Jan Erik Sundermann, Jolanta Bubeliene, Ludmilla Obholz, Andreas Petzold STEINBUCH CENTRE FOR COMPUTING (SCC) KIT The Research University in the Helmholtz Association www.kit.edu

GridKa Online Storage System New storage system in production since 2017 Storage system design allows for scalability both in size and performance Spectrum Scale (v4.2) software defined storage Few large file systems for better scalability and manageability Allows to adapt to various use cases (4 experiments, tape buffer, ) Total capacity: 34PB 2 July 2018

Technical Design NEC GxFS Storage Appliance with 22 redundant storage servers 90 disk enclosures 5400 HDDs for data (8+10TB) 58 SSDs (1.6 TB) in separate enclosures for file system metadata Two redundant IB fabrics (Mellanox 56G FDR) 64 protocol servers with 40GE 8x 100GE uplink to GridKa backbone 3 July 2018

Storage Building Blocks (SBB) Each SBB with 2 servers (NEC Express 5800 R120g-2M) Redundantly connected to 2x5 disk enclosures (NetApp E5600 / DE6600) 60 disks per enclosure (8TB / 10TB NL SAS) Setup in Dynamic Disk Pools (DDP) with 50 disks per DDP Data stored as raid6-stripes (8+2P) Disk space corresponding to 3 disks as reserved capacity for fast rebuild NEC Express 5800 R120g E5600 DE6600 4 July 2018

File Servers: xrootd File servers separated form storage in remote protocol clusters Native xrootd setup for ALICE Each xrootd file server is able to serve all files Directly profits from parallel file system Easy to scale throughput by adding additional file servers Spectrum Scale Storage Cluster with file sytem Protocol Server Cluster xrootd 1 /export/alice/... xrootd 2 xrootd x 5 July 2018

File Servers: dcache File servers separated form storage in remote protocol clusters Separate dcache instances for ATLAS, CMS, LHCb, Belle II Each dcache pool lives in sub-directory of large file system No automatic fail over but easy to deploy pool on new hardware Work in progress to improve pool startup times Spectrum Scale Storage Cluster with file sytem /export/atlas/pool1 /export/atlas/pool2 Protocol Server Cluster dcache 1 dcache 2 /export/atlas/poolx dcache x 6 July 2018

Deployment and Configuration Centralized host management for different infrastructures All configuration in GitLab Rolled out via continuous integration system Puppet / Foreman host configuration & deployment Spectrum Scale integration New nodes join existing Spectrum Scale clusters automatically Reinstalled nodes recover Spectrum Scale config automatically Nodes join existing CES clusters automatically 7 July 2018

Monitoring and Performance Data Started to use ELK stack to collect log files from storage servers Helps to correlate errors events of cluster nodes Example: RDMA connection errors due to broken IB cable Collect performance and health data with Telegraf / InfluxDB / Grafana 8 July 2018

Performance User requested maximum read/write performance so far: 50GB/s Usage statistics (2017 average rates per month) 7.5 / 4.2 PB read from compute farm / remote 1.1 / 3.0 PB written from compute farm / remote 40 30 20 10 0 GPFS r/w throughput read write GPFS read volume GPFS write volume 9 July 2018

Experience with Storage Expansion GridKa storage expansion for 2018/2019 has already been setup Capacity: 23 à 34 PB # protocol servers: 44 à 64, # NSD servers: 16 à 22 Expected combined r/w performance: 70GB/s à ~ 100GB/s Transparently included in existing IB fabrics and GPFS file systems During upgrade changed setup of IB fabrics from 1:1 à 1:2 blocking factor Spine 1:1 1:2 Leaf 10 July 2018

Summary and Outlook New online storage system successful in production @GridKa WLCG Tier-1 center Transparently expanded system with 2018/19 pledges Started investigating technological options for time beyond 2019 ~20% per year resource increase envisaged ~49PB disk storage till 2021 Ready for scaling system for requirements towards HL-LHC Disk capacity [TB] 11 July 2018