Tape Data Storage in Practice Minnesota Supercomputing Institute

Similar documents
Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

<Insert Picture Here> Tape Technologies April 4, 2011

DX-Series: Disk Archive to 320 TB with Continuous Backup to LTO, Optical Disc or the Cloud

SXL-8500 Series: LTO Digital Video Archives

By 2014, World-Wide file based

XenData Product Brief: SX-550 Series Servers for LTO Archives

Organizational Update: December 2015

XenData SX-10. LTO Video Archive Appliance. Managed by XenData6 Server Software. Overview. Compatibility

SXL-6500 Series: LTO Digital Video Archives

Deep Storage for Exponential Data. Nathan Thompson CEO, Spectra Logic

I/O Challenges: Todays I/O Challenges for Big Data Analysis. Henry Newman CEO/CTO Instrumental, Inc. April 30, 2013

SXL-8: LTO-8 Archive System. managed by XenData6 Server software. Eight cartridge library simplifies management of your offline files.

Global Headquarters: 5 Speen Street Framingham, MA USA P F

Automating Information Lifecycle Management with

Providing a first class, enterprise-level, backup and archive service for Oxford University

Compact Muon Solenoid: Cyberinfrastructure Solutions. Ken Bloom UNL Cyberinfrastructure Workshop -- August 15, 2005

Sun Lustre Storage System Simplifying and Accelerating Lustre Deployments

HPSS Treefrog Summary MARCH 1, 2018

NCAR Globally Accessible Data Environment (GLADE) Updated: 15 Feb 2017

TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 13 th CALL (T ier-0)

How do Design a Cluster

XenData SX-10 V2 LTO Video Archive Appliance

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

CSD3 The Cambridge Service for Data Driven Discovery. A New National HPC Service for Data Intensive science

Presents. Functionality. Overview. Great Compatibility. LTO-7 Archive System. managed by XenData6 Server software

The New Reliability Requirements: Tape Libraries

HPCMP Requirements for Metadata and Archiving at Scale

Low-cost BYO Mass Storage Project

Leonhard: a new cluster for Big Data at ETH

Spectra Stack FAQ March 2018

Data Archiving Managing the Mountain. Andrew Quinn 17/01/2012

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

RESEARCH DATA DEPOT AT PURDUE UNIVERSITY

IBM Storwize V7000 Unified

Cold Storage: The Road to Enterprise Ilya Kuznetsov YADRO

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

Emerging Technologies for HPC Storage

The Stampede is Coming Welcome to Stampede Introductory Training. Dan Stanzione Texas Advanced Computing Center

Effizientes Speichern von Cold-Data

Parallel File Systems. John White Lawrence Berkeley National Lab

Managing HPC Active Archive Storage with HPSS RAIT at Oak Ridge National Laboratory

Comet Virtualization Code & Design Sprint

Rio-2 Hybrid Backup Server

XtreemStore A SCALABLE STORAGE MANAGEMENT SOFTWARE WITHOUT LIMITS YOUR DATA. YOUR CONTROL

Data Movement & Tiering with DMF 7

XenData SXL- 4200N Archive System: Getting Started & Library Operation Basics. Document last updated: November 30, 2017.

Updating the HPC Bill Punch, Director HPCC Nov 17, 2017

SXL-4205Q LTO-8 Digital Archive

IBM IBM Storage Sales Version 8. Practice Test. Version QQ:

Vendor must indicate at what level its proposed solution will meet the College s requirements as delineated in the referenced sections of the RFP:

Oracle StorageTek. Dr. James Cates Senior Vice President Archive Development. Public

BlackPearl Converged Storage System Software Composition

StorageTek ACSLS Manager Software

Backup & Recovery on AWS

Product Brief: XenData X2500-SAS LTO-6 Digital Video Archive System

BlackPearl Customer Created Clients Using Free & Open Source Tools

Diamond Networks/Computing. Nick Rees January 2011

PART-I (B) (TECHNICAL SPECIFICATIONS & COMPLIANCE SHEET) Supply and installation of High Performance Computing System

IT Certification Exams Provider! Weofferfreeupdateserviceforoneyear! h ps://

SC17 - Overview

Computing and Networking at Diamond Light Source. Mark Heron Head of Control Systems

i Scalar 2000 The Intelligent Library for the Enterprise FEATURES AND BENEFITS

AN INTRODUCTION TO CLUSTER COMPUTING

Creating BlackPearl S3 Clients Using Free and Open Source Tools. Easily move data to low-cost deep storage

Chapter 7. GridStor Technology. Adding Data Paths. Data Paths for Global Deduplication. Data Path Properties

designed. engineered. results. Parallel DMF

Deduplication has been around for several

Storageflex HA3969 High-Density Storage: Key Design Features and Hybrid Connectivity Benefits. White Paper

HPC and IT Issues Session Agenda. Deployment of Simulation (Trends and Issues Impacting IT) Mapping HPC to Performance (Scaling, Technology Advances)

Virtualization Selling with IBM Tape

AWS Services for Data Migration Luke Anderson Head of Storage, AWS APAC

1. Introduction. Outline

Unified Computing System Launch. Welcome to Yas Island

(software agnostic) Computational Considerations

BUCKNELL S SCIENCE DMZ

Part 2: Computing and Networking Capacity (for research and instructional activities)

VMAX ALL FLASH FAMILY

All-Flash High-Performance SAN/NAS Solutions for Virtualization & OLTP

Vendor: IBM. Exam Code: C Exam Name: Power Systems with POWER7 and IBM i Technical Sales Skills -v2. Version: Demo

Costefficient Storage with Dataprotection

Storage Optimization with Oracle Database 11g

Enterprise Backup Architecture. Richard McClain Senior Oracle DBA

TechTour. Winning Technology Roadmaps. Sig Knapstad. Cutting Edge

White paper Version 3.10

The term "physical drive" refers to a single hard disk module. Figure 1. Physical Drive

The INFN Tier1. 1. INFN-CNAF, Italy

TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 11th CALL (T ier-0)

XenData Product Brief: XenData6 Server Software

Fujifilm 2015 Conference Into Tomorrow with Tape Technology Investing in the Future. Nathan Thompson CEO & Founder Spectra Logic Corporation

System i and System p. Creating a virtual computing environment

Anil Vasudeva IMEX Research. (408)

Data storage services at KEK/CRC -- status and plan

Executive Summary. The Need for Shared Storage. The Shared Storage Dilemma for the SMB. The SMB Answer - DroboElite. Enhancing your VMware Environment

PRACE Project Access Technical Guidelines - 19 th Call for Proposals

SUPERMICRO NEXENTASTOR 5.0 REFERENCE ARCHITECTURE

LTO-8 The Future of Storage is Here

The Future of Archiving

Unveiling the new QM2 M.2 SSD/10GbE PCIe Expansion cards

The Cambridge Bio-Medical-Cloud An OpenStack platform for medical analytics and biomedical research

POWEREDGE RACK SERVERS

Transcription:

Tape Data Storage in Practice Minnesota Supercomputing Institute GlobusWorld 208 Jeffrey McDonald, PhD Associate Director for Operations

HPC Resources MSI Users PI Accounts: 843 Users: > 4600 Mesabi Cores: > 8,000 Memory: 67 TB Accelerators: 80 K40 gpgpus Peak: 675 TF 320 Gbps to Storage New Technologies FPGAs Nvidia GPUs Intel PHI Storage (Intel NVME)

Storage High Performance Storage 7.2 PB Usable 48GB/s read/write Available on HPC resources Tier-2 Storage 3. PB Usable Available via Amazon s S3 interface Available anywhere in the world Archive Storage > 3.5 PB tape-based storage offline storage All available via Globus

Who Uses MSI?

System: Spectra T950, three frames. Blackpearl Enterprise model, LTO-7 model. Blackpearl has two dedicated LTO-7 drives. Tape Library Logically partition between TSM and Blackpearl. Dedicated 40 GbE network link to the blackpearl.

Tape Storage as a Service Launched Summer of 207. Three PIs purchased (four) LTO-7 tape buckets. Each unit is for a 5-year term, is by-tape, with LTO-7 (assume 2 TB of data storage) and include redundancy. Purchase rate for $456.2 per unit Cost estimate based upon occupancy of a full frame.

Policies Considerations One bucket per group. Tools to chunk data to adequate sizes Duplicity Tar Service model always; MSI never relinquishes the tape and we will not export. Globus provides data movement channel.

Tape Considerations LTO-7 error rates /200,000 tapes/.25 Exabytes (5-9s of reliability), 000x better than disk (LTO-consortium) MSI is conservative with storage, so we elected to make two copies of all user data Data can be deleted; space is not recovered unless bucket is re-initialized Configuration uses redundant copy configuration. Placement of data is controlled by the redundant copy configuration. Prevents mixing of user s data Forces the redundant copy.

Breaking down the hardware costs Item Quantity Item cost Extended cost % to Service Term Library Frame 25,000 25,000 00.00% 2 2,083 $0.3245067 Robotics from initial purchase 0 0 00.00% 2 0 $0.0000000 Robotics expansion 6,000 6,000 00.00% 2 500 $0.077886 Library Maintenance 0,000 0,000 00.00% 0,000 $.5576324 070 06.00 3,420 00.00% 5 22,684 $3.5333333 LTO-7 drives 2 0,500.00 2,000 00.00% 7 3,000 $0.4672897 BlackPearl server 00,000.00 00,000 00.00% 5 20,000 $3.52648 BlackPearl maintenance contract 4,800.00 4,800 00.00% 4,800 $0.7476636 070 65.00 69,550 00.00% 2 5,796 LTO-7 tape Slot activation Total Hardware 349,770 Cost /Year TB 68,863 Cost/TB/Year 6,420 $0.9027778 $0.7263499

Personnel Costing 0% of FTE for Technical Person. 0% of FTE for User Support Person. 5% FTE for management. This ads a base of $4.48 cost per TB per Year. The total cost per TB (no utilities) is $5.20 per TB per year (w/redundancy) or 456.2 per 5-years.

Power Considerations The archive and blackpearl use about 500W (averaged over duty cycle and share of T950). Independent of Capacity--Tapes don t use power. Compare to 6 PB spinning archive, 500-2 TB drives (no redundancy) 5W/drive = 2.5kW of power. Scales with capacity, no servers included here. 5 year lifetime savings: $20k power + $60k cooling. Saving grow with archive size (assuming at least 5 servers for spinning).

Globus Transfers Data set between 207-06-30 and 208-04-09. 3 Successful transfers. 28 Failed transfers (various reasons). 40 TB transferred. Rates as high as 95 MB/sec and average 30 MB/sec. Highly directional, one in 28 from archive.

Users/File Profiles These are the users who purchased tapes in 207 and their utilization. The primary concern for us was that the users aggregate their data to limit object counts on the Blackpearl system. MSI staff would spend a couple of hours with each user. Most transfers occur between our tier2 and tier3 storage for these users. User Dept/Type Tapes Files (Size) Average Size Comments Geneticist 2 3862 (6 TB) 4.3 GB mostly large dna sequence files Biologist 4 (50 GB) 36 GB project archives Biologist 56 (600 GB) GB large dna sequence files

Conclusions MSI should spend some money on Marketing! Globus is providing MSI with important tools toward an HSM-like framework for storage. Expanding the framework for some other appropriate use cases. Tape continues to offer a compelling use case for storage.

Backup Slides