Shared Object-Based Storage and the HPC Data Center

Similar documents
StorNext SAN on an X1

X1 StorNext SAN. Jim Glidewell Information Technology Services Boeing Shared Services Group

HPC File Systems and Storage. Irena Johnson University of Notre Dame Center for Research Computing

SGI Overview. HPC User Forum Dearborn, Michigan September 17 th, 2012

Parallel File Systems Compared

EMC Celerra CNS with CLARiiON Storage

XenData Product Brief: SX-550 Series Servers for LTO Archives

Coordinating Parallel HSM in Object-based Cluster Filesystems

Introducing Panasas ActiveStor 14

Sun Lustre Storage System Simplifying and Accelerating Lustre Deployments

StorageCraft OneXafe and Veeam 9.5

An Introduction to GPFS

Early X1 Experiences at Boeing. Jim Glidewell Information Technology Services Boeing Shared Services Group

Design a Remote-Office or Branch-Office Data Center with Cisco UCS Mini

Veritas NetBackup on Cisco UCS S3260 Storage Server

Architecting Storage for Semiconductor Design: Manufacturing Preparation

The Last Bottleneck: How Parallel I/O can improve application performance

Storage for High-Performance Computing Gets Enterprise Ready

The World s Fastest Backup Systems

DELL POWERVAULT NX3500. A Dell Technical Guide Version 1.0

See what s new: Data Domain Global Deduplication Array, DD Boost and more. Copyright 2010 EMC Corporation. All rights reserved.

StorageCraft OneBlox and Veeam 9.5 Expert Deployment Guide

A GPFS Primer October 2005

Is There Any Alternative To Your Enterprise UNIX Platform? Andrej Gursky PosAm TechDays EAST, March 2015

IBM IBM Open Systems Storage Solutions Version 4. Download Full Version :

Design a Remote-Office or Branch-Office Data Center with Cisco UCS Mini

IOPStor: Storage Made Easy. Key Business Features. Key Business Solutions. IOPStor IOP5BI50T Network Attached Storage (NAS) Page 1 of 5

ACCRE High Performance Compute Cluster

An introduction to GPFS Version 3.3

EMC VPLEX Geo with Quantum StorNext

Xcellis Technical Overview: A deep dive into the latest hardware designed for StorNext 5

The Last Bottleneck: How Parallel I/O can attenuate Amdahl's Law

HP Dynamic Deduplication achieving a 50:1 ratio

HPC at UZH: status and plans

EMC Backup and Recovery for Microsoft SQL Server

XenData Product Brief: SX-550 Series Servers for Sony Optical Disc Archives

LUG 2012 From Lustre 2.1 to Lustre HSM IFERC (Rokkasho, Japan)

Surveillance Dell EMC Storage with Verint Nextiva

Dell EMC Storage with the Avigilon Control Center System

Data Movement & Tiering with DMF 7

Product Release Notes

Video Surveillance EMC Storage with LENSEC Perspective VMS

Microsoft Office SharePoint Server 2007

Optimizing and Managing File Storage in Windows Environments

Rio-2 Hybrid Backup Server

i Scalar 2000 The Intelligent Library for the Enterprise FEATURES AND BENEFITS

February 15, 2012 FAST 2012 San Jose NFSv4.1 pnfs product community

Executive Summary. The Need for Shared Storage. The Shared Storage Dilemma for the SMB. The SMB Answer - DroboElite. Enhancing your VMware Environment

designed. engineered. results. Parallel DMF

TECHNICAL OVERVIEW OF NEW AND IMPROVED FEATURES OF EMC ISILON ONEFS 7.1.1

Introducing SUSE Enterprise Storage 5

Surveillance Dell EMC Storage with Synectics Digital Recording System

Shared Parallel Filesystems in Heterogeneous Linux Multi-Cluster Environments

IBM Storwize V7000 Unified

XenData Product Brief: SX-250 Server for Optical Disc Archive Libraries

Dell Fluid Data solutions. Powerful self-optimized enterprise storage. Dell Compellent Storage Center: Designed for business results

Dell EMC Storage with Panasonic Video Insight

DEDUPLICATION BASICS

Lustre overview and roadmap to Exascale computing

Hitachi HQT-4210 Exam

System input-output, performance aspects March 2009 Guy Chesnot

DELL EQUALLOGIC FS7500 INTEGRATION WITHIN A LINUX ENVIRONMENT

EMC Integrated Infrastructure for VMware. Business Continuity

Is it a good idea to create an HSM without tapes?

Protect enterprise data, achieve long-term data retention

HPC Growing Pains. IT Lessons Learned from the Biomedical Data Deluge

A M L/2 AML/2. Automated Mixed Media Library

Provisioning with SUSE Enterprise Storage. Nyers Gábor Trainer &

WHY DO I NEED FALCONSTOR OPTIMIZED BACKUP & DEDUPLICATION?

Configuring a Single Oracle ZFS Storage Appliance into an InfiniBand Fabric with Multiple Oracle Exadata Machines

Surveillance Dell EMC Storage with IndigoVision Control Center

Dell EMC Storage with the Avigilon Control Center System

Symantec NetBackup PureDisk Compatibility Matrix Created August 26, 2010

VSTOR Vault Mass Storage at its Best Reliable Mass Storage Solutions Easy to Use, Modular, Scalable, and Affordable

IBM System Storage DCS3700

Red Hat Global File System

DELL POWERVAULT NX3500 INTEGRATION WITHIN A MICROSOFT WINDOWS ENVIRONMENT

Object Storage: Redefining Bandwidth for Linux Clusters

p5 520 server Robust entry system designed for the on demand world Highlights

Enterprise power with everyday simplicity

Customer Success Story Los Alamos National Laboratory

Copyright 2010 EMC Corporation. Do not Copy - All Rights Reserved.

EMC Backup and Recovery for Microsoft Exchange 2007 SP1. Enabled by EMC CLARiiON CX4-120, Replication Manager, and VMware ESX Server 3.

HOW DATA DEDUPLICATION WORKS A WHITE PAPER

Vendor: IBM. Exam Code: Exam Name: IBM Midrange Storage Technical Support V3. Version: Demo

SPECIFICATION FOR NETWORK ATTACHED STORAGE (NAS) TO BE FILLED BY BIDDER. NAS Controller Should be rack mounted with a form factor of not more than 2U

STOREONCE OVERVIEW. Neil Fleming Mid-Market Storage Development Manager. Copyright 2010 Hewlett-Packard Development Company, L.P.

Symantec Design of DP Solutions for UNIX using NBU 5.0. Download Full Version :

My First SAN solution guide

Open Source Storage. Ric Wheeler Architect & Senior Manager April 30, 2012

IBM Spectrum Protect Version Introduction to Data Protection Solutions IBM

Dell DR4100. Disk based Data Protection and Disaster Recovery. April 3, Birger Ferber, Enterprise Technologist Storage EMEA

Parallel File Systems for HPC

SXL-6500 Series: LTO Digital Video Archives

NAS When, Why and How?

EMC Symmetrix DMX Series The High End Platform. Tom Gorodecki EMC

The storage challenges of virtualized environments

Backup Solution Testing on UCS for Small-Medium Range Customers- FC (Disk to Tape Library)

Surveillance Dell EMC Storage with Infinova 2217 Security Management System

Backing Up and Restoring Multi-Terabyte Data Sets

Transcription:

Shared Object-Based Storage and the HPC Data Center Jim Glidewell High Performance Computing BOEING is a trademark of Boeing Management Company.

Computing Environment Cray X1 2 Chassis, 128 MSPs, 1TB memory 46 TB storage managed by ADIC StorNext HSM (5.5 TB online) 8 TB of direct-attached short-term storage Linux Clusters Systems: 2 128 node dual-xeon (32 bit) clusters 2 128 node dual-opteron clusters 3 256 node dual-opteron clusters More on the way All Clusters share access to a second ADIC StorNext HSM

History and Current Issues Used DMF (Data Migration Facility) since the early-90 s to manage disk space With the Cray X1, DMF was not an option An HSM was deemed essential Selected ADIC StorNext based on Cray support and recommendation Initially for the Cray X1 only Soon after, chosen for Linux cluster as well The I/O demands of the cluster were severely underestimated, as was the cluster growth rate As our clusters have grown, StorNext has developed significant performance problems

Hierarchical Storage Management - Pros & Cons Pros Reduces storage costs Makes highly efficient use of disk space Allows users to view the storage available as unlimited Eliminates need for user quotas Reduces day to day storage maintenance issues Simplifies detailed storage capacity decisions Reduces backup requirements Cons Administration is complex and time-consuming User delays waiting for file retrieval Data tends to build without bounds Serious cleanup only occurs when a system is retired Moving data from one HSM to a new one is very time consuming

Strategy for Shared Storage Situation HPC storage was tied to computing platform No common storage for all HPC systems Duplication of data as user processes use multiple platforms Current Cluster SAN unable to deal with increasing load Needed a storage system To serve as a shared repository for HPC data Preferred direct access from cluster, NFS option High-performance NFS from Cray X1 To serve as a high-performance replacement for cluster SAN Wanted a solution to serve both functions Shared HPC permanent directory Cluster home directory Shared HPC temporary storage (7-30 days) Cluster temporary storage (7-30 days

Storage System Selection Criteria Accessibility from all HPC systems NFS from the Cray X1 Direct client access from Linux clusters preferred Availability 24 by 7 uptime Concurrent storage system maintenance Reliability, resiliency, and redundancy Performance Ability to operate with a large number of clients High single-node performance High aggregate bandwidth Scalable performance Manageability Ability to grow volumes seamlessly No dump & reload No performance penalty Simple interface for management

Utilizing Panasas in Boeing HPC Used for multiple functions Linux user home directories Shared HPC user storage Linux high-speed temporary storage Shared temporary storage Panasas directory for each user Linux home directory is a subdirectory Cray home directory remains on X1 Shared home directory between systems not desirable Different binaries, shell init scripts, etc. Common absolute path for permanent and temporary storage on all HPC systems DirectFlow access from Linux clusters, NFS from Cray X1

Panasas Access Methods NFS DirectFlow

User Directory Structure Linux Clusters Cray X1 /share/joe /acct/joe (home directory) /share/joe /acct/joe (home directory) /joe linux cray origin... Panasas

Temporary Directory Structure Linux Clusters /stmp /ptmp Cray X1 /stmp /ptmp /stmp /linux...... Panasas

What is Shared Object-Based Storage? ANSI Standard OSD-1 r10 defines the Object-based Storage Device (OSD) interface Multiple Vendors and Options Lustre Panasas EMC HP Files exist as one or more objects, rather than groups of blocks Storage is intelligent and can move these objects around for redundancy and/or performance Design goals are robustness, scalability, flexibility Storage interface is standardized, but metadata handling is proprietary

The Panasas Storage System Realms, Bladesets, Volumes Logically, Panasas presents itself as: a single realm, containing one or more bladesets, each containing one or more volumes Shelves, Blades Panasas hardware is delivered in shelves (rack-mounted), which each contain 11 blades Blades come in two types: Director blades - manage metadata traffic, NFS access Storage blades - contain drives & intelligent controller Access DirectFlow client on Linux NFS and SMB from other clients

Panasas Hardware

Panasas Hardware (continued )

Our Panasas Installation Production - 52 Shelves of Panasas 3000 Storage Each shelf contains 2 director blades and 9 storage blades (2+9) 500 gigabytes per storage blade 4.5 terabytes per shelf raw capacity Seven racks, total of 234 terabytes raw capacity Evaluation System 3 1+10 shelves, 800 GB blades Used for initial evaluation Administrator training and familiarization Validated bladeset expansion process Very rigorous testing Retained to test Panasas 3.x software

HPC Panasas Storage

Panasas Performance Performance is a function of multiple factors Network speed Concurrent usage Number of shelves in bladeset Access method DirectFlow for Linux clients NFS/CIFS for other clients NFS speed from Cray X1 35 Mbytes/second Single Stream from dual-opteron node (gigabit link) Up to 85 MBytes/second Single shelf bandwidth ~ 300MBytes/second 20 clients, 4 shelves 1.2 GBytes/second (60 MBytes/sec. average per client) Total aggregate bandwidth Over 10GBytes/second - limited by network bandwidth

Panasas Issues Bugs reported and resolved Evaluation system was extensively tested A large number of support cases were opened Gathering needed debug data was time-consuming The vast majority of these cases were closed quickly System limitations Needed to split realm - too many director blades Unable to mix blade disk sizes within a bladeset Scaling issues regarding administration Time to reboot realm with new software Outstanding enhancement requests Management of multiple realms by a single GUI Site-defined metadata Tool to get stat() data in bulk (similar to SGI_FS_BULKSTAT) ACLs

Backup Issues Storage growth is having a big effect on backup Disk and RAID systems capacity growth exceeds that of tape Traditional Base + incrementals backup strategy is becoming impractical Evaluated using the enterprise backup service Adding our storage would double weekly backup Required significant upgrade to their hardware Weekly base dumps were not practical Synthetic base dumps were an untried option Analysis showed that after 12 months, >75% of all data being written to tape was data that had already been backed up HSM as a backup server

HSM as a Backup Server Basic Backup Strategy User storage is not managed by HSM HSM contains volumes and directories that match that of user storage One-way file synchronization is done nightly From user storage to HSM Can be done on a volume or directory basis Disk to disk copy Uses rsync command HSM migrates data to tape over time HSM-aware backup facility xfsdump -a Backs up inode information only Data is on HSM-managed tapes HSM is not directly user accessible

The Boeing HPC HSM Backup System - Specs Hardware SGI Altix 450 16 cores 48 gigabytes memory 24 fibre channel ports 40 Terabytes of DDN-based storage (InfiniteStorage 6700) SUN/STK SL8500 Automated Tape Library 1500 tape slots 6 T10000 Drives Software SLES 9 + SGI ProPack 4 DMF 3.6 TMF

The Boeing HPC HSM Backup System - Hardware SGI Altix 450 STK SL8500

The Boeing HPC HSM Backup System Backup Server Storage RSYNC RSYNC RSYNC DMF Tape Library Panasas Storage Backup Server

HSM as a Backup Server - Benefits HSMs have proven functionality Mature and robust products HPC group has years of experience with DMF Data is written once to tape Optimized usage of tape media, drives HSM manages tape merges and soft-deleted data Fast recovery option in case of catastrophic failure of primary storage Suspend all work Mount HSM system in place of production storage Resume production Option to use (part of) the backup server as a true HSM

Summary Panasas has met our needs for a central HPC storage facility Performance via DirectFlow client is very good, NFS access from the Cray is more than adequate Panasas has provided very good support, and was very responsive to bug reports Evaluation system was very helpful tool for familiarization and testing The use of an HSM as a backup server has been a great success for us Users have been very happy with performance