Parallel File Systems Compared

Similar documents
Experiences with HP SFS / Lustre in HPC Production

Filesystems on SSCK's HP XC6000

Lessons learned from Lustre file system operation

Parallel File Systems for HPC

High Performance Storage Solutions

Shared Parallel Filesystems in Heterogeneous Linux Multi-Cluster Environments

Beyond Petascale. Roger Haskin Manager, Parallel File Systems IBM Almaden Research Center

Lustre overview and roadmap to Exascale computing

Voltaire Making Applications Run Faster

Using file systems at HC3

GPFS on a Cray XT. Shane Canon Data Systems Group Leader Lawrence Berkeley National Laboratory CUG 2009 Atlanta, GA May 4, 2009

DVS, GPFS and External Lustre at NERSC How It s Working on Hopper. Tina Butler, Rei Chi Lee, Gregory Butler 05/25/11 CUG 2011

Rechenzentrum HIGH PERFORMANCE SCIENTIFIC COMPUTING

Coordinating Parallel HSM in Object-based Cluster Filesystems

Lustre A Platform for Intelligent Scale-Out Storage

A Comparative Experimental Study of Parallel File Systems for Large-Scale Data Processing

Sun Lustre Storage System Simplifying and Accelerating Lustre Deployments

Symantec NetBackup PureDisk Compatibility Matrix Created August 26, 2010

LustreFS and its ongoing Evolution for High Performance Computing and Data Analysis Solutions

Future Trends in Hardware and Software for use in Simulation

Shared File System Requirements for SAS Grid Manager. Table Talk #1546 Ben Smith / Brian Porter

Linux Clustering Technologies. Mark Spencer November 8, 2005

Shared Object-Based Storage and the HPC Data Center

LUG 2012 From Lustre 2.1 to Lustre HSM IFERC (Rokkasho, Japan)

Andreas Dilger. Principal Lustre Engineer. High Performance Data Division

Data Management. Parallel Filesystems. Dr David Henty HPC Training and Support

Challenges in making Lustre systems reliable

An Introduction to GPFS

LLNL Lustre Centre of Excellence

Introduction The Project Lustre Architecture Performance Conclusion References. Lustre. Paul Bienkowski

A GPFS Primer October 2005

Lustre at the OLCF: Experiences and Path Forward. Galen M. Shipman Group Leader Technology Integration

Red Hat Global File System

CSCS HPC storage. Hussein N. Harake

Introducing Panasas ActiveStor 14

Parallel File Systems. John White Lawrence Berkeley National Lab

An introduction to GPFS Version 3.3

LCE: Lustre at CEA. Stéphane Thiell CEA/DAM

System input-output, performance aspects March 2009 Guy Chesnot

Introduction to HPC Parallel I/O

Feedback on BeeGFS. A Parallel File System for High Performance Computing

Computer Science Section. Computational and Information Systems Laboratory National Center for Atmospheric Research

The GUPFS Project at NERSC Greg Butler, Rei Lee, Michael Welcome. NERSC Lawrence Berkeley National Laboratory One Cyclotron Road Berkeley CA USA

HIGH PERFORMANCE COMPUTING: MODELS, METHODS, & MEANS PARALLEL FILE I/O 1

IT Certification Exams Provider! Weofferfreeupdateserviceforoneyear! h ps://

IBM Storwize V7000 Unified

HPC File Systems and Storage. Irena Johnson University of Notre Dame Center for Research Computing

Storage and Storage Access

InfiniBand based storage target

LUSTRE NETWORKING High-Performance Features and Flexible Support for a Wide Array of Networks White Paper November Abstract

Xyratex ClusterStor6000 & OneStor

Scaling a Global File System to the Greatest Possible Extent, Performance, Capacity, and Number of Users

Database Services at CERN with Oracle 10g RAC and ASM on Commodity HW

Data Movement & Storage Using the Data Capacitor Filesystem

Scalable I/O, File Systems, and Storage Networks R&D at Los Alamos LA-UR /2005. Gary Grider CCN-9

Experience the GRID Today with Oracle9i RAC

HPC Solution. Technology for a New Era in Computing

RAIDIX Data Storage Solution. Clustered Data Storage Based on the RAIDIX Software and GPFS File System

Birds of a Feather Presentation

Exam : Title : Storage Sales V2. Version : Demo

Mellanox Technologies Maximize Cluster Performance and Productivity. Gilad Shainer, October, 2007

Object Storage: Redefining Bandwidth for Linux Clusters

An ESS implementation in a Tier 1 HPC Centre

The ASCI/DOD Scalable I/O History and Strategy Run Time Systems and Scalable I/O Team Gary Grider CCN-8 Los Alamos National Laboratory LAUR

The advantages of architecting an open iscsi SAN

EMC VPLEX Geo with Quantum StorNext

InfiniBand Networked Flash Storage

Datura The new HPC-Plant at Albert Einstein Institute

Data Movement & Tiering with DMF 7

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance

HPC and IT Issues Session Agenda. Deployment of Simulation (Trends and Issues Impacting IT) Mapping HPC to Performance (Scaling, Technology Advances)

Embedded Filesystems (Direct Client Access to Vice Partitions)

Mission-Critical Lustre at Santos. Adam Fox, Lustre User Group 2016

IBM IBM Open Systems Storage Solutions Version 4. Download Full Version :

SFA12KX and Lustre Update

Store Process Analyze Collaborate Archive Cloud The HPC Storage Leader Invent Discover Compete

Future Routing Schemes in Petascale clusters

INTEGRATING HPFS IN A CLOUD COMPUTING ENVIRONMENT

NAS When, Why and How?

Three Generations of Linux Dr. Alexander Dunaevskiy

Sami Saarinen Peter Towers. 11th ECMWF Workshop on the Use of HPC in Meteorology Slide 1

1Z0-433

PeerStorage Arrays Unequalled Storage Solutions

STORAGE CONSOLIDATION WITH IP STORAGE. David Dale, NetApp

Assistance in Lustre administration

Server Networking e Virtual Data Center

Sun N1: Storage Virtualization and Oracle

System that permanently stores data Usually layered on top of a lower-level physical storage medium Divided into logical units called files

SGI Overview. HPC User Forum Dearborn, Michigan September 17 th, 2012

GPFS on a Cray XT. 1 Introduction. 2 GPFS Overview

System Requirements. Version 8.2 May 2, For the most recent version of this document, visit our documentation website.

IBM FileNet Content Manager and IBM GPFS

VERITAS SANPoint Storage Appliance Overview of an Open Platform for the Implementation of Intelligent Storage Server

2008 International ANSYS Conference

System Requirements. Version 8.1 March 2, For the most recent version of this document, visit our documentation website.

Communication has significant impact on application performance. Interconnection networks therefore have a vital role in cluster systems.

HPC Customer Requirements for OpenFabrics Software

HP Data Protector 8.00 Platform and Integration Support Matrix

Manufacturing Bringing New Levels of Performance to CAE Applications

Unleashing Clustered ECMWF-Workshop 2004

Transcription:

Parallel File Systems Compared Computing Centre (SSCK) University of Karlsruhe, Germany Laifer@rz.uni-karlsruhe.de page 1

Outline» Parallel file systems (PFS) Design and typical usage Important features Comparison of the most important products Deployment at the computing centre page 2

Introduction» What is a distributed file system? File system data is usable at the same time from different clients With multiple servers applications see separate file systems» What is a parallel file system (PFS)? Distributed file system with parallel data paths from clients to disks Even with multiple servers applications typically see one file system page 3

Current trends» Storage needs increase and disk costs decrease steadily Storage systems are rapidly growing Trend towards RAID6 because of growing chance of multiple disk failures Storage consolidation in order to reduce administrative costs Also allows to dynamically allocate storage New trend to have one parallel file system for multiple clusters» Nearly no improvement in disk access times Increased speed by striping data over multiple disks/disk subsystems» Frequently need for high transfer rates Trend towards parallel file systems Several new parallel file systems were recently developed Existing parallel file systems were greatly enhanced» Number of clients in HPC systems is heavily increasing Scalability becomes more and more important page 4

SAN based parallel file systems Interconnect MDS MDS C C C file SAN» Striping over disk subsystems» Needs a storage area network (SAN) Traditionally FC based, alternatives are iscsi or InfiniBand protocols» Examples: ADIC SNFS, SGI CXFS, RedHat GFS, IBM GPFS (without NSD servers) page 5

Network based parallel file systems Interconnect MDS MDS S S S S C C file» Striping over servers» Uses low level and fast communication (RDMA) over interconnect if possible» Examples: Lustre, IBM GPFS (with NSD servers), Panasas ActiveScale Storage Cluster page 6

Typical PFS usage (1): Cluster file system Cluster nodes Interconnect or SAN» File system and cluster usually from same vendor Good parallel file system is important for cluster selection» Benefit is increased throughput, scalability and easy usability page 7

Typical PFS usage (2): Workflow file system Pre-processing Computing Post-processing 1 2 3» Typical customers: Oil & gas, digital media» Usually moderate number of heterogeneous clients SAN based PFS are used in most cases» Benefit is accelerated workflow and easy usability page 8

Typical PFS usage (3): Global file system... Interconnect A... SMP system Visualisation Cluster A Cluster B... Interconnect B...» New concept with additional requirements: Lots of clients (scalability) Minimal downtime» Examples at LLNL, ORNL, NERSC, TeraGrid, DEISA page 9

Common PFS properties» Throughput performance mainly depends on available hardware Most PFS can reach accumulated > 10 GB/s for sequential read/write More than 100 GB/s have been demontsrated with GPFS and Lustre» Metadata performance of one file system is limited Maximum is usually 5000-10000 operations per second May be lower for deletes or if files are stored in a single directory» Possible configurations without single point of failure Requires dedicated hardware and failover support of software» User level security is not available Root on all clients has full file system access» Linux kernel dependencies» NFS or CIFS gateways to connect unsupported clients» POSIX file system semantics page 10

Main PFS differences (1)» Scalability, i.e. number of supported clients SAN based file systems are often limited to 100 clients» Heterogeneity Supported operating system and kernel versions SAN based file systems often support more operating systems» Reliability Number of installed systems of similar size Expect software problems if file system is new Quality of software support» Costs Supported storage subsystems and disks Requirement for special or additional hardware Software and maintenance Complexity of administration page 11

Main PFS differences (2)» Metadata and lock management implementation Is most critical and complicated part of each PFS Usually a PFS is not well suited for mail or database servers For MPI-IO parameters have to be carefully chosen» Network options Supported networks, protocols and speed Examples: GigE, 10 GigE, 4x DDR InfiniBand, 4 Gb FC, iscsi Support for multiple network connections or low level gateways» Adequate backup solution Very fast or parallel restore is required Snapshots help to create consistent backup or to restore user data» HSM support Usually a PFS supports only a dedicated HSM system Archiving by users is an alternative to HSM page 12

PFS products (1): Lustre» Status User base is rapidly growing E.g. SSCK, U of Dresden, LLNL, PNNL, Sandia, ORNL, CEA, TITECH, PSC Roadmap, FAQs and source code from Cluster Filesystems Inc. (CFS) http://www.clusterfs.com/ Lustre products available from many vendors CFS, HP, Cray, Bull, LinuxNetworx, Transtec, Sun, DDN» Pros and Cons + Runs pretty stable Experiences at SSCK: http://www.rz.uni-karlsruhe.de/dienste/lustretalks + Open source + Scalable up to 10000 s of clients + High throughput with multiple network protocols InfiniBand, Quadrics, Myrinet, TCP/IP LNET routers provide gateways with high performance and failover Currently supports only Linux clients Patchless client will hopefully remove kernel dependencies page 13

PFS products (2): IBM GPFS» Status Large user base E.g. DEISA sites, FZK, NERSC, SDSC, LLNL, TeraGrid IBM GPFS Concepts, Planning, and Installation Guide provides good introduction http://publib.boulder.ibm.com/infocenter/clresctr/vxrx/index.jsp?topic=/com.ibm.cluster.g pfs.doc/gpfsbooks.html Also available via OEMs, e.g. LinuxNetworx» Pros and Cons + Runs pretty stable + Offers many useful features Snapshots, data and metadata replication, online disk removal + Scalable up to 1000 s of clients + Feature list is permanently improved Currently supports only Linux and AIX clients Limited support for different network protocols InfiniBand RDMA support is still missing page 14

PFS products (3): Panasas ActiveScale Storage Cluster» Status Medium user base E.g. U of Cologne (RRZK), LANL, Walt Disney, Paradigm Further information http://www.panasas.com/products_overview.html» Pros and Cons + Easy installation and administration + Supplies good performance for random IO + Offers additional useful features Snapshots, dynamic load balancing + Scalable up to 1000 s of clients Currently supports only Linux clients Supports only Gigabit Ethernet Throughput per client is limited to 80-100 MB/s Needs dedicated storage hardware from Panasas page 15

PFS products (4): ADIC StorNext File System (SNFS)» Status Medium user base E.g. FZK, CGG, Digital FilmWorks, Air Force Research Lab Further information http://www.adic.com/stornext ADIC is now owned by Quantum» Pros and Cons + Support for many different clients Linux, Irix, Solaris, Windows 2000/XP/2003, MAC OS X, AIX, HP-UX, UNICOS + Good HSM and backup integration + Easy installation + Offers additional useful features Snapshots, data replication, guaranteed bandwidth, multipathing Scalable up to 128 clients Needs a storage area network page 16

PFS products (5): SGI CXFS» Status Medium user base E.g. LRZ, U of Dresden, SARA, NASA, BBC, Ford, John Deere Further information http://www.sgi.com/products/storage/tech/file_systems.html» Pros and Cons + Support for many different clients Linux, Irix, Altix, Solaris, Windows 2000/XP/2003, MAC OS X, AIX + Good HSM and backup integration + Offers additional useful features Guaranteed bandwidth Scalable up to 64 clients Needs a storage area network InfiniBand is also supported Needs dedicated hardware for MDS page 17

PFS products (6): RedHat GFS» Status Possibly small commercial user base E.g. Secure-24, CD-adapco Further information http://www.redhat.com/software/rha/gfs/» Pros and Cons + Tightly integrated with RedHat Linux + Open source + Supports Oracle RAC database clustering + Scalable up to 256 clients Supports only Linux clients Needs a storage area network Needs HP Serviceguard for HA solution Not sure if stability is good Lock manager was redesigned due to performance problems page 18

Example: HP SFS/Lustre at SSCK s HP XC6000 120 clients (Itanium) C C C C C C... C C C C C C Quadrics QSNet II Interconnect Admin MDS OSS OSS OSS OSS OSS OSS EVA EVA EVA EVA EVA EVA EVA $HOME $WORK Capacity 3.8 TB 7.6 TB Write performance 240 MB/s 480 MB/s Read performance 380 MB/s 760 MB/s page 19

Example: HP SFS/Lustre at SSCK s HP XC4000 760 clients (Opteron)...... C C C C C C C C C C C C InfiniBand 4X DDR Interconnect MDS MDS OSS OSS OSS OSS OSS OSS OSS OSS $HOME $WORK Capacity 8 TB 48 TB Write performance 360 MB/s 2100 MB/s Read performance 600 MB/s 3600 MB/s page 20

Example: SSCK s plan for a global parallel file system... C C C C Visualisation C C... C C Infiniband... Infiniband Infiniband Switch MDS MDS... S S S S S S Parallel file system Ethernet Switch C... C NFS Server CIFS Server Ethernet... C C C C Campus network page 21