The Last Bottleneck: How Parallel I/O can improve application performance

Similar documents
The Last Bottleneck: How Parallel I/O can attenuate Amdahl's Law

pnfs Update A standard for parallel file systems HPC Advisory Council Lugano, March 2011

February 15, 2012 FAST 2012 San Jose NFSv4.1 pnfs product community

System input-output, performance aspects March 2009 Guy Chesnot

Introducing Panasas ActiveStor 14

Brent Welch. Director, Architecture. Panasas Technology. HPC Advisory Council Lugano, March 2011

Crossing the Chasm: Sneaking a parallel file system into Hadoop

Crossing the Chasm: Sneaking a parallel file system into Hadoop

HPC Considerations for Scalable Multidiscipline CAE Applications on Conventional Linux Platforms. Author: Correspondence: ABSTRACT:

STAR-CCM+ Performance Benchmark and Profiling. July 2014

Shared Object-Based Storage and the HPC Data Center

pnfs, parallel storage for grid and enterprise computing Joshua Konkle, NetApp, Inc.

pnfs BOF SC

pnfs and Linux: Working Towards a Heterogeneous Future

Coordinating Parallel HSM in Object-based Cluster Filesystems

NFSv4.1 Using pnfs PRESENTATION TITLE GOES HERE. Presented by: Alex McDonald CTO Office, NetApp

High-Performance Lustre with Maximum Data Assurance

Microsoft Exchange Server 2010 workload optimization on the new IBM PureFlex System

Panasas Corporate & Product Overview

Cluster Scalability of Implicit and Implicit-Explicit LS-DYNA Simulations Using a Parallel File System

Company. Intellectual Property. Headquartered in the Silicon Valley

HPC File Systems and Storage. Irena Johnson University of Notre Dame Center for Research Computing

Shared Parallel Filesystems in Heterogeneous Linux Multi-Cluster Environments

NFS: What s Next. David L. Black, Ph.D. Senior Technologist EMC Corporation NAS Industry Conference. October 12-14, 2004

FUSION1200 Scalable x86 SMP System

NFS-Ganesha w/gpfs FSAL And Other Use Cases

pnfs: Blending Performance and Manageability

NAS for Server Virtualization Dennis Chapman Senior Technical Director NetApp

IBM Power Systems solution for SugarCRM

Research on Implement Snapshot of pnfs Distributed File System

Dell EMC Ready Bundle for HPC Digital Manufacturing Dassault Systѐmes Simulia Abaqus Performance

Customer Success Story Los Alamos National Laboratory

Storage Aggregation for Performance & Availability:

Mission-Critical Enterprise Linux. April 17, 2006

Veritas NetBackup on Cisco UCS S3260 Storage Server

Storage Performance Through Standards

Experiences with HP SFS / Lustre in HPC Production

Future Trends in Hardware and Software for use in Simulation

IME (Infinite Memory Engine) Extreme Application Acceleration & Highly Efficient I/O Provisioning

designed. engineered. results. Parallel DMF

Performance & Scalability Testing in Virtual Environment Hemant Gaidhani, Senior Technical Marketing Manager, VMware

Parallel File Systems Compared

Storage for High-Performance Computing Gets Enterprise Ready

Parallel NFS (pnfs) Garth Gibson CTO & co-founder, Panasas Assoc. Professor, Carnegie Mellon University.

Making Supercomputing More Available and Accessible Windows HPC Server 2008 R2 Beta 2 Microsoft High Performance Computing April, 2010

IBM Emulex 16Gb Fibre Channel HBA Evaluation

Write a technical report Present your results Write a workshop/conference paper (optional) Could be a real system, simulation and/or theoretical

MAHA. - Supercomputing System for Bioinformatics

Object-based Storage (OSD) Architecture and Systems

Object Storage: Redefining Bandwidth for Linux Clusters

IBM IBM Open Systems Storage Solutions Version 4. Download Full Version :

SONAS Performance: SPECsfs benchmark publication

HPC Growing Pains. IT Lessons Learned from the Biomedical Data Deluge

Coming Changes in Storage Technology. Be Ready or Be Left Behind

Deploying Celerra MPFSi in High-Performance Computing Environments

Architecting Storage for Semiconductor Design: Manufacturing Preparation

QLogic TrueScale InfiniBand and Teraflop Simulations

Parallel File Systems for HPC

PRESENTATION TITLE GOES HERE

Backup Appliances. Geir Aasarmoen og Kåre Juvkam

Dell EMC Ready Bundle for HPC Digital Manufacturing ANSYS Performance

An Introduction to GPFS

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance

Scality RING on Cisco UCS: Store File, Object, and OpenStack Data at Scale

Open Source Storage. Ric Wheeler Architect & Senior Manager April 30, 2012

HPC and IT Issues Session Agenda. Deployment of Simulation (Trends and Issues Impacting IT) Mapping HPC to Performance (Scaling, Technology Advances)

SAS Enterprise Miner Performance on IBM System p 570. Jan, Hsian-Fen Tsao Brian Porter Harry Seifert. IBM Corporation

dcache NFSv4.1 Tigran Mkrtchyan Zeuthen, dcache NFSv4.1 Tigran Mkrtchyan 4/13/12 Page 1

Lustre overview and roadmap to Exascale computing

EMC SYMMETRIX VMAX 40K STORAGE SYSTEM

ANSYS HPC Technology Leadership

Vblock Architecture. Andrew Smallridge DC Technology Solutions Architect

IBM s Data Warehouse Appliance Offerings

W H I T E P A P E R. Comparison of Storage Protocol Performance in VMware vsphere 4

Scalable I/O, File Systems, and Storage Networks R&D at Los Alamos LA-UR /2005. Gary Grider CCN-9

EMC ISILON HARDWARE PLATFORM

pnfs, parallel storage for grid, virtualization and database computing Joshua Konkle, Chair NFS SIG

2008 International ANSYS Conference

Benchmarking computers for seismic processing and imaging

Scalable x86 SMP Server FUSION1200

Sun N1: Storage Virtualization and Oracle

MELLANOX MTD2000 NFS-RDMA SDK PERFORMANCE TEST REPORT

MSC Nastran Explicit Nonlinear (SOL 700) on Advanced SGI Architectures

2008 International ANSYS Conference

IBM B2B INTEGRATOR BENCHMARKING IN THE SOFTLAYER ENVIRONMENT

RAIDIX Data Storage Solution. Clustered Data Storage Based on the RAIDIX Software and GPFS File System

Webinar Series: Triangulate your Storage Architecture with SvSAN Caching. Luke Pruen Technical Services Director

IBM i supports additional IBM POWER6 hardware

Effective Use of CSAIL Storage

Structuring PLFS for Extensibility

IOmark- VDI. IBM IBM FlashSystem V9000 Test Report: VDI a Test Report Date: 5, December

Next-Generation NVMe-Native Parallel Filesystem for Accelerating HPC Workloads

State of the Linux Kernel

ANSYS HPC. Technology Leadership. Barbara Hutchings ANSYS, Inc. September 20, 2011

Provisioning with SUSE Enterprise Storage. Nyers Gábor Trainer &

Store Process Analyze Collaborate Archive Cloud The HPC Storage Leader Invent Discover Compete

GROMACS Performance Benchmark and Profiling. September 2012

Data Management. Parallel Filesystems. Dr David Henty HPC Training and Support

IBM Scale Out Network Attached Storage (SONAS) using the Acuo Universal Clinical Platform

DAFS Storage for High Performance Computing using MPI-I/O: Design and Experience

Transcription:

The Last Bottleneck: How Parallel I/O can improve application performance HPC ADVISORY COUNCIL STANFORD WORKSHOP; DECEMBER 6 TH 2011 REX TANAKIT DIRECTOR OF INDUSTRY SOLUTIONS

AGENDA Panasas Overview Who we are, what we do pnfs (Parallel NFS) What is it? How does it work? Benefits of Parallel I/O Higher Performance COMPANY CONFIDENTIAL 2

PANASAS OVERVIEW Founded by Dr. Garth Gibson in 1999. First Customer Ship in 2004 Over 330+ WW customers; many with petabytes of data HQ Silicon Valley, CA, USA Market Focus: o Energy o Academia o Government o Life Sciences o Manufacturing o Finance Technologies: parallel file system and parallel storage appliance Worldwide support with over 25 global resellers COMPANY CONFIDENTIAL 3

WHAT DO WE DO? We build high performance scalable storage systems using parallel FS From a single Panasas Shelf To Hundreds of Terabytes To Multiple Petabytes 40 or 60TB up to 88GB of cache Per Chassis Multiple chassis combine to create a single pool of storage And everything in between with a large coherent cache Virtually unlimited scaling capability COMPANY CONFIDENTIAL 4

PARALLEL SYSTEM GROWTH (TOP500) Source: http://top500.org/overtime/list/37/procclass COMPANY CONFIDENTIAL 5

PARALLEL SYSTEM System: Hardware + File system + Storage + Application +.. Amdahl s law: The speedup of a program using multiple processors in parallel computing is limited by the time needed for the sequential fraction of the program. Source: Wikipedia Multi-core and cpu are the norm Parallel File System: 10+ years with Panasas + pnfs Parallel programming: OpenMP, MPI, 15+years Applications are moving from serial to parallel I/O A standard like pnfs encourages moving to parallel I/O COMPANY CONFIDENTIAL 6

KEY PNFS PARTICIPANTS Panasas (Objects) ORNL and ESSC/DoD funding Linux pnfs development Network Appliance (Files over NFSv4) IBM (Files, based on GPFS) BlueArc (Files over NFSv4) EMC (Blocks, HighRoad MPFSi) Sun/Oracle (Files over NFSv4) U of Michigan/CITI (Linux maint., EMC and Microsoft contracts) DESY Java-based implementation COMPANY CONFIDENTIAL 7

PNFS KEY CONCEPTS Separate metadata and data Compute clients can access data directly and in parallel Add metadata server Introduce the concept of data layout Layout is a map for clients to access data on the storage Backend storage Supports multiple types of backend storage systems, including traditional block storage, other file servers, and object storage systems Metadata Server Parallel NFS Storage: Files, blocks, objects Parallel Storage: File server not in data path. Performance bottleneck eliminated. COMPANY CONFIDENTIAL 8

THE PNFS STANDARD The pnfs standard defines the NFSv4.1 protocol extensions between the server and client The I/O protocol between the client and storage is specified elsewhere, for example: SCSI Object-based Storage Device (OSD) over iscsi SCSI Block Commands (SBC) over Fibre Channel (FC) Network File System (NFS) The control protocol between the server and storage devices is also specified elsewhere, for example: SCSI Object-based Storage Device (OSD) over iscsi Clients Storage MetaData Server COMPANY CONFIDENTIAL 9

PNFS LAYOUTS Client gets a layout from the NFS Server The layout maps the file onto storage devices and addresses The client uses the layout to perform direct I/O to storage Client commits changes and returns the layout when it s done At any time the server can recall the layout pnfs is optional, the client can always use regular NFSv4 I/O Layout Clients NFSv4.1 Server Storage COMPANY CONFIDENTIAL 10

PNFS CLIENT Common client for different storage back ends Wider availability across operating systems Fewer support issues for storage vendors Client Apps NFSv4.1 pnfs Client Layout Driver 1. SBC (blocks) 2. OSD (objects) 3. NFS (files) pnfs Server Cluster Filesystem Layout metadata grant & revoke COMPANY CONFIDENTIAL 11

LINUX CLIENT RELEASE CYCLE 2011 Kernel Merge Window Date What s New 2.6.38 Jan 2011 More generic pnfs code, still disabled, not fully functional 2.6.39 Apr 2011 Files-based back end, read, write, commit on the client. Linux server is read-only via pnfs. 2.6.40 3.0 (Fedora 15) Jun 2011 RAID0/1 Object-based back end 3.1 Oct 22 nd 2011 Block-based back end 3.2 Dec 2011 (?) RAID5 Object (Already in Linus s Tree) RHEL 6 and SLES 11 based on 2.6.32 Backporting pnfs for files will be attempted RHEL 7 and SLES 12 based on 3.* Integrated pnfs of all flavors timeline 2012+ COMPANY CONFIDENTIAL 12

PNFS SERVER TODAY Up-to-date GIT tree from Linux pnfsd server git://linux-nfs.org/~bhalevy/linux-pnfs.git Files and Blocks Simple Server (spnfs) pnfs Object Open-source Server: http://open-osd.org Useful to get to OSD target, the user level program Exofs uses kernel initiator, need the target Questions: NFS mailing lists linux-nfs@kernel.org, nfsv4@ietf.org COMPANY CONFIDENTIAL 13

PNFS TODAY WITH PANASAS Panasas System Compute Clients Parallel NFS Director Blades Parallel NFS Parallel Storage: File server not in data path. Performance bottleneck eliminated. Storage Blades Panasas is designed to help remove the last bottleneck Compute clients access storage blades directly Full drive performance is delivered to clients COMPANY CONFIDENTIAL 14

BASIC BUILDING BLOCKS 4U Shelf 1 + 10 Configuration DirectorBlade StorageBlade COMPANY CONFIDENTIAL 15

Time (seconds) ANSYS FLUENT V 13.0 PARALLEL (N-1 WRITE) 128 Processes running on 16 Compute Nodes (8 PPN) Lower better Benchmark (Test cases) Cavity_100m (synthetic benchmark) Truck_111m (realistic workload) 160 140 Parallel N-1 Write > 2.5 X over serial I/O with Truck_111m data set 120 100 80 Serial Clients: 16 Compute Nodes, each with Dual Socket Intel Xeon X5650 24GB memory 10GigE network Storage Single ActiveStor-11 shelf 10GigE 60 40 20 0 cavity_100m Truck_111m 128 Processes (8 PPN) Parallel COMPANY CONFIDENTIAL 16

Write (MB/s) higher better STAR-CCM+ WRITE RESULTS (N-1 ACCESS) Serial I/O: constant as process count increases Parallel I/O: performance increases > 3X serial I/O at 64P Low np limited by # of clients High np limited by # of disks 800 700 600 500 Write Rates (Lemans 17 million cell case) Storage subsystem: PAS8: 4 shelves, 33 SBs 10GigE PanFS version 3.5 400 300 200 Serial I/O Parallel I/O 100 Courtesy of CD-adapco, Inc 0 16 32 64 128 Process Count COMPANY CONFIDENTIAL 17

Read (MB/s) higher better STAR-CCM+ READ RESULTS (N-1 ACCESS) Similar trends to writes: Serial I/O: constant as process count increases Parallel I/O: performance increases Low np limited by # of clients High np limited by # of disks 1600 1400 1200 Read Rates (Lemans 17 million cell case) Read rates > write rates N-1 write has higher overhead due to coherency 1000 800 600 Serial I/O Parallel I/O 400 200 Courtesy of CD-adapco, Inc 0 16 32 64 128 Process Count COMPANY CONFIDENTIAL 18

MB/s LANDMARK PROMAX/SEISSPACE Parallel I/O throughput: multiple concurrent jobs Hardware: 4 Panasas AS12 shelves 32 8-core compute nodes Landmark ProMAX/SeisSpace generating geoseismic trace data Parallel I/O is done using JavaSeis (Opensource) 6000 5000 4000 3000 2000 1000 0 Scaling Concurrent Jobs (Serial I/O) 32 64 128 256 # Processes COMPANY CONFIDENTIAL 19

SUMMARY As cluster computing (HW) continues to grow, software is catching up to avoid slowest link (Amdahl s law) Software = File system, applications, I/O ISVs are coming Parallel system provides higher applications performance Faster time to market Increase ROI Panasas gives you parallel file system and parallel I/O today. COMPANY CONFIDENTIAL 20