Store Process Analyze Collaborate Archive Cloud The HPC Storage Leader Invent Discover Compete

Similar documents
HPC Storage Use Cases & Future Trends

朱义普. Resolving High Performance Computing and Big Data Application Bottlenecks with Application-Defined Flash Acceleration. Director, North Asia, HPC

DDN. DDN Updates. Data DirectNeworks Japan, Inc Shuichi Ihara. DDN Storage 2017 DDN Storage

IME Infinite Memory Engine Technical Overview

DDN About Us Solving Large Enterprise and Web Scale Challenges

IME (Infinite Memory Engine) Extreme Application Acceleration & Highly Efficient I/O Provisioning

DDN. DDN Updates. DataDirect Neworks Japan, Inc Nobu Hashizume. DDN Storage 2018 DDN Storage 1

Improved Solutions for I/O Provisioning and Application Acceleration

SFA12KX and Lustre Update

Infinite Memory Engine Freedom from Filesystem Foibles

Leveraging Software-Defined Storage to Meet Today and Tomorrow s Infrastructure Demands

Application Performance on IME

Using DDN IME for Harmonie

DDN and Flash GRIDScaler, Flashscale Infinite Memory Engine

2012 HPC Advisory Council

Next-Generation NVMe-Native Parallel Filesystem for Accelerating HPC Workloads

RAIDIX Data Storage Solution. Clustered Data Storage Based on the RAIDIX Software and GPFS File System

Isilon Scale Out NAS. Morten Petersen, Senior Systems Engineer, Isilon Division

DDN s Vision for the Future of Lustre LUG2015 Robert Triendl

Short Talk: System abstractions to facilitate data movement in supercomputers with deep memory and interconnect hierarchy

High Capacity network storage solutions

Applying DDN to Machine Learning

FLASHARRAY//M Business and IT Transformation in 3U

The Oracle Database Appliance I/O and Performance Architecture

The Datacentered Future Greg Huff CTO, LSI Corporation

NetApp: Solving I/O Challenges. Jeff Baxter February 2013

Isilon Performance. Name

Coordinating Parallel HSM in Object-based Cluster Filesystems

LustreFS and its ongoing Evolution for High Performance Computing and Data Analysis Solutions

Building Self-Healing Mass Storage Arrays. for Large Cluster Systems

Scaling Internet TV Content Delivery ALEX GUTARIN DIRECTOR OF ENGINEERING, NETFLIX

Lustre2.5 Performance Evaluation: Performance Improvements with Large I/O Patches, Metadata Improvements, and Metadata Scaling with DNE

Introducing Panasas ActiveStor 14

Preparing GPU-Accelerated Applications for the Summit Supercomputer

Managing HPC Active Archive Storage with HPSS RAIT at Oak Ridge National Laboratory

Data Management. Parallel Filesystems. Dr David Henty HPC Training and Support

All-Flash High-Performance SAN/NAS Solutions for Virtualization & OLTP

Exa Scale FSIO Can we get there? Can we afford to?

Lustre* is designed to achieve the maximum performance and scalability for POSIX applications that need outstanding streamed I/O.

High Performance Storage Solutions

DDN Annual High Performance Computing Trends Survey Reveals Rising Deployment of Flash Tiers & Private/Hybrid Clouds vs.

Deploy a High-Performance Database Solution: Cisco UCS B420 M4 Blade Server with Fusion iomemory PX600 Using Oracle Database 12c

SAP HANA IBM x3850 X6

Storage for HPC, HPDA and Machine Learning (ML)

IBM FlashSystem. IBM FLiP Tool Wie viel schneller kann Ihr IBM i Power Server mit IBM FlashSystem 900 / V9000 Storage sein?

Nimble Storage vs HPE 3PAR: A Comparison Snapshot

Architecting Storage for Semiconductor Design: Manufacturing Preparation

Running VMware vsan Witness Appliance in VMware vcloudair First Published On: April 26, 2017 Last Updated On: April 26, 2017

SGI Overview. HPC User Forum Dearborn, Michigan September 17 th, 2012

libhio: Optimizing IO on Cray XC Systems With DataWarp

DELL EMC ISILON F800 AND H600 I/O PERFORMANCE

A ClusterStor update. Torben Kling Petersen, PhD. Principal Architect, HPC

Design a Remote-Office or Branch-Office Data Center with Cisco UCS Mini

Scalability Testing with Login VSI v16.2. White Paper Parallels Remote Application Server 2018

IBM Spectrum Scale IO performance

Forget IOPS: A Proper Way to Characterize & Test Storage Performance Peter Murray SwiftTest

Isilon: Raising The Bar On Performance & Archive Use Cases. John Har Solutions Product Manager Unstructured Data Storage Team

BIG DATA AND HADOOP ON THE ZFS STORAGE APPLIANCE

The Fusion Distributed File System

UCS Invicta: A New Generation of Storage Performance. Mazen Abou Najm DC Consulting Systems Engineer

NERSC Site Update. National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory. Richard Gerber

THE ADVANCEMENT OF STORAGE SYSTEM DESIGNS FOR DIGITAL INDIA. Dana Kammersgard February 2017

IBM CORAL HPC System Solution

Introducing Tegile. Company Overview. Product Overview. Solutions & Use Cases. Partnering with Tegile

Parallel File Systems. John White Lawrence Berkeley National Lab

Parallels Remote Application Server. Scalability Testing with Login VSI

HIGH PERFORMANCE SANLESS CLUSTERING THE POWER OF FUSION-IO THE PROTECTION OF SIOS

Webinar Series: Triangulate your Storage Architecture with SvSAN Caching. Luke Pruen Technical Services Director

Data Movement & Tiering with DMF 7

Enhancing Lustre Performance and Usability

Deep Storage for Exponential Data. Nathan Thompson CEO, Spectra Logic

Extreme I/O Scaling with HDF5

Lessons from Post-processing Climate Data on Modern Flash-based HPC Systems

Increasing Performance of Existing Oracle RAC up to 10X

Refining and redefining HPC storage

DVS, GPFS and External Lustre at NERSC How It s Working on Hopper. Tina Butler, Rei Chi Lee, Gregory Butler 05/25/11 CUG 2011

AFM Use Cases Spectrum Scale User Meeting

Analyzing the High Performance Parallel I/O on LRZ HPC systems. Sandra Méndez. HPC Group, LRZ. June 23, 2016

HIGH PERFORMANCE STORAGE SOLUTION PRESENTATION All rights reserved RAIDIX

Storage Optimization with Oracle Database 11g

Create a Flexible, Scalable High-Performance Storage Cluster with WekaIO Matrix

Toward portable I/O performance by leveraging system abstractions of deep memory and interconnect hierarchies

Universal Storage. Innovation to Break Decades of Tradeoffs VASTDATA.COM

Write a technical report Present your results Write a workshop/conference paper (optional) Could be a real system, simulation and/or theoretical

Accelerating Spectrum Scale with a Intelligent IO Manager

A Breakthrough in Non-Volatile Memory Technology FUJITSU LIMITED

INTEGRATING HPFS IN A CLOUD COMPUTING ENVIRONMENT

SAP Applications on IBM XIV System Storage

SUPERMICRO, VEXATA AND INTEL ENABLING NEW LEVELS PERFORMANCE AND EFFICIENCY FOR REAL-TIME DATA ANALYTICS FOR SQL DATA WAREHOUSE DEPLOYMENTS

IBM Spectrum NAS, IBM Spectrum Scale and IBM Cloud Object Storage

CSCS HPC storage. Hussein N. Harake

All-Flash High-Performance SAN/NAS Solutions for Virtualization & OLTP

Software Defined Storage for the Evolving Data Center

Tuning I/O Performance for Data Intensive Computing. Nicholas J. Wright. lbl.gov

Red Hat Ceph Storage and Samsung NVMe SSDs for intensive workloads

Campaign Storage. Peter Braam Co-founder & CEO Campaign Storage

Oak Ridge National Laboratory Computing and Computational Sciences

Data Movement & Storage Using the Data Capacitor Filesystem

FLASHARRAY//M Smart Storage for Cloud IT

API and Usage of libhio on XC-40 Systems

Transcription:

Store Process Analyze Collaborate Archive Cloud The HPC Storage Leader Invent Discover Compete 1

DDN Who We Are 2 We Design, Deploy and Optimize Storage Systems Which Solve HPC, Big Data and Cloud Business Challenges at Scale Main Office: Sunnyvale, California, USA Go To Market: Partner & Reseller Assisted, Direct DDN: World s Largest Private Storage Company Only Storage Company with Long-Term on Big Data Focus World-Renowned & Award-Winning

An Elite Collection Of HPC s Finest... Some of our 1000+ Customers 3

DDN 15 Years of HPC Innovation 4 DDN FOUNDED 1 st CUSTOMER NASA 10GB/s NCSA 100GB/s CEA, LLNL LARGEST PRIVATE STORAGE CO. (IDC) 1TB/s ORNL 500+ EMPLOYES 5 BP / Rack 1998 1999 2000 2001 2002 2003 2004 2005 2006 2007 2008 2009 2010 2011 2012 2013 2014 1 st Real-Time Appliance for High-Scale Big Data EXAScaler DDN s 1 st Parallel File System Offering ft. Lustre 1 st in Data Center Density DDN Leads On The List of Lists: 80% of the Top 10 67% of the Top 100 32% of the Top 500 SFA Storage Fusion Architecture 1 st in Bandwidth + IOPS 1 st In-Storage Processing SW-Only, Portable Architecture 1 st Hyperscale Object Storage Web-Scale Computing and HPC Collaboration SFX Flash Tiering 1 st Application-Aware Hybrid Caching Revolutionizing HPC

Our Unwavering Commitment to HPC 5 Investments in Exascale Real Engineering Is Needed To Scale 1000x Fast Forward

Exascale I/O Challenges - Cost 6 LANL Trinity Hybrid Scratch Cost Analysis Hybrid approach is necessary to meet bandwidth & capacity requirements

# of HDDs Exascale I/O Challenges Power Consumption 7 70000 60000 NERSC-8 Cost Comparison Power 768KW 26 SFA Controllers + BB 7 Power 1792KW 99 SFA Controllers 50000 40000 Power 470KW 26 SFA Controllers 30000 Hybrid HDDs 20000 10000 0 0.76 2.96 Burst Throughput (TB/sec)

Exascale I/O Challenges Efficiency Analysis: Argonne s LCF production storage system (circa 2010) 99% of the time, storage BW utilization < 33% of max 70% of the time, storage BW utilization < 5% of max Burst Buffer Absorbs the Peak Load 25 MB/s 50 GB/s 4 GB/s IME SC 13 Demo Cluster Burst Buffer Tier 8 1) Separation of bandwidth and capacity is required 2) Utilization efficiency must be improved Filesystem Handles the Remaining Load Persistent Storage Tier Archival Storage Tier

Why is today s I/O efficiency so poor? 9 Serialization at various points in the I/O path Stripe and block alignment (PFS and RAID) o Read-modify-writes to underlying storage Lock contention o Exacerbated by poor I/O structuring in applications Compute Node1 Compute Node2 100000000 10000000 1000000 100000 10000 1000 2015 2018 Performance (TF) 20000 1000000 Concurrency 5000000 1000000000 1.2E+09 1E+09 800000000 600000000 400000000 200000000 0 Source: http://storageconference.org/2011/presentations/snapi/1.grider.pdf Lock Contention Worsens with 1000s of nodes File Server 1 File Server 2 File Server 3 File Server 4 Storage

Why is today s I/O efficiency so poor? 10 Poor horizontal scaling characteristics in the PFS weakest link PFS are only as fast as the slowest I/O component Oversubscribed or crippled I/O components affect the entire system performance As I/O sections get larger and # of components increases the problem worsens (congestion) This weakest link can be all the way down to disks (RAID rebuilds) A single overloaded server can slow down the entire system File Server 1 File Server 2 File Server 3 File Server 4 Storage

Performance Efficiency (Percent) Percentage of Stripe Size Throughput (MB/s) PFS Efficiency as a Function of I/O Size 11 100 90 80 70 60 50 40 30 20 10 Performance & Efficiency of Non- Mergeable Writes as a Function of I/O Size 0 4096 409600 I/O Size (Log-Scale) 100 Aligned, full-stripe-width IO required for maximum PFS I/O performance 90 80 70 60 50 40 30 20 10 0 Performance I/O Size (bytes) 2000 1800 1600 1400 1200 1000 800 600 400 200 0 Parallel Filesystem on IME Demo Cluster SSDs (50GB/s available) 1 8 64 512 IO Request SIze (KB) Faster media (SSDs) may not address the underlying PFS performance limitations Avg

What is Infinite Memory Engine (IME )? 12 High performance I/O system based on parallel log structuring Massive concurrency regardless of application I/O pattern Dynamically load balancing helps steer clear of oversubscribed and handicapped components Innovative lookup mechanism enables immediate availability of data Distributed fault tolerance 12

The Infinite Memory Advantage 13 Designed for Scalability Patented DDN Algorithms Fully POSIX & HPC Compatible No Application Modifications Scale-Out Data Protection Distributed Erasure Coding Non-Deterministic System Write Anywhere, No Layout Needed Integrated With File Systems Designed to Accelerate Lustre*, GPFS No Code Modification Needed Writes: Fast. Reads: They re Fast Too. No other system offers both at scale.

SC 13 Demo Comparative Testing: Shared Writes 14 IME Clients one per compute node; 98 node-local MLC SSDs Linear Cluster Scaling 14 CLUSTER LEVEL TESTING DDN GRIDScaler IME( overall) 6,225 Concurrent Write Requests (8 MB) 12,250,000 Concurrent, Interleaved Write Requests (4 KB) 49 GB/s 49 GB/s 17 MB/s 49 GB/s DISK LEVEL TESTING DDN GRIDScaler (per SSD) IME (per SSD) 62.5 Concurrent Write Requests 438 MB/s 500 MB/s 125,000 Concurrent Write Requests 170 KB/s 500 MB/s SSDs behind a PFS don t help IME is at line rate to scale with SSD rates Avg. 2018 Top500 Cluster Concurrency 57,772,000 Cores (est)

IME Checkpoint / Migration Workload Demo Achieves >90% of Available Storage Bandwidth 15 Checkpoint I/O directed at IME (emulated with IOR) File #1 (49 50 GB/s) File #2 (49 50 GB/s) File #3 (49 50 GB/s) Migration of File #3 from IME to PFS (4-5 GB/s)

ISC 14 IME Demo Server 16 Off the shelf 2U Server Chassis Dual Socket Ivy Bridge with 128 GB RAM Up to 24 SSDs per IME Server 2 FDR IB Ports Expected Burst Bandwidth per IME Server: ~10GB/s

ISC 14 Demo System in DDN Booth 17 16U (servers) Total Peak BW: ~80GB/s