Jessica Popp Director of Engineering, High Performance Data Division

Similar documents
Andreas Dilger. Principal Lustre Engineer. High Performance Data Division

Community Release Update

Community Release Update

Andreas Dilger, Intel High Performance Data Division Lustre User Group 2017

Andreas Dilger, Intel High Performance Data Division SC 2017

Lustre * Features In Development Fan Yong High Performance Data Division, Intel CLUG

Brent Gorda. General Manager, High Performance Data Division

Andreas Dilger, Intel High Performance Data Division LAD 2017

HPE Scalable Storage with Intel Enterprise Edition for Lustre*

LustreFS and its ongoing Evolution for High Performance Computing and Data Analysis Solutions

The current status of the adoption of ZFS* as backend file system for Lustre*: an early evaluation

Lustre * 2.12 and Beyond Andreas Dilger, Intel High Performance Data Division LUG 2018

Lustre overview and roadmap to Exascale computing

UK LUG 10 th July Lustre at Exascale. Eric Barton. CTO Whamcloud, Inc Whamcloud, Inc.

DDN s Vision for the Future of Lustre LUG2015 Robert Triendl

DELL EMC ISILON F800 AND H600 I/O PERFORMANCE

Lustre 2.12 and Beyond. Andreas Dilger, Whamcloud

Data Movement & Tiering with DMF 7

Emerging Technologies for HPC Storage

Lustre on ZFS. Andreas Dilger Software Architect High Performance Data Division September, Lustre Admin & Developer Workshop, Paris, 2012

IME (Infinite Memory Engine) Extreme Application Acceleration & Highly Efficient I/O Provisioning

Introduction to Lustre* Architecture

LUSTRE NETWORKING High-Performance Features and Flexible Support for a Wide Array of Networks White Paper November Abstract

An ESS implementation in a Tier 1 HPC Centre

SFA12KX and Lustre Update

Andreas Dilger High Performance Data Division RUG 2016, Paris

LUG 2012 From Lustre 2.1 to Lustre HSM IFERC (Rokkasho, Japan)

Open SFS Roadmap. Presented by David Dillow TWG Co-Chair

WITH INTEL TECHNOLOGIES

Data Analytics and Storage System (DASS) Mixing POSIX and Hadoop Architectures. 13 November 2016

An Oracle Technical White Paper October Sizing Guide for Single Click Configurations of Oracle s MySQL on Sun Fire x86 Servers

High-Performance Lustre with Maximum Data Assurance

Multi-Rail LNet for Lustre

Intel Enterprise Edition Lustre (IEEL-2.3) [DNE-1 enabled] on Dell MD Storage

The modules covered in this course are:

Kubernetes Integration with Virtuozzo Storage

Parallel File Systems Compared

Fan Yong; Zhang Jinghai. High Performance Data Division

HPSS Treefrog Summary MARCH 1, 2018

Lustre HSM at Cambridge. Early user experience using Intel Lemur HSM agent

Why software defined storage matters? Sergey Goncharov Solution Architect, Red Hat

DataON and Intel Select Hyper-Converged Infrastructure (HCI) Maximizes IOPS Performance for Windows Server Software-Defined Storage

Small File I/O Performance in Lustre. Mikhail Pershin, Joe Gmitter Intel HPDD April 2018

Feedback on BeeGFS. A Parallel File System for High Performance Computing

The Leading Parallel Cluster File System

BeeGFS. Parallel Cluster File System. Container Workshop ISC July Marco Merkel VP ww Sales, Consulting

DDN. DDN Updates. Data DirectNeworks Japan, Inc Shuichi Ihara. DDN Storage 2017 DDN Storage

Next Generation Storage for The Software-Defned World

Introduction The Project Lustre Architecture Performance Conclusion References. Lustre. Paul Bienkowski

An Exploration into Object Storage for Exascale Supercomputers. Raghu Chandrasekar

6.5 Collective Open/Close & Epoch Distribution Demonstration

mos: An Architecture for Extreme Scale Operating Systems

Sun Lustre Storage System Simplifying and Accelerating Lustre Deployments

朱义普. Resolving High Performance Computing and Big Data Application Bottlenecks with Application-Defined Flash Acceleration. Director, North Asia, HPC

A New Key-value Data Store For Heterogeneous Storage Architecture Intel APAC R&D Ltd.

BIG DATA AND HADOOP ON THE ZFS STORAGE APPLIANCE

InfiniBand Networked Flash Storage

Parallel File Systems for HPC

Olaf Weber Senior Software Engineer SGI Storage Software. Amir Shehata Lustre Network Engineer Intel High Performance Data Division

Introducing SUSE Enterprise Storage 5

HPSS Treefrog Introduction.

Using DDN IME for Harmonie

Designing Next Generation FS for NVMe and NVMe-oF

Cluster Setup and Distributed File System

Hyper-Convergence De-mystified. Francis O Haire Group Technology Director

FlashGrid Software Enables Converged and Hyper-Converged Appliances for Oracle* RAC

DDN. DDN Updates. DataDirect Neworks Japan, Inc Nobu Hashizume. DDN Storage 2018 DDN Storage 1

Deploying Software Defined Storage for the Enterprise with Ceph. PRESENTATION TITLE GOES HERE Paul von Stamwitz Fujitsu

LLNL Lustre Centre of Excellence

Scalability Testing of DNE2 in Lustre 2.7 and Metadata Performance using Virtual Machines Tom Crowe, Nathan Lavender, Stephen Simms

MapR Enterprise Hadoop

Rethink Storage: The Next Generation Of Scale- Out NAS

RE-IMAGINING THE DATACENTER. Lynn Comp Director of Datacenter Solutions and Technologies

SGI Overview. HPC User Forum Dearborn, Michigan September 17 th, 2012

DataDirect Networks Japan, Inc.

GPFS on a Cray XT. Shane Canon Data Systems Group Leader Lawrence Berkeley National Laboratory CUG 2009 Atlanta, GA May 4, 2009

INTEGRATING HPFS IN A CLOUD COMPUTING ENVIRONMENT

Introducing Panasas ActiveStor 14

Lustre / ZFS at Indiana University

Quobyte The Data Center File System QUOBYTE INC.

GlusterFS Architecture & Roadmap

BeeGFS Solid, fast and made in Europe

Cloudian Sizing and Architecture Guidelines

Data Management. Parallel Filesystems. Dr David Henty HPC Training and Support

Shared Object-Based Storage and the HPC Data Center

MongoDB on Kaminario K2

A ClusterStor update. Torben Kling Petersen, PhD. Principal Architect, HPC

Lustre Beyond HPC. Presented to the Lustre* User Group Beijing October 2013

Jim Pappas Director of Technology Initiatives, Intel Vice-Chair, Storage Networking Industry Association (SNIA) December 07, 2018

Fast Forward I/O & Storage

Running Databases in Containers.

Challenges in making Lustre systems reliable

An Introduction to GPFS

Next-Generation NVMe-Native Parallel Filesystem for Accelerating HPC Workloads

SUSE Manager and Salt

Munara Tolubaeva Technical Consulting Engineer. 3D XPoint is a trademark of Intel Corporation in the U.S. and/or other countries.

Xyratex ClusterStor6000 & OneStor

Introduction to HPE ProLiant Servers HE643S

Application Acceleration Beyond Flash Storage

AUTOMATING IBM SPECTRUM SCALE CLUSTER BUILDS IN AWS PROOF OF CONCEPT

Transcription:

Jessica Popp Director of Engineering, High Performance Data Division

Legal Disclaimers 2016 Intel Corporation. Intel and the Intel logo are trademarks of Intel Corporation in the U.S. and/or other countries. *Other names and brands may be claimed as the property of others This document contains information on products, services and/or processes in development. All information provided here is subject to change without notice. Contact your Intel representative to obtain the latest forecast, schedule, specifications and roadmaps. FTC Optimization Notice: Optimization Notice Intel's compilers may or may not optimize to the same degree for non-intel microprocessors for optimizations that are not unique to Intel microprocessors. These optimizations include SSE2, SSE3, and SSE3 instruction sets and other optimizations. Intel does not guarantee the availability, functionality, or effectiveness of any optimization on microprocessors not manufactured by Intel. Microprocessor-dependent optimizations in this product are intended for use with Intel microprocessors. Certain optimizations not specific to Intel microarchitecture are reserved for Intel microprocessors. Please refer to the applicable product User and Reference Guides for more information regarding the specific instruction sets covered by this notice. 2

Intel Solutions for Lustre* Roadmap Q4 16 Q1 17 Q2 17 Q3 17 Q4 17 Enterprise HPC, commercial technical computing, and analytics EE 3.0 Lustre 2.7.x Client m data perf. OpenZFS perf. Snapshots SELinux, Kerberos OPA support RHEL 7 OpenZFS 0.6.5.3 Expanded OpenZFS support in IML IML graphical UI enhancements Bulk IO Performance (16MB IO) OpenZFS metadata performance Subdirectory mounts Enterprise Edition 3.1 Lustre 2.7.x DNE 2 Multi-rail LNET IML asset mgmt. ZED updates EE 4.0 Lustre <TBA> Cloud HPC, commercial technical computing and analytics delivered on cloud services Cloud Edition 1.3 Lustre 2.8.x RHEL 7 Server Support Lustre release update Cloud Edition 1.4 Lustre 2.9.x Updated Lustre and base OS builds New instance support SW updates New instances CE 1.5 Lustre <TBA> Key: Foundation Community Lustre software coupled with Intel support Shipping Development Planning FE 2.8 Lustre 2.8.x DNE 2 Client m data perf. LFSCK perf. SELinux, Kerberos UID/GID Mapping Shared key crypto Large Block IO Subdirectory mounts Foundation Edition 2.9 Lustre 2.9.x OpenZFS Snapshots Multi-rail LNET Project quotas Foundation Edition 2.10 Lustre 2.10.x Product placement not representative of final launch date within the specified quarter * Some names and brands may be claimed as the property of others 3

Intel Enterprise Edition for Lustre* software 3.1 Target GA Q4 2016 Based on community 2.7 release Latest RHEL 7.x server/client support IML managed mode for OpenZFS* GUI enhancements Bulk IO Performance for ldiskfs (16MB IO) Metadata performance improvements for Lustre on OpenZFS Subdirectory Mounts 4

Intel Cloud Edition for Lustre* software 1.3 Software defined storage cluster available via AWS Provides a performant, scalable filesystem on demand Powered by Lustre 2.8 Provides secure clients, IPSec and EBS encryption Available through AWS Cloud, AWS GovCloud, Microsoft Azure *Features may vary by platform 5

Intel Foundation Edition for Lustre* software 2.9 GA November 2016 Features RHEL* 7.2 servers/clients; SLES12* SP1 client support ZFS* Enhancements (Intel, LLNL) UID/GID mapping (IU, OpenSFS*) Shared Secret Key Encryption (IU, OpenSFS) Subdirectory mounts (DDN*) Server IO advice and hinting (DDN) Large Block IO 6

Development Community Growth Unique Developer and Organization Count IU LLNL Other SGI Clogeny Seagate ORNL ANU Atos Canonical CEA Cray DDN Fujitsu GSI 100 90 80 70 60 50 40 30 20 10 0 92 76 69 70 65 55 52 41 42 19 7 10 13 15 14 15 17 1 1.8.0 2.1.0 2.2.0 2.3.0 2.4.0 2.5.0 2.6.0 2.7.0 2.8.0 Intel Lustre 2.8 Commits Statistics courtesy of Lawrence Livermore National Laboratories (Chris Morrone) Developers Organizations Source: http://git.whamcloud.com/fs/lustre_release.git/shortlog/refs/heads/b2_8 7

Lustre 2.9 Contributions So Far Purdue 43 Other 879 SGI 305 Seagate 5061 ANU 545 Atos 449 CEA 210 Clogeny Cray 15 3617 DDN 1841 Fujitsu 1 ORNL 138 IU 2 Other 2 LLNL 17 GSI Seagate 1 SGI 8 72 ANU 4Atos 6 CrayCEA 17 89 DDN 52 LLNL 2786 ORNL 7873 Lines of Code IU 9164 Intel 37152 Reviews Intel 1759 Statistics courtest of Dustin Leverman (ORNL) Source: http://git.whamcloud.com/fs/lustre-release.git Aggregated data by organization between 2.8.50 and 2.8.60 8

Advanced Lustre Research Intel Parallel Computing Centers Uni Hamburg + German Client Research Centre (DKRZ) Client-side data compression Adaptive optimized ZFS data compression GSI Helmholtz Centre for Heavy Ion Research TSM * HSM copytool University of California Santa Cruz Automated client-side load balancing Johannes Gutenberg University Mainz Global adaptive IO scheduler Lawrence Berkeley National Laboratory Spark and Hadoop on 9

Lustre Feature Development Multi-Rail LNet (SGI *, Intel) Improve performance by aggregating bandwidth Improve resiliency by trying all interfaces before declaring msg undeliverable HSM Data Mover Copy tool to interface between Lustre and 3 rd party storage Early Access availability of agent with POSIX and S3 data movers via GitHub Additional data movers can be licensed as desired by developers 10

Lustre Feature Development Data on MDT (Intel) Optimize performance of small file IO Small files (as defined by administrator) are stored on MDT Client Small file IO directly to MDS Open, write, attributes layout, lock, attributes, read MDS OSS OSS OSS File Level Replication Robustness against data loss/corruption Provides higher availability for server/network failures Redundant layouts can be defined on per-file or directory basis Replicate/migrate between storage classes Replica 0 Replica 1 Object j (primary, preferred) Object k (stale) Overlapping (mirror) layout delayed resync 11

File Creates/sec ZFS Metadata Performance Improvements 120000 MDS-Survey - File Creates/Sec over 8 Directories 100000 80000 master+ ldiskfs 60000 master+zfs-0.7.0 +dacct master +ZFS-0.7.0 40000 2.8+ZFS-0.6.5 20000 0 8 16 32 64 128 Threads *Based on test platform in slide 20 12

Data Security Available in Intel EE for Lustre* software 3.0 Secure Clients SELinux8 support to enforce access control policies, including Multi- Level Security (MLS) on Lustre clients Secure Authentication and Encryption Kerberos* functionality can be applied to Intel EE for Lustre* software to establish trust between Lustre servers and clients and to support encrypted network communications In Development (2.9+) Shared Secret Key Crypto data encryption for networks including RDMA Node Maps UID/GID mapping for WAN clients Data isolation via filesystem containers subdirectory mounts with client authentication 13

Upstream Linux Kernel Client Refocused efforts on in-kernel Lustre client LNET virtually up-to-date with master Bug fixes to 2.5.51 landed, 2.5.58 pending Challenges Features currently at 2.3.54 (now DNE, HSM) Significant code cleanups still needed Lustre community developers cited among most active for Linux 4.6 kernel Oleg Droking (Intel): #5 by lines of code James Simmons (ORNL): #16 by lines of code 14

Lustre Development for CORAL draid Optimizations for large streaming IO Distributes data, parity, and spare blocks evenly across all disks Optimized and scalable RAID rebuild limits exposure to cascading disk failures and minimally impacts application IO Lustre Streaming ZFS Declustered Parity End-to-end large block support Block allocation improvements All work will be landed in appropriate upstream trees (Lustre and OpenZFS) 15

DAOS: The Future of Storage Open source, scalable, software-defined userspace storage system providing object file system storage on Intel hardware Applications can realize immediate benefit w/ little to no change via middleware modifications Fine-grained IO and Versioning provide greater speed and control over data DAOS exploits 3D-Xpoint DIMM & NVMe and Intel OmniPath Architecture Intel technology improves performance by reducing latency & keeping data closer to computing resources New Applications DAOS API Allows applications to directly access native Open Source Object Storage APIs Existing Applications HDF5, NetCDF Modern HPC data management framework makes porting apps easy Analytics Hadoop, Spark HDFS could access DAOS Storage easily to support popular analytics tools Legacy App POSIX-lite Full POSIX via Lustre Posix-compliant distributed file system for legacy application access DAOS - Distributed Async Object Storage Framework Reliable, Fast, Flexible storage built on Intel Technology 16

The Future is both Evolutionary & Revolutionary Lustre evolving in response to: A Growing Customer Base Changing use cases Emerging HW capabilities DAOS exploring new territory: What may lay beyond POSIX Use new HW capabilities as storage Object storage model exposes new capability for scalable consistency DAOS Tier IML Pool Pool Container Object DKey Akey[i] 17

Test Configuration 2 x Intel(R) Xeon(R) CPU E5-2660 v2 @ 2.20GHz 20 cores 64GB RAM 3 x 500GB local SATA HDD 7200 RPM CentOS 7.2.1511 3.10.0-327.28.2.el7 kernel (RHEL7) No remote clients, just local MDS testing (mds-survey script) 20

HPE Scalable Storage with Intel Enterprise Edition for Lustre* High Performance Storage Solution Meets Demanding I/O requirements Performance measured for an Apollo 4520 building block Up to 17 GB/s Read/15 GB/s Writes with EDR 1 Up to 16 GB/s Reads and Writes with OPA 1 Up to 21GB/s Reads and 15GB/s Writes with all SSD s² Designed for PB-Scale Data Sets Density Optimized Design For Scale Dense Storage Design Translates to Lower $/GB Limitless Scale Capability Solution Grows Linearly in Built on Apollo 4520 Appliance Like Delivery Model Pre-Configured Flexible Solution Deployment Services for Installation Simplified Wizard Driven Management thru Intel Manager for Lustre Innovative Software Features Leading Edge Yet Enterprise Ready Solution ZFS software RAID provides Snapshot, Compression & Error Correction ZFS reduces hardware costs with uncompromised performance Rigorously Tested for Stability & Efficiency 1: Different Conditions & Workloads can affect the Performance 2: 24 x1.6tb MU SSD s in A4520 no JBOD HPE Scalable Storage with Intel EE for Lustre* * Some names and brands may be claimed as the property of others. HPC Clients Apollo 6000 DL360 + MSA 2040 FDR/EDR* InfiniBand DL360 (Intel Management Server) 21