Andreas Dilger. Principal Lustre Engineer. High Performance Data Division

Similar documents
Andreas Dilger, Intel High Performance Data Division Lustre User Group 2017

Lustre * Features In Development Fan Yong High Performance Data Division, Intel CLUG

Andreas Dilger, Intel High Performance Data Division LAD 2017

Andreas Dilger, Intel High Performance Data Division SC 2017

Jessica Popp Director of Engineering, High Performance Data Division

Andreas Dilger High Performance Data Division RUG 2016, Paris

Lustre * 2.12 and Beyond Andreas Dilger, Intel High Performance Data Division LUG 2018

Community Release Update

Community Release Update

Lustre 2.12 and Beyond. Andreas Dilger, Whamcloud

Olaf Weber Senior Software Engineer SGI Storage Software. Amir Shehata Lustre Network Engineer Intel High Performance Data Division

Fan Yong; Zhang Jinghai. High Performance Data Division

Small File I/O Performance in Lustre. Mikhail Pershin, Joe Gmitter Intel HPDD April 2018

Introduction The Project Lustre Architecture Performance Conclusion References. Lustre. Paul Bienkowski

Lustre on ZFS. Andreas Dilger Software Architect High Performance Data Division September, Lustre Admin & Developer Workshop, Paris, 2012

The current status of the adoption of ZFS* as backend file system for Lustre*: an early evaluation

Open SFS Roadmap. Presented by David Dillow TWG Co-Chair

Zhang, Hongchao

Introduction to Lustre* Architecture

DL-SNAP and Fujitsu's Lustre Contributions

Lustre HSM at Cambridge. Early user experience using Intel Lemur HSM agent

Intel & Lustre: LUG Micah Bhakti

Multi-Rail LNet for Lustre

Intel Enterprise Edition Lustre (IEEL-2.3) [DNE-1 enabled] on Dell MD Storage

Optimization of Lustre* performance using a mix of fabric cards

Lustre HPCS Design Overview. Andreas Dilger Senior Staff Engineer, Lustre Group Sun Microsystems

DDN s Vision for the Future of Lustre LUG2015 Robert Triendl

The modules covered in this course are:

Fujitsu s Contribution to the Lustre Community

Parallel File Systems Compared

Scalability Testing of DNE2 in Lustre 2.7 and Metadata Performance using Virtual Machines Tom Crowe, Nathan Lavender, Stephen Simms

Lustre File System. Proseminar 2013 Ein-/Ausgabe - Stand der Wissenschaft Universität Hamburg. Paul Bienkowski Author. Michael Kuhn Supervisor

LustreFS and its ongoing Evolution for High Performance Computing and Data Analysis Solutions

Re-Architecting Cloud Storage with Intel 3D XPoint Technology and Intel 3D NAND SSDs

Data life cycle monitoring using RoBinHood at scale. Gabriele Paciucci Solution Architect Bruno Faccini Senior Support Engineer September LAD

SFA12KX and Lustre Update

Ben Walker Data Center Group Intel Corporation

Lustre TM. Scalability

Lustre on ZFS. At The University of Wisconsin Space Science and Engineering Center. Scott Nolin September 17, 2013

Johann Lombardi High Performance Data Division

OpenZFS Performance Improvements

Enhancing Lustre Performance and Usability

Lustre / ZFS at Indiana University

Integration Path for Intel Omni-Path Fabric attached Intel Enterprise Edition for Lustre (IEEL) LNET

LUSTRE NETWORKING High-Performance Features and Flexible Support for a Wide Array of Networks White Paper November Abstract

Feedback on BeeGFS. A Parallel File System for High Performance Computing

Brent Gorda. General Manager, High Performance Data Division

LUG 2012 From Lustre 2.1 to Lustre HSM IFERC (Rokkasho, Japan)

Lustre overview and roadmap to Exascale computing

InfiniBand Networked Flash Storage

Mission-Critical Lustre at Santos. Adam Fox, Lustre User Group 2016

An Overview of Fujitsu s Lustre Based File System

LLNL Lustre Centre of Excellence

Lustre Metadata Fundamental Benchmark and Performance

Architecting a High Performance Storage System

High Performance Computing. NEC LxFS Storage Appliance

Challenges in making Lustre systems reliable

Upstreaming of Features from FEFS

Extremely Fast Distributed Storage for Cloud Service Providers

Intel Clear Containers. Amy Leeland Program Manager Clear Linux, Clear Containers And Ciao

FhGFS - Performance at the maximum

An Introduction to GPFS

LNet Roadmap & Development. Amir Shehata Lustre * Network Engineer Intel High Performance Data Division

Data Management. Parallel Filesystems. Dr David Henty HPC Training and Support

LCOC Lustre Cache on Client

HPE Scalable Storage with Intel Enterprise Edition for Lustre*

Applying DDN to Machine Learning

Multi-tenancy: a real-life implementation

A ClusterStor update. Torben Kling Petersen, PhD. Principal Architect, HPC

WITH INTEL TECHNOLOGIES

Lustre usages and experiences

Introducing Panasas ActiveStor 14

Store Process Analyze Collaborate Archive Cloud The HPC Storage Leader Invent Discover Compete

Lessons learned from Lustre file system operation

Parallel File Systems for HPC

TGCC OVERVIEW. 13 février 2014 CEA 10 AVRIL 2012 PAGE 1

Architecting Storage for Semiconductor Design: Manufacturing Preparation

5.4 - DAOS Demonstration and Benchmark Report

Fujitsu's Lustre Contributions - Policy and Roadmap-

Efficient Object Storage Journaling in a Distributed Parallel File System

HPSS Treefrog Summary MARCH 1, 2018

Intel Unite Plugin Guide for VDO360 Clearwater

PowerVault MD3 SSD Cache Overview

High-Performance Lustre with Maximum Data Assurance

DELL EMC ISILON F800 AND H600 I/O PERFORMANCE

Entry-level Intel RAID RS3 Controller Family

Orchestrating an OpenStack* based IoT Smart Home

Olaf Weber Senior Software Engineer SGI Storage Software

Milestone Solution Partner IT Infrastructure Components Certification Report

NPTEL Course Jan K. Gopinath Indian Institute of Science

The Transition to PCI Express* for Client SSDs

Application Performance on IME

Open Source Storage. Ric Wheeler Architect & Senior Manager April 30, 2012

Implementing a Hierarchical Storage Management system in a large-scale Lustre and HPSS environment

Clear CMOS after Hardware Configuration Changes

A GPFS Primer October 2005

Andreas Schneider. Markus Leberecht. Senior Cloud Solution Architect, Intel Deutschland. Distribution Sales Manager, Intel Deutschland

Intel Unite Solution. Linux* Release Notes Software version 3.2

Lustre at the OLCF: Experiences and Path Forward. Galen M. Shipman Group Leader Technology Integration

Lustre* is designed to achieve the maximum performance and scalability for POSIX applications that need outstanding streamed I/O.

Transcription:

Andreas Dilger Principal Lustre Engineer High Performance Data Division

Focus on Performance and Ease of Use Beyond just looking at individual features... Incremental but continuous improvements Performance and scalability enhancements Usability and manageability Access control and data security Improved filesystem availability 2

Overview of Upcoming Lustre * Features Features underway or landed for 2.9 ZFS Enhancements (Intel, LLNL) UID/GID mapping, Shared Secret Key (IU, OpenSFS * ) Subdirectory mounts, Server IO advice (DDN * ) File lockahead (Cray * ) Features starting or underway for 2.10+ Multi-Rail LNet (SGI *, Intel) Project quotas (DDN) Data on MDT (Intel) Composite File Layouts (Intel, ORNL) 3

ZFS Enhancements (Intel, LLNL 2.9+) Changes for using ZFS better 1MB+ ZFS blocksize (IO performance, LLNL) Improved file create performance (Intel) Snapshots of whole filesystem (Intel) Changes to core ZFS code Inode quota accounting (Intel) Large dnodes to improve xattr performance (LLNL) Declustered parity & distributed hot spaces to improve resilvering (Intel) Metadata allocation class to store all metadata on SSD/NVRAM (Intel) Reduce CPU with hardware-assisted checksums, compression (Intel) 4

Data Security for All Environments UID/GID Mapping and Shared Secret Key Crypto (IU, OpenSFS * 2.9) Data encryption for networks including RDMA (IB, OPA) Strong client node authentication into administrative node groups UID/GID mapping for WAN clients Block unauthorized clients by network Data isolation via filesystem containers (DDN 2.9) Subdirectory mounts with client authentication Usable with hosted, isolated environments 5

Miscellaneous Features Code cleanups (Cray, Intel, ORNL 2.9+) Update to match upstream kernel, port patches to/from kernel Patchless server kernel, ldiskfs patch cleanup Dead code removal for maintainability, security Server-side IO advice - ladvise (DDN 2.9) Tunables per file/extent to manage internal cache on OSS Object readahead and random IO blocksize Client write lockahead (Cray 2.9) Project Quotas (DDN * 2.10) Allow quota tracking on directory subtrees independent of UID/GID Not strictly hierarchical, can be multiple trees with the same project 6

Networking Improvements (Intel, SGI 2.10) Improved networking capabilities Support for EDR and FDR InfiniBand, MLX5 Intel OmniPath Architecture network support RPC crypto for RDMA networks like IB and OPA Multi-Rail support for all network types 7

Improved Small File Performance (Intel 2.10+) Data-on-MDT optimizes small file IO Avoid OST overhead (data, lock RPCs) High-IOPS MDTs (mirrored SSD vs. RAID-6 HDD) Avoid contention with streaming IO to OSTs Prefetch file data with metadata Size on MDT for files Manage MDT usage by quota Complementary with DNE 2 striped directories Scale small file IOPS with multiple MDTs open, write data, read, attr Client MDS layout, lock, size, read data Small file IO directly to MDS 8

Composite File Layouts (Intel, ORNL 2.10+) Innovation in Storage Usage Progressive File Layouts simplify usage and provide new options Optimize performance for diverse users/applications Low overhead for small files, high bandwidth for large files Lower new user usage barrier and administrative burden Multiple storage classes within a single file HDD or SSD, mirror or RAID Example progressive file layout with 3 components 1 stripe [0, 32MB) 4 stripes [32MB, 1GB) 128 stripes [1GB, ) 9

Improved Data Availability (Intel 2.11+) File Level Redundancy provides significant value and functionality for HPC Configure on a per-file/dir basis (e.g. mirror input files and one daily checkpoint) Higher availability for server/network failure - finally better than HA failover Robustness against data loss/corruption - mirror or M+N erasure coding for stripes Increased read speed for widely shared files - mirror input data across many OSTs Replicate/migrate files between storage classes - NVRAM->SSD->HDD Local vs. remote replicas Partial HSM file restore File versioning,... Replica 0 Object j (primary, preferred) Replica 1 Object k (stale) delayed resync 10

Advanced Lustre * Research Intel Parallel Computing Centers Uni Hamburg + German Client Research Centre (DKRZ) Adaptive optimized ZFS data compression Client-side data compression GSI Helmholtz Centre for Heavy Ion Research TSM HSM copytool for Lustre University of California Santa Cruz Automated client-side load balancing Johannes Gutenberg University Mainz Global adaptive Network Request Scheduler Lawrence Berkeley National Laboratory Spark and Hadoop on Lustre 11

Legal Notices and Disclaimers Intel technologies features and benefits depend on system configuration and may require enabled hardware, software or service activation. Learn more at intel.com, or from the OEM or retailer. No computer system can be absolutely secure. Tests document performance of components on a particular test, in specific systems. Differences in hardware, software, or configuration will affect actual performance. Consult other sources of information to evaluate performance as you consider your purchase. For more complete information about performance and benchmark results, visit http://www.intel.com/performance. Intel disclaims all express and implied warranties, including, without limitation, the implied warranties and merchantability, fitness for a particular purpose, and non-infringement, as well as any warranty arising from course of performance, course of dealing or usage in trade. Intel, the Intel logo and others are trademarks of Intel Corporation in the U.S. and/or other countries. *Other names and brands may be claimed as the property of others. 2016 Intel Corporation.