Implementierung eines Dynamic Remote Storage Systems (DRS) für Applikationen mit hohen IO Anforderungen
|
|
- Roy Powers
- 5 years ago
- Views:
Transcription
1 Implementierung eines Dynamic Remote Storage Systems (DRS) für Applikationen mit hohen IO Anforderungen Jürgen Salk, Christian Mosch, Matthias Neuer, Karsten Siegmund, Volodymyr Kushnarenko, Stefan Kombrink, Thomas Nau, Stefan Wesner Steinbuch Centre for Computing (SCC) Founded by: Funded by: SC New Orleans
2 HPC Tier classification in Baden-Württemberg / bwhpc European High- European high performance computing center Gauss Center for Supercomputing National high performance computing center Hazel Hen at HLRS Stuttgart Research high perfomance computing ForHLR at SCC KIT Karlsruhe bwunicluster and bwforclusters performance Computing Centers (Tier 0) National Highperformance Computing Centers (Tier 1) Supraregional state-wide HPC-Centers (Tier 2) Regional HPC-Enablers (Tier 3) bwforcluster JUSTUS Page 2
3 Introduction Economy and social science General purpose, teaching Molecular life science Bioinformatics Mannheim Heidelberg Karlsruhe Neurosciences Astrophysics Tübingen Ulm Freiburg Microsystems engineering Page 3 Elementary particle physics Computational Chemistry
4 Introduction Economy and social science General purpose, teaching Molecular life science Bioinformatics Mannheim Heidelberg Karlsruhe Neurosciences Astrophysics Tübingen Ulm Freiburg Microsystems engineering Page 4 Elementary particle physics Computational Chemistry
5 Introduction Page 5 Motivation: JUSTUS Cluster in Ulm for Computational Chemistry. Demand for high I/O for coupled cluster calculations (Molpro, Gaussian) Presented techniques not restricted to computational chemistry Deciding which I/O system is the best for specific application: How high is the I/O demand and what is the read / write pattern?
6 Tracing Page 6
7 iostat iostat gives many metrics for device usage Statistics are per block device and partition Useful: %util: Percentage of time the device spent doing I/O iostat -dkx 1 -d -k -x 1 Page 7 Display the device utilization report. Display statistics in kilobytes per second. Display extended statistics. 1 second interval between reports
8 iostat Watch out for caching effects: Same job 16 GB RAM Page 8
9 iostat Watch out for caching effects: Same job 16 GB RAM Page 9 48 GB RAM
10 strace strace provides insight from the applications point of view Intercepts and records system calls strace -T -ttt -f -e trace=file,desc -o trace.out prog -T -ttt -f -e trace=file,desc -o Page 10 report elapsed time microseconds timing include child processes limit the trace to I/O system calls write output to file
11 strace Page 11 Large output Post-processing and visualization via custom scripts Result example: IOPS over time, access pattern
12 strace Blocksize per read operation (bytes) Effective seek offsets (bytes) Effective seek offsets (bytes)many seeks are noops... Many seeks are noops... but not all of them.... but not... all of them. time (sec) Page 12 J. Salk
13 blktrace Page 13 blktrace generates traces on blocklayer level Output is binary, can be converted to human-readable via blkparse blktrace -d /dev/sda -o - blkparse -i -d -o - device to trace (multiple devices possible) redirect output to stdout -i - read from stdin
14 blktrace Page 14 Prints device and accessed block number Useful for finding out access pattern on device level What are the effects of filesystem, raid, lvm,? Example: Used dd to write 10 test files (100 MB) one after another (no delete) Test case 1: Write all files into a single shared folders Test case 2: Write all files into individual sub folders Simultaneously run blktrace in live mode Record timestamp and location (as block numbers) of write operations on device Plot results for xfs filesystem
15 blktrace Page 15
16 Hybrid Storage Systems Page 16
17 Typical behavior of QC-jobs Informations gathered using the tracing methods: More read than write IOPs => read performance more important => cache-like solutions might be interesting IOPs beyond capabilities of typical hard disk => probably SSDs Coverage of current and (expected) future demands on scratch space purely by SSDs is too expensive => hybrid (SSD + HDD) solution economically reasonable for very large jobs Domains with and without I/O on each node => combination of central block storage (shared) and node local SSD => highest overall throughput Page 17
18 JUSTUS: Remote attached block storage NEC SNA460/060 Storage Array (+ extension unit) Access by means of SCSI RDMA Protocol (SRP) Page x 4 TB HDDs (Nearline SAS, 7.2k rpm) 2 Controllers, each one with 2 IB QDR ports higher throughput and lower latency than with TCP/IP communication protocol Local OS page caching still available just like with local disks No overhead introduced by cache coherency How to combine SSDs and remote attached storage?
19 Plain Filesystem Program SSD Page 19 Disk Separate use of SSD and remote scratch, two mount points SSD filesystem for hot files Ideally: Program itself makes decision Alternatively: Can be configured (e.g. Molpro), but Coarse splitting policy Needs a deep understanding of the algorithms used
20 Cache Based Approach Program Page 20 SSD Disk Use local SSDs as cache for remote attached backend Should work automatically Solutions: Intel CAS - Cache Acceleration SW Flash Cache - Facebook Bcache - Block layer Cache (Kernel > 3.10)
21 Cache Based Approach Page 21 Comparison between solutions Name Version Intel CAS 2.8 Flashcache Performance Footprint per GB cache Commercial/G Best 14 MB PL GPL 90% of CAS 6 MB bcache GPL License 90% of CAS 3 MB
22 Concatenated Storage System Program Disk SSD Page 22 LVM concatenated hybrid FS SSD contributes to storage capacity Write with preference for physical extents from SSD
23 30 % of I/O 70 % of I/O 2nd job 1st job history effect: shift of block baseline after file deletion Time [sec] Page 23 Remote attached device (50% of total size) Write and read operations Local SSD (50% of total size) Location on concat. filesystem [block number] blktrace of 2 * Molpro LCCSD-j2-c8 LVM compound of 3*RAID0-SSD + remote block storage J. Salk
24 Concatenated Storage System Ensure that pvcreate vgcreate lvcreate lvextend /dev/ssd /dev/hdd vg0 /dev/ssd /dev/hdd -l 100%PVS -n lv0 vg0 /dev/ssd -L $LVSIZE /dev/vg0/lv0 Choose the right filesystem: Page 24 Blocks with small numbers are placed on the fast storage system Blocks with high numbers are placed on the slow storage system ext4: doesn't fill from beginning after file deletion xfs: scatter files when placed in different directories
25 Implementation Page 25
26 Implementation How it should work from a users perspective: msub -l nodes=1:ppn=16 -l gres=scratch:100%drs_concat:5 job.sh Submits a job with MOAB which requests Page 26 1 node 16 processes per node 100 GB SSD space per process 5 TB space from remote attached storage Combine the two storage systems via lvm When job starts there is a mountpoint with 6.6 TB storage space
27 Implementation Components Moab: job scheduler Torque: resource manager drs_client: used for sending queries and commands to drs_broker drs_broker: keeps track of remote storage resource allocations and configures the storage targets Moab scheduler Torque batchsystem drs_client staticclient dynamicclient Mounthelper drs_broker staticbroker dynamicbroker Unmounthelper Dummy NetAppSRP Target Target ISER Target XFS filesystem Kernel SRP Initiator Kernel iser Initiator Page 27 Go RPC IB SRP IB iser NetApp SRP Target NetApp iser Target
28 Implementation Workflow 1: ask if resources are available 2: start job 3: request the volumes 4: create volume, register host 5: create lvm- or cache-device, create filesystem, mount Moab Scheduler 1a. resource inquiry 2. selection + job start 1b. resource 3b. resource inquiry request DRS broker Page 28 Node Prolog/ Epilog Prolog/ Epilog 5 SRP configuration + LVM configuration + filesystem preparation 3a. resource request DRS client Node volume 4. host registration + volume mapping return of identifier volume SRP target
29 Moab Implementation (and pitfalls) DRS volumes are treated as shared (floating) cluster resource Similar to integrating floating software licenses in scheduler Used Moab's existing interface for external FLEXlm licence managers The DRS Broker acts as license server in order to keep track of configured volumes and volumens allocated by running jobs. (Just like lmgrd does for FLEXlm license tokens) The DRS Client queries the DRS Broker for that information. (Just like lmstat does for FLEXlm license tokens) Straightforward we initially thought... but... as soon as licensing interface enabled in configuration, Moab fails to schedule jobs correctly to nodes according to the memory requirements of the jobs: Dedicated memory were always off by a factor of ppn for all submitted jobs. Turned out to be bug in Moab (at least in our somewhat ancient Moab 7 version) Needs workaround by means of Moab's submitfilter facility to always multiply requested mem/pmem by a factor of ppn for all jobs. Page 29
30 Moab Implementation (and pitfalls) At job submission the user can not only specify the amount of dynamic remote scratch to allocate, but also how the tiered scratch space shall be assembled for the job. Has been achieved by introduction of alias names for requested DRS resources, e.g.: -l gres=drs_concat:5 use 5 TB DRS space, concatenate hybrid FS with local SSDs -l gres=drs_cache:5 use 5 TB DRS space, cached by local SSDs Alias names can be easily be configured in DRS Broker Requested DRS resources are passed to job environment (as 5 th argument to prologue script) by their alias name, such that prologue script also knows what to do with the remote attached storage. Page 30
31 Moab Implementation (and pitfalls) Autoadjust Node Access Policy of job whenever DRS resource is requested Default Node Access Policy of JUSTUS cluster is SINGLEUSER (i.e. multiple jobs of the same user may run side by side on one node). For jobs with dedicated DRS resource, we need to build, assemble and format the local scratch filesystem in the job prologue (and also revert these changes in the job epilogue). In order to prevent interferences with other jobs, any job that requests DRS resources must run on a dedicated node with no other job running on that very same node, i.e. node access policy attribute of job must be automatically adjusted to SINGLEJOB. This has also been achieved by implementing appropriate rules in Moab's submitfilter facility. Page 31
32 Moab Implementation (and pitfalls) Final provisioning of DRS resources at compute node: A customized torque prologue, which runs under with root privileges on the job execution nodes, decides which remote storage resources are requested for the job (amount of DRS storage and what to do with it, according to its alias name). The prologue script runs the DRS Client, which in turn queries the DRS Broker, which makes those resources available on the remote storage device and keeps track of the allocation. Finally, the SRP initiator is configured and the scratch filesystem is created and mounted. When the job ends, a customized torque epilogue is used to unmount the scratch filesystem, clean up and to signal the broker that resources are no longer in use. Broker also takes care of unprovisioning DRS volumens on storage device side Page 32
33 Benchmarks Page 33
34 Cache Based Approach Page 34 Synthetic fio-benchmark modeling real quantum chemistry job Single process run System: 128 GB RAM, a RAID0 array of 4 SSDs with a total of 850 GB
35 Concatenated Storage System Page 35 Molpro LCCSD benchmark
36 Conclusions and outlook Page 36 Comprehensive I/O analysis possible with standard linux tools Combination of local and remote storage can be a solution Caching solutions didn't work well in our case LVM reveals good performance Flexible implementation possible with Moab Ready for production mode in Q Investigation and improvement with NEC continues
37 Thank you for your attention. Questions? Matthias Neuer. Communication and Information Center (kiz) Infrastructure Dept. Scientific Software & Compute Services Albert-Einstein-Allee Ulm Germany Acknowledgements: Installation * Thanks to the Deutsche Forschungsgemeinschaft (DFG) and the Ministry for Science, Research and Arts Baden-Württemberg for funding the project. * Thanks to NEC for their support and cooperation. * Thanks to the bwhpc-c5 team Baden-Württemberg. * Thanks to all of my colleagues from the SSCS team for their contributions to this talk. Page 37
Hands-On Workshop bwunicluster June 29th 2015
Hands-On Workshop bwunicluster June 29th 2015 Agenda Welcome Introduction to bwhpc and the bwunicluster Modules - Software Environment Management Job Submission and Monitoring Interactive Work and Remote
More informationbwfortreff bwhpc user meeting
bwfortreff bwhpc user meeting bwhpc Competence Center MLS&WISO Universitätsrechenzentrum Heidelberg Rechenzentrum der Universität Mannheim Steinbuch Centre for Computing (SCC) Funding: www.bwhpc-c5.de
More informationVirtualization of the ATLAS Tier-2/3 environment on the HPC cluster NEMO
Virtualization of the ATLAS Tier-2/3 environment on the HPC cluster NEMO Ulrike Schnoor (CERN) Anton Gamel, Felix Bührer, Benjamin Rottler, Markus Schumacher (University of Freiburg) February 02, 2018
More informationAccess: bwunicluster, bwforcluster, ForHLR
Access: bwunicluster, bwforcluster, ForHLR Shamna Shamsudeen, SCC, KIT Steinbuch Centre for Computing (SCC) Funding: www.bwhpc-c5.de Outline Introduction Registration Processes bwunicluster bwforcluster
More informationOperating two InfiniBand grid clusters over 28 km distance
Operating two InfiniBand grid clusters over 28 km distance Sabine Richling, Steffen Hau, Heinz Kredel, Hans-Günther Kruse IT-Center University of Heidelberg, Germany IT-Center University of Mannheim, Germany
More informationExtraordinary HPC file system solutions at KIT
Extraordinary HPC file system solutions at KIT Roland Laifer STEINBUCH CENTRE FOR COMPUTING - SCC KIT University of the State Roland of Baden-Württemberg Laifer Lustre and tools for ldiskfs investigation
More informationNEC Express5800 A2040b 22TB Data Warehouse Fast Track. Reference Architecture with SW mirrored HGST FlashMAX III
NEC Express5800 A2040b 22TB Data Warehouse Fast Track Reference Architecture with SW mirrored HGST FlashMAX III Based on Microsoft SQL Server 2014 Data Warehouse Fast Track (DWFT) Reference Architecture
More informationTriton file systems - an introduction. slide 1 of 28
Triton file systems - an introduction slide 1 of 28 File systems Motivation & basic concepts Storage locations Basic flow of IO Do's and Don'ts Exercises slide 2 of 28 File systems: Motivation Case #1:
More informationAccelerate Applications Using EqualLogic Arrays with directcache
Accelerate Applications Using EqualLogic Arrays with directcache Abstract This paper demonstrates how combining Fusion iomemory products with directcache software in host servers significantly improves
More informationNow SAML takes it all:
Now SAML takes it all: Federation of non Web-based Services in the State of Baden-Württemberg Sebastian Labitzke Karlsruhe Institute of Technology (KIT) Steinbuch Centre for Computing (SCC) labitzke@kit.edu
More informationAssistance in Lustre administration
Assistance in Lustre administration Roland Laifer STEINBUCH CENTRE FOR COMPUTING - SCC KIT University of the State of Baden-Württemberg and National Laboratory of the Helmholtz Association www.kit.edu
More informationSystem Administration. Storage Systems
System Administration Storage Systems Agenda Storage Devices Partitioning LVM File Systems STORAGE DEVICES Single Disk RAID? RAID Redundant Array of Independent Disks Software vs. Hardware RAID 0, 1,
More informationFeedback on BeeGFS. A Parallel File System for High Performance Computing
Feedback on BeeGFS A Parallel File System for High Performance Computing Philippe Dos Santos et Georges Raseev FR 2764 Fédération de Recherche LUmière MATière December 13 2016 LOGO CNRS LOGO IO December
More informationLessons learned from Lustre file system operation
Lessons learned from Lustre file system operation Roland Laifer STEINBUCH CENTRE FOR COMPUTING - SCC KIT University of the State of Baden-Württemberg and National Laboratory of the Helmholtz Association
More informationToward SLO Complying SSDs Through OPS Isolation
Toward SLO Complying SSDs Through OPS Isolation October 23, 2015 Hongik University UNIST (Ulsan National Institute of Science & Technology) Sam H. Noh 1 Outline Part 1: FAST 2015 Part 2: Beyond FAST 2
More informationA Long-distance InfiniBand Interconnection between two Clusters in Production Use
A Long-distance InfiniBand Interconnection between two Clusters in Production Use Sabine Richling, Steffen Hau, Heinz Kredel, Hans-Günther Kruse IT-Center, University of Heidelberg, Germany IT-Center,
More informationWhite Paper. File System Throughput Performance on RedHawk Linux
White Paper File System Throughput Performance on RedHawk Linux By: Nikhil Nanal Concurrent Computer Corporation August Introduction This paper reports the throughput performance of the,, and file systems
More informationIBM V7000 Unified R1.4.2 Asynchronous Replication Performance Reference Guide
V7 Unified Asynchronous Replication Performance Reference Guide IBM V7 Unified R1.4.2 Asynchronous Replication Performance Reference Guide Document Version 1. SONAS / V7 Unified Asynchronous Replication
More informationMAHA. - Supercomputing System for Bioinformatics
MAHA - Supercomputing System for Bioinformatics - 2013.01.29 Outline 1. MAHA HW 2. MAHA SW 3. MAHA Storage System 2 ETRI HPC R&D Area - Overview Research area Computing HW MAHA System HW - Rpeak : 0.3
More informationPresented by: Nafiseh Mahmoudi Spring 2017
Presented by: Nafiseh Mahmoudi Spring 2017 Authors: Publication: Type: ACM Transactions on Storage (TOS), 2016 Research Paper 2 High speed data processing demands high storage I/O performance. Flash memory
More informationbwfdm Communities - a Research Data Management Initiative in the State of Baden-Wuerttemberg
bwfdm Communities - a Research Data Management Initiative in the State of Baden-Wuerttemberg Karlheinz Pappenberger Tromsø, 9th Munin Conference on Scholarly Publishing, 27/11/2014 Overview 1) Federalism
More informationVirtual Storage Tier and Beyond
Virtual Storage Tier and Beyond Manish Agarwal Sr. Product Manager, NetApp Santa Clara, CA 1 Agenda Trends Other Storage Trends and Flash 5 Min Rule Issues for Flash Dedupe and Flash Caching Architectural
More informationFhGFS - Performance at the maximum
FhGFS - Performance at the maximum http://www.fhgfs.com January 22, 2013 Contents 1. Introduction 2 2. Environment 2 3. Benchmark specifications and results 3 3.1. Multi-stream throughput................................
More informationData storage on Triton: an introduction
Motivation Data storage on Triton: an introduction How storage is organized in Triton How to optimize IO Do's and Don'ts Exercises slide 1 of 33 Data storage: Motivation Program speed isn t just about
More informationManage your disk space... for free :)
Manage your disk space... for free :) Julien Wallior Plug Central Agenda Background RAID LVM Some basics Practice Booting on a raid device What is that? How it works Hardware raid... if you really want
More informationNVM Express over Fabrics Storage Solutions for Real-time Analytics
NVM Express over Fabrics Storage Solutions for Real-time Analytics Presented by Paul Prince, CTO Santa Clara, CA 1 NVMe Over Fabrics NVMf Why do we need NVMf? What is it? How does it fit in the Market?
More informationEarly Evaluation of the "Infinite Memory Engine" Burst Buffer Solution
Early Evaluation of the "Infinite Memory Engine" Burst Buffer Solution Wolfram Schenck Faculty of Engineering and Mathematics, Bielefeld University of Applied Sciences, Bielefeld, Germany Salem El Sayed,
More informationOptimize Storage Performance with Red Hat Enterprise Linux
Optimize Storage Performance with Red Hat Enterprise Linux Mike Snitzer Senior Software Engineer, Red Hat 09.03.2009 2 Agenda Block I/O Schedulers Linux DM Multipath Readahead I/O
More informationMFT / Linux Setup Documentation May 25, 2008
MFT / Linux Setup Documentation May 25, 2008 1. Loading the MFT software. The MFT software actually uses a driver called Fast Block Device or fbd. The MFT software is designed to run from /usr/local/fbd.
More informationImproving Performance using the LINUX IO Scheduler Shaun de Witt STFC ISGC2016
Improving Performance using the LINUX IO Scheduler Shaun de Witt STFC ISGC2016 Role of the Scheduler Optimise Access to Storage CPU operations have a few processor cycles (each cycle is < 1ns) Seek operations
More informationUsers and utilization of CERIT-SC infrastructure
Users and utilization of CERIT-SC infrastructure Equipment CERIT-SC is an integral part of the national e-infrastructure operated by CESNET, and it leverages many of its services (e.g. management of user
More informationEvaluation Report: Improving SQL Server Database Performance with Dot Hill AssuredSAN 4824 Flash Upgrades
Evaluation Report: Improving SQL Server Database Performance with Dot Hill AssuredSAN 4824 Flash Upgrades Evaluation report prepared under contract with Dot Hill August 2015 Executive Summary Solid state
More informationRapid Prototyping and Evaluation of Intelligence Functions of Active Storage Devices
Rapid Prototyping and Evaluation of Intelligence Functions of Active Storage Devices Yongsoo Joo Embedded Software Research Center Ewha Womans University This research was supported by Basic Science Research
More informationDesigning Next Generation FS for NVMe and NVMe-oF
Designing Next Generation FS for NVMe and NVMe-oF Liran Zvibel CTO, Co-founder Weka.IO @liranzvibel Santa Clara, CA 1 Designing Next Generation FS for NVMe and NVMe-oF Liran Zvibel CTO, Co-founder Weka.IO
More informationDeploy a High-Performance Database Solution: Cisco UCS B420 M4 Blade Server with Fusion iomemory PX600 Using Oracle Database 12c
White Paper Deploy a High-Performance Database Solution: Cisco UCS B420 M4 Blade Server with Fusion iomemory PX600 Using Oracle Database 12c What You Will Learn This document demonstrates the benefits
More informationReliable Storage for HA, DR, Clouds and Containers Philipp Reisner, CEO LINBIT
Reliable Storage for HA, DR, Clouds and Containers Philipp Reisner, CEO LINBIT LINBIT - the company behind it COMPANY OVERVIEW TECHNOLOGY OVERVIEW Developer of DRBD 100% founder owned Offices in Europe
More informationHPC/Cloud Hybrids for Efficient Resource Allocation and Throughput. Multicore World, Wellington, New Zealand, Feb 2017
HPC/Cloud Hybrids for Efficient Resource Allocation and Throughput Multicore World, Wellington, New Zealand, Feb 2017 lev@levlafayette.com It All Begins at Multicore World At the last Multicore World A
More informationDDN About Us Solving Large Enterprise and Web Scale Challenges
1 DDN About Us Solving Large Enterprise and Web Scale Challenges History Founded in 98 World s Largest Private Storage Company Growing, Profitable, Self Funded Headquarters: Santa Clara and Chatsworth,
More informationPerformance Modeling and Analysis of Flash based Storage Devices
Performance Modeling and Analysis of Flash based Storage Devices H. Howie Huang, Shan Li George Washington University Alex Szalay, Andreas Terzis Johns Hopkins University MSST 11 May 26, 2011 NAND Flash
More informationChoosing Hardware and Operating Systems for MySQL. Apr 15, 2009 O'Reilly MySQL Conference and Expo Santa Clara,CA by Peter Zaitsev, Percona Inc
Choosing Hardware and Operating Systems for MySQL Apr 15, 2009 O'Reilly MySQL Conference and Expo Santa Clara,CA by Peter Zaitsev, Percona Inc -2- We will speak about Choosing Hardware Choosing Operating
More informationAzor: Using Two-level Block Selection to Improve SSD-based I/O caches
Azor: Using Two-level Block Selection to Improve SSD-based I/O caches Yannis Klonatos, Thanos Makatos, Manolis Marazakis, Michail D. Flouris, Angelos Bilas {klonatos, makatos, maraz, flouris, bilas}@ics.forth.gr
More informationData Movement & Tiering with DMF 7
Data Movement & Tiering with DMF 7 Kirill Malkin Director of Engineering April 2019 Why Move or Tier Data? We wish we could keep everything in DRAM, but It s volatile It s expensive Data in Memory 2 Why
More informationFile System Performance Tuning For Gdium Example of general methods. Coly Li Software Engineer SuSE Labs, Novell.inc
File System Performance Tuning For Gdium Example of general methods Coly Li Software Engineer SuSE Labs, Novell.inc Content Brief Introduction to Gdium Storage Module of Gdium I/O Profiling Methods Key
More informationChallenges in making Lustre systems reliable
Challenges in making Lustre systems reliable Roland Laifer STEINBUCH CENTRE FOR COMPUTING - SCC KIT University of the State Roland of Baden-Württemberg Laifer Challenges and in making Lustre systems reliable
More informationArch Linux with an SSD Cache using LVM on BCache. Jeremy Runyan
Arch Linux with an SSD Cache using LVM on BCache Jeremy Runyan 1 Table of Contents Introduction 3 Materials. 3 Prepare.... 4 Create Partitions.. 4-6 Format and Mount Partitions.. 6 Install Arch Linux.
More informationRed Hat Gluster Storage performance. Manoj Pillai and Ben England Performance Engineering June 25, 2015
Red Hat Gluster Storage performance Manoj Pillai and Ben England Performance Engineering June 25, 2015 RDMA Erasure Coding NFS-Ganesha New or improved features (in last year) Snapshots SSD support Erasure
More informationREQUEST FOR PROPOSAL FOR PROCUREMENT OF
REQUEST FOR PROPOSAL FOR PROCUREMENT OF Upgrade of department RFP No.: SBI/GITC/ATM/2018-19/481 : 18/05/2018 Corrigendum II dated 30/05/2018 to Ref: SBI/GITC/ATM/2018-19/481 : 18/05/2018 State Bank of
More informationNew Storage Technologies First Impressions: SanDisk IF150 & Intel Omni-Path. Brian Marshall GPFS UG - SC16 November 13, 2016
New Storage Technologies First Impressions: SanDisk IF150 & Intel Omni-Path Brian Marshall GPFS UG - SC16 November 13, 2016 Presenter Background Brian Marshall Computational Scientist at Virginia Tech
More informationStorage Technologies - 3
Storage Technologies - 3 COMP 25212 - Lecture 10 Antoniu Pop antoniu.pop@manchester.ac.uk 1 March 2019 Antoniu Pop Storage Technologies - 3 1 / 20 Learning Objectives - Storage 3 Understand characteristics
More informationData center requirements
Prerequisites, page 1 Data center workflow, page 2 Determine data center requirements, page 2 Gather data for initial data center planning, page 2 Determine the data center deployment model, page 3 Determine
More informationDDN. DDN Updates. DataDirect Neworks Japan, Inc Nobu Hashizume. DDN Storage 2018 DDN Storage 1
1 DDN DDN Updates DataDirect Neworks Japan, Inc Nobu Hashizume DDN Storage 2018 DDN Storage 1 2 DDN A Broad Range of Technologies to Best Address Your Needs Your Use Cases Research Big Data Enterprise
More informationPractice of Software Development: Dynamic scheduler for scientific simulations
Practice of Software Development: Dynamic scheduler for scientific simulations @ SimLab EA Teilchen STEINBUCH CENTRE FOR COMPUTING - SCC KIT Universität des Landes Baden-Württemberg und nationales Forschungszentrum
More informationOncilla - a Managed GAS Runtime for Accelerating Data Warehousing Queries
Oncilla - a Managed GAS Runtime for Accelerating Data Warehousing Queries Jeffrey Young, Alex Merritt, Se Hoon Shon Advisor: Sudhakar Yalamanchili 4/16/13 Sponsors: Intel, NVIDIA, NSF 2 The Problem Big
More informationA Comparative Study of Microsoft Exchange 2010 on Dell PowerEdge R720xd with Exchange 2007 on Dell PowerEdge R510
A Comparative Study of Microsoft Exchange 2010 on Dell PowerEdge R720xd with Exchange 2007 on Dell PowerEdge R510 Incentives for migrating to Exchange 2010 on Dell PowerEdge R720xd Global Solutions Engineering
More informationOutline. March 5, 2012 CIRMMT - McGill University 2
Outline CLUMEQ, Calcul Quebec and Compute Canada Research Support Objectives and Focal Points CLUMEQ Site at McGill ETS Key Specifications and Status CLUMEQ HPC Support Staff at McGill Getting Started
More informationSPECIFICATION FOR NETWORK ATTACHED STORAGE (NAS) TO BE FILLED BY BIDDER. NAS Controller Should be rack mounted with a form factor of not more than 2U
SPECIFICATION FOR NETWORK ATTACHED STORAGE (NAS) TO BE FILLED BY BIDDER S.No. Features Qualifying Minimum Requirements No. of Storage 1 Units 2 Make Offered 3 Model Offered 4 Rack mount 5 Processor 6 Memory
More informationNext-Generation NVMe-Native Parallel Filesystem for Accelerating HPC Workloads
Next-Generation NVMe-Native Parallel Filesystem for Accelerating HPC Workloads Liran Zvibel CEO, Co-founder WekaIO @liranzvibel 1 WekaIO Matrix: Full-featured and Flexible Public or Private S3 Compatible
More informationOperational characteristics of a ZFS-backed Lustre filesystem
Operational characteristics of a ZFS-backed Lustre filesystem Daniel Kobras science + computing ag IT-Dienstleistungen und Software für anspruchsvolle Rechnernetze Tübingen München Berlin Düsseldorf science+computing
More informationPerformance Sentry VM Provider Objects April 11, 2012
Introduction This document describes the Performance Sentry VM (Sentry VM) Provider performance data objects defined using the VMware performance groups and counters. This version of Performance Sentry
More informationIsilon: Raising The Bar On Performance & Archive Use Cases. John Har Solutions Product Manager Unstructured Data Storage Team
Isilon: Raising The Bar On Performance & Archive Use Cases John Har Solutions Product Manager Unstructured Data Storage Team What we ll cover in this session Isilon Overview Streaming workflows High ops/s
More informationForensic Toolkit System Specifications Guide
Forensic Toolkit System Specifications Guide February 2012 When it comes to performing effective and timely investigations, we recommend examiners take into consideration the demands the software, and
More informationSHRD: Improving Spatial Locality in Flash Storage Accesses by Sequentializing in Host and Randomizing in Device
SHRD: Improving Spatial Locality in Flash Storage Accesses by Sequentializing in Host and Randomizing in Device Hyukjoong Kim 1, Dongkun Shin 1, Yun Ho Jeong 2 and Kyung Ho Kim 2 1 Samsung Electronics
More informationAssessing performance in HP LeftHand SANs
Assessing performance in HP LeftHand SANs HP LeftHand Starter, Virtualization, and Multi-Site SANs deliver reliable, scalable, and predictable performance White paper Introduction... 2 The advantages of
More informationPivot3 Acuity with Microsoft SQL Server Reference Architecture
Pivot3 Acuity with Microsoft SQL Server 2014 Reference Architecture How to Contact Pivot3 Pivot3, Inc. General Information: info@pivot3.com 221 West 6 th St., Suite 750 Sales: sales@pivot3.com Austin,
More informationECE 598 Advanced Operating Systems Lecture 14
ECE 598 Advanced Operating Systems Lecture 14 Vince Weaver http://www.eece.maine.edu/~vweaver vincent.weaver@maine.edu 19 March 2015 Announcements Homework #4 posted soon? 1 Filesystems Often a MBR (master
More informationOn-demand provisioning of HEP compute resources on cloud sites and shared HPC centers
On-demand provisioning of HEP compute resources on cloud sites and shared HPC centers CHEP 2016 - San Francisco, United States of America Gunther Erli, Frank Fischer, Georg Fleig, Manuel Giffels, Thomas
More informationPlexistor SDM User Guide
Plexistor SDM User Guide Latest version: m1fs v1.7.1 February 2016 The latest version of this document is available online at www.plexistor.com/sdm_user_guide Plexistor SDM Community Edition (CE) is provided
More informationEmerging Technologies for HPC Storage
Emerging Technologies for HPC Storage Dr. Wolfgang Mertz CTO EMEA Unstructured Data Solutions June 2018 The very definition of HPC is expanding Blazing Fast Speed Accessibility and flexibility 2 Traditional
More informationPowerVault MD3 SSD Cache Overview
PowerVault MD3 SSD Cache Overview A Dell Technical White Paper Dell Storage Engineering October 2015 A Dell Technical White Paper TECHNICAL INACCURACIES. THE CONTENT IS PROVIDED AS IS, WITHOUT EXPRESS
More informationData Management. Parallel Filesystems. Dr David Henty HPC Training and Support
Data Management Dr David Henty HPC Training and Support d.henty@epcc.ed.ac.uk +44 131 650 5960 Overview Lecture will cover Why is IO difficult Why is parallel IO even worse Lustre GPFS Performance on ARCHER
More informationThe Optimal CPU and Interconnect for an HPC Cluster
5. LS-DYNA Anwenderforum, Ulm 2006 Cluster / High Performance Computing I The Optimal CPU and Interconnect for an HPC Cluster Andreas Koch Transtec AG, Tübingen, Deutschland F - I - 15 Cluster / High Performance
More informationRAIDIX Data Storage Solution. Clustered Data Storage Based on the RAIDIX Software and GPFS File System
RAIDIX Data Storage Solution Clustered Data Storage Based on the RAIDIX Software and GPFS File System 2017 Contents Synopsis... 2 Introduction... 3 Challenges and the Solution... 4 Solution Architecture...
More informationIntroduction The Project Lustre Architecture Performance Conclusion References. Lustre. Paul Bienkowski
Lustre Paul Bienkowski 2bienkow@informatik.uni-hamburg.de Proseminar Ein-/Ausgabe - Stand der Wissenschaft 2013-06-03 1 / 34 Outline 1 Introduction 2 The Project Goals and Priorities History Who is involved?
More informationStorage Protocol Offload for Virtualized Environments Session 301-F
Storage Protocol Offload for Virtualized Environments Session 301-F Dennis Martin, President August 2016 1 Agenda About Demartek Offloads I/O Virtualization Concepts RDMA Concepts Overlay Networks and
More informationIntel Enterprise Edition Lustre (IEEL-2.3) [DNE-1 enabled] on Dell MD Storage
Intel Enterprise Edition Lustre (IEEL-2.3) [DNE-1 enabled] on Dell MD Storage Evaluation of Lustre File System software enhancements for improved Metadata performance Wojciech Turek, Paul Calleja,John
More informationThe RAMDISK Storage Accelerator
The RAMDISK Storage Accelerator A Method of Accelerating I/O Performance on HPC Systems Using RAMDISKs Tim Wickberg, Christopher D. Carothers wickbt@rpi.edu, chrisc@cs.rpi.edu Rensselaer Polytechnic Institute
More informationDDN. DDN Updates. Data DirectNeworks Japan, Inc Shuichi Ihara. DDN Storage 2017 DDN Storage
DDN DDN Updates Data DirectNeworks Japan, Inc Shuichi Ihara DDN A Broad Range of Technologies to Best Address Your Needs Protection Security Data Distribution and Lifecycle Management Open Monitoring Your
More informationSun Lustre Storage System Simplifying and Accelerating Lustre Deployments
Sun Lustre Storage System Simplifying and Accelerating Lustre Deployments Torben Kling-Petersen, PhD Presenter s Name Principle Field Title andengineer Division HPC &Cloud LoB SunComputing Microsystems
More informationGetting Started with Pentaho and Cloudera QuickStart VM
Getting Started with Pentaho and Cloudera QuickStart VM This page intentionally left blank. Contents Overview... 1 Before You Begin... 1 Prerequisites... 1 Use Case: Development Sandbox for Pentaho and
More informationNetwork Request Scheduler Scale Testing Results. Nikitas Angelinas
Network Request Scheduler Scale Testing Results Nikitas Angelinas nikitas_angelinas@xyratex.com Agenda NRS background Aim of test runs Tools used Test results Future tasks 2 NRS motivation Increased read
More informationEffective Use of CSAIL Storage
Effective Use of CSAIL Storage How to get the most out of your computing infrastructure Garrett Wollman, Jonathan Proulx, and Jay Sekora The Infrastructure Group Introduction Outline of this talk 1. Introductions
More informationHosted Microsoft Exchange Server 2003 Deployment Utilizing Network Appliance Storage Solutions
Hosted Microsoft Exchange Server 23 Deployment Utilizing Network Appliance Storage Solutions Large-Scale, 68,-Mailbox Exchange Server Proof of Concept Lee Dorrier, Network Appliance, Inc. Eric Johnson,
More informationIBM FlashSystem. IBM FLiP Tool Wie viel schneller kann Ihr IBM i Power Server mit IBM FlashSystem 900 / V9000 Storage sein?
FlashSystem Family 2015 IBM FlashSystem IBM FLiP Tool Wie viel schneller kann Ihr IBM i Power Server mit IBM FlashSystem 900 / V9000 Storage sein? PiRT - Power i Round Table 17 Sep. 2015 Daniel Gysin IBM
More informationThe Software Defined Online Storage System at the GridKa WLCG Tier-1 Center
The Software Defined Online Storage System at the GridKa WLCG Tier-1 Center CHEP 2018, Sofia Jan Erik Sundermann, Jolanta Bubeliene, Ludmilla Obholz, Andreas Petzold STEINBUCH CENTRE FOR COMPUTING (SCC)
More informationCluster Setup and Distributed File System
Cluster Setup and Distributed File System R&D Storage for the R&D Storage Group People Involved Gaetano Capasso - INFN-Naples Domenico Del Prete INFN-Naples Diacono Domenico INFN-Bari Donvito Giacinto
More informationBT Cloud Compute. Adding a Volume to an existing VM running Linux. The power to build your own cloud solutions to serve your specific business needs
Adding a Volume to an existing VM running Linux BT Cloud Compute The power to build your own cloud solutions to serve your specific business needs Issue 3 Introduction This guide has been designed to walk
More informationSTORAGE SYSTEMS. Operating Systems 2015 Spring by Euiseong Seo
STORAGE SYSTEMS Operating Systems 2015 Spring by Euiseong Seo Today s Topics HDDs (Hard Disk Drives) Disk scheduling policies Linux I/O schedulers Secondary Storage Anything that is outside of primary
More informationOptimizing Local File Accesses for FUSE-Based Distributed Storage
Optimizing Local File Accesses for FUSE-Based Distributed Storage Shun Ishiguro 1, Jun Murakami 1, Yoshihiro Oyama 1,3, Osamu Tatebe 2,3 1. The University of Electro-Communications, Japan 2. University
More informationIsilon Performance. Name
1 Isilon Performance Name 2 Agenda Architecture Overview Next Generation Hardware Performance Caching Performance Streaming Reads Performance Tuning OneFS Architecture Overview Copyright 2014 EMC Corporation.
More informationIBM InfoSphere Streams v4.0 Performance Best Practices
Henry May IBM InfoSphere Streams v4.0 Performance Best Practices Abstract Streams v4.0 introduces powerful high availability features. Leveraging these requires careful consideration of performance related
More informationMass-Storage Structure
Operating Systems (Fall/Winter 2018) Mass-Storage Structure Yajin Zhou (http://yajin.org) Zhejiang University Acknowledgement: some pages are based on the slides from Zhi Wang(fsu). Review On-disk structure
More informationUsing file systems at HC3
Using file systems at HC3 Roland Laifer STEINBUCH CENTRE FOR COMPUTING - SCC KIT University of the State of Baden-Württemberg and National Laboratory of the Helmholtz Association www.kit.edu Basic Lustre
More informationHPC Architectures. Types of resource currently in use
HPC Architectures Types of resource currently in use Reusing this material This work is licensed under a Creative Commons Attribution- NonCommercial-ShareAlike 4.0 International License. http://creativecommons.org/licenses/by-nc-sa/4.0/deed.en_us
More informationA Container On a Virtual Machine On an HPC? Presentation to HPC Advisory Council. Perth, July 31-Aug 01, 2017
A Container On a Virtual Machine On an HPC? Presentation to HPC Advisory Council Perth, July 31-Aug 01, 2017 http://levlafayette.com Necessary and Sufficient Definitions High Performance Computing: High
More informationSFS: Random Write Considered Harmful in Solid State Drives
SFS: Random Write Considered Harmful in Solid State Drives Changwoo Min 1, 2, Kangnyeon Kim 1, Hyunjin Cho 2, Sang-Won Lee 1, Young Ik Eom 1 1 Sungkyunkwan University, Korea 2 Samsung Electronics, Korea
More informationResilient and Fast Persistent Container Storage Leveraging Linux s Storage Functionalities Philipp Reisner, CEO LINBIT
Resilient and Fast Persistent Container Storage Leveraging Linux s Storage Functionalities Philipp Reisner, CEO LINBIT LINBIT - the company behind it COMPANY OVERVIEW TECHNOLOGY OVERVIEW Developer of DRBD
More informationHigh Performance Solid State Storage Under Linux
High Performance Solid State Storage Under Linux Eric Seppanen, Matthew T. O Keefe, David J. Lilja Electrical and Computer Engineering University of Minnesota April 20, 2010 Motivation SSDs breaking through
More informationCSCS HPC storage. Hussein N. Harake
CSCS HPC storage Hussein N. Harake Points to Cover - XE6 External Storage (DDN SFA10K, SRP, QDR) - PCI-E SSD Technology - RamSan 620 Technology XE6 External Storage - Installed Q4 2010 - In Production
More informationThe Why and How of HPC-Cloud Hybrids with OpenStack
The Why and How of HPC-Cloud Hybrids with OpenStack OpenStack Australia Day Melbourne June, 2017 Lev Lafayette, HPC Support and Training Officer, University of Melbourne lev.lafayette@unimelb.edu.au 1.0
More informationThe Btrfs Filesystem. Chris Mason
The Btrfs Filesystem Chris Mason The Btrfs Filesystem Jointly developed by a number of companies Oracle, Redhat, Fujitsu, Intel, SUSE, many others All data and metadata is written via copy-on-write CRCs
More information