Integration Path for Intel Omni-Path Fabric attached Intel Enterprise Edition for Lustre (IEEL) LNET

Size: px
Start display at page:

Download "Integration Path for Intel Omni-Path Fabric attached Intel Enterprise Edition for Lustre (IEEL) LNET"

Transcription

1 Integration Path for Intel Omni-Path Fabric attached Intel Enterprise Edition for Lustre (IEEL) LNET

2 Table of Contents Introduction 3 Architecture for LNET 4 Integration 5 Proof of Concept routing for multiple fabrics 5 Ko2iblnd settings 6 Client Mounting 6 Production LNET routing for CSD3 7 Performance Tuning 8 Performance Tests and Results 8 Summary 12 Glossary 13 Appendix A 14 LNET Networks and corresponding network types 14 LNET Routing Configuration: 14 Lustre Server Configuration: 14 Lustre Client Configuration ( EDR ): 14 Lustre Client Configuration ( OPA ): 14

3 Introduction As High Performance Computing centres grow, data centre infrastructure becomes more complex as new and older services are integrated. This can increase the number of server, storage and network technologies that are connected together, making it critical for the successful operation of services that they work seamlessly. A centre s growth can also be impacted in the extension of its service portfolio, which puts pressure on the provision of flexible and scalable platforms, especially storage. Storage requirements increase, often doubling the existing storage with each new service adopted by HPC centres. Across each service provided it is often desirable to have a common storage infrastructure that can be accessed from each of the services provided to users. Allowing users to effectively migrate data across different systems can be challenging, and creates a risk of duplicating data and wasting storage space, together with placing undue stress on network resources. In the case of the University of Cambridge Research Computing Service (RCS), a new set of supercomputing resources has recently been procured and installed for the growing needs of science, both at the University as well as nationally within the UK. The Cambridge Service for Data Driven Discovery (CSD3) provides three new supercomputing resources, along with the existing Intel and dedicated GPU supercomputers. The RCS has made use of Lustre parallel file systems for most of its main resources, and they have been the backbone for providing high performance scalable storage across all research computing platforms. Lustre filesystems support high performance networks such as Ethernet, Infiniband and Intel Omni-Path Fabric. The older HPC service has five Lustre storage filesystems providing access through 5PB of storage and CSD3 introduces an additional five Lustre filesystems, to provide the service with another 5PB of storage space. The new storage platform has been designed and deployed with the intention of allowing both old and new systems to mount these new filesystems, allowing users to migrate and consume data as they switch between CSD3 and the existing resources. In order to take advantage of platform-specific features at the time of acquisition the CSD3 GPU system (Wilkes-2 in Figure 1 below) uses Mellanox EDR Infiniband and the new Intel Xeon Phi and Intel Xeon Gold 6142 CPU resources use the Intel Omni-Path Fabric. The goal of building a common Lustre storage system that can be accessed over the HPC fabrics of different generations and technology can be achieved through the use of LNET routing. LNET routing allows the RCS to expand beyond the confines of the existing FDR InfiniBand fabric, facilitating the translation between fabrics. Services on Intel Omni-Path Fabric, EDR/FDR InfiniBand and Ethernet can now consume existing and new Lustre storage. For example, a user on CSD3 can now write files to Lustre and launch a visualisation instance in the RCS OpenStack cloud, seamlessly accessing Lustre storage concurrently without the user being aware of the underlying infrastructure and placement. LNET routing can not only be used to join dispersed supercomputing resources; LNET routers can be used in the same way as conventional Ethernet routing by setting up multiple routers, thus Lustre traffic can traverse multiple hops of a complicated networking infrastructure, allowing for fine grain routing as scientific computing progresses beyond Petascale systems.

4 Architecture for LNET Skylake & Intel Xeon Phi x200 Wilkes-2 EDR Fabric Darwin Intel Omni-Path EDR Fabric Research Data Store CSD3 Storage LNET Router General Lustre1-5 Ethernet OpenStack Mix of Interconnects Wilkes Hosted Clusters Figure 1 High Level Diagram of the University of Cambridge Research Computing Services estate integrating LNET routers CSD3 incorporates two distinct processor technologies, Intel Xeon Gold 6142, internally referred to as Skylake, and Intel Xeon Phi X200 for Intel architecture together with NVIDIA P100 GPGPU, known as Wilkes-2, underpinned by multiple Lustre filesystems attached to Intel Omni-Path. While the Intel systems use Intel Omni-Path directly themselves, Wilkes-2 uses EDR InfiniBand as its fabric and this presents. Figure 1 shows a high-level view of the current RCS estate. LNET routers shown in the centre of the diagram provide a translation layer between different types of network fabrics, allowing for Lustre access across all systems within the RCS where convergence on one type of interconnect is not possible. Figure 2 shows the LNET routers that connect storage and servers that use Intel Omni-Path Fabric to provide Wilkes-2 with access to common storage. The LNET router does not mount the Lustre file system but merely forwards traffic from one network to the other. In this example two LNET routers load balance traffic and act as an active-active failover pair for Lustre traffic. Additional nodes can be added throughout the network topology to balance network routes across systems. Details on load balancing can be found in the Intel user guide for LNET [1]. Current production services, such as the Darwin CPU cluster and the existing Wilkes GPU connect to the LNET routers over FDR. This flexibility now allows for users to migrate their data over a high speed interconnect as they transition to the new service.

5 Figure 2 LNET Detail showing a pair of routers between Peta-4 and Wilkes-2 Integration Before progressing with the deployment of a production LNET service, an initial experimental routing set-up was completed. This concept demonstrator was then used to aid the construction of the production LNET routing within CSD3. Proof of Concept routing for multiple fabrics When integrating LNET, it is best to map out the LNET for each of the fabric or TCP networks that will connect to the router. Each fabric must have its own o2ib network in order to distinguish between each of the fabrics. Table 1 shows an example from the concept demonstrator system: Fabric Type LNET Network tag Router IP IP Subnet Intel Omni-Path Fabric o2ib /24 InfiniBand o2ib /16 Ethernet tcp /24 Table 1 Example LNET layout All Lustre clients must have defined a list of LNET tags and the address of the router on the respective fabric. A compute node on o2ib0 would have the following router definition within its /etc/modprob.d/lustre.conf options lnet networks= o2ib0(ib0) routes= tcp @o2ib0 \ live_router_check_interval=60 dead_router_check_interval=60 router_ping_timeout=60 Figure 3 Example compute node router configuration Lustre Servers MDS/MGS and OSS nodes should define similar configurations in reverse. These must know about all available LNET fabrics that would wish to mount Lustre. The test system used a set of Lustre storage servers built within an OpenStack Tenant to allow for quick development. Again the server /etc/modprob.d/lustre.conf is shown. options lnet networks= tcp0(em2) routes= o2ib @tcp0, \ o2ib @tcp0 live_router_check_interval=60 dead_router_check_interval=60 router_ping_timeout=60 Figure 4 Example Lustre server route configuration

6 Each node can define multiple routes using a bracketed list of IP addresses within the module configuration. routes= o2ib @tcp0; o2ib [251,252,253,254]2tcp0 Figure 5 LNET route expressing multiple routers This shows the LNET server/client that in order to access the fabric o2ib1 the address to 254@tcp0 can be used. Further settings tell the nodes how to treat the router in the event that Lustre RPCs cannot be successfully routed. When implementing LNET routing it is important to think in the context of Lustre traffic as opposed to a standard ICMP package. While the network port might be up in the traditional sense, if a lctl ping command fails, or if there is no endpoint, each LNET router will mark the route as down. The status of an available path for routing can be viewed using lctl route_list as shown below. net o2ib0 hops gw @tcp up pri 0 net o2ib0 hops gw @tcp up pri 0 [rot@lnet-mds-0 ~]# Figure 6 Screenshot of lctl route_list showing the status of available routing paths for LNET traffic. Router nodes will receive the following configuration to set the node s LNET to route traffic between fabrics. options lnet networks= o2ib0(ib0),o2ib1(ib1),tcp0(em2) forwarding=enabled Figure 7 LNET router node configuration The routing options shown after each network is defined are presented as sensible defaults. This ensures that should a router go down, a client and server can mitigate against the issue while the system administrator remediates the situation. Ko2iblnd settings The Ko2iblnd module should have the same settings for all participating LNET nodes. Due to compatibility issues between mlx and Intel Omni-Path Fabric drivers, users may need to increase the value of the map on demand option to 256, depending on the version of Lustre used. From Lustre 2.10 this can be varied with dynamic LNET configuration. options ko2iblnd-opa peer_credits=128 peer_credits_hiw=64 credits 1024 concurrent_sends=256 ntx=2048 map_on_demand=256 fmr_pool_size=2048 fmr_flush_trigger=512 fmr_cache=1 Figure 8 ko2iblnd settings for LNET estates that mix Intel Omni-Path Fabric and MLX IB Client Mounting Clients on the respective fabrics will mount the address of the MGS, Lustre Management Server, as normal. Any errors can be found in the kernel log messages. Figure 9 CSD3 Storage networking for LNET

7 Figure 9 shows the topology of the LNET storage network as it is currently deployed for the CSD3 early access release. On the left-hand side, the systems are connected to an Intel Omni-Path network. On the right side of the LNET routers, the GPU cluster, Wilkes-2 and the existing HPC infrastructure are connected. At present only two of the routers contain all three types of InfiniBand card, so that the existing FDR network can reach the new CSD3 lustre storage. The current implementation for CSD3 deploys Lustre 2.7 as provided by Intel Enterprise Lustre 3.1 with backports from 2.10 applied. For future deployments it is recommended that users deploy Lustre 2.10, as this will contain the most current patches and features required for future deployments. Migrating from an existing FDR infrastructure to Intel Omni-Path Fabric and EDR requires some careful configuration. During the testing of different configurations, converging the FDR and EDR into a single fabric resulted in a loss of both fabrics. The exact cause of the loss of network when converging is not currently understood, and therefore is not advised. This requires, therefore, that two servers are able to use both Intel Omni-Path Fabric and EDR-FDR mlx4 and mlx5 drivers concurrently. Each LNET router uses the latest Red Hat Linux release 7.3 at time of writing and installs the Lustre client kernel upgrades and the Intel Omni-Path Fabric installation package. It is not advised to run the Mellanox OFED package when combining cards - each package requires libraries that can be replaced by either package, preventing optimal operation. The router must therefore only make use of the Intel Omni-Path Fabric-provided software and the RDMA packages supplied by a supported Linux distribution. The standard ibtools package can be installed to ensure the correct operation of the EDR-FDR cards. EDR and FDR endpoints on compute infrastructure use the latest packages provided by Mellanox. By applying a patch from Lustre 2.10 (don t page-align a remote address with FastReg) when applied to IEEL 3.1 (Lustre 2.7) the routers make use of new lnetctl dynamic network configuration. Support for this is improved in Lustre 2.10 and makes integrating mlx5-edr cards with Intel Omni-Path smoother. This new method of configuring LNET replaces the k2oibconf and lnet.conf files found in the modprobe.d directory. For CSD3 this is only used on the LNET routers - lustre client and servers continue to make use of the existing modprobe.d configuration files. An example of the full dynamic LNET configuration for an LNET Router using all three fabric types is provided in Appendix A. Two of the eight servers require the configuration of all three cards (in the Cambridge configuration). Due to the length of these files, the important configurations that are tuneable are shown below.

8 Performance Tuning Performance Tuning can be achieved firstly through the use of the dynamic LNET configuration found in Lustre 2.10 and above. Second is the introduction of accelerated RDMA within Intel Omni-Path Fabric, which should be enabled to give the best performance from Intel Omni-Path network links. Further Intel Omni-Path performance enhancements, such as adaptive routing, may be considered but were not tested during the writing of this paper. In this deployment the existing ko2iblnd settings are used to provide defaults to LNET and the dynamic LNET configuration yaml file is loaded by a systemd unit file that then loads as per interface parameters. Tests and Results The tests results, for IOR, are shown for two API configurations, POSIX and MPI-IO. A baseline reference test was performed on the Intel Xeon Phi X200 system to provide a result for applications running on the same fabric, and then on Wilkes-2 to compare any reduction in performance introduced by the LNET router. Each test comprises node counts from 1,6,12,24,48,72, with two sets of these tests for one MPI rank per node and then twelve MPI ranks one per x86 core of the Wilkes-2 node. Tests were performed for a shared file striped across all OSTs and File Writer Per Process to an unstriped directory. These tests were performed using the IOR benchmark programme, with each node writing 64GB or 5GB in the case of 12 MPI ranks. 64GB is a close synthetic to how much data all four GPUs could store. Thus, should the user wish to write the contents of all the GPUs, this test makes for a reasonable approximation. The results shown in Figures 10 to 13 are for each of the APIs over the same number of nodes for the Intel Xeon Phi X200 and GPU systems. Dashed lines show the read performance, and solid lines indicate write. Both APIs show a similar performance for the same number of MPI ranks for similar core counts. For higher numbered core counts and nodes, MPI-IO performs better than the standard POSIX API. For both APIs, the GPU and Intel Xeon Phi X200 systems each achieved an approximate read performance of 12GiB/s, and a write performance of 8GiB/s. As each OSS provides 3GiB/s and a single lustre filesystem contains 24 OSTs, the maximum bandwidth for each Lustre filesystem is limited. A second test was performed from the GPU system to see if the LNET routers limited the underlying SAS performance of the Lustre storage nodes. IOR is capable of performing a test over multiple Lustre file systems. Two of the Lustre systems attached to CSD3 were used; directories were set to unstriped and initially the same tests as before were run using FilePerProcess. An initial test showed improved performance over one filesystem. However, on larger node counts the results were unrealistic, as the division of data between each MPI process was too small. Doubling the size of data to be written, the test was re-done, and the results shown in Figure 14 and 15 present an improved performance over one Lustre filesystem. The Intel Xeon Phi X200 system may appear slower, which, however, is due to a slower clock speed than that found within the processors of Wilkes-2. Figure 10 Intel Xeon Phi X200 Shared File Performance in MiB/s

9 Figure 11 Intel Xeon Phi X200 File Per Process Performance in MiB/s Figure 12 GPU Shared File Performance in MiB/s

10 Figure 13 GPU File Per Process Performance in MiB/s Figure 14 GPU Multiple Lustre Performance in MiB/s

11 Figure 15 KNL Multiple Lustre Performance in MiB/s

12 Summary The introduction of LNET routing to Research Computing Services has performed within the expectations of what can be delivered by the current design. LNET routing has not been shown to impact I/O performance to such a degree that it would impact on programmes run on the GPU services, viz routed via InfiniBand to Intel Omni- Path networked Lustre servers, compared with the Intel Xeon Phi X200 I/O performance. Research Computing Services can now offer greater flexibility when more services consume Lustre irrespective of fabric network type. Multiple filesystem performance tests show there was no degradation in application I/O from the LNET routers, and Lustre storage administrators are limited only by the chosen disk technologies, disk host adapters (e.g. SAS) and the size of each Lustre filesystem. To extract more performance from Lustre storage, SSD disks are being looked at, to improve capacity Lustre performance through converged or tiered solutions. Glossary Term EDR FDR KNL lnetctl lctl MDS MGS o2ib Description Enhanced Data Rate InfiniBand Fourteen Data Rate InfiniBand Intel Xeon Phi x200 processor LNET command and configuration program Lustre control utility Metadata Server for Lustre Management Server for Lustre. Usually run with a Metadata server Named identifier of the LNET subnet. Further development is planned to test the overall performance of older InfiniBand and Ethernet fabrics for application I/O, which should help to provide training materials for maximising the best use of Lustre when considering improvements to application I/O. References [1] Intel Omni-Path IP and Storage Router. Intel, 2017.

13 Appendix A LNET Networks and corresponding network types: o2ib0 FDR network for Existing HPC services o2ib1 OPA network for Peta4 and new Lustre storage o2ib2 EDR network for Wilkes-2 LNET Routing Configuration: o2ib0 peer_credits=128 peer_credits_hiw=127 credits=1024 concurrent_sends=64 map_on_demand=0 fmr_pool_size=512 fmr_flush_ trigger=384 fmr_cache=1 o2ib1 peer_credits=128 peer_credits_hiw=127 credits=1024 concurrent_sends=256 map_on_demand=32 fmr_pool_size=2048 fmr_ flush_trigger=512 fmr_cache=1 o2ib2 peer_credits=128 peer_credits_hiw=127 credits=1024 concurrent_sends=256 map_on_demand=32 fmr_pool_size=2048 fmr_ flush_trigger=512 fmr_cache=1 Lustre Server Configuration: options ko2iblnd-opa peer_credits=128 peer_credits_hiw=64 credits=1024 concurrent_sends=256 ntx=2048 map_on_ demand=32 fmr_pool_size=2048 fmr_flush_trigger=512 fmr_cache=1 options lnet networks= o2ib1(ib0), tcp2(em1.43) routes= o2ib [ ]@o2ib1; o2ib [ ]@ o2ib1 auto_down=1 avoid_asym_router_failure=1 check_routers_before_use=1 dead_router_check_interval=60 live_router_ check_interval=60 router_ping_timeout=60 Lustre Client Configuration (EDR): options ko2iblnd-mlx5 peer_credits=128 peer_credits_hiw=127 credits=1024 concurrent_sends=256 ntx=2048 map_on_ demand=32 fmr_pool_size=2048 fmr_flush_trigger=512 fmr_cache=1 options lnet networks=o2ib2(ib0) routes= o2ib [ ]@o2ib2; o2ib [ ]@o2ib2 auto_down=1 avoid_asym_router_failure=1 check_routers_before_use=1 dead_router_check_interval=60 live_router_check_interval=60 router_ping_timeout=60 Lustre Client Configuration (OPA): options ko2iblnd-opa peer_credits=128 peer_credits_hiw=64 credits=1024 concurrent_sends=256 ntx=2048 map_on_ demand=32 fmr_pool_size=2048 fmr_flush_trigger=512 fmr_cache=1 options lnet networks=o2ib1(ib0) routes= o2ib [ ]@o2ib1; o2ib [ ]@o2ib1 auto_down=1 avoid_asym_router_failure=1 check_routers_before_use=1 dead_router_check_interval=60 live_router_check_interval=60 router_ping_timeout=60

14 Intel technologies features and benefits depend on system configuration and may require enabled hardware, software or service activation. Performance varies depending on system configuration. No computer system can be absolutely secure. Check with your system manufacturer or retailer or learn more at intel.com. Tests document performance of components on a particular test, in specific systems. Differences in hardware, software, or configuration will affect actual performance. Consult other sources of information to evaluate performance as you consider your purchase. For more complete information about performance and benchmark results, visit com/performance. Intel does not control or audit third-party benchmark data or the web sites referenced in this document. You should visit the referenced web site and confirm whether referenced data are accurate. Intel, the Intel logo, Intel Xeon, Intel SSD DC S3610, Intel SSD DC S3710, and Intel SSD DC P3600 are trademarks of Intel Corporation in the U.S. and/or other countries.

15 2017 Dell EMC, All rights reserved. Dell EMC, the Dell EMC logo and products as identified in this document are registered trademarks of Dell, Inc. in the U.S.A. and/or other countries. No part of this document may be reproduced or transmitted in any form or by any means, electronic or mechanical, including photocopying and recording for any purpose without the written permission of Dell, Inc. ( Dell ). Dell EMC disclaims proprietary interest in the marks and names of others. Dell EMC service offerings do not affect customer s statutory rights. Availability and terms of Dell EMC Services vary by region. Terms and Conditions of Sales, Service and Finance apply and are available on request or at Dell.co.uk/terms. THIS WHITE PAPER IS FOR INFORMATIONAL PURPOSES ONLY, AND MAY CONTAIN TYPOGRAPHICAL ERRORS AND TECHNICAL INACCURACIES. THE CONTENT IS PROVIDED AS IS, WITHOUT EXPRESS OR IMPLIED WARRANTIES OF ANY KIND. Dell Corporation Limited. Registered in England. Reg. No Dell House, The Boulevard, Cain Road, Bracknell, Berkshire, RG12 1LF, UK.

Optimization of Lustre* performance using a mix of fabric cards

Optimization of Lustre* performance using a mix of fabric cards * Some names and brands may be claimed as the property of others. Optimization of Lustre* performance using a mix of fabric cards Dmitry Eremin Agenda High variety of RDMA solutions Network optimization

More information

Olaf Weber Senior Software Engineer SGI Storage Software. Amir Shehata Lustre Network Engineer Intel High Performance Data Division

Olaf Weber Senior Software Engineer SGI Storage Software. Amir Shehata Lustre Network Engineer Intel High Performance Data Division Olaf Weber Senior Software Engineer SGI Storage Software Amir Shehata Lustre Network Engineer Intel High Performance Data Division Intel and the Intel logo are trademarks or registered trademarks of Intel

More information

Dell EMC Ready Bundle for HPC Digital Manufacturing Dassault Systѐmes Simulia Abaqus Performance

Dell EMC Ready Bundle for HPC Digital Manufacturing Dassault Systѐmes Simulia Abaqus Performance Dell EMC Ready Bundle for HPC Digital Manufacturing Dassault Systѐmes Simulia Abaqus Performance This Dell EMC technical white paper discusses performance benchmarking results and analysis for Simulia

More information

LNet Roadmap & Development. Amir Shehata Lustre * Network Engineer Intel High Performance Data Division

LNet Roadmap & Development. Amir Shehata Lustre * Network Engineer Intel High Performance Data Division LNet Roadmap & Development Amir Shehata Lustre * Network Engineer Intel High Performance Data Division Outline LNet Roadmap Non-contiguous buffer support Map-on-Demand re-work 2 LNet Roadmap (2.12) LNet

More information

Andreas Dilger. Principal Lustre Engineer. High Performance Data Division

Andreas Dilger. Principal Lustre Engineer. High Performance Data Division Andreas Dilger Principal Lustre Engineer High Performance Data Division Focus on Performance and Ease of Use Beyond just looking at individual features... Incremental but continuous improvements Performance

More information

CSD3 The Cambridge Service for Data Driven Discovery. A New National HPC Service for Data Intensive science

CSD3 The Cambridge Service for Data Driven Discovery. A New National HPC Service for Data Intensive science CSD3 The Cambridge Service for Data Driven Discovery A New National HPC Service for Data Intensive science Dr Paul Calleja Director of Research Computing University of Cambridge Problem statement Today

More information

Intel Enterprise Edition Lustre (IEEL-2.3) [DNE-1 enabled] on Dell MD Storage

Intel Enterprise Edition Lustre (IEEL-2.3) [DNE-1 enabled] on Dell MD Storage Intel Enterprise Edition Lustre (IEEL-2.3) [DNE-1 enabled] on Dell MD Storage Evaluation of Lustre File System software enhancements for improved Metadata performance Wojciech Turek, Paul Calleja,John

More information

Mission-Critical Lustre at Santos. Adam Fox, Lustre User Group 2016

Mission-Critical Lustre at Santos. Adam Fox, Lustre User Group 2016 Mission-Critical Lustre at Santos Adam Fox, Lustre User Group 2016 About Santos One of the leading oil and gas producers in APAC Founded in 1954 South Australia Northern Territory Oil Search Cooper Basin

More information

Multi-Rail LNet for Lustre

Multi-Rail LNet for Lustre Multi-Rail LNet for Lustre Rob Mollard September 2016 The SGI logos and SGI product names used or referenced herein are either registered trademarks or trademarks of Silicon Graphics International Corp.

More information

An Introduction to GPFS

An Introduction to GPFS IBM High Performance Computing July 2006 An Introduction to GPFS gpfsintro072506.doc Page 2 Contents Overview 2 What is GPFS? 3 The file system 3 Application interfaces 4 Performance and scalability 4

More information

Lustre HSM at Cambridge. Early user experience using Intel Lemur HSM agent

Lustre HSM at Cambridge. Early user experience using Intel Lemur HSM agent Lustre HSM at Cambridge Early user experience using Intel Lemur HSM agent Matt Rásó-Barnett Wojciech Turek Research Computing Services @ Cambridge University-wide service with broad remit to provide research

More information

DELL EMC ISILON F800 AND H600 I/O PERFORMANCE

DELL EMC ISILON F800 AND H600 I/O PERFORMANCE DELL EMC ISILON F800 AND H600 I/O PERFORMANCE ABSTRACT This white paper provides F800 and H600 performance data. It is intended for performance-minded administrators of large compute clusters that access

More information

Small File I/O Performance in Lustre. Mikhail Pershin, Joe Gmitter Intel HPDD April 2018

Small File I/O Performance in Lustre. Mikhail Pershin, Joe Gmitter Intel HPDD April 2018 Small File I/O Performance in Lustre Mikhail Pershin, Joe Gmitter Intel HPDD April 2018 Overview Small File I/O Concerns Data on MDT (DoM) Feature Overview DoM Use Cases DoM Performance Results Small File

More information

Olaf Weber Senior Software Engineer SGI Storage Software

Olaf Weber Senior Software Engineer SGI Storage Software Olaf Weber Senior Software Engineer SGI Storage Software Intel and the Intel logo are trademarks or registered trademarks of Intel Corporation or its subsidiaries in the United States and other countries.

More information

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance 11 th International LS-DYNA Users Conference Computing Technology LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance Gilad Shainer 1, Tong Liu 2, Jeff Layton

More information

Implementing Storage in Intel Omni-Path Architecture Fabrics

Implementing Storage in Intel Omni-Path Architecture Fabrics white paper Implementing in Intel Omni-Path Architecture Fabrics Rev 2 A rich ecosystem of storage solutions supports Intel Omni- Path Executive Overview The Intel Omni-Path Architecture (Intel OPA) is

More information

Chelsio Communications. Meeting Today s Datacenter Challenges. Produced by Tabor Custom Publishing in conjunction with: CUSTOM PUBLISHING

Chelsio Communications. Meeting Today s Datacenter Challenges. Produced by Tabor Custom Publishing in conjunction with: CUSTOM PUBLISHING Meeting Today s Datacenter Challenges Produced by Tabor Custom Publishing in conjunction with: 1 Introduction In this era of Big Data, today s HPC systems are faced with unprecedented growth in the complexity

More information

High-Performance Lustre with Maximum Data Assurance

High-Performance Lustre with Maximum Data Assurance High-Performance Lustre with Maximum Data Assurance Silicon Graphics International Corp. 900 North McCarthy Blvd. Milpitas, CA 95035 Disclaimer and Copyright Notice The information presented here is meant

More information

unleashed the future Intel Xeon Scalable Processors for High Performance Computing Alexey Belogortsev Field Application Engineer

unleashed the future Intel Xeon Scalable Processors for High Performance Computing Alexey Belogortsev Field Application Engineer the future unleashed Alexey Belogortsev Field Application Engineer Intel Xeon Scalable Processors for High Performance Computing Growing Challenges in System Architecture The Walls System Bottlenecks Divergent

More information

Dell EMC Ready Bundle for HPC Digital Manufacturing ANSYS Performance

Dell EMC Ready Bundle for HPC Digital Manufacturing ANSYS Performance Dell EMC Ready Bundle for HPC Digital Manufacturing ANSYS Performance This Dell EMC technical white paper discusses performance benchmarking results and analysis for ANSYS Mechanical, ANSYS Fluent, and

More information

HPC Innovation Lab Update. Dell EMC HPC Community Meeting 3/28/2017

HPC Innovation Lab Update. Dell EMC HPC Community Meeting 3/28/2017 HPC Innovation Lab Update Dell EMC HPC Community Meeting 3/28/2017 Dell EMC HPC Innovation Lab charter Design, develop and integrate Heading HPC systems Lorem ipsum Flexible reference dolor sit amet, architectures

More information

Data life cycle monitoring using RoBinHood at scale. Gabriele Paciucci Solution Architect Bruno Faccini Senior Support Engineer September LAD

Data life cycle monitoring using RoBinHood at scale. Gabriele Paciucci Solution Architect Bruno Faccini Senior Support Engineer September LAD Data life cycle monitoring using RoBinHood at scale Gabriele Paciucci Solution Architect Bruno Faccini Senior Support Engineer September 2015 - LAD Agenda Motivations Hardware and software setup The first

More information

Sun Lustre Storage System Simplifying and Accelerating Lustre Deployments

Sun Lustre Storage System Simplifying and Accelerating Lustre Deployments Sun Lustre Storage System Simplifying and Accelerating Lustre Deployments Torben Kling-Petersen, PhD Presenter s Name Principle Field Title andengineer Division HPC &Cloud LoB SunComputing Microsystems

More information

Performance Optimizations via Connect-IB and Dynamically Connected Transport Service for Maximum Performance on LS-DYNA

Performance Optimizations via Connect-IB and Dynamically Connected Transport Service for Maximum Performance on LS-DYNA Performance Optimizations via Connect-IB and Dynamically Connected Transport Service for Maximum Performance on LS-DYNA Pak Lui, Gilad Shainer, Brian Klaff Mellanox Technologies Abstract From concept to

More information

Welcome! Virtual tutorial starts at 15:00 BST

Welcome! Virtual tutorial starts at 15:00 BST Welcome! Virtual tutorial starts at 15:00 BST Parallel IO and the ARCHER Filesystem ARCHER Virtual Tutorial, Wed 8 th Oct 2014 David Henty Reusing this material This work is licensed

More information

Dell Reference Configuration for Large Oracle Database Deployments on Dell EqualLogic Storage

Dell Reference Configuration for Large Oracle Database Deployments on Dell EqualLogic Storage Dell Reference Configuration for Large Oracle Database Deployments on Dell EqualLogic Storage Database Solutions Engineering By Raghunatha M, Ravi Ramappa Dell Product Group October 2009 Executive Summary

More information

An ESS implementation in a Tier 1 HPC Centre

An ESS implementation in a Tier 1 HPC Centre An ESS implementation in a Tier 1 HPC Centre Maximising Performance - the NeSI Experience José Higino (NeSI Platforms and NIWA, HPC Systems Engineer) Outline What is NeSI? The National Platforms Framework

More information

LUSTRE NETWORKING High-Performance Features and Flexible Support for a Wide Array of Networks White Paper November Abstract

LUSTRE NETWORKING High-Performance Features and Flexible Support for a Wide Array of Networks White Paper November Abstract LUSTRE NETWORKING High-Performance Features and Flexible Support for a Wide Array of Networks White Paper November 2008 Abstract This paper provides information about Lustre networking that can be used

More information

Application Performance on IME

Application Performance on IME Application Performance on IME Toine Beckers, DDN Marco Grossi, ICHEC Burst Buffer Designs Introduce fast buffer layer Layer between memory and persistent storage Pre-stage application data Buffer writes

More information

HPC Architectures. Types of resource currently in use

HPC Architectures. Types of resource currently in use HPC Architectures Types of resource currently in use Reusing this material This work is licensed under a Creative Commons Attribution- NonCommercial-ShareAlike 4.0 International License. http://creativecommons.org/licenses/by-nc-sa/4.0/deed.en_us

More information

New Storage Architectures

New Storage Architectures New Storage Architectures OpenFabrics Software User Group Workshop Replacing LNET routers with IB routers #OFSUserGroup Lustre Basics Lustre is a clustered file-system for supercomputing Architecture consists

More information

Best Practices for Deploying a Mixed 1Gb/10Gb Ethernet SAN using Dell EqualLogic Storage Arrays

Best Practices for Deploying a Mixed 1Gb/10Gb Ethernet SAN using Dell EqualLogic Storage Arrays Dell EqualLogic Best Practices Series Best Practices for Deploying a Mixed 1Gb/10Gb Ethernet SAN using Dell EqualLogic Storage Arrays A Dell Technical Whitepaper Jerry Daugherty Storage Infrastructure

More information

JÜLICH SUPERCOMPUTING CENTRE Site Introduction Michael Stephan Forschungszentrum Jülich

JÜLICH SUPERCOMPUTING CENTRE Site Introduction Michael Stephan Forschungszentrum Jülich JÜLICH SUPERCOMPUTING CENTRE Site Introduction 09.04.2018 Michael Stephan JSC @ Forschungszentrum Jülich FORSCHUNGSZENTRUM JÜLICH Research Centre Jülich One of the 15 Helmholtz Research Centers in Germany

More information

Virtuozzo Hyperconverged Platform Uses Intel Optane SSDs to Accelerate Performance for Containers and VMs

Virtuozzo Hyperconverged Platform Uses Intel Optane SSDs to Accelerate Performance for Containers and VMs Solution brief Software-Defined Data Center (SDDC) Hyperconverged Platforms Virtuozzo Hyperconverged Platform Uses Intel Optane SSDs to Accelerate Performance for Containers and VMs Virtuozzo benchmark

More information

LustreFS and its ongoing Evolution for High Performance Computing and Data Analysis Solutions

LustreFS and its ongoing Evolution for High Performance Computing and Data Analysis Solutions LustreFS and its ongoing Evolution for High Performance Computing and Data Analysis Solutions Roger Goff Senior Product Manager DataDirect Networks, Inc. What is Lustre? Parallel/shared file system for

More information

InfiniBand Networked Flash Storage

InfiniBand Networked Flash Storage InfiniBand Networked Flash Storage Superior Performance, Efficiency and Scalability Motti Beck Director Enterprise Market Development, Mellanox Technologies Flash Memory Summit 2016 Santa Clara, CA 1 17PB

More information

Xyratex ClusterStor6000 & OneStor

Xyratex ClusterStor6000 & OneStor Xyratex ClusterStor6000 & OneStor Proseminar Ein-/Ausgabe Stand der Wissenschaft von Tim Reimer Structure OneStor OneStorSP OneStorAP ''Green'' Advancements ClusterStor6000 About Scale-Out Storage Architecture

More information

MELLANOX EDR UPDATE & GPUDIRECT MELLANOX SR. SE 정연구

MELLANOX EDR UPDATE & GPUDIRECT MELLANOX SR. SE 정연구 MELLANOX EDR UPDATE & GPUDIRECT MELLANOX SR. SE 정연구 Leading Supplier of End-to-End Interconnect Solutions Analyze Enabling the Use of Data Store ICs Comprehensive End-to-End InfiniBand and Ethernet Portfolio

More information

Becca Paren Cluster Systems Engineer Software and Services Group. May 2017

Becca Paren Cluster Systems Engineer Software and Services Group. May 2017 Becca Paren Cluster Systems Engineer Software and Services Group May 2017 Clusters are complex systems! Challenge is to reduce this complexity barrier for: Cluster architects System administrators Application

More information

Short Talk: System abstractions to facilitate data movement in supercomputers with deep memory and interconnect hierarchy

Short Talk: System abstractions to facilitate data movement in supercomputers with deep memory and interconnect hierarchy Short Talk: System abstractions to facilitate data movement in supercomputers with deep memory and interconnect hierarchy François Tessier, Venkatram Vishwanath Argonne National Laboratory, USA July 19,

More information

NetApp High-Performance Storage Solution for Lustre

NetApp High-Performance Storage Solution for Lustre Technical Report NetApp High-Performance Storage Solution for Lustre Solution Design Narjit Chadha, NetApp October 2014 TR-4345-DESIGN Abstract The NetApp High-Performance Storage Solution (HPSS) for Lustre,

More information

HPC NETWORKING IN THE REAL WORLD

HPC NETWORKING IN THE REAL WORLD 15 th ANNUAL WORKSHOP 2019 HPC NETWORKING IN THE REAL WORLD Jesse Martinez Los Alamos National Laboratory March 19 th, 2019 [ LOGO HERE ] LA-UR-19-22146 ABSTRACT Introduction to LANL High Speed Networking

More information

Active-Active LNET Bonding Using Multiple LNETs and Infiniband partitions

Active-Active LNET Bonding Using Multiple LNETs and Infiniband partitions April 15th - 19th, 2013 LUG13 LUG13 Active-Active LNET Bonding Using Multiple LNETs and Infiniband partitions Shuichi Ihara DataDirect Networks, Japan Today s H/W Trends for Lustre Powerful server platforms

More information

OpenStack and Hadoop. Achieving near bare-metal performance for big data workloads running in a private cloud ABSTRACT

OpenStack and Hadoop. Achieving near bare-metal performance for big data workloads running in a private cloud ABSTRACT OpenStack and Hadoop Achieving near bare-metal performance for big data workloads running in a private cloud ABSTRACT IT organizations are increasingly turning to the open source Apache Hadoop software

More information

SIMPLIFYING HPC SIMPLIFYING HPC FOR ENGINEERING SIMULATION WITH ANSYS

SIMPLIFYING HPC SIMPLIFYING HPC FOR ENGINEERING SIMULATION WITH ANSYS SIMPLIFYING HPC SIMPLIFYING HPC FOR ENGINEERING SIMULATION WITH ANSYS THE DELL WAY We are an acknowledged leader in academic supercomputing including major HPC systems installed at the Cambridge University

More information

Performance and Energy Efficiency of the 14 th Generation Dell PowerEdge Servers

Performance and Energy Efficiency of the 14 th Generation Dell PowerEdge Servers Performance and Energy Efficiency of the 14 th Generation Dell PowerEdge Servers This white paper details the performance improvements of Dell PowerEdge servers with the Intel Xeon Processor Scalable CPU

More information

The Future of High Performance Interconnects

The Future of High Performance Interconnects The Future of High Performance Interconnects Ashrut Ambastha HPC Advisory Council Perth, Australia :: August 2017 When Algorithms Go Rogue 2017 Mellanox Technologies 2 When Algorithms Go Rogue 2017 Mellanox

More information

Fast and Easy Persistent Storage for Docker* Containers with Storidge and Intel

Fast and Easy Persistent Storage for Docker* Containers with Storidge and Intel Solution brief Intel Storage Builders Storidge ContainerIO TM Intel Xeon Processor Scalable Family Intel SSD DC Family for PCIe*/NVMe Fast and Easy Persistent Storage for Docker* Containers with Storidge

More information

NVIDIA COLLECTIVE COMMUNICATION LIBRARY (NCCL)

NVIDIA COLLECTIVE COMMUNICATION LIBRARY (NCCL) NVIDIA COLLECTIVE COMMUNICATION LIBRARY (NCCL) RN-08645-000_v01 September 2018 Release Notes TABLE OF CONTENTS Chapter Chapter Chapter Chapter Chapter Chapter Chapter Chapter Chapter Chapter 1. NCCL Overview...1

More information

Scaling to Petaflop. Ola Torudbakken Distinguished Engineer. Sun Microsystems, Inc

Scaling to Petaflop. Ola Torudbakken Distinguished Engineer. Sun Microsystems, Inc Scaling to Petaflop Ola Torudbakken Distinguished Engineer Sun Microsystems, Inc HPC Market growth is strong CAGR increased from 9.2% (2006) to 15.5% (2007) Market in 2007 doubled from 2003 (Source: IDC

More information

FlashGrid Software Enables Converged and Hyper-Converged Appliances for Oracle* RAC

FlashGrid Software Enables Converged and Hyper-Converged Appliances for Oracle* RAC white paper FlashGrid Software Intel SSD DC P3700/P3600/P3500 Topic: Hyper-converged Database/Storage FlashGrid Software Enables Converged and Hyper-Converged Appliances for Oracle* RAC Abstract FlashGrid

More information

DELL Terascala HPC Storage Solution (DT-HSS2)

DELL Terascala HPC Storage Solution (DT-HSS2) DELL Terascala HPC Storage Solution (DT-HSS2) A Dell Technical White Paper Dell Li Ou, Scott Collier Terascala Rick Friedman Dell HPC Solutions Engineering THIS WHITE PAPER IS FOR INFORMATIONAL PURPOSES

More information

Lustre Networking at Cray. Chris Horn

Lustre Networking at Cray. Chris Horn Lustre Networking at Cray Chris Horn hornc@cray.com Agenda Lustre Networking at Cray LNet Basics Flat vs. Fine-Grained Routing Cost Effectiveness - Bandwidth Matching Connection Reliability Dealing with

More information

Fan Yong; Zhang Jinghai. High Performance Data Division

Fan Yong; Zhang Jinghai. High Performance Data Division Fan Yong; Zhang Jinghai High Performance Data Division How Can Lustre * Snapshots Be Used? Undo/undelete/recover file(s) from the snapshot Removed file by mistake, application failure causes data invalid

More information

Meltdown and Spectre Interconnect Performance Evaluation Jan Mellanox Technologies

Meltdown and Spectre Interconnect Performance Evaluation Jan Mellanox Technologies Meltdown and Spectre Interconnect Evaluation Jan 2018 1 Meltdown and Spectre - Background Most modern processors perform speculative execution This speculation can be measured, disclosing information about

More information

Parallel File Systems for HPC

Parallel File Systems for HPC Introduction to Scuola Internazionale Superiore di Studi Avanzati Trieste November 2008 Advanced School in High Performance and Grid Computing Outline 1 The Need for 2 The File System 3 Cluster & A typical

More information

Enhancing Oracle VM Business Continuity Using Dell Compellent Live Volume

Enhancing Oracle VM Business Continuity Using Dell Compellent Live Volume Enhancing Oracle VM Business Continuity Using Dell Compellent Live Volume Wendy Chen, Roger Lopez, and Josh Raw Dell Product Group February 2013 This document is for informational purposes only and may

More information

Lustre at Scale The LLNL Way

Lustre at Scale The LLNL Way Lustre at Scale The LLNL Way D. Marc Stearman Lustre Administration Lead Livermore uting - LLNL This work performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory

More information

Intel Omni-Path Fabric Manager GUI Software

Intel Omni-Path Fabric Manager GUI Software Intel Omni-Path Fabric Manager GUI Software Release Notes for 10.6 October 2017 Order No.: J82663-1.0 You may not use or facilitate the use of this document in connection with any infringement or other

More information

Lenovo Database Configuration Guide

Lenovo Database Configuration Guide Lenovo Database Configuration Guide for Microsoft SQL Server OLTP on ThinkAgile SXM Reduce time to value with validated hardware configurations up to 2 million TPM in a DS14v2 VM SQL Server in an Infrastructure

More information

HPE Scalable Storage with Intel Enterprise Edition for Lustre*

HPE Scalable Storage with Intel Enterprise Edition for Lustre* HPE Scalable Storage with Intel Enterprise Edition for Lustre* HPE Scalable Storage with Intel Enterprise Edition For Lustre* High Performance Storage Solution Meets Demanding I/O requirements Performance

More information

VM Migration Acceleration over 40GigE Meet SLA & Maximize ROI

VM Migration Acceleration over 40GigE Meet SLA & Maximize ROI VM Migration Acceleration over 40GigE Meet SLA & Maximize ROI Mellanox Technologies Inc. Motti Beck, Director Marketing Motti@mellanox.com Topics Introduction to Mellanox Technologies Inc. Why Cloud SLA

More information

An Overview of Fujitsu s Lustre Based File System

An Overview of Fujitsu s Lustre Based File System An Overview of Fujitsu s Lustre Based File System Shinji Sumimoto Fujitsu Limited Apr.12 2011 For Maximizing CPU Utilization by Minimizing File IO Overhead Outline Target System Overview Goals of Fujitsu

More information

Feedback on BeeGFS. A Parallel File System for High Performance Computing

Feedback on BeeGFS. A Parallel File System for High Performance Computing Feedback on BeeGFS A Parallel File System for High Performance Computing Philippe Dos Santos et Georges Raseev FR 2764 Fédération de Recherche LUmière MATière December 13 2016 LOGO CNRS LOGO IO December

More information

HPC and AI Solution Overview. Garima Kochhar HPC and AI Innovation Lab

HPC and AI Solution Overview. Garima Kochhar HPC and AI Innovation Lab HPC and AI Solution Overview Garima Kochhar HPC and AI Innovation Lab 1 Dell EMC HPC and DL team charter Design, develop and integrate HPC and DL Heading systems Lorem ipsum dolor sit amet, consectetur

More information

Computer Science Section. Computational and Information Systems Laboratory National Center for Atmospheric Research

Computer Science Section. Computational and Information Systems Laboratory National Center for Atmospheric Research Computer Science Section Computational and Information Systems Laboratory National Center for Atmospheric Research My work in the context of TDD/CSS/ReSET Polynya new research computing environment Polynya

More information

An Oracle Technical White Paper October Sizing Guide for Single Click Configurations of Oracle s MySQL on Sun Fire x86 Servers

An Oracle Technical White Paper October Sizing Guide for Single Click Configurations of Oracle s MySQL on Sun Fire x86 Servers An Oracle Technical White Paper October 2011 Sizing Guide for Single Click Configurations of Oracle s MySQL on Sun Fire x86 Servers Introduction... 1 Foundation for an Enterprise Infrastructure... 2 Sun

More information

TALK THUNDER SOFTWARE FOR BARE METAL HIGH-PERFORMANCE SOFTWARE FOR THE MODERN DATA CENTER WITH A10 DATASHEET YOUR CHOICE OF HARDWARE

TALK THUNDER SOFTWARE FOR BARE METAL HIGH-PERFORMANCE SOFTWARE FOR THE MODERN DATA CENTER WITH A10 DATASHEET YOUR CHOICE OF HARDWARE DATASHEET THUNDER SOFTWARE FOR BARE METAL YOUR CHOICE OF HARDWARE A10 Networks application networking and security solutions for bare metal raise the bar on performance with an industryleading software

More information

Interconnect Your Future

Interconnect Your Future Interconnect Your Future Gilad Shainer 2nd Annual MVAPICH User Group (MUG) Meeting, August 2014 Complete High-Performance Scalable Interconnect Infrastructure Comprehensive End-to-End Software Accelerators

More information

The Future of Interconnect Technology

The Future of Interconnect Technology The Future of Interconnect Technology Michael Kagan, CTO HPC Advisory Council Stanford, 2014 Exponential Data Growth Best Interconnect Required 44X 0.8 Zetabyte 2009 35 Zetabyte 2020 2014 Mellanox Technologies

More information

Create a Flexible, Scalable High-Performance Storage Cluster with WekaIO Matrix

Create a Flexible, Scalable High-Performance Storage Cluster with WekaIO Matrix Solution brief Intel Storage Builders WekaIO Matrix Intel eon Processor E5-2600 Product Family Intel Ethernet Converged Network Adapter 520 Intel SSD Data Center Family Data Plane Development Kit Create

More information

Lustre * Features In Development Fan Yong High Performance Data Division, Intel CLUG

Lustre * Features In Development Fan Yong High Performance Data Division, Intel CLUG Lustre * Features In Development Fan Yong High Performance Data Division, Intel CLUG 2017 @Beijing Outline LNet reliability DNE improvements Small file performance File Level Redundancy Miscellaneous improvements

More information

Extremely Fast Distributed Storage for Cloud Service Providers

Extremely Fast Distributed Storage for Cloud Service Providers Solution brief Intel Storage Builders StorPool Storage Intel SSD DC S3510 Series Intel Xeon Processor E3 and E5 Families Intel Ethernet Converged Network Adapter X710 Family Extremely Fast Distributed

More information

Andreas Schneider. Markus Leberecht. Senior Cloud Solution Architect, Intel Deutschland. Distribution Sales Manager, Intel Deutschland

Andreas Schneider. Markus Leberecht. Senior Cloud Solution Architect, Intel Deutschland. Distribution Sales Manager, Intel Deutschland Markus Leberecht Senior Cloud Solution Architect, Intel Deutschland Andreas Schneider Distribution Sales Manager, Intel Deutschland Legal Disclaimers 2016 Intel Corporation. Intel, the Intel logo, Xeon

More information

Four-Socket Server Consolidation Using SQL Server 2008

Four-Socket Server Consolidation Using SQL Server 2008 Four-Socket Server Consolidation Using SQL Server 28 A Dell Technical White Paper Authors Raghunatha M Leena Basanthi K Executive Summary Businesses of all sizes often face challenges with legacy hardware

More information

Open storage architecture for private Oracle database clouds

Open storage architecture for private Oracle database clouds Open storage architecture for private Oracle database clouds White Paper rev. 2016-05-18 2016 FlashGrid Inc. 1 www.flashgrid.io Abstract Enterprise IT is transitioning from proprietary mainframe and UNIX

More information

Future Routing Schemes in Petascale clusters

Future Routing Schemes in Petascale clusters Future Routing Schemes in Petascale clusters Gilad Shainer, Mellanox, USA Ola Torudbakken, Sun Microsystems, Norway Richard Graham, Oak Ridge National Laboratory, USA Birds of a Feather Presentation Abstract

More information

Purchasing Services SVC East Fowler Avenue Tampa, Florida (813)

Purchasing Services SVC East Fowler Avenue Tampa, Florida (813) Purchasing Services SVC 1073 4202 East Fowler Avenue Tampa, Florida 33620 (813) 974-2481 Web Address: http://www.usf.edu/business-finance/purchasing/staff-procedures/index.aspx April 25, 2018 Invitation

More information

Birds of a Feather Presentation

Birds of a Feather Presentation Mellanox InfiniBand QDR 4Gb/s The Fabric of Choice for High Performance Computing Gilad Shainer, shainer@mellanox.com June 28 Birds of a Feather Presentation InfiniBand Technology Leadership Industry Standard

More information

NEMO Performance Benchmark and Profiling. May 2011

NEMO Performance Benchmark and Profiling. May 2011 NEMO Performance Benchmark and Profiling May 2011 Note The following research was performed under the HPC Advisory Council HPC works working group activities Participating vendors: HP, Intel, Mellanox

More information

The BioHPC Nucleus Cluster & Future Developments

The BioHPC Nucleus Cluster & Future Developments 1 The BioHPC Nucleus Cluster & Future Developments Overview Today we ll talk about the BioHPC Nucleus HPC cluster with some technical details for those interested! How is it designed? What hardware does

More information

IBM Power Systems HPC Cluster

IBM Power Systems HPC Cluster IBM Power Systems HPC Cluster Highlights Complete and fully Integrated HPC cluster for demanding workloads Modular and Extensible: match components & configurations to meet demands Integrated: racked &

More information

A GPFS Primer October 2005

A GPFS Primer October 2005 A Primer October 2005 Overview This paper describes (General Parallel File System) Version 2, Release 3 for AIX 5L and Linux. It provides an overview of key concepts which should be understood by those

More information

Comet Virtualization Code & Design Sprint

Comet Virtualization Code & Design Sprint Comet Virtualization Code & Design Sprint SDSC September 23-24 Rick Wagner San Diego Supercomputer Center Meeting Goals Build personal connections between the IU and SDSC members of the Comet team working

More information

AMD EPYC Processors Showcase High Performance for Network Function Virtualization (NFV)

AMD EPYC Processors Showcase High Performance for Network Function Virtualization (NFV) White Paper December, 2018 AMD EPYC Processors Showcase High Performance for Network Function Virtualization (NFV) Executive Summary Data centers and cloud service providers are creating a technology shift

More information

ORACLE FABRIC MANAGER

ORACLE FABRIC MANAGER ORACLE FABRIC MANAGER MANAGE CONNECTIVITY IN REAL TIME KEY BENEFITS Control connectivity across servers from a single screen. Instantly replicate connectivity configurations across a group of servers with

More information

Technical Computing Suite supporting the hybrid system

Technical Computing Suite supporting the hybrid system Technical Computing Suite supporting the hybrid system Supercomputer PRIMEHPC FX10 PRIMERGY x86 cluster Hybrid System Configuration Supercomputer PRIMEHPC FX10 PRIMERGY x86 cluster 6D mesh/torus Interconnect

More information

Lustre File System. Proseminar 2013 Ein-/Ausgabe - Stand der Wissenschaft Universität Hamburg. Paul Bienkowski Author. Michael Kuhn Supervisor

Lustre File System. Proseminar 2013 Ein-/Ausgabe - Stand der Wissenschaft Universität Hamburg. Paul Bienkowski Author. Michael Kuhn Supervisor Proseminar 2013 Ein-/Ausgabe - Stand der Wissenschaft Universität Hamburg September 30, 2013 Paul Bienkowski Author 2bienkow@informatik.uni-hamburg.de Michael Kuhn Supervisor michael.kuhn@informatik.uni-hamburg.de

More information

Introduction The Project Lustre Architecture Performance Conclusion References. Lustre. Paul Bienkowski

Introduction The Project Lustre Architecture Performance Conclusion References. Lustre. Paul Bienkowski Lustre Paul Bienkowski 2bienkow@informatik.uni-hamburg.de Proseminar Ein-/Ausgabe - Stand der Wissenschaft 2013-06-03 1 / 34 Outline 1 Introduction 2 The Project Goals and Priorities History Who is involved?

More information

A New Key-value Data Store For Heterogeneous Storage Architecture Intel APAC R&D Ltd.

A New Key-value Data Store For Heterogeneous Storage Architecture Intel APAC R&D Ltd. A New Key-value Data Store For Heterogeneous Storage Architecture Intel APAC R&D Ltd. 1 Agenda Introduction Background and Motivation Hybrid Key-Value Data Store Architecture Overview Design details Performance

More information

Toward portable I/O performance by leveraging system abstractions of deep memory and interconnect hierarchies

Toward portable I/O performance by leveraging system abstractions of deep memory and interconnect hierarchies Toward portable I/O performance by leveraging system abstractions of deep memory and interconnect hierarchies François Tessier, Venkatram Vishwanath, Paul Gressier Argonne National Laboratory, USA Wednesday

More information

Interconnect Your Future

Interconnect Your Future Interconnect Your Future Smart Interconnect for Next Generation HPC Platforms Gilad Shainer, August 2016, 4th Annual MVAPICH User Group (MUG) Meeting Mellanox Connects the World s Fastest Supercomputer

More information

QLogic TrueScale InfiniBand and Teraflop Simulations

QLogic TrueScale InfiniBand and Teraflop Simulations WHITE Paper QLogic TrueScale InfiniBand and Teraflop Simulations For ANSYS Mechanical v12 High Performance Interconnect for ANSYS Computer Aided Engineering Solutions Executive Summary Today s challenging

More information

LAMMPS-KOKKOS Performance Benchmark and Profiling. September 2015

LAMMPS-KOKKOS Performance Benchmark and Profiling. September 2015 LAMMPS-KOKKOS Performance Benchmark and Profiling September 2015 2 Note The following research was performed under the HPC Advisory Council activities Participating vendors: Intel, Dell, Mellanox, NVIDIA

More information

Lustre overview and roadmap to Exascale computing

Lustre overview and roadmap to Exascale computing HPC Advisory Council China Workshop Jinan China, October 26th 2011 Lustre overview and roadmap to Exascale computing Liang Zhen Whamcloud, Inc liang@whamcloud.com Agenda Lustre technology overview Lustre

More information

University at Buffalo Center for Computational Research

University at Buffalo Center for Computational Research University at Buffalo Center for Computational Research The following is a short and long description of CCR Facilities for use in proposals, reports, and presentations. If desired, a letter of support

More information

Cost and Performance benefits of Dell Compellent Automated Tiered Storage for Oracle OLAP Workloads

Cost and Performance benefits of Dell Compellent Automated Tiered Storage for Oracle OLAP Workloads Cost and Performance benefits of Dell Compellent Automated Tiered Storage for Oracle OLAP This Dell technical white paper discusses performance and cost benefits achieved with Dell Compellent Automated

More information

Lustre Interface Bonding

Lustre Interface Bonding Lustre Interface Bonding Olaf Weber Sr. Software Engineer 1 Interface Bonding A long-standing wish list item known under a variety of names: Interface bonding Channel bonding Multi-rail Fujitsu implemented

More information

Dell EMC. VxBlock Systems for VMware NSX 6.3 Architecture Overview

Dell EMC. VxBlock Systems for VMware NSX 6.3 Architecture Overview Dell EMC VxBlock Systems for VMware NSX 6.3 Architecture Overview Document revision 1.1 March 2018 Revision history Date Document revision Description of changes March 2018 1.1 Updated the graphic in Logical

More information

A Cloud WHERE PHYSICAL ARE TOGETHER AT LAST

A Cloud WHERE PHYSICAL ARE TOGETHER AT LAST A Cloud WHERE PHYSICAL AND VIRTUAL STORAGE ARE TOGETHER AT LAST Not all Cloud solutions are the same so how do you know which one is right for your business now and in the future? NTT Communications ICT

More information