CS500 SMARTER CLUSTER SUPERCOMPUTERS

Similar documents
Smarter Clusters from the Supercomputer Experts

Designed for Maximum Accelerator Performance

HPC Software Requirements to Support an HPC Cluster Supercomputer

HIGH-PERFORMANCE STORAGE FOR DISCOVERY THAT SOARS

Introducing the next generation of affordable and productive massively parallel processing (MPP) computing the Cray XE6m supercomputer.

Preparing GPU-Accelerated Applications for the Summit Supercomputer

Scaling Across the Supercomputer Performance Spectrum

Refining and redefining HPC storage

Cray XD1 Supercomputer Release 1.3 CRAY XD1 DATASHEET

Scaling to Petaflop. Ola Torudbakken Distinguished Engineer. Sun Microsystems, Inc

IBM Power Systems HPC Cluster

Xyratex ClusterStor6000 & OneStor

Managing complex cluster architectures with Bright Cluster Manager

Scalable Computing at Work

IBM Power AC922 Server

Altair OptiStruct 13.0 Performance Benchmark and Profiling. May 2015

Huawei KunLun Mission Critical Server. KunLun 9008/9016/9032 Technical Specifications

Dell EMC Ready Bundle for HPC Digital Manufacturing Dassault Systѐmes Simulia Abaqus Performance

Oak Ridge National Laboratory Computing and Computational Sciences

Huawei KunLun Mission Critical Server. KunLun 9008/9016/9032 Technical Specifications

Portable and Productive Performance with OpenACC Compilers and Tools. Luiz DeRose Sr. Principal Engineer Programming Environments Director Cray Inc.

CAS 2K13 Sept Jean-Pierre Panziera Chief Technology Director

BeeGFS. Parallel Cluster File System. Container Workshop ISC July Marco Merkel VP ww Sales, Consulting

SUSE. High Performance Computing. Eduardo Diaz. Alberto Esteban. PreSales SUSE Linux Enterprise

HPC and AI Solution Overview. Garima Kochhar HPC and AI Innovation Lab

ABySS Performance Benchmark and Profiling. May 2010

SUSE. High Performance Computing. Kai Dupke. Meike Chabowski. Senior Product Manager SUSE Linux Enterprise

Overview of Tianhe-2

GROMACS (GPU) Performance Benchmark and Profiling. February 2016

LAMMPS-KOKKOS Performance Benchmark and Profiling. September 2015

SUN CUSTOMER READY HPC CLUSTER: REFERENCE CONFIGURATIONS WITH SUN FIRE X4100, X4200, AND X4600 SERVERS Jeff Lu, Systems Group Sun BluePrints OnLine

IBM Power Advanced Compute (AC) AC922 Server

Making Supercomputing More Available and Accessible Windows HPC Server 2008 R2 Beta 2 Microsoft High Performance Computing April, 2010

Illinois Proposal Considerations Greg Bauer

I/O Monitoring at JSC, SIONlib & Resiliency

SGI Overview. HPC User Forum Dearborn, Michigan September 17 th, 2012

INCREASE IT EFFICIENCY, REDUCE OPERATING COSTS AND DEPLOY ANYWHERE

Design and Evaluation of a 2048 Core Cluster System

Pedraforca: a First ARM + GPU Cluster for HPC

19. prosince 2018 CIIRC Praha. Milan Král, IBM Radek Špimr

Readme for Platform Open Cluster Stack (OCS)

The FlashStack Data Center

NVDIA DGX Data Center Reference Design

QuickSpecs HPE SGI 8600

An Introduction to GPFS

VERTIV SERVICE CAPABILITY

Real Parallel Computers

STAR-CCM+ Performance Benchmark and Profiling. July 2014

User Training Cray XC40 IITM, Pune

Interconnect Your Future

InfoBrief. Platform ROCKS Enterprise Edition Dell Cluster Software Offering. Key Points

Sun Lustre Storage System Simplifying and Accelerating Lustre Deployments

HPC Saudi Jeffrey A. Nichols Associate Laboratory Director Computing and Computational Sciences. Presented to: March 14, 2017

The Stampede is Coming Welcome to Stampede Introductory Training. Dan Stanzione Texas Advanced Computing Center

Feedback on BeeGFS. A Parallel File System for High Performance Computing

in Action Fujitsu High Performance Computing Ecosystem Human Centric Innovation Innovation Flexibility Simplicity

Barcelona Supercomputing Center

InfiniBand-based HPC Clusters

IBM Z servers running Oracle Database 12c on Linux

An Oracle White Paper December Accelerating Deployment of Virtualized Infrastructures with the Oracle VM Blade Cluster Reference Configuration

Chelsio Communications. Meeting Today s Datacenter Challenges. Produced by Tabor Custom Publishing in conjunction with: CUSTOM PUBLISHING

ARCHER/RDF Overview. How do they fit together? Andy Turner, EPCC

Deep Learning mit PowerAI - Ein Überblick

Architecting Storage for Semiconductor Design: Manufacturing Preparation

HPC Solution. Technology for a New Era in Computing

NetApp High-Performance Storage Solution for Lustre

IBM CORAL HPC System Solution

Bright Cluster Manager

Technical Computing Suite supporting the hybrid system

HPC Architectures. Types of resource currently in use

IBM Intelligent Cluster portfolio now includes IBM Flex Systems and enhancements to DataDirect Networks, Mellanox, and NetApp offerings

Productive Performance on the Cray XK System Using OpenACC Compilers and Tools

IBM Spectrum Scale IO performance

NVIDIA DGX SYSTEMS PURPOSE-BUILT FOR AI

The Red Storm System: Architecture, System Update and Performance Analysis

Present and Future Leadership Computers at OLCF

Leonhard: a new cluster for Big Data at ETH

Store Process Analyze Collaborate Archive Cloud The HPC Storage Leader Invent Discover Compete

The Stampede is Coming: A New Petascale Resource for the Open Science Community

UCX: An Open Source Framework for HPC Network APIs and Beyond

The following NVIDIA accelerators are available from HPE, for use in certain HPE ProLiant DL-series, ML-series and SL-series servers.

Introducing Panasas ActiveStor 14

Innodisk FlexiArray Storage Appliance. Accelerates Your I/O-Intensive Applications with High IOPS.

IBM Virtual Fabric Architecture

DELIVERABLE D5.5 Report on ICARUS visualization cluster installation. John BIDDISCOMBE (CSCS) Jerome SOUMAGNE (CSCS)

The following NVIDIA accelerators are available from HPE, for use in certain HPE ProLiant DL-series, ML-series and SL-series servers.

PART-I (B) (TECHNICAL SPECIFICATIONS & COMPLIANCE SHEET) Supply and installation of High Performance Computing System

Storage + VDI: the results speak for themselves at Nuance Communications

Linux HPC Software Stack

ECS APPLIANCES. Specification Sheet

DESCRIPTION GHz, 1.536TB shared memory RAM, and 20.48TB RAW internal storage teraflops About ScaleMP

SNAP Performance Benchmark and Profiling. April 2014

The following NVIDIA accelerators are available from HPE, for use in certain HPE ProLiant DL-series, ML-series and SL-series servers.

University at Buffalo Center for Computational Research

Boundless Computing Inspire an Intelligent Digital World

Purchasing Services SVC East Fowler Avenue Tampa, Florida (813)

TECHNICAL OVERVIEW ACCELERATED COMPUTING AND THE DEMOCRATIZATION OF SUPERCOMPUTING

HYCOM Performance Benchmark and Profiling

IBM SpectrumAI with NVIDIA Converged Infrastructure Solutions for AI workloads

MELLANOX EDR UPDATE & GPUDIRECT MELLANOX SR. SE 정연구

Transcription:

CS500 SMARTER CLUSTER SUPERCOMPUTERS

OVERVIEW Extending the boundaries of what you can achieve takes reliable computing tools matched to your workloads. That s why we tailor the Cray CS500 cluster supercomputer to your requirements. The CS500 system is customizable, easy to manage, and based on industry standards. It can handle the broadest range of modeling, simulation, analytics, and artificial intelligence workloads. And we configure it to meet your needs. We optimize components and integrate them into a powerful highperformance compute environment that can scale to over 11,000 nodes and 60+ peak petaflops. Flexible node configurations feature the latest processors and interconnect technology. Expertly engineered software simplifies system administration and maintenance. And it s all backed by renowned Cray expertise and support. Flexible computing performance Designed for your workload Reliable and energy efficient Expert Cray support

DESIGNED FOR YOUR WORKLOAD SCALABLE CONFIGURATIONS EXPERTLY OPTIMIZED FOR MULTIPLE USES: ALL-PURPOSE, COMPUTE- OR DATA-INTENSIVE, AND HYBRID WORKLOADS

Cray CS500 cluster supercomputers are designed for your workloads. Choose from a wide range of flexible configurations with the latest processors and accelerators from AMD. Select your chassis, blades, nodes, interconnect topology, storage, and system management capabililties. We optimize industry-standard server nodes and components and pair them with a comprehensive software stack. What you get is a unified HPC system that excels at compute, data, or hybrid workloads. And it s all designed, installed, and supported by Cray supercomputing experts.

CS500 FLEXIBLE COMPUTING PERFORMANCE SCALABLE CONFIGURATIONS The CS500 cluster supercomputer can be customized for multiple uses from an allpurpose, massively parallel HPC system to a cluster optimized for hybrid compute- and data-intensive workloads. Each system has compute nodes and management nodes. Compute nodes run parallel MPI and OpenMP tasks with maximum efficiency. Management nodes provide I/O connectivity, function as login nodes, and can be configured for unique workloads requiring large memory configurations. RELIABLE AND ENERGY EFFICIENT Reliability starts with careful board and component selection, testing, and validation. We add multiple layers of redundancy and fault tolerance to boost your uptime. Your CS500 cluster has redundant power, cooling, management servers, and networks all with failover capabilities. Each system is backed by our team of experts. Options to increase energy savings include highefficiency, load-balancing power options that can reduce energy loss and overall TCO. Cray rackmount servers are highly configurable in an HPC-optimized, industrystandard package. Each configuration can be replicated to create a reliable and powerful large-scale system.

HPC & DEEP LEARNING ENVIRONMENT Develop and optimize applications and efficiently monitor system resources with our complete software environment. The CS500 system s customizable software stack is compatible with most open-source and commercial applications. It includes key Crayauthored components, popular open-source software, and industry-leading applications from Cray partners. You get: Red Hat Enterprise Linux or CentOS Cray Programming Environment, a complete, fully integrated suite of HPCoptimized compilers, libraries, and tools for programming, debugging, performance analysis, workload management, environment setup, and more Bright Cluster Manager for HPC or Cray s Advanced Cluster Engine (ACE) for system management Bright for Deep Learning, a comprehensive deep learning environment that includes a choice of machine learning frameworks including Caffe, Torch, Tensorflow, and Theano; and machine learning libraries including MLPython, NVIDIA CUDA Deep Neural Network library (cudnn), Deep Learning GPU Training System (DIGITS), and CaffeOnSpark EXPERT CRAY SUPPORT Mission-critical systems around the globe rely on our service teams for world-class support. Cray support engineers, field engineers, and field analysts are experienced, highly trained, and backed by subject matter experts. We can program, install, troubleshoot, maintain and administer your system, and optimize your applications. Plus you get access to product support groups and in-house systems to help with diagnostics. CS500 CONFIGURATION OPTIONS Cray rackmount servers Choice of AMD processors and accelerators Multiple interconnect options: 3D torus/fat tree, single/dual rail, QDR/FDR InfiniBand Local hard drives in each server Choice of network-attached file systems and parallel file systems Server management options

CS500 CRAY CS500 SPECIFICATIONS ARCHITECTURE PROCESSORS AND ACCELERATORS Air cooled, up to 72 nodes per rack cabinet Support for AMD EPYC processors MEMORY Up to 2TB depending on processor SKU FDR or EDR InfiniBand with Connect-IB INTERCONNECT AND NETWORKS Options for single- or dual-rail fat tree or 3D torus (configurations vary by chassis) 1 GbE and 10 GbE for management Redundant networks (InfiniBand, 1 GbE and 10 GbE) with failover SYSTEM MANAGEMENT RELIABLE, AVAILABLE, SERVICEABLE (RAS) RESOURCE MANAGEMENT AND JOB SCHEDULING Bright Cluster Manager for HPC (optional) Redundant power, cooling, and management servers with failover capabilities All critical components easily accessible Options for SLURM, Altair PBS Professional, IBM Platform LSF, Adaptive Computing Torque, Maui and Moab, and Grid Engine FILE SYSTEM Cray ClusterStor, NFS, Local FS (Ext3, Ext4 XFS), Lustre, GPFS and Panasas PanFS available as global file systems DISK STORAGE OPERATING SYSTEM PERFORMANCE MONITORING TOOLS COMPILERS, LIBRARIES AND TOOLS POWER COOLING FEATURES CABINET DIMENSIONS (HxWxD) CABINET WEIGHT Full line of FC-attached disk arrays with support for FC, SATA disk drives and SSDs Red Hat, SUSE or CentOS available on compute nodes ACE Management Servers delivered with Red Hat Linux Open source packages such as HPCC, Perfctr, IOR, PAPI/IPM, netperf Cray Programming Environment on Cluster Systems: Cray Compiling Environment (CCE), Cray Scientific Libraries (LibSci and LibSci_ACC), Cray Debugging Support Tools (LGDB and CCDB), and Cray Performance Measurement, Analysis and Porting Tools (CPMAT). Optional: PGI compiler, NVIDIA CUDA, CUSA C/C++, OpenCL, Direct Compute Toolkits, OFED programming tools, Open MPI, MVAPICH2, and IBM Platform MPI Power supplies deliver up to 38 kw per cabinet, with actual consumption based upon configuration Optional 480V power distribution Air cooled Airflow: up to 3,000 cfm in densest configuration; Intake: front; Exhaust: back Optional passive or active chilled cooling rear door heat exchangers 42U/19": 78.39" (1,991 mm) x 23.62" (600 mm) x 47.24" (1,200 mm) standard rack cabinet 42U/19": up to 1,856.3 lbs.; 232 lbs./sq. ft. per cabinet SUPPORT AND SERVICES Turnkey installation services with worldwide support and service options WWW.CRAY.COM 2019 Cray Inc. All rights reserved. www.cray.com, Cray, the Cray logo and ClusterStor are registered trademarks and CS and CS500 are trademarks of Cray Inc. All other trademarks mentioned herein are the properties of their respective owners. 20190110KL