Rechenzentrum HIGH PERFORMANCE SCIENTIFIC COMPUTING

Similar documents
Experiences with HP SFS / Lustre in HPC Production

Filesystems on SSCK's HP XC6000

Parallel File Systems Compared

Lessons learned from Lustre file system operation

Altix Usage and Application Programming

BlueGene/L. Computer Science, University of Warwick. Source: IBM

Our new HPC-Cluster An overview

Intel Enterprise Edition Lustre (IEEL-2.3) [DNE-1 enabled] on Dell MD Storage

Brutus. Above and beyond Hreidar and Gonzales

Sami Saarinen Peter Towers. 11th ECMWF Workshop on the Use of HPC in Meteorology Slide 1

User Training Cray XC40 IITM, Pune

Communication has significant impact on application performance. Interconnection networks therefore have a vital role in cluster systems.

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance

Using file systems at HC3

Performance Tools for Technical Computing

Dell EMC Ready Bundle for HPC Digital Manufacturing Dassault Systѐmes Simulia Abaqus Performance

Users and utilization of CERIT-SC infrastructure

Performance Analysis and Prediction for distributed homogeneous Clusters

Windows-HPC Environment at RWTH Aachen University

Operating two InfiniBand grid clusters over 28 km distance

ACCRE High Performance Compute Cluster

Cray RS Programming Environment

The Cray Programming Environment. An Introduction

Habanero Operating Committee. January

Introduction...2. Executive summary...2. Test results...3 IOPs...3 Service demand...3 Throughput...4 Scalability...5

An Overview of Fujitsu s Lustre Based File System

Cluster Network Products

The AMD64 Technology for Server and Workstation. Dr. Ulrich Knechtel Enterprise Program Manager EMEA

Intel High-Performance Computing. Technologies for Engineering

Intel Math Kernel Library

Scalable Computing at Work

Datura The new HPC-Plant at Albert Einstein Institute

Computer Science Section. Computational and Information Systems Laboratory National Center for Atmospheric Research

IBM Power Systems HPC Cluster

Intel Many Integrated Core (MIC) Architecture

Hands-On Workshop bwunicluster June 29th 2015

6.1 Multiprocessor Computing Environment

The Optimal CPU and Interconnect for an HPC Cluster

HOKUSAI System. Figure 0-1 System diagram

Assessing performance in HP LeftHand SANs

GPUs and Emerging Architectures

IME (Infinite Memory Engine) Extreme Application Acceleration & Highly Efficient I/O Provisioning

Performance Monitoring in an HP SFS Environment

HP solutions for mission critical SQL Server Data Management environments

A Study of High Performance Computing and the Cray SV1 Supercomputer. Michael Sullivan TJHSST Class of 2004

High Performance Computing. What is it used for and why?

SAS Enterprise Miner Performance on IBM System p 570. Jan, Hsian-Fen Tsao Brian Porter Harry Seifert. IBM Corporation

SuperMike-II Launch Workshop. System Overview and Allocations

2008 International ANSYS Conference

Portable and Productive Performance with OpenACC Compilers and Tools. Luiz DeRose Sr. Principal Engineer Programming Environments Director Cray Inc.

Maximize Performance and Scalability of RADIOSS* Structural Analysis Software on Intel Xeon Processor E7 v2 Family-Based Platforms

HP Integrity rx2600 server

Basic Specification of Oakforest-PACS

Arm Processor Technology Update and Roadmap

Feedback on BeeGFS. A Parallel File System for High Performance Computing

General Purpose GPU Computing in Partial Wave Analysis

Accelerating Implicit LS-DYNA with GPU

NERSC Site Update. National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory. Richard Gerber

Intel Performance Libraries

Introducing the next generation of affordable and productive massively parallel processing (MPP) computing the Cray XE6m supercomputer.

Introduction to High Performance Computing and an Statistical Genetics Application on the Janus Supercomputer. Purpose

Outline. March 5, 2012 CIRMMT - McGill University 2

Current Status of the Next- Generation Supercomputer in Japan. YOKOKAWA, Mitsuo Next-Generation Supercomputer R&D Center RIKEN

Sun Lustre Storage System Simplifying and Accelerating Lustre Deployments

Introduction to the SHARCNET Environment May-25 Pre-(summer)school webinar Speaker: Alex Razoumov University of Ontario Institute of Technology

The Stampede is Coming: A New Petascale Resource for the Open Science Community

GOING ARM A CODE PERSPECTIVE

Dell EMC Ready Bundle for HPC Digital Manufacturing ANSYS Performance

The Oracle Database Appliance I/O and Performance Architecture

The Cray Rainier System: Integrated Scalar/Vector Computing

CP2K Performance Benchmark and Profiling. April 2011

Experiences with the Parallel Virtual File System (PVFS) in Linux Clusters

A Comprehensive Study on the Performance of Implicit LS-DYNA

HPC on Windows. Visual Studio 2010 and ISV Software

FUJITSU PHI Turnkey Solution

: A new version of Supercomputing or life after the end of the Moore s Law

Hewlett Packard Enterprise HPE GEN10 PERSISTENT MEMORY PERFORMANCE THROUGH PERSISTENCE

Fujitsu s Approach to Application Centric Petascale Computing

High-Performance Lustre with Maximum Data Assurance

Developing, Debugging, and Optimizing GPU Codes for High Performance Computing with Allinea Forge

Monitoring and Trouble Shooting on BioHPC

Workshop on High Performance Computing (HPC08) School of Physics, IPM February 16-21, 2008 HPC tools: an overview

Intel Math Kernel Library 10.3

Performance Benefits of OpenVMS V8.4 Running on BL8x0c i2 Server Blades

The Use of Cloud Computing Resources in an HPC Environment

SAP HANA Inspirience Day Workshop SAP HANA Infra. René Witteveen Master ASE Converged Infrastructure, HP

NetApp High-Performance Storage Solution for Lustre

MSC Nastran Explicit Nonlinear (SOL 700) on Advanced SGI Architectures

The Arm Technology Ecosystem: Current Products and Future Outlook

Lecture 3: Intro to parallel machines and models

New User Seminar: Part 2 (best practices)

Performance Analysis of Memory Transfers and GEMM Subroutines on NVIDIA TESLA GPU Cluster

Technical Computing Suite supporting the hybrid system

High Performance Computing. Leopold Grinberg T. J. Watson IBM Research Center, USA

MOHA: Many-Task Computing Framework on Hadoop

10 Gbit/s Challenge inside the Openlab framework

Sybase Adaptive Server Enterprise on Linux

Introduction of Oakforest-PACS

High Performance Computing and Data Resources at SDSC

HPC and IT Issues Session Agenda. Deployment of Simulation (Trends and Issues Impacting IT) Mapping HPC to Performance (Scaling, Technology Advances)

Transcription:

Rechenzentrum HIGH PERFORMANCE SCIENTIFIC COMPUTING

Contents Scientifi c Supercomputing Center Karlsruhe (SSCK)... 4 Consultation and Support... 5 HP XC 6000 Cluster at the SSC Karlsruhe... 6 Architecture of the XC6000 Cluster... 7 Confi guration of the Phase 1 System xc1... 8 The Parallel File System HP SFS... 10 Applications on the XC6000 Cluster... 12 Program Development Environment... 13 High Performance Technical Computing Competence Center (HPTC 3 )... 14 Universität Karlsruhe (TH) Rechenzentrum Scientifi c Supercomputing Center Karlsruhe Zirkel 2 D-76128 Karlsruhe Phone: +49 721 608 8011 Fax: +49 721 32550 xc-hotline@uni-karlsruhe.de http://www.rz.uni-karlsruhe.de/ssc Editor: Prof. Dr. Wilfried Juling June 2005

4 Scientific Supercomputing Center Karlsruhe (SSCK) The Scientifi c Supercomputing Center Karlsruhe is the service unit of the University s Computing Center that cares for supercomputer customers. The SSCK provides support for experts as well as for novices in parallel computing. Our goal is to help our customers in all problems related to scientifi c supercomputing. Our services are not only confi ned to advice on how to use supercomputers effi ciently but you will also get qualifi ed help if you are looking for appropriate mathematical methods or simply having problems to login. Since the installation of the fi rst supercomputer in Karlsruhe in 1983 the Computing Center of the University has been continuously engaged in supercomputing. Besides the operating of the machines Karlsruhe has always been establishing expert groups for scientifi c supercomputing. In the 1980s and 1990s experts of the Computing Center in Karlsruhe tuned numerical libraries for vector computers as well as microprocessor based parallel supercomputers. From this time on close cooperation with other institutes of the university and industrial companies has been initiated. At the Computing Center solvers for arbitrary systems of nonlinear partial differential equations and iterative linear solvers have been developed. Thus, the experts at the SSCK know about the needs and problems of their customers from their own experience. The SSCK has allied with the HLRS from Stuttgart University and the IWR from Heidelberg University. The High Performance Computing Competence Center Baden-Württemberg (hkz-bw) combines HPC resources in the State of Baden-Württemberg in order to provide leading edge HPC service, support and research.

Consultation and Support According to the mission of the Scientific Supercomputing Center Karlsruhe you will have support on different levels: Hotline The SSCK has set up a helpline where you rapidly get support from Monday to Friday from 9:00 to 17:00. Simply call +49(0)721/608-8011. If your problem cannot immediately be solved, it will be pursued by our experts. You can contact us as well by e-mail: xc-hotline@uni-karlsruhe.de. Basic support For parallel computers there are rules that should be obeyed in order to get efficient codes based on the hardware design. We will provide you with all our knowledge how to use massively parallel and vector computers efficiently. Standard solutions At the SSCK you will find experts who know both: the application software packages and the necessary libraries you need to solve your problem. This means that you obtain assistance from the beginning in selecting your appropriate solution and later in using it. Individual solutions: If you want to solve your particular problem for the first time with the aid of a high performance computer, then you will find at the SSCK even experts for modelling and state-of-theart solvers. This unique feature is highly considered by all our customers. The SSCK offers several ways to communicate with its customers using the HP XC6000 cluster: 5 Mailing list for communication between SSCK staff and users xc-users-l@lists.uni-karlsruhe.de E-mail hotline for all inquiries, comments etc. xc-hotline@uni-karlsruhe.de Website of SSCK with lots of information on XC6000, its hardware and software environment and project submission interface http://www.rz.uni-karlsruhe.de/ssc

6 HP XC 6000 Cluster at the SSC Karlsruhe The largest system that is currently operated by SSCK is an HP XC6000 cluster. This supercomputer is provided for projects from the universities of Baden-Württemberg requiring computing power which cannot be satisfi ed by local resources. The system is embedded into the infrastructure of the High Performance Computing Competence Center Baden-Württemberg (hkz-bw). Through the private-public partnership hww the system is also available to industrial customers. The installation of the HP XC6000 Cluster began in April 2004 with a small 16 node test system named xc0. In December 2004 the fi rst production system (xc1) with 108 HP Integrity rx2600 servers followed. After a short installation and internal test phase fi rst customers were able to access the system by the end of 2004. Full production started on March 1, 2005. In May 2005 another six 16-way rx8620 servers were installed and integrated into the xc1 system. By early 2006 the overall system will be upgraded to about 340 nodes, 1200 processor cores, a peak performance of 11 TFlops, 7 TB of main memory and more than 40 TB of global shared disk space. The three steps of the XC6000 installation at SSCK are: Development and test cluster xc0 (April 2004): 12 two-way rx2600 nodes with 4 GB main memory 4 file server nodes Single rail QsNet II interconnect 2 TB shared storage Production cluster xc1 (December 2004 / Mai 2005): 108 two-way rx2600 nodes with 12 GB main memory 6 sixteen-way nodes with 128 GB main memory (configured as 12 eight-way nodes) 8 file server nodes Single rail QsNet II interconnect 11 TB global disk space Production cluster xc2 (Early 2006): 218 four-way nodes o Two sockets o Dual core Itanium2 processor (codenamed Montecito) Single or dual rail QsNet II 30 TB global disk space

Architecture of the XC6000 Cluster The HP XC6000 Cluster is a distributed memory parallel computer where each node has two or more Intel Itanium2 processors, local memory, disks and network adapters. Special file server nodes are added to the HP XC6000 Cluster to provide a fast and scalable parallel fi le system. All nodes are connected by a Quadrics QsNet II interconnect. The HP XC software environment is based on HP XC Linux for HPC, a Linux implementation which is compatible to Redhat AS 3.0. On top of this basic operating system a set of open source as well as proprietary software components constitute the XC software stack. This architecture implements a production class environment for scalable clusters based on open standards which enables the application programmer to port easily applications to this system. The nodes of the HP XC6000 Cluster may have different roles and are separated into disjoint groups according to the services supplied by each node. From an end users point of view the different groups of nodes are login nodes, compute nodes, fi le server nodes and cluster management nodes. Login Nodes The login nodes are the only nodes that are directly accessible by end users. These nodes are used for interactive login, file management, program development and interactive pre- and post-processing. Several nodes are dedicated to this service. But the Linux Virtual Server (LVS) provides a single login to the whole cluster. LVS will distribute the login sessions to the different login nodes. Compute Nodes The majority of nodes are compute nodes which are managed by the batch system. Users submit their jobs to this batch system. A job is executed depending on its priority, when the required resources become available. Several queues with specific characteristics for development and production can be implemented. File Server Nodes Special dedicated nodes are used as servers for the HP StorageWorks Scalable File Share (HP SFS). HP SFS is a parallel and scalable file system product based on the Lustre file system. In addition to shared file space there is also local storage on the disks of each node. Management Nodes Certain nodes are dedicated to additional services within the HP XC6000 cluster. This includes resource management, external network connection, administration etc. All the management tasks are served by dedicated management nodes. So there is a clear separation between resources that are used for management or administration and resources used for computing, i.e. the compute nodes are mostly freed from administrative tasks. 7

8 Configuration of the Phase 1 System xc1 The HP XC6000 cluster at the SSC Karlsruhe for the first production phase (xc1) consists of 108 two-way HP Integrity rx2600 nodes: Each of these nodes contains two Intel Itanium2 processors which run at a clock speed of 1.5 GHz and have 6 MB of level 3 cache on the processor chip. Each node has 12 GB of main memory, 146 GB local disk space and an adapter to connect to the Quadrics QsNet II interconnect. 6 16-way HP Integrity rx8620 nodes: For a transition period each of these nodes is partitioned into two units, with 8 processors, 6 GB of main memory and 500 GB local disk space. The Intel Itanium2 processors of the rx8620 nodes run at a speed of 1.6 GHz and have 64 MB of level 3 cache. Each eight CPU partition has its own connection to the QsNet II interconnect. 8 HP Proliant DL 360 file server nodes: Through a storage area network these nodes are connected to seven HP EVA5000 disk systems. This global shared storage has a capacity of 11 TB. It is subdivided into a part used for home directories and a larger part for non permanent files. Two eight-way partitions of the rx8620 nodes are used as login nodes while the other eight-way nodes as well as the majority of the two-way nodes are compute nodes which run parallel user jobs. Quadrics QsNet II FC networks application nodes login nodes service nodes Figure 1: General layout of the HP XC6000 Cluster at the SSC Karlsruhe

The heart of the xc1 cluster is the fast Quadrics QsNet II interconnect. All nodes are attached to this interconnect which is characterized by its very low latency of less than 3 microseconds and a point to point bandwidth of more than 800 MB/s between any pair of nodes. Both values, latency as well as bandwidth, have been measured at the MPI level. The sustained bisection bandwidth of the 128 node network of the xc1 is more than 51 GB/s. These excellent performance figures of the cluster interconnect makes the xc1 ideal for communication intensive applictions. 9 time [µsec] 8 7 6 5 4 3 2 1 0 MPI latency between nodes 0 4 8 16 32 64 128 256 512 1024 2048 message length [Bytes] transfer speed [MB/s] MPI bandwidth between nodes 900 800 700 600 500 400 300 200 100 0 0 32 64 96 128 160 192 224 256 message length [kb] Figure 2: Performance of the cluster interconnect The key figures of the xc1 are: 2-way rx2600 nodes 16-way rx8620 nodes Number of nodes 108 6 CPUs per node 2 16 Peak performance per CPU 6 GFlops 6.4 GFlops Memory per Node 12 GB 128 GB Local disk space per node 146 GB 1060 GB Peak network bandwidth 1.3 GB/s 2.6 GB/s Sustained network bandwidth 800 MB/s 1600 MB/s Total peak performance Total main memory Total local disk space Shared disk space Bisection bandwidth Sustained bisection bandwidth 1.9 TFlops 2 TB 22 TB 11 TB 83 GB/s 51 GB/s

10 The Parallel File System HP SFS In modern compute clusters the CPU performance, the number of nodes and the available memory is steadily increasing, and the applications I/O requirements are often increasing in a similar way. In order to avoid loosing lots of compute cycles by applications waiting for I/O compute clusters have to offer an efficient parallel file system. The parallel file system product on HP XC6000 clusters is HP Storage- Works Scalable File Share (HP SFS). It is based on Lustre technology from Cluster Filesystems Inc. (see www.lustre.org). Architecture of HP SFS In addition to the standard Lustre product HP SFS includes additional software for failover and management and is restricted to certain hardware components. c c c c c c c c c c c c c c Fast Interconnect (QSNet) Admin MDS OSS OSS OSS OSS OSS OSS EVA EVA EVA EVA EVA EVA EVA Figure 3: SSCK s HP SFS system Figure 3 shows the structure of the HP SFS system at SSCK. The servers are configured in failover pairs: The administration node and the Meta Data Server (MDS) node build one pair and the Object Storage Server (OSS) nodes are grouped into pairs so that for each OSS there is another one that can act as its backup. If the MDS node fails a takeover of the corresponding services is done by the Admin node and vice versa. At SSCK s HP XC6000 system eight HP SFS servers and two file systems are configured. One file system called data, is used for home directories and software, has 3.8 TB storage capacity, and uses two designated OSS. The other file system called work is used for scratch data, e.g. intermediate files between different job runs. It has 7.6 TB capacity and uses four OSS. Features of HP SFS The most important requirement for parallel file systems is performance. Parallel file systems typically offer completely parallel data paths from clients to disks even if the files are stored in the same subdirectory. In HP SFS data is transferred from the clients via the fast cluster interconnect to servers and from there to the attached storage. The data is striped over multiple servers and multiple storage systems.

11 Figure 4: Block-wise write and read performance of the file system work Figure 4 shows measurements for the sequential block-wise write and read performance of the file system work. A single client can reach 400 MB/s for writes and 300 MB/s for reads. On the server side the bottleneck for writes is the storage subsystem. For reads the FC adapter on the servers is the restricting hardware component. Since the file system data uses half the number of OSS, the throughput performance is half as high as for the file system work. Besides throughput the metadata performance is important. In fact many parallel file systems are restricted to allow only tens or few hundreds of file creations per second. On the xc1 cluster at SSCK more than 5000 file creates per second have been measured. This metadata performance is shared between the two file systems work and data. In a cluster the scalability of the file system is required: The parallel file system must work even if all clients are heavily doing I/O. Figure 5 shows that the performance stays pretty much at the same rate when lots of clients are doing I/O into the file system data. It also shows that character-wise I/O only has an impact on the performance of clients. This is because Lustre gathers small I/O packets into large buffers. Figure 5: Character-wise and block-wise write and read performance of the file system data

12 Applications on the XC6000 Cluster With its different types of nodes the XC6000 cluster at SSCK can meet the requirements of a broad range of applications: applications that are parallelized by the message passing paradigm and use high numbers of processors will run on a subset of the twoway rx2600 nodes and exchange messages over the Quadrics inter connect. applications that are parallelized using shared memory either by OpenMP or by explicit multithreading with Pthreads can run on the 8-way nodes or soon on 16-way nodes. In the future even applications combining both parallelization paradigms may be supported. As all nodes have at least 6 GB of main memory for each CPU the system is especially suited for applications with high memory requirements. The Intel Itanium2 processors of the HP XC6000 cluster are characterized in particular by a high floating point performance and a very large cache, which is located on the processor chip. Therefore it can be accessed with very short latency and at extremely high bandwidth. Optimization of application programs for the architecture of the Itanium2 processor have demonstrated that an increase of performance by a factor of two can be reached. Only a few months after start of operation of the XC cluster at the SSCK user projects from different research fields have been initiated including: Numerical Analysis Chemistry Life Sciences Computational Fluid Dynamics Structural Mechanics Computer Science Electrical Engineering Solid State Physics Instead of developing and porting own programs software from independent software vendors (ISV) may be used on XC6000 clusters. Many ISV codes are available today and SSCK is working together with HP, to increase the list of codes that are enabled for XC6000 clusters. Programs that are available for the XC6000 cluster include CFX-5, Fluent, StarCD, MSC.Nastran, ABAQUS and LS-DYNA. More information, on which codes have been installed already at the SSCK and at which conditions they can be used are published on the SSCK website.

Program Development Environment Since the operating system of the XC6000 is Linux and many parts of the XC system software are based on open source products a rich set of program development tools is available or can be installed within the XC environment. Compilers Different versions of the Intel C/C++ and Fortran compilers, the latest GNU C/C++ and Fortran compilers as well as the NAG Fortran95 compiler have been installed on xc1. All compilers can compile programs that are parallelized with the message passing paradigm. The Intel compilers and the NAG compiler also support the OpenMP standard. Parallelization Environments For distributed memory parallelism HP MPI is available. It supports the full MPI 1.2 specification but also includes many functions of MPI 2. MPI IO is supported for the parallel file system HP SFS and shows high parallel performance. Specific interface libraries allow object code compatibility with MPICH. Debuggers Besides the serial debugger idb form Intel and the GNU debugger gdb the Distributed Debugging Tool (DDT), a parallel debugger from Allinea Ltd. is available on xc1. DDT supports a graphical user interface and is able to debug serial and MPI-parallelized as well as thread-parallelized programs. Performance Analysis Tools The distribution of communication and computation can be visualized and analyzed with Intel s tracecollector and traceanalyzer package. Some other performance anlysis features are offered by the runtime environment of HP MPI. In order to do a detailed analysis of resource usage different profiling tools and interfaces to the performance counters of the Itanium2 processor are being studied by the SSCK and will be made available to end users. This includes tools like Intel s VTune, oprofile, perfmon and others. Mathematical Libraries Numerical subprogram libraries that have been installed on the xc1 cluster include Intel s Mathematical Kernel Library (MKL), HP s mathematical library MLIB, the NAG C and Fortran libraries as well as the linear solver package LINSOL from SSCK. Tuned implementations of well established open source libraries like BLAS, LAPACK, ScaLAPACK or Metis are part of MKL or MLIB. More detailed information on the software environment can be found on the web (see http://www.rz.uni-karlsruhe.de/ssc/software). 13

14 High Performance Technical Computing Competence Center (HPTC 3 ) Together with Hewlett Packard and Intel, the SSCK has established the High Performance Technical Computing Competence Center (HPTC 3 ). The main goals of this partnership are the further development of the HP XC systems to make it even more usable and reliable for the customers of the SSCK. This includes development, testing and early deployment of advanced tools supporting the application development process. In close cooperation with end users and independent software vendors the HPTC 3 will extend the portfolio of application packages that is available on HP XC6000 clusters. The tuning and optimization of applications to reach highest possible performance is another challenging goal of HPTC 3. Projects that have been started in the framework of HPTC 3 include: Integration of LDAP into the XC software environment Improvements of high availability of critical resources of the XC cluster Tuning of application codes Early test of new tools in a production environment Organization of workshops