High Performance Computing with Fujitsu Ivo Doležel 0 2017 FUJITSU
FUJITSU Software HPC Cluster Suite A complete HPC software stack solution HPC cluster general characteristics HPC clusters consist primarily of compute nodes with exactly the same hardware Clusters can have a few to thousands of compute nodes The software used on each compute node is exactly the same Compute nodes have no keyboards or displays attached to them Fundamental operational requirements Bare metal deployment with no intervention Central management and deployment of all needed software components A way to control resource usage across the compute nodes Ability to run many applications (serial & parallel) simultaneously on the cluster High-speed inter node communication and access to large data storage areas Some sort of shared storage is needed Monitoring and management of nodes NCI cluster in Australia 3592 CX250 nodes ~ 1.2 PFlop/s 1 2017 FUJITSU
FUJITSU Software HPC Cluster Suite Main Features (I) Bare metal deployment of compute nodes Central management of node installation images Central management of additional software packages Automatic management of key configuration files Parallel Middleware Management of cluster resources Application programs Fujitsu HPC Cluster Suite Graphical end-user interface Scientific Libraries Workload manager Manage serial and parallel jobs Compilers, performancee and profiling tools Cluster deployment and management Parallel File System Fair share usage between users Central management of node configuration files Automated installation and configuration Administrator interface Operation and monitoring User environment management Cluster checker Central management of users/passwords Operating System Support for LDAP/AD Red Hat Linux OS Drivers CentOS GPGPU and XEON Phi software support Fujitsu PRIMERGY HPC Clusters 2 2017 FUJITSU
FUJITSU Software HPC Cluster Suite Main Features (II) Application programs Central management of NFS settings Fujitsu HPC Cluster Suite Graphical end-user interface Selection of Open source or commercial workload managers Parallel Middleware Scientific Libraries Compilers, performancee and profiling tools Parallel File System Selection of message passing environments Variation of software configuration within a node group using package groups Web based monitoring for node health/usage Web based interface for application execution, data management Management of cluster resources Automated installation and configuration Red Hat Linux Workload manager Manage serial and parallel jobs Cluster deployment and management Administrator interface Operation and monitoring OS Drivers CentOS User environment management Operating System Fair share usage between users Cluster checker GPGPU and XEON Phi software support Fujitsu PRIMERGY HPC Clusters 3 2017 FUJITSU
FUJITSU Software HPC Cluster Suite HPC Cluster Suite (HCS) Bare metal deployment Software management User management Node configuration Deployment/Management Monitoring and Alerting Comprehensive & Flexible options Flexible choice of Workload Manager Libraries, Compilers Support for Parallel File Systems HPC Gateway - Integrated intuitive WEB interface - Simplicity in using the HPC Cluster and Applications More effective use of resources Broaden HPC and process reuse Share and exchange data more widely FEFS - Parallel File System - Single file namespace across all nodes Increases Storage performance Required in large or high load I/O configurations Fujitsu Exabyte File System Fujitsu developed (Lustre based) 4 2017 FUJITSU
Comparison with the manual installation Knowledge/Skills/Commands needed for installing HCS HCS Manual install Using the HCS Installer Actions/Commands Skills required Actions/Commands Skills required Install Linux OS Upload ISO images Mount iso s Setup SNMP Setup SMTP Run rpm Run CDM installer Run ifconfig Run fjkit-mgr Run cdm-kitops Run cdm-repoman Run cdm-ngedit Run cdm-nfsedit Run cdm-mpedit Run cdm-addhost Reboot a server OS install experience How to copy files How to mount ISO s Editing text files Configuration of SNMP Configuration of SMTP How to install/del RPM s Basic HPC architecture How to run CDM How to configure network interfaces How to install CDM kits, update CDM Repositories, update/create/configure nodegroups, create CDM NFS export and mount definitions, add new hosts to the configuration Reboot a Linux server Approximate install time: 2-3 days Install Linux OS Upload ISO images Edit the hcs.cfg file Run hcs-installer --unattended Turn compute nodes on OS install experience How to copy files Basic HPC architecture Editing text files How to run HCS installer Approximate install time: 2-3 hours 5 2017 FUJITSU
HPC found at all Scales Different systems for varying MESH SIZE Meshing balance between accuracy/quality and turnaround time Mesh size = 0.1m Mesh size = 0.02m Mesh size = 0.005m Grain Conveyor simulation with Discrete Element Method in STAR-CCM+ 6 2017 FUJITSU
HPC found in many Products Different systems utilized for different MODELS Adapted to users segment and capability Bicycle helmet modelling Car beam optimisation Full car offset impact 7 2017 FUJITSU
HPC Cluster User expectations stable working environment hide cluster complexity more time for creativity raising productivity increase innovation LAN ease of use eliminate waste reliable and predictable results transferable best practice workflows optimize the development process Head Node (Management Node) HPC Cluster Distributed / Parallel processing Compute Cluster (Compute Nodes) increase project throughput maximize application effectiveness migrate more projects and new users into HPC Jobs are run here User submits jobs here Jobs are queued here Management network High-speed network (Interconnect) 8 2017 FUJITSU
Fujitsu HPC Gateway Demo Centre Fujitsu HPC Gateway is the end-user interface in PRIMEFLEX for HPC An online demonstrator is available for trial Users can sign up on dedicated web page Login and secure private area are assigned by return Initial trial period of 2 weeks 9 2017 FUJITSU
PRIMEFLEX for HPC First Application Appliances Industry: Customers: CAE Product Manufacturing, Engineering Application: ANSYS Fluent ANSYS CFX Models: CFD (Computational Fluid Dynamics) Industry: Customers: CAE, Physics-based simulation Product design, Engineering, Geophysics Application: COMSOL Multiphysics Model: Geomechanics, subsurface flow, mechanics, chemical Industry: Customers: Automotive, Creative Application: VRED Models: Automotive OEM, Creative agencies 3D Visualisation, real-time digital prototyping 10 2017 FUJITSU
PRIMEFLEX for HPC Appliance Building blocks Platform Entry PRIMERGY RX PRIMERGY CX Base No switch direct interconnect Increment Not available Single-switch maximum Single-switch maximum Rack None Single cabinet Single cabinet 11 2017 FUJITSU
Select Your Preferred Hardware Platform CELSIUS PRIMERGY Rack Server Capacity Scalability, Compute density PRIMERGY Blade Server Capability Scalability, Infrastructure density Flexibility to address all kinds of customer requirements PRIMERGY CX400 skinless server Massive scale-out due ultra dense server GPU coprocessor support PRIMERGY blade server Industry leading blade server density PRIMERGY rack server CELSIUS workstations PRIMERGY Scale-out Server 12 2017 FUJITSU
Supercomputers since 1977, PRIMERGY in HPC for more than 10 Years! FX100 Exascale World s Fastest Vector Processor (1999) VPP5000 FX10 F230-75APU Japan s First Vector (Array) Supercomputer(1977) No.1 in Top500 (Nov. 1993) Gordon Bell Prize (1994, 95, 96) VP Series NWT* Developed with NAL VPP500 CJAXA AP1000 VPP300/700 AP3000 SPARC Enterprise PRIMEPOWER HPC2500 World s Most Scalable Supercomputer (2003) FX1 K computer Most Efficient Performance in Top500 (Nov. 2008) PRIMEQUEST Cluster node Japan s Largest Cluster in Top500 (July 2004) HX600 Cluster node PRIMERGY RX200 No.1 in Top500 (Jun / Nov 2011) PRIMERGY CX400 Scale-out server PRIMERGY BX400/900 Cluster node next x86 generation PRIMERGY CX600 Scale-out server *NWT: Numerical Wind Tunnel 13 2017 FUJITSU
FUJITSU Server PRIMERGY CX600 M1 Multi-node chassis Platform for highly parallel computing Maximum density with 8 nodes in 2U Specialized for parallel workloads Compliant to conventional datacenter environment Optimized software stack CX600 HPC optimized scale-out server platform based on Intel Xeon Phi 7200 ( Knights Landing ) technology CX1640 Single socket Xeon Phi server node for significant performance boost in parallel-processing 14 2017 FUJITSU
FUJITSU Server PRIMERGY CX600 M1 Server Node Intel Xeon Phi processor 7200 product family 16 GB high-bandwidth on-package MCDRAM memory, >500GB/sec Additional 6x DDR4 memory DIMMs, up to 384 GB, 2,400 MHz At air cooling: 1x SATADOM or 1x 2.5 non hotplug HDD / SSD, At liquid cooling: 1x SATADOM Fan less server node with shared power and cooling PRIMERGY CX1640 M1 Parallel computing node Condensed half-width-1u server node 8x CX1640 M1 per chassis HPC Usage Scenarios Head Node File Server Node NAS Compute Node Login Node Accelerator Card Node 15 2017 FUJITSU
Higher... Faster... Further... A path towards Exascale enforces a deployment of parallelism at each level to the ultimate extent: Node level (distributed memory) Multi socket (shared memory on nodes) CPU level (number of cores) Instruction level (SIMD) Intel Xeon Phi Processor Challenges Node parallelism ultra-high-speed interconnect CPU parallelism higher memory bandwidth greater complexity of memory hierarchy Core parallelism increase of system errors Amdahl s Law every portion of serial code lowers the overall performance Name FLOPS exaflops 10 18 petaflops 10 15 floating-point operations per second Intel Xeon E5 Processor 22 cores, 44 threads Towards Many Core Architectures e.g. Intel Xeon Phi 7200 product family (up to 72 cores) 16 2017 FUJITSU
17 2017 FUJITSU