Introduction of Oakforest-PACS

Similar documents
Basic Specification of Oakforest-PACS

Oakforest-PACS (OFP) Taisuke Boku

Towards Efficient Communication and I/O on Oakforest-PACS: Large-scale KNL+OPA Cluster

Oakforest-PACS and PACS-X: Present and Future of CCS Supercomputers

Overview of Supercomputer Systems. Supercomputing Division Information Technology Center The University of Tokyo

Overview of Reedbush-U How to Login

Overview of Supercomputer Systems. Supercomputing Division Information Technology Center The University of Tokyo

Overview of Supercomputer Systems. Supercomputing Division Information Technology Center The University of Tokyo

Supercomputers in ITC/U.Tokyo 2 big systems, 6 yr. cycle

IHK/McKernel: A Lightweight Multi-kernel Operating System for Extreme-Scale Supercomputing

Update of Post-K Development Yutaka Ishikawa RIKEN AICS

Japan s post K Computer Yutaka Ishikawa Project Leader RIKEN AICS

Results from TSUBAME3.0 A 47 AI- PFLOPS System for HPC & AI Convergence

User Training Cray XC40 IITM, Pune

Fujitsu HPC Roadmap Beyond Petascale Computing. Toshiyuki Shimizu Fujitsu Limited

Fujitsu s Technologies to the K Computer

The Architecture and the Application Performance of the Earth Simulator

2018/9/25 (1), (I) 1 (1), (I)

HOKUSAI System. Figure 0-1 System diagram

Accelerated Computing Activities at ITC/University of Tokyo

T2K & HA-PACS Projects Supercomputers at CCS

Cray XC Scalability and the Aries Network Tony Ford

Brand-New Vector Supercomputer

Site Update for Oakforest-PACS at JCAHPC

HPC Storage Use Cases & Future Trends

Current Status of the Next- Generation Supercomputer in Japan. YOKOKAWA, Mitsuo Next-Generation Supercomputer R&D Center RIKEN

Oakforest-PACS (OFP) : Japan s Fastest Supercomputer

Overview of Tianhe-2

Introduction of Fujitsu s next-generation supercomputer

NERSC Site Update. National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory. Richard Gerber

Kengo Nakajima Information Technology Center, The University of Tokyo. SC15, November 16-20, 2015 Austin, Texas, USA

Post-K: Building the Arm HPC Ecosystem

DDN and Flash GRIDScaler, Flashscale Infinite Memory Engine

Umeå University

Umeå University

High Performance Computing with Fujitsu

in Action Fujitsu High Performance Computing Ecosystem Human Centric Innovation Innovation Flexibility Simplicity

Intel Many Integrated Core (MIC) Architecture

InfiniBand Strengthens Leadership as the Interconnect Of Choice By Providing Best Return on Investment. TOP500 Supercomputers, June 2014

Introduction to the K computer

Introduction to High Performance Computing. Shaohao Chen Research Computing Services (RCS) Boston University

HPC Architectures. Types of resource currently in use

Fujitsu s Approach to Application Centric Petascale Computing

Fujitsu s new supercomputer, delivering the next step in Exascale capability

Joint Usage/Research Center for Interdisciplinary Large-scale Information Infrastructures 2018 Call for Proposal of Joint Research Projects

The Arm Technology Ecosystem: Current Products and Future Outlook

Pedraforca: a First ARM + GPU Cluster for HPC

HPC Architectures evolution: the case of Marconi, the new CINECA flagship system. Piero Lanucara

Overview of Reedbush-U How to Login

Arm in HPC. Toshinori Kujiraoka Sales Manager, APAC HPC Tools Arm Arm Limited

Tianhe-2, the world s fastest supercomputer. Shaohua Wu Senior HPC application development engineer

How to run applications on Aziz supercomputer. Mohammad Rafi System Administrator Fujitsu Technology Solutions

Fujitsu Petascale Supercomputer PRIMEHPC FX10. 4x2 racks (768 compute nodes) configuration. Copyright 2011 FUJITSU LIMITED

GOING ARM A CODE PERSPECTIVE

Post-K Supercomputer Overview. Copyright 2016 FUJITSU LIMITED

An Introduction to the Intel Xeon Phi. Si Liu Feb 6, 2015

Designed for Maximum Accelerator Performance

Design and Evaluation of a 2048 Core Cluster System

ARM High Performance Computing

IME (Infinite Memory Engine) Extreme Application Acceleration & Highly Efficient I/O Provisioning

Benchmark results on Knight Landing architecture

IBM CORAL HPC System Solution

Intel Xeon Phi архитектура, модели программирования, оптимизация.

IT4Innovations national supercomputing center. Branislav Jansík

Lustre2.5 Performance Evaluation: Performance Improvements with Large I/O Patches, Metadata Improvements, and Metadata Scaling with DNE

Resources Current and Future Systems. Timothy H. Kaiser, Ph.D.

The Earth Simulator Current Status

Rechenzentrum HIGH PERFORMANCE SCIENTIFIC COMPUTING

Technical Computing Suite supporting the hybrid system

Innovative Alternate Architecture for Exascale Computing. Surya Hotha Director, Product Marketing

Communication-Computation Overlapping with Dynamic Loop Scheduling for Preconditioned Parallel Iterative Solvers on Multicore/Manycore Clusters

FUJITSU PHI Turnkey Solution

System Software Stack for the Next Generation High-Performance Computers

The way toward peta-flops

The next generation supercomputer. Masami NARITA, Keiichi KATAYAMA Numerical Prediction Division, Japan Meteorological Agency

INTRODUCTION TO THE ARCHER KNIGHTS LANDING CLUSTER. Adrian

Managing HPC Active Archive Storage with HPSS RAIT at Oak Ridge National Laboratory

Barcelona Supercomputing Center

Key Technologies for 100 PFLOPS. Copyright 2014 FUJITSU LIMITED

Trends in HPC (hardware complexity and software challenges)

DDN About Us Solving Large Enterprise and Web Scale Challenges

I/O and Scheduling aspects in DEEP-EST

Tightly Coupled Accelerators Architecture

Update on LRZ Leibniz Supercomputing Centre of the Bavarian Academy of Sciences and Humanities. 2 Oct 2018 Prof. Dr. Dieter Kranzlmüller

Infinite Memory Engine Freedom from Filesystem Foibles

FUJITSU HPC and the Development of the Post-K Supercomputer

Short Talk: System abstractions to facilitate data movement in supercomputers with deep memory and interconnect hierarchy

TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 13 th CALL (T ier-0)

Preparing GPU-Accelerated Applications for the Summit Supercomputer

Oak Ridge National Laboratory Computing and Computational Sciences

Cheyenne NCAR s Next-Generation Data-Centric Supercomputing Environment

The Stampede is Coming Welcome to Stampede Introductory Training. Dan Stanzione Texas Advanced Computing Center

Japan HPC Programs - The Japanese national project of the K computer -

Intel Xeon Phi Coprocessor

Intel Xeon Phi архитектура, модели программирования, оптимизация.

Fujitsu and the HPC Pyramid

HIGH PERFORMANCE COMPUTING FROM SUN

White paper Advanced Technologies of the Supercomputer PRIMEHPC FX10

Titan - Early Experience with the Titan System at Oak Ridge National Laboratory

LBRN - HPC systems : CCT, LSU

Transcription:

Introduction of Oakforest-PACS Hiroshi Nakamura Director of Information Technology Center The Univ. of Tokyo (Director of JCAHPC)

Outline Supercomputer deployment plan in Japan What is JCAHPC? Oakforest-PACS system Application Summary Impacts of extreme scale computing (2017/11/2) 2

Computational Resource Providers of HPCI Tier1: K Computer at RIKEN Tier2: Supercomputers of 9 universities and 2 research institutes : Oakforest-PACS Oakforest-PACS Impacts of extreme scale computing (2017/11/2) 3

Deployment plan of Tier-2 Supercomputers (as of May. 2017) available at HPCI Consortium (www.hpci-c.jp) http://www.hpci-c.jp/news/hpci-infra-summary.pdf Power consumption indicates maximum of power supply (includes cooling facility) Impacts of extreme scale computing (2017/11/2) 4

PF 1000 100 Towards Exascale Computing Tier-1 and Tier-2 supercomputers move forward to Exascale computing like two wheels Future Exascale Post K Computer RIKEN AICS 10 Oakforest-PACS JCAHPC (The Univ. of Tokyo and Univ. of Tsukuba) 1 Tokyo Tech. TSUBAME2.0 T2K U. of Tsukuba U. of Tokyo Kyoto U. 2008 2010 2012 2014 2016 2018 2020 Impacts of extreme scale computing (2017/11/2) 5

JCAHPC Joint Center for Advanced High Performance Computing (http://jcahpc.jp) director: Hiroshi Nakamura @ ITC, U-Tokyo established in 2013 under agreement between Information Technology Center (ITC) at The University of Tokyo Center for Computational Sciences (CCS) at University of Tsukuba Design, operate and manage next-generation supercomputer system for researchers Community of advanced HPC research Impacts of extreme scale computing (2017/11/2) 6

Procurement Policy of JCAHPC joint procurement by two universities uniform specification, single shared system Each university is financially responsible to introduce the machine and its operation first attempt in Japan the largest class of budget as national universities supercomputer in Japan Oakforest-PACS : largest scale in Japan investment ratio: U. Tokyo : U. Tsukuba = 2:1 Impacts of extreme scale computing (2017/11/2) 7

Oakforest-PACS Impacts of extreme scale computing (2017/11/2) 8 Full operation started Dec. 2016 Official Program started on April 2017 25 PFLOPS peak 8208 KNL CPUs Fat-Tree (full bisection BW) by OmniPath HPL 13.55 PFLOPS: #1 in Japan (2017/6) WW #6(2016/11) #7(2017/6) HPCG WW #3(2016/11) #5(2017/6)

HPCG on Nov. 2016 Impacts of extreme scale computing (2017/11/2) 9

Location of Oakforest-PACS : Kashiwa Campus of U. Tokyo Univ. of Tsukuba Kashiwa Campus Univ. of Tokyo Hongo Campus of U. Tokyo Impacts of extreme scale computing (2017/11/2) 10

Oakforest-PACS in the Room 2 nd floor of Kashiwa Research Complex http://news.mynavi.jp/news/2016/12/02/035/ Impacts of extreme scale computing (2017/11/2) 11

Specification of Oakforest-PACS Total peak performance Total number of compute nodes Compute node Interconnect Product 25 PFLOPS 8,208 Fujitsu PRIMERGY CX600 M1 (2U) + CX1640 M1 x 8node Processor Intel Xeon Phi 7250 (Code name: Knights Landing), 68 cores, 1.4 GHz Memory High BW 16 GB, 490 GB/sec (MCDRAM, effective rate) Product Link speed Topology Low BW 96 GB, 115.2 GB/sec (peak rate) Intel Omni-Path Architecture 100 Gbps Fat-tree with (completely) full-bisection bandwidth Impacts of extreme scale computing (2017/11/2) 12

Computation node & chassis Water cooling wheel & pipe Chassis with 8 nodes, 2U size Computation node (Fujitsu next generation PRIMERGY) with single chip Intel Xeon Phi (Knights Landing, 3+TFLOPS) and Intel Omni-Path Architecture card (100Gbps) Impacts of extreme scale computing (2017/11/2) 13

Rack 15 Chassis with 120 nodes per Rack rear panel radiator Impacts of extreme scale computing (2017/11/2) 14 water cooling pipe

Full bisection bandwidth Fat-tree by Intel Omni-Path Architecture 12 of 768 port Director Switch (Source by Intel) Uplink: 24 Downlink: 24 2......... 1 24 25 48 49 72 All the nodes are connected with FBB Fat-tree globally full bisection bandwidth is preferable for flexible job management. 2/3 of system : University of Tokyo 1/3 of system : University of Tsukuba but job assignment is flexible (no boudary) 362 of 48 port Edge Switch Impacts of extreme scale computing (2017/11/2) 15 2 Compute Nodes 8208 Login Nodes 20 Parallel FS 64 IME 300 Mgmt, etc. 8 Total 8600

Specification of Oakforest-PACS (I/O) Parallel File System File Cache System Type Total Capacity Product Aggregate BW Type Total capacity Product Aggregate BW Power consumption Lustre File System 26.2 PB # of racks 102 DataDirect Networks SFA14KE 500 GB/sec Burst Buffer, Infinite Memory Engine (by DDN) 940 TB (NVMe SSD, including parity data by erasure coding) DataDirect Networks IME14K 1,560 GB/sec 4.2 MW (including cooling) actually ~3.0MW Impacts of extreme scale computing (2017/11/2) 16

Software of Oakforest-PACS Compute node Login node OS CentOS 7, McKernel Red Hat Enterprise Linux 7 Compiler gcc, Intel compiler (C, C++, Fortran), XcalbleMP MPI Intel MPI, MVAPICH2 Library Intel MKL Application Distributed FS Job Scheduler Debugger Profiler LAPACK, FFTW, SuperLU, PETSc, METIS, Scotch, ScaLAPACK, GNU Scientific Library, NetCDF, Parallel netcdf, Xabclib, ppopen-hpc, ppopen-at, MassiveThreads mpijava, XcalableMP, OpenFOAM, ABINIT-MP, PHASE system, FrontFlow/blue, FrontISTR, REVOCAP, OpenMX, xtapp, AkaiKKR, MODYLAS, ALPS, feram, GROMACS, BLAST, R packages, Bioconductor, BioPerl, BioRuby Globus Toolkit, Gfarm Fujitsu Technical Computing Suite Allinea DDT Intel VTune Amplifier, Trace Analyzer & Collector Impacts of extreme scale computing (2017/11/2) 17

Post-K Computer and Oakforest-PACS as the two wheels of HPCI in Japan Oakforest-PACS fills blank period between K Computer and Post-K Computer Installation of Post-K Computer is planned in 2020-2021 Shutdown of K Computer is planned in 2018-2019?? System software in Oakforest-PACS developed for Post-K McKernel OS for Many-core era, for a number of thin-cores without OS jitter and core binding Primary OS (based on Linux) on Post-K, and application development goes ahead XcalableMP (XMP) Parallel programming language for directive-base easy coding on distributed memory system Not like explicit message passing with MPI Impacts of extreme scale computing (2017/11/2) 18

Oakforest-PACS resource sharing program (nation-wide) As JCAHPC (20%) HPCI HPC Infrastructure program in Japan to share all supercomputers (free!) Big challenge special use (full system size) As U. Tokyo (56.7%) Interdisciplinary Joint Research Program General use Industrial trial use Educational use Young & Female special use As U. Tsukuba (23.3%) Interdisciplinary Academic Program Large scale general use Impacts of extreme scale computing (2017/11/2) 19

Applications on Oakforest-PACS 20 ARTED (SALMON) Electron Dynamics Lattice QCD Quantum Chrono Dynamics NICAM & COCO Atmosphere & Ocean Coupling GHYDRA Earthquake Simulations Seism3D Seismic Wave Propagation Impacts of extreme scale computing (2017/11/2)

Summary JCAHPC : joint resource center for advanced HPC by Univ. of Tokyo and Univ. of Tsukuba for community for advanced HPC research Oakforest-PACS is currently #1 supercomputer in Japan and available for nation-wide resource sharing programs Oakforest-PACS and Post-K : two wheels of HPCI Oakforest-PACS is also a testbed for McKernel and XcalableMP system software to support Post-K development Full system scale applications are under development with extreme scale and getting new results fundamental physics, global science, disaster simulation, material science, etc. Impacts of extreme scale computing (2017/11/2) 21