UAntwerpen, 24 June 2016

Similar documents
Genius Quick Start Guide

Cerebro Quick Start Guide

Genius - introduction

Genius - introduction

Before We Start. Sign in hpcxx account slips Windows Users: Download PuTTY. Google PuTTY First result Save putty.exe to Desktop

Introduction to PICO Parallel & Production Enviroment

Our new HPC-Cluster An overview

Transitioning to Leibniz and CentOS 7

Introduction to GALILEO

Introduc)on to Hyades

OBTAINING AN ACCOUNT:

Introduction to HPC2N

Leibniz Supercomputer Centre. Movie on YouTube

TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 13 th CALL (T ier-0)

Introduction to GALILEO

Introduction to the NCAR HPC Systems. 25 May 2018 Consulting Services Group Brian Vanderwende

LS-DYNA Performance Benchmark and Profiling. October 2017

KISTI TACHYON2 SYSTEM Quick User Guide

IT4Innovations national supercomputing center. Branislav Jansík

Introduction to NCAR HPC. 25 May 2017 Consulting Services Group Brian Vanderwende

GRID Testing and Profiling. November 2017

Gateways to Discovery: Cyberinfrastructure for the Long Tail of Science

Cluster Clonetroop: HowTo 2014

The JANUS Computing Environment

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

How to run applications on Aziz supercomputer. Mohammad Rafi System Administrator Fujitsu Technology Solutions

CENTER FOR HIGH PERFORMANCE COMPUTING. Overview of CHPC. Martin Čuma, PhD. Center for High Performance Computing

Habanero Operating Committee. January

Introduction to Cheyenne. 12 January, 2017 Consulting Services Group Brian Vanderwende

Belnet and BrENIAC. High Performance Network for High Performance Computing. Jan Ooghe, Herman Moons KU Leuven BNC Oktober

Altix Usage and Application Programming

Introduction to CINECA HPC Environment

Guillimin HPC Users Meeting February 11, McGill University / Calcul Québec / Compute Canada Montréal, QC Canada

XSEDE New User Tutorial

Introduction to HPC2N

HPCF Cray Phase 2. User Test period. Cristian Simarro User Support. ECMWF April 18, 2016

Vienna Scientific Cluster: Problems and Solutions

Introduction to GALILEO

Our Workshop Environment

TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 11th CALL (T ier-0)

Comet Virtualization Code & Design Sprint

Introduction to High Performance Computing Using Sapelo2 at GACRC

n N c CIni.o ewsrg.au

SuperMike-II Launch Workshop. System Overview and Allocations

Computing with the Moore Cluster

TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 16 th CALL (T ier-0)

NCAR s Data-Centric Supercomputing Environment Yellowstone. November 28, 2011 David L. Hart, CISL

Introduction to HPC Using the New Cluster at GACRC

Lisa User Day Lisa architecture. John Donners

NAMD Performance Benchmark and Profiling. January 2015

Tech Computer Center Documentation

Introduction to HPC Numerical libraries on FERMI and PLX

Quick Start Guide. by Burak Himmetoglu. Supercomputing Consultant. Enterprise Technology Services & Center for Scientific Computing

Introduction to HPC2N

Advanced Topics in High Performance Scientific Computing [MA5327] Exercise 1

Compiling applications for the Cray XC

UF Research Computing: Overview and Running STATA

TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 6 th CALL (Tier-0)

High Performance Computing and Data Resources at SDSC

Supercomputing environment TMA4280 Introduction to Supercomputing

Resources Current and Future Systems. Timothy H. Kaiser, Ph.D.

Working on the NewRiver Cluster

LBRN - HPC systems : CCT, LSU

CalcUA. Agenda. UAntwerpen-HPC. CalcUA. Stefan Becuwe Kurt Lust. Franky Backeljauw Bert Tijskens. 1. Introduction to UAntwerpen-HPC

Overview of Reedbush-U How to Login

LS-DYNA Performance Benchmark and Profiling. October 2017

The Stampede is Coming Welcome to Stampede Introductory Training. Dan Stanzione Texas Advanced Computing Center

Introduction to CINECA Computer Environment

SCALABLE HYBRID PROTOTYPE

Knights Landing production environment on MARCONI

XSEDE New User Tutorial

Cornell Theory Center 1

User Training Cray XC40 IITM, Pune

Introduction to Unix Environment: modules, job scripts, PBS. N. Spallanzani (CINECA)

First steps on using an HPC service ARCHER

NCAR s Data-Centric Supercomputing Environment Yellowstone. November 29, 2011 David L. Hart, CISL

Batch Systems. Running calculations on HPC resources

Introduction to High-Performance Computing (HPC)

HPC Architectures. Types of resource currently in use

The Why and How of HPC-Cloud Hybrids with OpenStack

Analyzing the High Performance Parallel I/O on LRZ HPC systems. Sandra Méndez. HPC Group, LRZ. June 23, 2016

Effective Use of CCV Resources

Batch Systems & Parallel Application Launchers Running your jobs on an HPC machine

Future of Enzo. Michael L. Norman James Bordner LCA/SDSC/UCSD

BRC HPC Services/Savio

Illinois Proposal Considerations Greg Bauer

Altair OptiStruct 13.0 Performance Benchmark and Profiling. May 2015

PRACE Project Access Technical Guidelines - 19 th Call for Proposals

Using Sapelo2 Cluster at the GACRC

NERSC. National Energy Research Scientific Computing Center

XSEDE New User Tutorial

Guillimin HPC Users Meeting. Bryan Caron

University of California, Riverside. Computing and Communications. Computational UCR. March 28, Introduction 2

Intel Manycore Testing Lab (MTL) - Linux Getting Started Guide

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

Introduction to High-Performance Computing (HPC)

TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 14 th CALL (T ier-0)

Outline. March 5, 2012 CIRMMT - McGill University 2

OpenFOAM Performance Testing and Profiling. October 2017

Using the Yale HPC Clusters

Transcription:

Tier-1b Info Session UAntwerpen, 24 June 2016

VSC HPC environment Tier - 0 47 PF Tier -1 623 TF Tier -2 510 Tf 16,240 CPU cores 128/256 GB memory/node IB EDR interconnect Tier -3 HOPPER/TURING STEVIN THINKING/CEREBRO HYDRA Laptop, workstation Specialized resources

VSC Staff 28 FTE

Tier-1 access Who Researchers of: University and association in the Flemish Community Research institute under authority or supervision of a university in the Flemish Community Strategic research center What Validated research project Experience on Tier1/Tier2 Technical feasibility Responsible for software licenses How www.vscentrum.be/en/access-and-infrastructure/project-access-tier1 When Proposals open continuously allocation at specified dates

Are you a Tier-1 candidate? Do you have already experience with HPC infrastructure? Is your code already running on a HPC infrastructure? Is your code optimized and scalable? Do you need lots of compute time? Will it take too long to get the wanted results using Tier-2? Do you want to simulate more complex, bigger systems? Do you need your results faster?

Project applications Computing resources Software Total node days: 500 5,000 Single job max walltime: 3 days Maximum for 6 months SCRATCH: 250 GB-100,000 files Memory: 128 GB/node 256GB/node Software already available at Tier-1 Software not yet available: is suitable for Tier-1 you have a valid license Installation on Tier-1 @ KUL enough licenses Funding Computer time is FREE

Starting grant/preparatory access Computing resources Submissions at any time Allocations are personal Total node days: 100 Maximum for 2 months Single job walltime: 3 days SCRATCH: 250 GB-100,000 files Memory: 128 GB/node 256GB/node Allocated time free Fast submission procedure Preparation for project application

Tier-1 application workflow TIER-1 CALL Eligible? SCIENTIFICAL EVALUATION Accepted? YES ON YES Tier-1 Experience? YES STANDARD APPLICATION admissible? YES Category 1? YES TECHNICAL EVALUATION NO STARTING GRANT Need more? ALLOCATED TIME YES Accepted? NO FEEDBACK

Evaluation committee a Policy Officer of FWO at least three experts appointed by the Board of Directors who: are not active within the Flemish Community have wide experience in the field of using large computing capacity

Pilot plan Call for open pilots May 27, 2016 Today Start Pilot Tier-1 current users Jun 28, 2016 End Pilot Phase Oct 17, 2016 May Jun Jul Aug Sep Oct Nov Dec 2017 Jun 27, 2016 Deadline for submission open pilots Jul 4, 2016 Start Open Pilot Open to all researchers affiliated with a Flemish university or research institute Proven HPC experience needed More information: https://www.vscentrum.be/assets/1087

Tier-1b system specs

BrENIAC Brain + ENIAC

Top500 list June 2016

BrENIAC architecture

System comparison MUK BrENIAC Total nodes 528 580 Processor type Sandy Bridge E5-2670 Broadwell E5-2680v4 Base Clock Speed 2.6 GHz 2.4 GHz Cores per node 16 28 Total cores 8,448 16,240 x2 Memory per node (GB) 64 435x128/145x256 Memory per core (GB) 4 4.5/9.1 x2 Local Disk 500 GB SATA 128GB SSD Peak performance (TF) 175 623 Network Infiniband FDR fat tree non-blocking Infiniband EDR 2:1 x2

System comparison(2) Hopper BrENIAC Total nodes 144/24 580 Processor type Ivibridge E5-2680v2 Broadwell E5-2680v4 Base Clock Speed 2.8 GHz 2.4 GHz Cores per node 20 28 Total cores 3,360 16,240 Memory per node (GB) 64/256 435x128/145x256 Memory per core (GB) 3.2/12.8 4.5/9.1 Local Disk 500GB SAS 128GB SSD Peak performance (TF) 75 623 Network Infiniband FDR-10 Infiniband EDR 2:1

Compute nodes: Broadwell processors Socket 0 Numa node 0 Socket 1 Numa node 1 core0 core7 core0 core7 DDR4 core1 core8 core1 core8 DDR4 DDR4 core2 core3 L3 core9 core0 core10 QPI QPI core2 core3 L3 core9 core10 DDR4 DDR4 core4 core11 core4 core11 DDR4 DDR4 core5 core6 core12 core13 core5 core6 core12 core13 DDR4 IB I/O

Processor comparison Broadwell DDR4 2400 MHz AVX2 16 DP FLOPs/cycle two 4-wide FMA instructions Cluster on Die 2 NUMA nodes x socket Sandy Bridge Ivybridge DDR3 1600 MHz AVX 8 DP FLOPs/cycle 4-wide AVX add +4-wide AVX mult DDR3 1866 MHz AVX 8 DP FLOPs/cycle 4-wide AVX add +4-wide AVX mult

Network Mellanox Infiniband EDR Fat tree with blocking factor 2:1

Infiniband Islands Edge 03 Edge 06 Edge 09 Edge 12 Edge 17 Edge 20 Edge 23 Edge 26 Edge 02 Edge 05 Edge 08 Edge 11 Spine 12 Edge 16 Edge 19 Edge 22 Edge 25 Edge 01 Edge 04 Edge 07 Edge 10 Spine 11 Edge 15 Edge 18 Edge 21 Edge 24 24n 672c 16n 448c 20n 560c

How to start? Do you have a VSC-number account? Do you have an active Tier-1 project? Your VSC-number credentials are known! ssh vscxxxxx@login1-tier1.hpc.kuleuven.be 2 login nodes: login1-tier1.hpc.kuleuven.be login2-tier1.hpc.kuleuven.be Adding visualization features? and you are in!

VSC Common User Environment Make user s life easy Unify VSC cluster environments Based in an aggregated layered approach

CUE base configuration Application Development Toolchains Cluster management queues Commands support scripts node features System Module system Environment variables: $VSC_XXXXl login.hpc.site.be

Storage areas Name Variable Type Access Backup Quota /user/antwerepn/20x/vsc20xxx $VSC_HOME NFS Global YES 3 GB /data/antwerpen/20x/vsc20xxx $VSC_DATA NFS Global YES 75 GB /scratch/leuven/20x/vsc20xxx $VSC_SCRATCH $VSC_SCRATCH_SITE GPFS Global NO XXX GB /node_scratch $VSC_SCRATCH_NODE ext4 local NO 128 GB Only data from $VSC_SCRATCH needs to be migrated SCRATCH is not removed until the end of the project After the project there is 14 days grace

Toolchains Intel compilers Name intel foss version 2016a 2016a Compilers Intel compilers icc, icpc, ifort Open source GNU compilers gcc, g++, gfortran MPI Library Intel MPI OpenMPI Math libraries Intel MKL OpenBLAS, LAPACK FFTW ScaLAPACK Intel Toolchain has newer versions! VSC BrENIAC Icc, icpc, ifort 16.0.1.150 Build 20151021 16.0.3.210 Build 20160415 Intel MKL 11.3.1.150 11.3.3.210 Intel MPI 5.1.2.150 5.1.3.181

Software Base software is already installed More to come!

Accounting Moab Accounting Manager (MAM) = credit card payment PEND RUN DONE REQ PREAUTH Requested: Walltime nodes BALANCE HOLD Requested: Walltime nodes CHARGE Used: Walltime nodes CREDITS POOL CREDITS POOL REQ CREDITS POOL USED

Accounting commands Projects : lt1_projectname Credits pool = assigned node/days per project Module load accounting gbalance -> check your account status gquote -> check how much credits the job will cost gstatement -> detailed overview glsjob -> account information about a job

Accounting: gbalance : gbalance Id Name Balance Reserved Effective CreditLimit Available ---- ----------------------- ------- -------- --------- ----------- --------- 894 lt1_sys 2630 0 2630 0 2630 918 4492 0 4492 0 4492 1550 Account=lt1_bench 9989 0 9989 0 9989

Submitting jobs Torque/Moab will be used for scheduling Single job policy Use worker framework for single core jobs https://www.vscentrum.be/cluster-doc/running-jobs/worker-framework As always: specify resources -lwalltime=4:30:00 (job will last 4h 30 min) -lnodes=2:ppn=28 (job needs 2 nodes and 20 cores per node) -lpmem=4gb (job request 4 GB of memory per core) Other node features: memory. -lfeature=mem128/mem256 qsub lnodes=10:ppn=28,walltime=1:00:00,pmem=4gb my_job.pbs

Migrating PBS scripts #!/bin/bash -l #PBS -N HPL-2.2-T1-SB-64-1n #PBS -lnodes=1:ppn=16 #PBS -lwalltime=24:00:00 #PBS -lvmem=55gb #PBS lnodes=1:ppn=28 #PBS -lwalltime=10:00:00 #PBS lpmem=4gb module load intel/2016a module load scripts cd $HOME/mydatadir OUTDIR=$VSC_SCRATCH/myresdir mkdir -p $OUTDIR cd $OUTDIR #execute mympirun --stats=3./xhpl.sb.intel2016a

Migrating PBS scripts #!/bin/bash -l #PBS -N HPL-2.2-T1-SB-64-1n #PBS -lnodes=1:ppn=16 #PBS -lwalltime=24:00:00 #PBS -lvmem=55gb #PBS lnodes=1:ppn=28 #PBS -lwalltime=10:00:00 #PBS lpmem=4gb module load intel/2016a module load scripts cd $HOME/mydatadir #PBS A lt1_project OUTDIR=$VSC_DATA/myresdir mkdir -p $OUTDIR cd $OUTDIR #execute mympirun --stats=3./xhpl.sb.intel2016a

Migrating PBS scripts #!/bin/bash -l #PBS -N HPL-2.2-T1-SB-64-1n #PBS -lnodes=1:ppn=16 #PBS -lwalltime=24:00:00 #PBS -lvmem=55gb #PBS lnodes=1:ppn=28 #PBS -lwalltime=10:00:00 #PBS lpmem=4gb module load intel/2016a module load scripts cd $HOME/mydatadir Not available! #PBS A lt1_project OUTDIR=$VSC_SCRATCH/myresdir mkdir -p $OUTDIR cd $OUTDIR #execute mympirun --stats=3./xhpl.sb.intel2016a

Migrating PBS scripts #!/bin/bash -l #PBS -N HPL-2.2-T1-SB-64-1n #PBS -lnodes=1:ppn=16 #PBS -lwalltime=24:00:00 #PBS -lvmem=55gb #PBS lnodes=1:ppn=28 #PBS -lwalltime=10:00:00 #PBS lpmem=4gb module load intel/2016a module load scripts cd $HOME/mydatadir Remove! #PBS A lt1_project OUTDIR=$VSC_SCRATCH/myresdir mkdir -p $OUTDIR cd $OUTDIR #execute mympirun --stats=3./xhpl.sb.intel2016a mpirun np $PBS_NP./xhpl.sb.intel2016a export I_MPI_STATS=ipm

Migrating PBS scripts #!/bin/bash -l #PBS -N HPL-2.2-T1-SB-64-1n #PBS -lnodes=1:ppn=16 #PBS -lwalltime=24:00:00 #PBS -lvmem=55gb #PBS lnodes=1:ppn=28 #PBS -lwalltime=10:00:00 #PBS lpmem=4gb module load intel/2016a module load scripts cd $HOME/mydatadir Not available! #PBS A lt1_project OUTDIR=$VSC_SCRATCH/myresdir mkdir -p $OUTDIR cd $OUTDIR #execute mympirun --stats=3./xhpl.sb.intel2016a export I_MPI_STATS=ipm mpirun np $PBS_NP./xhpl.sb.intel2016a

Migrating PBS scripts #!/bin/bash -l #PBS -N HPL-2.2-T1-SB-64-1n #PBS -lnodes=2:ppn=16 #PBS -lwalltime=24:00:00 #PBS -lvmem=55gb #PBS lnodes=1:ppn=28 #PBS -lwalltime=10:00:00 #PBS lpmem=4gb module load intel/2016a module load scripts cd $HOME/mydatadir #PBS A lt1_project Hybrid MPI + OpenMP OUTDIR=$VSC_SCRATCH/myresdir mkdir -p $OUTDIR cd $OUTDIR #execute Export OMP_NUM_THREADS=16 mympirun universe= 2 --stats=3./xhpl.sb.intel2016a export I_MPI_JOB_RESPECT_PROCESS_PLACEMENT=disable export I_MPI_PIN_DOMAIN=core export KMP_AFFINITY=compact,1 export I_MPI_STATS=ipm mpirun np $PBS_NP./xhpl.sb.intel2016a

What could you expect? 4 Single node performance comparison 3 2 1 0 HPL T1 A TK-HSW T1 B STREAM Triad

What could you expect? time (seconds) 9,000 8,000 7,000 6,000 5,000 4,000 3,000 2,000 1,000 Node performance comparison AMRVAC T1-A TK-HSW T1-B 0 16 32 48 96 192 #nodes

What could you expect? time (seconds) 9,000 8,000 Node performance comparison AMRVAC 7,000 6,000 + 61% 5,000 4,000 3,000 2,000 1,000 + 64% + 64% + 64% T1-A T1-B + 51% 0 16 32 48 96 192 #nodes T1B +75% cores T1A +8% freq System size

What could you expect? 1 Parallel efficiency 2688 AMRVAC 1536 3072 1 T1A T1B 0 32 48 96 192 #nodes

What could you expect? time (seconds) 5,000 4,000 Node performance comparison + 25% AMRVAC 3,000 2,000 1,000 + 24% + 25% 0 #nodes 16 32 48 TK-HSW T1-B T1B +16% cores TK-HSW +5% freq Intel2016a vs intel2015a

Stefan Becuwe Jan Ooghe Ingrid Barcena hpcinfo@kuleuven.be