bwfortreff bwhpc user meeting

Size: px
Start display at page:

Download "bwfortreff bwhpc user meeting"

Transcription

1 bwfortreff bwhpc user meeting bwhpc Competence Center MLS&WISO Universitätsrechenzentrum Heidelberg Rechenzentrum der Universität Mannheim Steinbuch Centre for Computing (SCC) Funding:

2 What is bwfortreff? Participants Users of bwgrid/bwhpc systems Students and scientists interested in HPC Members of all bwhpc partner universities Scope System status bwgrid/bwhpc HPC related lectures and workshops Questions and Discussions User contributions 2

3 bwfortreff Agenda 3 Time Topic 16:15 bwhpc and bwhpc-c5 (H. Kredel, MA) 16:30 bwunicluster (S. Richling, HD) 17:30 bwfilestorage (T. Kienzle, MA) 17:45 Q&A 18:00 End

4 bwhpc and bwhpc-c5 Steinbuch Centre for Computing (SCC) Funding:

5 What is bwhpc/bwhpc-c5? 5 bw = Baden-Würtemberg bwhpc = Strategy for high perfomance computing in BW for 2013 to 2018 (in particular for Tier 3) bwhpc-c5 = Federated user and IT support activities for bwhpc

6 bwhpc - Tier Classification 6

7 bwhpc: Tier 3 ( ) MA/HD: bwforcluster MLS&WISO Mannheim Heidelberg KA: bwunicluster TU: bwforcluster BinAC Karlsruhe Tübingen Ulm Freiburg FR: bwforcluster NME 7 UL: bwforcluster TheoChem

8 bwhpc - covered research areas Economy and social science General purpose, Teaching Molecular life science Mannheim Heideberg Astrophysics Karlsruhe Neurosciences Bioinformatics Tübingen Ulm Freiburg Theoretical chemistry Microsystems engineering Elementary particle physics Independent of location, users of research area X use science cluster X 8

9 Federated (1) bwunicluster: Uni = universal or University Financed by Baden-Württemberg's ministry of science, research and arts and the shareholders: Freiburg, Tübingen, KIT, Heidelberg, Ulm, Hohenheim, Konstanz, Mannheim, Stuttgart Usage: Free of charge General purpose, teaching & education Technical computing (sequential & weak parallel) & parallel computing Access / limitations: For all members of shareholder's university Quota and computation share based on university's share 9

10 Federated (2) bwforcluster: For = Forschung (research) Financed by German Research Society (DFG) and Baden-Württemberg's ministry of science, research and arts Access: All university members in Baden-Württemberg For science communities according to DFG proposal Usage, limitations: Free of charge Access only to bwforcluster matching users field of research Access requires approval of compute proposal 10

11 What is bwhpc-c5? C5 = Coordinated compute cluster competence centers Federated user support and IT services activities for bwhpc For User: Send your support requests to competence centers (CC) CC are not local, but distributed over whole BW CC are community specific For BW: Bridging Science & HPC Bridging HPC tiers Embedding services Science 11 HPC infrastructure bw services bwhpc-c5

12 bwhpc-c5: Location of project partners Uni Heidelberg Uni Mannheim Uni Stuttgart HTF Stuttgart KIT Uni Hohenheim HS Esslingen Uni Tübingen Uni Freiburg Uni Konstanz 12 Uni Ulm

13 bwhpc-c5: Federated science support HPC competence centers Formation of a HPC expert pool (related to field of research and knowledge in parallel software development) Coordination of tiger team activities (high level support teams) Buildup of a best practice repository Coordination of teaching activities Evaluation of courses Generate offline and online material (elearning, MOOC) Cluster Innovations New Technology: Accelerators etc. HPC virtualization, Compute cloud Innovative cluster access, pre- and post-processing tools 13

14 bwhpc-c5: What kind of support? Information seminars, hands-on, HPC specific workshops Documentation + best practices repository Providing/maintaining/developing: simplified access to all bwhpc resources software portfolio cluster independent & unified user environment tools for data management trouble ticket system cluster information system Migration support: code adaptation, e.g. MPI or OpenMP parallelisation code porting (from desktop or old HPC clusters) to tier 2 and 1 14

15 bwhpc-lna: Scientific Steering Committee LNA = Landesnutzerausschuss: Scientific steering of bwhpc and bwdata Website: Tasks: Set bwhpc access formalities Assessment of bwhpc workload Regulation of bwhpc cluster expansion Assignment of science community to science clusters Representation of HPC user interests concerning: resource demands, HPC technologies and software licenses, adjustment of resource quota 15

16 Status bwgrid/bwhpc bwgrid Cluster: Stuttgart, Karlsruhe and Ulm already offline Heidelberg/Mannheim: will close 2014 (running without support) Freiburg, Tübingen, Esslingen: will close 2015 or later Storage: closed in 12/2013 bwhpc (Tier 3) bwcluster: bwunicluster: available since Q1/2014 bwforcluster TheoChem and MLS&WISO: in Q3/2014 bwforcluster BinAC and MNE: in 2015 bwfilestorage: open since 12/

17 bwunicluster First steps (HD/MA) Steinbuch Centre for Computing (SCC) Funding:

18 1. Intro 18

19 Documentation/Literature Intro Intro bwunicluster Wiki bwfortreff Slides Introduction to Unix/Linux commands Bash scripting: (intro) (advanced) Environment modulefiles: MOAB queueing system: 19

20 2. Hardware 20

21 bwunicluster - Hardware Compute Nodes: 512 thin nodes: 16 cores (2 x 8-core Intel Xeon Sandy Bridge), 64 GB RAM, 2 TB local disk 8 fat nodes: 32 cores (4 x 8-core Intel Xeon Sandy Bridge), 1 TB RAM, 7 TB local disk Interconnect: InfiniBand 4X FDR Parallel filesystem Lustre for $HOME (469 TB) and $WORK (938 TB) 21

22 2. Access 22

23 bwunicluster Registration (1) Access Access Access to bwunicluster: Step A: Application for bwunicluster Entitlement Step B: Web Registration for bwunicluster Documentation: Step A is different for each University University of Heidelberg: Apply for bwunicluster entitlement at University of Mannheim: Apply for bwgrid Mannheim/Heidelberg at 23

24 2. 2. Access Access bwunicluster Registration (2) Step B Web Registration: Choose your organization 24

25 2. 2. Access Access bwunicluster Registration (3a) Step B Web Registration: University of Heidelberg: Login with Uni-ID 25

26 2. 2. Access Access bwunicluster Registration (3b) Step B Web Registration: University of Mannheim: Login with RUM account 26

27 2. 2. Access Access bwunicluster Registration (4) Select bwunicluster Service Description 27

28 2. 2. Access Access bwunicluster Registration (5) Read terms of usage and accept 28

29 2. 2. Access Access bwunicluster Registration (6) Read registry information Lookup localuid Deregistration possible 29

30 2. 2. Access Access bwunicluster Registration (7) Set service password (required): 30

31 bwunicluster - Login Access Access Name of login server: bwunicluster.scc.kit.edu Login with localuid + service password via SSH Examples: HD: $ ssh hd_ab123@bwunicluster.scc.kit.edu MA: $ ssh ma_amuster@bwunicluster.scc.kit.edu 31

32 3. Usage 32

33 Software modules Usage Usage Environment modules dynamic modification of the session environment instructions stored in modulefiles Why environment modules? multiple versions of the same software can be installed and be used in a controlled manner, i.e., by load and unload modulefiles How to use modulefiles in general? $ module help More information: 33

34 Modulefiles (1) Usage Usage Display all modulefiles: $ module avail Display all modulefiles in category devel: $ module avail devel Show help of modulefiles: $ module help <modulefile> List all instructions of modulefile: $ module show <modulefile> Display all loaded modules: $ module list Modulefiles are sorted in categories, software name and versions: $ module load <category>/<software_name>/<version> Load a default software: $ module load <category>/<software_name> e.g. Intel compiler: $ module load compiler/intel To remove module: $ module unload <modulefile> 34 $ module remove <modulefile>

35 Modulefiles (2) Usage Usage Conflicts: load different software version in the same session, e.g. Intel: compiler/intel/12.1(376):error:150: Module 'compiler/intel/12.1' conflicts with the currently loaded module(s) 'compiler/intel/13.1' load module with dependencies on other modules: $ module load mpi/openmpi/1.6.5-intel-13.1 Loading module dependency 'compiler/intel/13.1'. compiler/intel/13.1(386):error:150: Module 'compiler/intel/13.1' conflicts with the currently loaded module(s) 'compiler/intel/12.1' Be aware that you can create inconsistencies: e.g. you can remove compiler/intel/13.1 while mpi/openmpi/1.6.5-intel-13.1 is still loaded swap = remove + load e.g.: $ module swap compiler/intel/13.1 compiler/intel/

36 3. 3. Usage Usage File System Characteristics of bwunicluster 36 $HOME, $WORK and workspaces are on the parallel file system Lustre

37 $HOME and $WORK Usage Usage $HOME: Quota: $ lfs quota -u $USER $HOME $WORK: Change to it via: $ cd $WORK Quota: $ lfs quota -u $USER $WORK files older than 28 days will be deleted guaranteed lifetime for files is 7 days 37

38 Workspaces Usage Usage Workspaces: allocated folders with lifetime Howto: $ ws_allocate foo 10 Allocate a workspace named foo for 10 days $ ws_list -a List all your workspaces $ ws_find foo Get absolute path of workspace foo $ ws_extend foo 5 Extend lifetime of your workspace foo by 5 days from now. $ ws_release foo Manually erase your workspace foo Maximum for lifetime: 60 days Number of extensions: 3 times 38

39 Workspaces Usage Usage Workspaces: allocated folders with lifetime Howto: $ ws_allocate foo 10 Allocate a workspace named foo for 10 days $ ws_list -a List all your workspaces $ ws_find foo Get absolute path of workspace foo $ ws_extend foo 5 Extend lifetime of your workspace foo by 5 days from now. $ ws_release foo Manually erase your workspace foo Maximum for lifetime: 60 days Number of extensions: 3 times 39

40 4. Batch System 40

41 4. 4. Batch Batch System System Resource and workload manager Job submission via MOAB commands Use of MOAB commands is planned for all bwforcluster Example job submission: $ msub <resource_options> <job_script> Compute nodes are shared according to resource requests Fairshare based Queues: waiting time depends on: your university's share your job demands your demand history 41

42 4. 4. Batch Batch System System msub options msub options: command line or in job script command line option overwrites script option 42

43 4. 4. Batch Batch System System msub -l resources Resources can combined, but must be separated by comma, e.g.: $ msub -l nodes=1:ppn=1,walltime=00:01:00,pmem=1gb <job_script> Request exclusive usage of nodes with option: -l naccesspolicy=singlejob 43

44 msub -q queues Batch Batch System System If queue not specified: Job is assigned to develop, singlenode and multinode based on requested walltime, nodes and processes. No automatic assignment to: verylong fat 44

45 4. 4. Batch Batch System System Environment variables MOAB adds the following variables to the job's environment MOAB variables can be used to generalize you job scripts, e.g.: ## add suffix to job output file./program > $program_${moab_jobid}.log 45

46 4. 4. Batch Batch System System Check/change status of your jobs After submission msub returns <job-id> $ msub job.sh uc Monitoring commands: 46 $ showq All your active, eligible, blocked, and/or recently completed jobs $ showstart <job-id> Get information about start time of job with <job-id> $ showstart 16@12:00:00 Get information about start time of 16 procs with run time of 12 hours $ checkjob <job-id> Get detailed information of your job explains why your job is pending $ showq -c -u $(whoami) Display completed job $ canceljob <job-id> Cancel the job with <job-id>

47 Bash job script Job Job scripts scripts Define workload manager options via #MSUB Job starts in submit directory Minimal job script: #!/bin/bash #MSUB -l nodes=1:ppn=4 #MSUB -l walltime=00:10:00 # Load required module files module load mpi/openmpi/1.6.5-gnu-4.4 # Jobs starts in submit directory, change if necessary cd $HOME/example # Start program mpiexec simple > simple.out 47

48 bwfilestorage Steinbuch Centre for Computing (SCC) Funding:

49 bwfilestorage Replacement for bwgrid central storage Location KIT Karlsruhe Starting Size 600 TB For users of bwgrid and bwhpc Requirements Entitlement bwfilestorage (granted to users with bwgrid and/or bwunicluster entitlement) Web Registration 49

50 bwfilestorage Limits 100 GB quota for new users 40 TB quota per organization Snapshots: 7 daily, 4 weekly, 2 monthly Backup: for disaster recovery only Temp: 7 days unchanged files (soft), files for 7 days (hard) 100 GB (soft), 200 GB for 7 days (hard) 50

51 bwfilestorage Registration Web Registration: Choose your organisation 51

52 bwfilestorage Registration Web Registration: University of Heidelberg: Login with Uni-ID 52

53 bwfilestorage Registration Web Registration: University of Mannheim: Login with RUM account 53

54 bwfilestorage Registration Web Registration: Service Description 54

55 bwfilestorage Summary Web Registration: Hosts bwfilestorage.lsdf.kit.edu / bwfilestorage-login.lsdf.kit.edu (SSH) Commands: scp -c aes128-cbc testfile hd_jsmith@bwfilestorage.lsdf.kit.edu: scp -c arcfour128 testfile hd_jsmith@bwfilestorage.lsdf.kit.edu: Performance you should expect: Test MB/s time (32GB file) uc1 -> bwfilestorage (aes128) 3-4 min. frbw4 -> bwfilestorage (arcfour128) 8-12 min. For more information please follow the user manual (german). 55

Hands-On Workshop bwunicluster June 29th 2015

Hands-On Workshop bwunicluster June 29th 2015 Hands-On Workshop bwunicluster June 29th 2015 Agenda Welcome Introduction to bwhpc and the bwunicluster Modules - Software Environment Management Job Submission and Monitoring Interactive Work and Remote

More information

bwunicluster Tutorial Access, Data Transfer, Compiling, Modulefiles, Batch Jobs

bwunicluster Tutorial Access, Data Transfer, Compiling, Modulefiles, Batch Jobs bwunicluster Tutorial Access, Data Transfer, Compiling, Modulefiles, Batch Jobs Frauke Bösert, SCC, KIT 1 Material: Slides & Scripts https://indico.scc.kit.edu/indico/event/263/ @bwunicluster/forhlr I/ForHLR

More information

bwunicluster Tutorial Access, Data Transfer, Compiling, Modulefiles, Batch Jobs

bwunicluster Tutorial Access, Data Transfer, Compiling, Modulefiles, Batch Jobs bwunicluster Tutorial Access, Data Transfer, Compiling, Modulefiles, Batch Jobs Frauke Bösert, SCC, KIT 1 Material: Slides & Scripts https://indico.scc.kit.edu/indico/event/263/ @bwunicluster/forhlr I/ForHLR

More information

Access: bwunicluster, bwforcluster, ForHLR

Access: bwunicluster, bwforcluster, ForHLR Access: bwunicluster, bwforcluster, ForHLR Shamna Shamsudeen, SCC, KIT Steinbuch Centre for Computing (SCC) Funding: www.bwhpc-c5.de Outline Introduction Registration Processes bwunicluster bwforcluster

More information

Operating two InfiniBand grid clusters over 28 km distance

Operating two InfiniBand grid clusters over 28 km distance Operating two InfiniBand grid clusters over 28 km distance Sabine Richling, Steffen Hau, Heinz Kredel, Hans-Günther Kruse IT-Center University of Heidelberg, Germany IT-Center University of Mannheim, Germany

More information

bwgrid Treff am URZ Sabine Richling, Heinz Kredel Universitätsrechenzentrum Heidelberg Rechenzentrum Universität Mannheim 29.

bwgrid Treff am URZ Sabine Richling, Heinz Kredel Universitätsrechenzentrum Heidelberg Rechenzentrum Universität Mannheim 29. bwgrid Treff am URZ Sabine Richling, Heinz Kredel Universitätsrechenzentrum Heidelberg Rechenzentrum Universität Mannheim 29. April 2010 Richling/Kredel (URZ/RUM) bwgrid Treff SS 2010 1 / 49 Course Organization

More information

A Long-distance InfiniBand Interconnection between two Clusters in Production Use

A Long-distance InfiniBand Interconnection between two Clusters in Production Use A Long-distance InfiniBand Interconnection between two Clusters in Production Use Sabine Richling, Steffen Hau, Heinz Kredel, Hans-Günther Kruse IT-Center, University of Heidelberg, Germany IT-Center,

More information

Performance Analysis and Prediction for distributed homogeneous Clusters

Performance Analysis and Prediction for distributed homogeneous Clusters Performance Analysis and Prediction for distributed homogeneous Clusters Heinz Kredel, Hans-Günther Kruse, Sabine Richling, Erich Strohmaier IT-Center, University of Mannheim, Germany IT-Center, University

More information

Implementierung eines Dynamic Remote Storage Systems (DRS) für Applikationen mit hohen IO Anforderungen

Implementierung eines Dynamic Remote Storage Systems (DRS) für Applikationen mit hohen IO Anforderungen Implementierung eines Dynamic Remote Storage Systems (DRS) für Applikationen mit hohen IO Anforderungen Jürgen Salk, Christian Mosch, Matthias Neuer, Karsten Siegmund, Volodymyr Kushnarenko, Stefan Kombrink,

More information

Virtualization of the ATLAS Tier-2/3 environment on the HPC cluster NEMO

Virtualization of the ATLAS Tier-2/3 environment on the HPC cluster NEMO Virtualization of the ATLAS Tier-2/3 environment on the HPC cluster NEMO Ulrike Schnoor (CERN) Anton Gamel, Felix Bührer, Benjamin Rottler, Markus Schumacher (University of Freiburg) February 02, 2018

More information

Now SAML takes it all:

Now SAML takes it all: Now SAML takes it all: Federation of non Web-based Services in the State of Baden-Württemberg Sebastian Labitzke Karlsruhe Institute of Technology (KIT) Steinbuch Centre for Computing (SCC) labitzke@kit.edu

More information

Before We Start. Sign in hpcxx account slips Windows Users: Download PuTTY. Google PuTTY First result Save putty.exe to Desktop

Before We Start. Sign in hpcxx account slips Windows Users: Download PuTTY. Google PuTTY First result Save putty.exe to Desktop Before We Start Sign in hpcxx account slips Windows Users: Download PuTTY Google PuTTY First result Save putty.exe to Desktop Research Computing at Virginia Tech Advanced Research Computing Compute Resources

More information

Outline. March 5, 2012 CIRMMT - McGill University 2

Outline. March 5, 2012 CIRMMT - McGill University 2 Outline CLUMEQ, Calcul Quebec and Compute Canada Research Support Objectives and Focal Points CLUMEQ Site at McGill ETS Key Specifications and Status CLUMEQ HPC Support Staff at McGill Getting Started

More information

Tutorial: Compiling, Makefile, Parallel jobs

Tutorial: Compiling, Makefile, Parallel jobs Tutorial: Compiling, Makefile, Parallel jobs Hartmut Häfner Steinbuch Centre for Computing (SCC) Funding: www.bwhpc-c5.de Outline Compiler + Numerical Libraries commands Linking Makefile Intro, Syntax

More information

UAntwerpen, 24 June 2016

UAntwerpen, 24 June 2016 Tier-1b Info Session UAntwerpen, 24 June 2016 VSC HPC environment Tier - 0 47 PF Tier -1 623 TF Tier -2 510 Tf 16,240 CPU cores 128/256 GB memory/node IB EDR interconnect Tier -3 HOPPER/TURING STEVIN THINKING/CEREBRO

More information

bwfdm Communities - a Research Data Management Initiative in the State of Baden-Wuerttemberg

bwfdm Communities - a Research Data Management Initiative in the State of Baden-Wuerttemberg bwfdm Communities - a Research Data Management Initiative in the State of Baden-Wuerttemberg Karlheinz Pappenberger Tromsø, 9th Munin Conference on Scholarly Publishing, 27/11/2014 Overview 1) Federalism

More information

Extraordinary HPC file system solutions at KIT

Extraordinary HPC file system solutions at KIT Extraordinary HPC file system solutions at KIT Roland Laifer STEINBUCH CENTRE FOR COMPUTING - SCC KIT University of the State Roland of Baden-Württemberg Laifer Lustre and tools for ldiskfs investigation

More information

Introduction to Discovery.

Introduction to Discovery. Introduction to Discovery http://discovery.dartmouth.edu The Discovery Cluster 2 Agenda What is a cluster and why use it Overview of computer hardware in cluster Help Available to Discovery Users Logging

More information

The JANUS Computing Environment

The JANUS Computing Environment Research Computing UNIVERSITY OF COLORADO The JANUS Computing Environment Monte Lunacek monte.lunacek@colorado.edu rc-help@colorado.edu What is JANUS? November, 2011 1,368 Compute nodes 16,416 processors

More information

Practice of Software Development: Dynamic scheduler for scientific simulations

Practice of Software Development: Dynamic scheduler for scientific simulations Practice of Software Development: Dynamic scheduler for scientific simulations @ SimLab EA Teilchen STEINBUCH CENTRE FOR COMPUTING - SCC KIT Universität des Landes Baden-Württemberg und nationales Forschungszentrum

More information

Workshop Agenda Feb 25 th 2015

Workshop Agenda Feb 25 th 2015 Workshop Agenda Feb 25 th 2015 Time Presenter Title 09:30 T. König Talk bwhpc Concept & bwhpc-c5 - Federated User Support Activities 09:45 R. Walter Talk bwhpc architecture (bwunicluster, bwforcluster

More information

Introduction to Discovery.

Introduction to Discovery. Introduction to Discovery http://discovery.dartmouth.edu The Discovery Cluster 2 Agenda What is a cluster and why use it Overview of computer hardware in cluster Help Available to Discovery Users Logging

More information

Using file systems at HC3

Using file systems at HC3 Using file systems at HC3 Roland Laifer STEINBUCH CENTRE FOR COMPUTING - SCC KIT University of the State of Baden-Württemberg and National Laboratory of the Helmholtz Association www.kit.edu Basic Lustre

More information

Assistance in Lustre administration

Assistance in Lustre administration Assistance in Lustre administration Roland Laifer STEINBUCH CENTRE FOR COMPUTING - SCC KIT University of the State of Baden-Württemberg and National Laboratory of the Helmholtz Association www.kit.edu

More information

Introduction to High-Performance Computing (HPC)

Introduction to High-Performance Computing (HPC) Introduction to High-Performance Computing (HPC) Computer components CPU : Central Processing Unit cores : individual processing units within a CPU Storage : Disk drives HDD : Hard Disk Drive SSD : Solid

More information

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved. Minnesota Supercomputing Institute Introduction to Job Submission and Scheduling Andrew Gustafson Interacting with MSI Systems Connecting to MSI SSH is the most reliable connection method Linux and Mac

More information

Using the IAC Chimera Cluster

Using the IAC Chimera Cluster Using the IAC Chimera Cluster Ángel de Vicente (Tel.: x5387) SIE de Investigación y Enseñanza Chimera overview Beowulf type cluster Chimera: a monstrous creature made of the parts of multiple animals.

More information

Introduction to High Performance Computing (HPC) Resources at GACRC

Introduction to High Performance Computing (HPC) Resources at GACRC Introduction to High Performance Computing (HPC) Resources at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? Concept

More information

High Performance Computing (HPC) Using zcluster at GACRC

High Performance Computing (HPC) Using zcluster at GACRC High Performance Computing (HPC) Using zcluster at GACRC On-class STAT8060 Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC?

More information

How to run applications on Aziz supercomputer. Mohammad Rafi System Administrator Fujitsu Technology Solutions

How to run applications on Aziz supercomputer. Mohammad Rafi System Administrator Fujitsu Technology Solutions How to run applications on Aziz supercomputer Mohammad Rafi System Administrator Fujitsu Technology Solutions Agenda Overview Compute Nodes Storage Infrastructure Servers Cluster Stack Environment Modules

More information

Introduction to HPC Using zcluster at GACRC

Introduction to HPC Using zcluster at GACRC Introduction to HPC Using zcluster at GACRC On-class PBIO/BINF8350 Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? What

More information

Introduction to HPC Using the New Cluster at GACRC

Introduction to HPC Using the New Cluster at GACRC Introduction to HPC Using the New Cluster at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? What is the new cluster

More information

Introduction to High-Performance Computing (HPC)

Introduction to High-Performance Computing (HPC) Introduction to High-Performance Computing (HPC) Computer components CPU : Central Processing Unit cores : individual processing units within a CPU Storage : Disk drives HDD : Hard Disk Drive SSD : Solid

More information

Introduction to HPC Using zcluster at GACRC

Introduction to HPC Using zcluster at GACRC Introduction to HPC Using zcluster at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? What is HPC Concept? What is

More information

OBTAINING AN ACCOUNT:

OBTAINING AN ACCOUNT: HPC Usage Policies The IIA High Performance Computing (HPC) System is managed by the Computer Management Committee. The User Policies here were developed by the Committee. The user policies below aim to

More information

CENTER FOR HIGH PERFORMANCE COMPUTING. Overview of CHPC. Martin Čuma, PhD. Center for High Performance Computing

CENTER FOR HIGH PERFORMANCE COMPUTING. Overview of CHPC. Martin Čuma, PhD. Center for High Performance Computing Overview of CHPC Martin Čuma, PhD Center for High Performance Computing m.cuma@utah.edu Spring 2014 Overview CHPC Services HPC Clusters Specialized computing resources Access and Security Batch (PBS and

More information

Introduction to High Performance Computing (HPC) Resources at GACRC

Introduction to High Performance Computing (HPC) Resources at GACRC Introduction to High Performance Computing (HPC) Resources at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu 1 Outline GACRC? High Performance

More information

New User Seminar: Part 2 (best practices)

New User Seminar: Part 2 (best practices) New User Seminar: Part 2 (best practices) General Interest Seminar January 2015 Hugh Merz merz@sharcnet.ca Session Outline Submitting Jobs Minimizing queue waits Investigating jobs Checkpointing Efficiency

More information

TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 13 th CALL (T ier-0)

TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 13 th CALL (T ier-0) TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 13 th CALL (T ier-0) Contributing sites and the corresponding computer systems for this call are: BSC, Spain IBM System x idataplex CINECA, Italy Lenovo System

More information

How to Use a Supercomputer - A Boot Camp

How to Use a Supercomputer - A Boot Camp How to Use a Supercomputer - A Boot Camp Shelley Knuth Peter Ruprecht shelley.knuth@colorado.edu peter.ruprecht@colorado.edu www.rc.colorado.edu Outline Today we will discuss: Who Research Computing is

More information

Guillimin HPC Users Meeting. Bart Oldeman

Guillimin HPC Users Meeting. Bart Oldeman June 19, 2014 Bart Oldeman bart.oldeman@mcgill.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Outline Compute Canada News Upcoming Maintenance Downtime in August Storage System

More information

The cluster system. Introduction 22th February Jan Saalbach Scientific Computing Group

The cluster system. Introduction 22th February Jan Saalbach Scientific Computing Group The cluster system Introduction 22th February 2018 Jan Saalbach Scientific Computing Group cluster-help@luis.uni-hannover.de Contents 1 General information about the compute cluster 2 Available computing

More information

Introduction to HPC Using zcluster at GACRC On-Class GENE 4220

Introduction to HPC Using zcluster at GACRC On-Class GENE 4220 Introduction to HPC Using zcluster at GACRC On-Class GENE 4220 Georgia Advanced Computing Resource Center University of Georgia Suchitra Pakala pakala@uga.edu Slides courtesy: Zhoufei Hou 1 OVERVIEW GACRC

More information

Introduction to GALILEO

Introduction to GALILEO Introduction to GALILEO Parallel & production environment Mirko Cestari m.cestari@cineca.it Alessandro Marani a.marani@cineca.it Domenico Guida d.guida@cineca.it Maurizio Cremonesi m.cremonesi@cineca.it

More information

Illinois Proposal Considerations Greg Bauer

Illinois Proposal Considerations Greg Bauer - 2016 Greg Bauer Support model Blue Waters provides traditional Partner Consulting as part of its User Services. Standard service requests for assistance with porting, debugging, allocation issues, and

More information

Introduction to High Performance Computing Using Sapelo2 at GACRC

Introduction to High Performance Computing Using Sapelo2 at GACRC Introduction to High Performance Computing Using Sapelo2 at GACRC Georgia Advanced Computing Resource Center University of Georgia Suchitra Pakala pakala@uga.edu 1 Outline High Performance Computing (HPC)

More information

Introduction to PICO Parallel & Production Enviroment

Introduction to PICO Parallel & Production Enviroment Introduction to PICO Parallel & Production Enviroment Mirko Cestari m.cestari@cineca.it Alessandro Marani a.marani@cineca.it Domenico Guida d.guida@cineca.it Nicola Spallanzani n.spallanzani@cineca.it

More information

Our new HPC-Cluster An overview

Our new HPC-Cluster An overview Our new HPC-Cluster An overview Christian Hagen Universität Regensburg Regensburg, 15.05.2009 Outline 1 Layout 2 Hardware 3 Software 4 Getting an account 5 Compiling 6 Queueing system 7 Parallelization

More information

Introduction to Discovery.

Introduction to Discovery. Introduction to Discovery http://discovery.dartmouth.edu March 2014 The Discovery Cluster 2 Agenda Resource overview Logging on to the cluster with ssh Transferring files to and from the cluster The Environment

More information

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved. Minnesota Supercomputing Institute Introduction to MSI for Physical Scientists Michael Milligan MSI Scientific Computing Consultant Goals Introduction to MSI resources Show you how to access our systems

More information

Grid Computing Competence Center Large Scale Computing Infrastructures (MINF 4526 HS2011)

Grid Computing Competence Center Large Scale Computing Infrastructures (MINF 4526 HS2011) Grid Computing Competence Center Large Scale Computing Infrastructures (MINF 4526 HS2011) Sergio Maffioletti Grid Computing Competence Centre, University of Zurich http://www.gc3.uzh.ch/

More information

LBRN - HPC systems : CCT, LSU

LBRN - HPC systems : CCT, LSU LBRN - HPC systems : CCT, LSU HPC systems @ CCT & LSU LSU HPC Philip SuperMike-II SuperMIC LONI HPC Eric Qeenbee2 CCT HPC Delta LSU HPC Philip 3 Compute 32 Compute Two 2.93 GHz Quad Core Nehalem Xeon 64-bit

More information

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved. Minnesota Supercomputing Institute Introduction to MSI Systems Andrew Gustafson The Machines at MSI Machine Type: Cluster Source: http://en.wikipedia.org/wiki/cluster_%28computing%29 Machine Type: Cluster

More information

Introduction to GACRC Storage Environment. Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer

Introduction to GACRC Storage Environment. Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer Introduction to GACRC Storage Environment Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? Overview of Linux Commands GACRC

More information

Graham vs legacy systems

Graham vs legacy systems New User Seminar Graham vs legacy systems This webinar only covers topics pertaining to graham. For the introduction to our legacy systems (Orca etc.), please check the following recorded webinar: SHARCNet

More information

Knights Landing production environment on MARCONI

Knights Landing production environment on MARCONI Knights Landing production environment on MARCONI Alessandro Marani - a.marani@cineca.it March 20th, 2017 Agenda In this presentation, we will discuss - How we interact with KNL environment on MARCONI

More information

Introduction to the NCAR HPC Systems. 25 May 2018 Consulting Services Group Brian Vanderwende

Introduction to the NCAR HPC Systems. 25 May 2018 Consulting Services Group Brian Vanderwende Introduction to the NCAR HPC Systems 25 May 2018 Consulting Services Group Brian Vanderwende Topics to cover Overview of the NCAR cluster resources Basic tasks in the HPC environment Accessing pre-built

More information

Using Sapelo2 Cluster at the GACRC

Using Sapelo2 Cluster at the GACRC Using Sapelo2 Cluster at the GACRC New User Training Workshop Georgia Advanced Computing Resource Center (GACRC) EITS/University of Georgia Zhuofei Hou zhuofei@uga.edu 1 Outline GACRC Sapelo2 Cluster Diagram

More information

XSEDE New User Training. Ritu Arora November 14, 2014

XSEDE New User Training. Ritu Arora   November 14, 2014 XSEDE New User Training Ritu Arora Email: rauta@tacc.utexas.edu November 14, 2014 1 Objectives Provide a brief overview of XSEDE Computational, Visualization and Storage Resources Extended Collaborative

More information

Introduction to GACRC Storage Environment. Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer

Introduction to GACRC Storage Environment. Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer Introduction to GACRC Storage Environment Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu 1 Outline What is GACRC? Overview of Linux Commands GACRC

More information

Choosing Resources Wisely Plamen Krastev Office: 38 Oxford, Room 117 FAS Research Computing

Choosing Resources Wisely Plamen Krastev Office: 38 Oxford, Room 117 FAS Research Computing Choosing Resources Wisely Plamen Krastev Office: 38 Oxford, Room 117 Email:plamenkrastev@fas.harvard.edu Objectives Inform you of available computational resources Help you choose appropriate computational

More information

Introduction to HPC Using zcluster at GACRC

Introduction to HPC Using zcluster at GACRC Introduction to HPC Using zcluster at GACRC On-class STAT8330 Georgia Advanced Computing Resource Center University of Georgia Suchitra Pakala pakala@uga.edu Slides courtesy: Zhoufei Hou 1 Outline What

More information

Guillimin HPC Users Meeting. Bryan Caron

Guillimin HPC Users Meeting. Bryan Caron July 17, 2014 Bryan Caron bryan.caron@mcgill.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Outline Compute Canada News Upcoming Maintenance Downtime in August Storage System

More information

Grid Computing Activities at KIT

Grid Computing Activities at KIT Grid Computing Activities at KIT Meeting between NCP and KIT, 21.09.2015 Manuel Giffels Karlsruhe Institute of Technology Institute of Experimental Nuclear Physics & Steinbuch Center for Computing Courtesy

More information

HPC at UZH: status and plans

HPC at UZH: status and plans HPC at UZH: status and plans Dec. 4, 2013 This presentation s purpose Meet the sysadmin team. Update on what s coming soon in Schroedinger s HW. Review old and new usage policies. Discussion (later on).

More information

RHRK-Seminar. High Performance Computing with the Cluster Elwetritsch - II. Course instructor : Dr. Josef Schüle, RHRK

RHRK-Seminar. High Performance Computing with the Cluster Elwetritsch - II. Course instructor : Dr. Josef Schüle, RHRK RHRK-Seminar High Performance Computing with the Cluster Elwetritsch - II Course instructor : Dr. Josef Schüle, RHRK Overview Course I Login to cluster SSH RDP / NX Desktop Environments GNOME (default)

More information

Introduction to HPC Using the New Cluster at GACRC

Introduction to HPC Using the New Cluster at GACRC Introduction to HPC Using the New Cluster at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? What is the new cluster

More information

Moab Workload Manager on Cray XT3

Moab Workload Manager on Cray XT3 Moab Workload Manager on Cray XT3 presented by Don Maxwell (ORNL) Michael Jackson (Cluster Resources, Inc.) MOAB Workload Manager on Cray XT3 Why MOAB? Requirements Features Support/Futures 2 Why Moab?

More information

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez Scientific data processing at global scale The LHC Computing Grid Chengdu (China), July 5th 2011 Who I am 2 Computing science background Working in the field of computing for high-energy physics since

More information

Guillimin HPC Users Meeting March 16, 2017

Guillimin HPC Users Meeting March 16, 2017 Guillimin HPC Users Meeting March 16, 2017 guillimin@calculquebec.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Please be kind to your fellow user meeting attendees Limit to

More information

Slurm basics. Summer Kickstart June slide 1 of 49

Slurm basics. Summer Kickstart June slide 1 of 49 Slurm basics Summer Kickstart 2017 June 2017 slide 1 of 49 Triton layers Triton is a powerful but complex machine. You have to consider: Connecting (ssh) Data storage (filesystems and Lustre) Resource

More information

Getting started with the CEES Grid

Getting started with the CEES Grid Getting started with the CEES Grid October, 2013 CEES HPC Manager: Dennis Michael, dennis@stanford.edu, 723-2014, Mitchell Building room 415. Please see our web site at http://cees.stanford.edu. Account

More information

KISTI TACHYON2 SYSTEM Quick User Guide

KISTI TACHYON2 SYSTEM Quick User Guide KISTI TACHYON2 SYSTEM Quick User Guide Ver. 2.4 2017. Feb. SupercomputingCenter 1. TACHYON 2 System Overview Section Specs Model SUN Blade 6275 CPU Intel Xeon X5570 2.93GHz(Nehalem) Nodes 3,200 total Cores

More information

Introduction to HPC Using the New Cluster at GACRC

Introduction to HPC Using the New Cluster at GACRC Introduction to HPC Using the New Cluster at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu 1 Outline What is GACRC? What is the new cluster

More information

Introduction to High-Performance Computing (HPC)

Introduction to High-Performance Computing (HPC) Introduction to High-Performance Computing (HPC) Computer components CPU : Central Processing Unit CPU cores : individual processing units within a Storage : Disk drives HDD : Hard Disk Drive SSD : Solid

More information

Introduction to HPC Resources and Linux

Introduction to HPC Resources and Linux Introduction to HPC Resources and Linux Burak Himmetoglu Enterprise Technology Services & Center for Scientific Computing e-mail: bhimmetoglu@ucsb.edu Paul Weakliem California Nanosystems Institute & Center

More information

CYFRONET SITE REPORT IMPROVING SLURM USABILITY AND MONITORING. M. Pawlik, J. Budzowski, L. Flis, P. Lasoń, M. Magryś

CYFRONET SITE REPORT IMPROVING SLURM USABILITY AND MONITORING. M. Pawlik, J. Budzowski, L. Flis, P. Lasoń, M. Magryś CYFRONET SITE REPORT IMPROVING SLURM USABILITY AND MONITORING M. Pawlik, J. Budzowski, L. Flis, P. Lasoń, M. Magryś Presentation plan 2 Cyfronet introduction System description SLURM modifications Job

More information

Genius Quick Start Guide

Genius Quick Start Guide Genius Quick Start Guide Overview of the system Genius consists of a total of 116 nodes with 2 Skylake Xeon Gold 6140 processors. Each with 18 cores, at least 192GB of memory and 800 GB of local SSD disk.

More information

INTRODUCTION TO THE CLUSTER

INTRODUCTION TO THE CLUSTER INTRODUCTION TO THE CLUSTER WHAT IS A CLUSTER? A computer cluster consists of a group of interconnected servers (nodes) that work together to form a single logical system. COMPUTE NODES GATEWAYS SCHEDULER

More information

Introduction to BioHPC

Introduction to BioHPC Introduction to BioHPC New User Training [web] [email] portal.biohpc.swmed.edu biohpc-help@utsouthwestern.edu 1 Updated for 2015-06-03 Overview Today we re going to cover: What is BioHPC? How do I access

More information

Choosing Resources Wisely. What is Research Computing?

Choosing Resources Wisely. What is Research Computing? Choosing Resources Wisely Scott Yockel, PhD Harvard - Research Computing What is Research Computing? Faculty of Arts and Sciences (FAS) department that handles nonenterprise IT requests from researchers.

More information

The GPU-Cluster. Sandra Wienke Rechen- und Kommunikationszentrum (RZ) Fotos: Christian Iwainsky

The GPU-Cluster. Sandra Wienke Rechen- und Kommunikationszentrum (RZ) Fotos: Christian Iwainsky The GPU-Cluster Sandra Wienke wienke@rz.rwth-aachen.de Fotos: Christian Iwainsky Rechen- und Kommunikationszentrum (RZ) The GPU-Cluster GPU-Cluster: 57 Nvidia Quadro 6000 (29 nodes) innovative computer

More information

Introduction to HPC Using zcluster at GACRC

Introduction to HPC Using zcluster at GACRC Introduction to HPC Using zcluster at GACRC Georgia Advanced Computing Resource Center University of Georgia Suchitra Pakala pakala@uga.edu Slides courtesy: Zhoufei Hou OVERVIEW GACRC High Performance

More information

TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 16 th CALL (T ier-0)

TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 16 th CALL (T ier-0) PRACE 16th Call Technical Guidelines for Applicants V1: published on 26/09/17 TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 16 th CALL (T ier-0) The contributing sites and the corresponding computer systems

More information

Compute Cluster Server Lab 1: Installation of Microsoft Compute Cluster Server 2003

Compute Cluster Server Lab 1: Installation of Microsoft Compute Cluster Server 2003 Compute Cluster Server Lab 1: Installation of Microsoft Compute Cluster Server 2003 Compute Cluster Server Lab 1: Installation of Microsoft Compute Cluster Server 2003... 1 Lab Objective... 1 Overview

More information

UL HPC Monitoring in practice: why, what, how, where to look

UL HPC Monitoring in practice: why, what, how, where to look C. Parisot UL HPC Monitoring in practice: why, what, how, where to look 1 / 22 What is HPC? Best Practices Getting Fast & Efficient UL HPC Monitoring in practice: why, what, how, where to look Clément

More information

Working on the NewRiver Cluster

Working on the NewRiver Cluster Working on the NewRiver Cluster CMDA3634: Computer Science Foundations for Computational Modeling and Data Analytics 22 February 2018 NewRiver is a computing cluster provided by Virginia Tech s Advanced

More information

Challenges in making Lustre systems reliable

Challenges in making Lustre systems reliable Challenges in making Lustre systems reliable Roland Laifer STEINBUCH CENTRE FOR COMPUTING - SCC KIT University of the State Roland of Baden-Württemberg Laifer Challenges and in making Lustre systems reliable

More information

Computing for LHC in Germany

Computing for LHC in Germany 1 Computing for LHC in Germany Günter Quast Universität Karlsruhe (TH) Meeting with RECFA Berlin, October 5th 2007 WLCG Tier1 & Tier2 Additional resources for data analysis - HGF ''Physics at the Terascale''

More information

High Performance Computing Cluster Basic course

High Performance Computing Cluster Basic course High Performance Computing Cluster Basic course Jeremie Vandenplas, Gwen Dawes 30 October 2017 Outline Introduction to the Agrogenomics HPC Connecting with Secure Shell to the HPC Introduction to the Unix/Linux

More information

Lessons learned from Lustre file system operation

Lessons learned from Lustre file system operation Lessons learned from Lustre file system operation Roland Laifer STEINBUCH CENTRE FOR COMPUTING - SCC KIT University of the State of Baden-Württemberg and National Laboratory of the Helmholtz Association

More information

PACE. Instructional Cluster Environment (ICE) Orientation. Research Scientist, PACE

PACE. Instructional Cluster Environment (ICE) Orientation. Research Scientist, PACE PACE Instructional Cluster Environment (ICE) Orientation Mehmet (Memo) Belgin, PhD Research Scientist, PACE www.pace.gatech.edu What is PACE A Partnership for an Advanced Computing Environment Provides

More information

Guillimin HPC Users Meeting March 17, 2016

Guillimin HPC Users Meeting March 17, 2016 Guillimin HPC Users Meeting March 17, 2016 guillimin@calculquebec.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Outline Compute Canada News System Status Software Updates Training

More information

RWTH GPU-Cluster. Sandra Wienke March Rechen- und Kommunikationszentrum (RZ) Fotos: Christian Iwainsky

RWTH GPU-Cluster. Sandra Wienke March Rechen- und Kommunikationszentrum (RZ) Fotos: Christian Iwainsky RWTH GPU-Cluster Fotos: Christian Iwainsky Sandra Wienke wienke@rz.rwth-aachen.de March 2012 Rechen- und Kommunikationszentrum (RZ) The GPU-Cluster GPU-Cluster: 57 Nvidia Quadro 6000 (29 nodes) innovative

More information

High Performance Computing (HPC) Club Training Session. Xinsheng (Shawn) Qin

High Performance Computing (HPC) Club Training Session. Xinsheng (Shawn) Qin High Performance Computing (HPC) Club Training Session Xinsheng (Shawn) Qin Outline HPC Club The Hyak Supercomputer Logging in to Hyak Basic Linux Commands Transferring Files Between Your PC and Hyak Submitting

More information

Introduction to HPCC at MSU

Introduction to HPCC at MSU Introduction to HPCC at MSU Chun-Min Chang Research Consultant Institute for Cyber-Enabled Research Download this presentation: https://wiki.hpcc.msu.edu/display/teac/2016-03-17+introduction+to+hpcc How

More information

Name Department/Research Area Have you used the Linux command line?

Name Department/Research Area Have you used the Linux command line? Please log in with HawkID (IOWA domain) Macs are available at stations as marked To switch between the Windows and the Mac systems, press scroll lock twice 9/27/2018 1 Ben Rogers ITS-Research Services

More information

University at Buffalo Center for Computational Research

University at Buffalo Center for Computational Research University at Buffalo Center for Computational Research The following is a short and long description of CCR Facilities for use in proposals, reports, and presentations. If desired, a letter of support

More information

Computing with the Moore Cluster

Computing with the Moore Cluster Computing with the Moore Cluster Edward Walter An overview of data management and job processing in the Moore compute cluster. Overview Getting access to the cluster Data management Submitting jobs (MPI

More information

Batch Usage on JURECA Introduction to Slurm. May 2016 Chrysovalantis Paschoulas HPS JSC

Batch Usage on JURECA Introduction to Slurm. May 2016 Chrysovalantis Paschoulas HPS JSC Batch Usage on JURECA Introduction to Slurm May 2016 Chrysovalantis Paschoulas HPS group @ JSC Batch System Concepts Resource Manager is the software responsible for managing the resources of a cluster,

More information