Monitoring and Trouble Shooting on BioHPC
|
|
- Alberta Lloyd
- 5 years ago
- Views:
Transcription
1 Monitoring and Trouble Shooting on BioHPC [web] [ ] portal.biohpc.swmed.edu 1 Updated for
2 Why Monitoring & Troubleshooting data code Monitoring jobs running on the cluster Understand how current HPC resource is used Optimize usage to maximum capacities 2
3 Why Monitoring & Troubleshooting Try to understand if the job is: CPU intensive Memory intensive I/O intensive A combination of above Try to figure out: Where are the bottlenecks How to boost the computational efficiency -Completing more tasks during available time window -Run an analysis with larger data set in the same amount of time 3
4 What to Monitor First, start by profiling the application on an interactive node. CPU Usage - lscpu - pstree - top Memory Usage - free - vmstat I/O Usage - iostat Network/Bandwidth - ifstat 4
5 CPU Usage Achieve speedup on HPC? Increased frequencies Increased scalability lscpu: display information about CPU architecture 5
6 CPU Usage: command line tools Job running on the compute node: astrocyte_cli test <workflow> align-bowtie-se.sh bowtie/ samples pstree: display a tree of processes * You may also use top and pstree command to verify if your job is running across multiple nodes 6
7 CPU Usage: command line tools top: display Linux tasks, provides a dynamic real-time view of a running system. 7
8 Memory Usage: The Memory Hierarchy 8
9 Memory Usage: command line tools free: displays the total amount of free and used physical and swap memory in the system, as well as the buffers used by the kernel Mem (RAM): can be used by currently-running process Swap (Virtual Memory): is used when the amount of physical memory (RAM) is full. Constant swapping should be avoided buffers: file system metadata cached: pages with actual contents of files for future faster access, not currently used memory 9
10 Memory Usage: command line tools vmstat: (Virtual Memory Statistics) outputs instantaneous reports about your system's processes, memory, paging, block I/O, interrupts and CPU activity. 10
11 Disk Usage & I/O Parallel Filesystems on BioHPC Advantages: scalability the capability to distribute large files across multiple nodes Issues Inadequate I/O capability can severely degrade overall cluster performance 11
12 Disk Usage & I/O: command line tools iostat: generates reports that can be used to change system configuration to better balance the input/output load between physical disks. %iowait is the percentage of time your processors are waiting on the disk 12
13 Network/Bandwidth Usage Minimizing communication 13
14 Network/Bandwidth Usage: command line tools ifstat: reports the network bandwidth in a batch style mode 14
15 All-in-One tools Too many tools? All-in-on tools - Dstat - Linux Collectl Profile - HPCTools 15
16 Dstat: Versatile resource statistics tool DAG:a versatile replacement for vmstat, iostat, netstat and ifstat. 16
17 Linux Collectl Profiler Information from monitoring an application can aid the user to run it optimally Collectl is a tool which monitors a broad set of subsystems of a server while user application is running on it Helpful to know your application s usage of cpu, memory, disk, etc to determine if system resources are being stressed or over utilized Many subsystems in summary or detail available to monitor, but initial interest to a user running an application. CPU Memory Disk Lustre InfiniBand NFS usage TCP summary 17
18 collectl --showsubsys Shows ALL subsystems that data can be collected for and plotted in Summary plots: b - buddy info (memory fragmentation) c - cpu d - disk f - nfs i - inodes j - interrupts by CPU m - memory n - network s - sockets t - tcp x - interconnect (currently supported: OFED/Infiniband) y - slabs 18
19 collectl --showsubsys Shows all subsystems that data collected can be shown in Detailed plots: C - individual CPUs, including interrupts if -sj or -sj D - individual Disks E - environmental (fan, power, temp) [requires ipmitool] F - nfs data J - interrupts by CPU by interrupt number M - memory numa/node N - individual Networks T - tcp details (lots of data!) X - interconnect ports/rails (Infiniband/Quadrics) Y - slabs/slubs Z - processes L - lustre 19
20 Why Monitoring - Linux Collectl Profiler Getting LUSTRE metrics In your script that you sbatch to run a job, execute collectl running in the background: #!/bin/bash module add collectl/4.1.2 cd /project/biohpcadmin/s mkdir test collectl -sclmx -P -f /project/biohpcadmin/s175049/test &>/dev/null & dd if=/dev/zero of=stripe4 bs=4m count=4096 kill %1 Data is collected for subsystems that are listed in s option Collectl data files are written to user directory test above 20
21 Why Monitoring - Linux Colplot Visualizer View data with Gnuplot either while job is running or after it is finished: % colplot dir /project/biohpcadmin/s175049/test plot cpu,mem,inter,cltdet % colplot showplot shows ALL the different args to plot to display the plots you want May need to refine timeline by specifying specific timeframe to view: % colplot dir /project/biohpcadmin/s175049/test plot \ cpu,mem,inter,cltdet -time 08:20-08:30 21
22 Why Monitoring - Linux Collectl & Colplot Documentation with examples and tutorials: collectl.sourceforge.net/documentation.html colplot.sourceforge.net/documentation.html Collectl and colplot man pages: linux.die.net/man/1/collectl collectl-utils.sourceforge.net/coplot.html 22
23 23 What s next
24 Optimization: Use appropriate compiler options Intel Math Kernel Library: a library of optimized math routines for science, engineering and financial applications. Basic Linear Algebra Subroutines LAPCK Fast Fourier Transform (FFT) Vector Math Library Build in OpenMP multithreading (set OMP_NUM_THREADS>1) Modules with MKL on BioHPC R/ Intel R/3.3.2-gccmkl julia/0.4.6 JAGS/4.2.0 Compile your own MKL using the mkl complier option (detailed options refer to: 24
25 Optimization : Load big data into memory to reduce I/O 8GB RAM 256GB RAM Significantly reduced I/O 25
26 Optimization : Single-Instruction, Multiple-Data Vector Processing Unit Scalar Loop for ( i = 0; i < n; i++) A[i] = A[i] + B[i]; SIMD Loop for ( i = 0; i < n; I += 8) A[I : (i+8)] = A[I : (i+8)] + B[i : (i+8)]; * Each SIMD addition operator acts on 8 numbers at a time Intel AVX data types allow packing of up to 32 elements in a register if bytes are used. The number of elements depends upon the element type: 8 single-precision floating point types or 4 doubleprecision floating point types. Another example is GPU 26
27 Optimization: GNU Parallel If all jobs are independent to each other... A shell tool for executing jobs in parallel using one or more computers. Make best use of CPU resource with balanced job load Predefined the job pool to match the total number of Cores Spawns a new process when one finishes module load parallel keeping the CPUs active and thus saving time 27
28 Optimization: Multithreading If communication between jobs are needed... Shared memory Advantages: user friendly programming fast data sharing between tasks Disadvantage: programmer s responsibility for synchronization construction that ensure correct access of shared memory libs pthread openmp tools phenix bowtie2 28
29 Optimization: Shared Memory concurrent read: Maybe concurrent write: No Modified from Figure 1 in Possible bottleneck: 29
30 Optimization: Message Passing Interface If communication between jobs are needed... e.g.: MPI job across multiple nodes slave node 1 master node slave node 2 slave node 3 30
31 Optimization: Message Passing Interface Possible bottleneck: communication cost unbalanced load Decompose dataset in a smart way to: Minimize the overlaps (proportion to What is the maximum speed-up you could achieve? communication cost) Balance the data between nodes Example: METIS Graph partition tool verview 31
32 Optimization: Multithreading & Message Passing MPI + pthread If you try to run relion job across 2 nodes on 256GB partition, 48*2 = 96 cores No. of MPI jobs No. of threads No. MPI * No. threads Q: Which one has the shortest computation time? 32
33 Demo: Project Gutenberg big data reader Data: books Size: 10 GB Type: plain/text Count the number of occurrences of the words: dog cat boy girl Goal: Complete as fast as possible by reducing bottlenecks and inefficiencies 33
34 Demo: Project Gutenberg big data reader: Solution I (single-processor, many files) file_00.txt file_01.txt file_02.txt LUSTRE LUSTRE LUSTRE Read text into node RAM Read text into node RAM Read text into node RAM CPU_00 count keywords CPU_00 count keywords CPU_00 count keywords 34
35 Demo: Project Gutenberg big data reader: Solution II (multi-processor, partition file set) file_00.txt file_01.txt file_02.txt file_03.txt file_04.txt file_05.txt LUS TRE LUS TRE LUS TRE LUS TRE LUS TRE LUS TRE Read line of text into node RAM Read line of text into node RAM Read line of text into node RAM Read line of text into node RAM Read line of text into node RAM Read line of text into node RAM CPU_00 count keywords CPU_00 count keywords CPU_00 count keywords CPU_01 count keywords CPU_01 count keywords CPU_01 count keywords file_06.txt file_07.txt file_08txt file_09.txt file_10.txt file_11.txt LUS TRE LUS TRE LUS TRE LUS TRE LUS TRE LUS TRE Read line of text into node RAM Read line of text into node RAM Read line of text into node RAM Read line of text into node RAM Read line of text into node RAM Read line of text into node RAM CPU_02 count keywords CPU_02 count keywords CPU_02 count keywords CPU_03 count keywords CPU_03 count keywords CPU_03 count keywords 35
36 Demo: Project Gutenberg big data reader: Solution III (single-processor, one large file, chunked) large_txt.bin (all text from all books in one large file) LUSTRE Distribute file chunks to RAM Distribute memory to CPU_00 in limited chunks chunk_00 chunk_01 chunk_02 CPU_00 count keywords CPU_00 count keywords CPU_00 count keywords
37 Demo: Project Gutenberg big data reader: Solution IV (multiple-processors, one large file, chunked) large_txt.bin (all text from all books in one large file) LUSTRE Load all text into node memory Partition memory to all procesors in chunks multiple chunks multiple chunks multiple chunks multiple chunks CPU_00 count keywords CPU_01 count keywords CPU_02 count keywords CPU_03 count keywords
38 Demo: Project Gutenberg big data reader: Results time python inefficient_reader.py: Solution I time python multithreaded_inefficient_reader.py Solution II time python efficient_reader.py: Solution III time python multithreaded_efficient_reader.py Solution IV 7.2 min 2.0 min 3.5 min 0.7 min 38
The BioHPC Nucleus Cluster & Future Developments
1 The BioHPC Nucleus Cluster & Future Developments Overview Today we ll talk about the BioHPC Nucleus HPC cluster with some technical details for those interested! How is it designed? What hardware does
More informationOur new HPC-Cluster An overview
Our new HPC-Cluster An overview Christian Hagen Universität Regensburg Regensburg, 15.05.2009 Outline 1 Layout 2 Hardware 3 Software 4 Getting an account 5 Compiling 6 Queueing system 7 Parallelization
More informationQuestion No: 1 In capacity planning exercises, which tools assist in listing and identifying processes of interest? (Choose TWO correct answers.
Volume: 129 Questions Question No: 1 In capacity planning exercises, which tools assist in listing and identifying processes of interest? (Choose TWO correct answers.) A. acpid B. lsof C. pstree D. telinit
More informationR on BioHPC. Rstudio, Parallel R and BioconductoR. Updated for
R on BioHPC Rstudio, Parallel R and BioconductoR 1 Updated for 2015-07-15 2 Today we ll be looking at Why R? The dominant statistics environment in academia Large number of packages to do a lot of different
More informationHigh Performance Computing Cluster Advanced course
High Performance Computing Cluster Advanced course Jeremie Vandenplas, Gwen Dawes 9 November 2017 Outline Introduction to the Agrogenomics HPC Submitting and monitoring jobs on the HPC Parallel jobs on
More informationChoosing Resources Wisely Plamen Krastev Office: 38 Oxford, Room 117 FAS Research Computing
Choosing Resources Wisely Plamen Krastev Office: 38 Oxford, Room 117 Email:plamenkrastev@fas.harvard.edu Objectives Inform you of available computational resources Help you choose appropriate computational
More informationIntroduction to High-Performance Computing (HPC)
Introduction to High-Performance Computing (HPC) Computer components CPU : Central Processing Unit cores : individual processing units within a CPU Storage : Disk drives HDD : Hard Disk Drive SSD : Solid
More informationIntroduction to parallel Computing
Introduction to parallel Computing VI-SEEM Training Paschalis Paschalis Korosoglou Korosoglou (pkoro@.gr) (pkoro@.gr) Outline Serial vs Parallel programming Hardware trends Why HPC matters HPC Concepts
More informationIntroduction to High-Performance Computing (HPC)
Introduction to High-Performance Computing (HPC) Computer components CPU : Central Processing Unit cores : individual processing units within a CPU Storage : Disk drives HDD : Hard Disk Drive SSD : Solid
More informationCSCI 402: Computer Architectures. Parallel Processors (2) Fengguang Song Department of Computer & Information Science IUPUI.
CSCI 402: Computer Architectures Parallel Processors (2) Fengguang Song Department of Computer & Information Science IUPUI 6.6 - End Today s Contents GPU Cluster and its network topology The Roofline performance
More informationIntroduction to High Performance Computing and an Statistical Genetics Application on the Janus Supercomputer. Purpose
Introduction to High Performance Computing and an Statistical Genetics Application on the Janus Supercomputer Daniel Yorgov Department of Mathematical & Statistical Sciences, University of Colorado Denver
More informationMunara Tolubaeva Technical Consulting Engineer. 3D XPoint is a trademark of Intel Corporation in the U.S. and/or other countries.
Munara Tolubaeva Technical Consulting Engineer 3D XPoint is a trademark of Intel Corporation in the U.S. and/or other countries. notices and disclaimers Intel technologies features and benefits depend
More informationLS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance
11 th International LS-DYNA Users Conference Computing Technology LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance Gilad Shainer 1, Tong Liu 2, Jeff Layton
More informationTools for Intel Xeon Phi: VTune & Advisor Dr. Fabio Baruffa - LRZ,
Tools for Intel Xeon Phi: VTune & Advisor Dr. Fabio Baruffa - fabio.baruffa@lrz.de LRZ, 27.6.- 29.6.2016 Architecture Overview Intel Xeon Processor Intel Xeon Phi Coprocessor, 1st generation Intel Xeon
More informationData storage on Triton: an introduction
Motivation Data storage on Triton: an introduction How storage is organized in Triton How to optimize IO Do's and Don'ts Exercises slide 1 of 33 Data storage: Motivation Program speed isn t just about
More informationWrite a technical report Present your results Write a workshop/conference paper (optional) Could be a real system, simulation and/or theoretical
Identify a problem Review approaches to the problem Propose a novel approach to the problem Define, design, prototype an implementation to evaluate your approach Could be a real system, simulation and/or
More informationTriton file systems - an introduction. slide 1 of 28
Triton file systems - an introduction slide 1 of 28 File systems Motivation & basic concepts Storage locations Basic flow of IO Do's and Don'ts Exercises slide 2 of 28 File systems: Motivation Case #1:
More informationCompiling applications for the Cray XC
Compiling applications for the Cray XC Compiler Driver Wrappers (1) All applications that will run in parallel on the Cray XC should be compiled with the standard language wrappers. The compiler drivers
More informationPerformance analysis tools: Intel VTuneTM Amplifier and Advisor. Dr. Luigi Iapichino
Performance analysis tools: Intel VTuneTM Amplifier and Advisor Dr. Luigi Iapichino luigi.iapichino@lrz.de Which tool do I use in my project? A roadmap to optimisation After having considered the MPI layer,
More informationHPC Input/Output. I/O and Darshan. Cristian Simarro User Support Section
HPC Input/Output I/O and Darshan Cristian Simarro Cristian.Simarro@ecmwf.int User Support Section Index Lustre summary HPC I/O Different I/O methods Darshan Introduction Goals Considerations How to use
More informationExam Guide COMPSCI 386
FOUNDATIONS We discussed in broad terms the three primary responsibilities of an operating system. Describe each. What is a process? What is a thread? What parts of a process are shared by threads? What
More informationIntroduction to BioHPC
Introduction to BioHPC New User Training [web] [email] portal.biohpc.swmed.edu biohpc-help@utsouthwestern.edu 1 Updated for 2015-06-03 Overview Today we re going to cover: What is BioHPC? How do I access
More informationIntel Knights Landing Hardware
Intel Knights Landing Hardware TACC KNL Tutorial IXPUG Annual Meeting 2016 PRESENTED BY: John Cazes Lars Koesterke 1 Intel s Xeon Phi Architecture Leverages x86 architecture Simpler x86 cores, higher compute
More informationGraham vs legacy systems
New User Seminar Graham vs legacy systems This webinar only covers topics pertaining to graham. For the introduction to our legacy systems (Orca etc.), please check the following recorded webinar: SHARCNet
More informationIntel Enterprise Edition Lustre (IEEL-2.3) [DNE-1 enabled] on Dell MD Storage
Intel Enterprise Edition Lustre (IEEL-2.3) [DNE-1 enabled] on Dell MD Storage Evaluation of Lustre File System software enhancements for improved Metadata performance Wojciech Turek, Paul Calleja,John
More informationHigh Performance Computing. Introduction to Parallel Computing
High Performance Computing Introduction to Parallel Computing Acknowledgements Content of the following presentation is borrowed from The Lawrence Livermore National Laboratory https://hpc.llnl.gov/training/tutorials
More informationProgramming Models for Multi- Threading. Brian Marshall, Advanced Research Computing
Programming Models for Multi- Threading Brian Marshall, Advanced Research Computing Why Do Parallel Computing? Limits of single CPU computing performance available memory I/O rates Parallel computing allows
More informationBright Cluster Manager Advanced HPC cluster management made easy. Martijn de Vries CTO Bright Computing
Bright Cluster Manager Advanced HPC cluster management made easy Martijn de Vries CTO Bright Computing About Bright Computing Bright Computing 1. Develops and supports Bright Cluster Manager for HPC systems
More informationApproaches to Parallel Computing
Approaches to Parallel Computing K. Cooper 1 1 Department of Mathematics Washington State University 2019 Paradigms Concept Many hands make light work... Set several processors to work on separate aspects
More informationSlurm and Abel job scripts. Katerina Michalickova The Research Computing Services Group SUF/USIT November 13, 2013
Slurm and Abel job scripts Katerina Michalickova The Research Computing Services Group SUF/USIT November 13, 2013 Abel in numbers Nodes - 600+ Cores - 10000+ (1 node->2 processors->16 cores) Total memory
More informationNew User Seminar: Part 2 (best practices)
New User Seminar: Part 2 (best practices) General Interest Seminar January 2015 Hugh Merz merz@sharcnet.ca Session Outline Submitting Jobs Minimizing queue waits Investigating jobs Checkpointing Efficiency
More informationIntroduction to GALILEO
Introduction to GALILEO Parallel & production environment Mirko Cestari m.cestari@cineca.it Alessandro Marani a.marani@cineca.it Domenico Guida d.guida@cineca.it Maurizio Cremonesi m.cremonesi@cineca.it
More informationPerformance Tools for Technical Computing
Christian Terboven terboven@rz.rwth-aachen.de Center for Computing and Communication RWTH Aachen University Intel Software Conference 2010 April 13th, Barcelona, Spain Agenda o Motivation and Methodology
More informationThe Stampede is Coming: A New Petascale Resource for the Open Science Community
The Stampede is Coming: A New Petascale Resource for the Open Science Community Jay Boisseau Texas Advanced Computing Center boisseau@tacc.utexas.edu Stampede: Solicitation US National Science Foundation
More informationRunning in parallel. Total number of cores available after hyper threading (virtual cores)
First at all, to know how many processors/cores you have available in your computer, type in the terminal: $> lscpu The output for this particular workstation is the following: Architecture: x86_64 CPU
More informationGPUs and Emerging Architectures
GPUs and Emerging Architectures Mike Giles mike.giles@maths.ox.ac.uk Mathematical Institute, Oxford University e-infrastructure South Consortium Oxford e-research Centre Emerging Architectures p. 1 CPUs
More informationSlurm and Abel job scripts. Katerina Michalickova The Research Computing Services Group SUF/USIT October 23, 2012
Slurm and Abel job scripts Katerina Michalickova The Research Computing Services Group SUF/USIT October 23, 2012 Abel in numbers Nodes - 600+ Cores - 10000+ (1 node->2 processors->16 cores) Total memory
More informationBig Data Analytics Performance for Large Out-Of- Core Matrix Solvers on Advanced Hybrid Architectures
Procedia Computer Science Volume 51, 2015, Pages 2774 2778 ICCS 2015 International Conference On Computational Science Big Data Analytics Performance for Large Out-Of- Core Matrix Solvers on Advanced Hybrid
More informationScheduling FFT Computation on SMP and Multicore Systems Ayaz Ali, Lennart Johnsson & Jaspal Subhlok
Scheduling FFT Computation on SMP and Multicore Systems Ayaz Ali, Lennart Johnsson & Jaspal Subhlok Texas Learning and Computation Center Department of Computer Science University of Houston Outline Motivation
More informationSherlock for IBIIS. William Law Stanford Research Computing
Sherlock for IBIIS William Law Stanford Research Computing Overview How we can help System overview Tech specs Signing on Batch submission Software environment Interactive jobs Next steps We are here to
More informationPerformance Analysis of Memory Transfers and GEMM Subroutines on NVIDIA TESLA GPU Cluster
Performance Analysis of Memory Transfers and GEMM Subroutines on NVIDIA TESLA GPU Cluster Veerendra Allada, Troy Benjegerdes Electrical and Computer Engineering, Ames Laboratory Iowa State University &
More informationMemory Management. q Basic memory management q Swapping q Kernel memory allocation q Next Time: Virtual memory
Memory Management q Basic memory management q Swapping q Kernel memory allocation q Next Time: Virtual memory Memory management Ideal memory for a programmer large, fast, nonvolatile and cheap not an option
More informationChapter 11: Implementing File Systems. Operating System Concepts 8 th Edition,
Chapter 11: Implementing File Systems, Silberschatz, Galvin and Gagne 2009 Chapter 11: Implementing File Systems File-System Structure File-System Implementation Directory Implementation Allocation Methods
More informationParallel Computing Ideas
Parallel Computing Ideas K. 1 1 Department of Mathematics 2018 Why When to go for speed Historically: Production code Code takes a long time to run Code runs many times Code is not end in itself 2010:
More informationHPC Middle East. KFUPM HPC Workshop April Mohamed Mekias HPC Solutions Consultant. Agenda
KFUPM HPC Workshop April 29-30 2015 Mohamed Mekias HPC Solutions Consultant Agenda 1 Agenda-Day 1 HPC Overview What is a cluster? Shared v.s. Distributed Parallel v.s. Massively Parallel Interconnects
More informationhigh performance medical reconstruction using stream programming paradigms
high performance medical reconstruction using stream programming paradigms This Paper describes the implementation and results of CT reconstruction using Filtered Back Projection on various stream programming
More informationParallel Applications on Distributed Memory Systems. Le Yan HPC User LSU
Parallel Applications on Distributed Memory Systems Le Yan HPC User Services @ LSU Outline Distributed memory systems Message Passing Interface (MPI) Parallel applications 6/3/2015 LONI Parallel Programming
More informationIntroduction to parallel computers and parallel programming. Introduction to parallel computersand parallel programming p. 1
Introduction to parallel computers and parallel programming Introduction to parallel computersand parallel programming p. 1 Content A quick overview of morden parallel hardware Parallelism within a chip
More informationUAntwerpen, 24 June 2016
Tier-1b Info Session UAntwerpen, 24 June 2016 VSC HPC environment Tier - 0 47 PF Tier -1 623 TF Tier -2 510 Tf 16,240 CPU cores 128/256 GB memory/node IB EDR interconnect Tier -3 HOPPER/TURING STEVIN THINKING/CEREBRO
More informationOpenMP Exercises. These exercises will introduce you to using OpenMP for parallel programming. There are four exercises:
OpenMP Exercises These exercises will introduce you to using OpenMP for parallel programming. There are four exercises: 1. OMP Hello World 2. Worksharing Loop 3. OMP Functions 4. Hand-coding vs. MKL To
More informationGPU ACCELERATED DATABASE MANAGEMENT SYSTEMS
CIS 601 - Graduate Seminar Presentation 1 GPU ACCELERATED DATABASE MANAGEMENT SYSTEMS PRESENTED BY HARINATH AMASA CSU ID: 2697292 What we will talk about.. Current problems GPU What are GPU Databases GPU
More informationComputing on GPUs. Prof. Dr. Uli Göhner. DYNAmore GmbH. Stuttgart, Germany
Computing on GPUs Prof. Dr. Uli Göhner DYNAmore GmbH Stuttgart, Germany Summary: The increasing power of GPUs has led to the intent to transfer computing load from CPUs to GPUs. A first example has been
More informationFeedback on BeeGFS. A Parallel File System for High Performance Computing
Feedback on BeeGFS A Parallel File System for High Performance Computing Philippe Dos Santos et Georges Raseev FR 2764 Fédération de Recherche LUmière MATière December 13 2016 LOGO CNRS LOGO IO December
More informationThe JANUS Computing Environment
Research Computing UNIVERSITY OF COLORADO The JANUS Computing Environment Monte Lunacek monte.lunacek@colorado.edu rc-help@colorado.edu What is JANUS? November, 2011 1,368 Compute nodes 16,416 processors
More informationOperating Systems Design Exam 2 Review: Spring 2012
Operating Systems Design Exam 2 Review: Spring 2012 Paul Krzyzanowski pxk@cs.rutgers.edu 1 Question 1 Under what conditions will you reach a point of diminishing returns where adding more memory may improve
More informationAdvanced Parallel Programming I
Advanced Parallel Programming I Alexander Leutgeb, RISC Software GmbH RISC Software GmbH Johannes Kepler University Linz 2016 22.09.2016 1 Levels of Parallelism RISC Software GmbH Johannes Kepler University
More informationSlurm basics. Summer Kickstart June slide 1 of 49
Slurm basics Summer Kickstart 2017 June 2017 slide 1 of 49 Triton layers Triton is a powerful but complex machine. You have to consider: Connecting (ssh) Data storage (filesystems and Lustre) Resource
More informationNEMO Performance Benchmark and Profiling. May 2011
NEMO Performance Benchmark and Profiling May 2011 Note The following research was performed under the HPC Advisory Council HPC works working group activities Participating vendors: HP, Intel, Mellanox
More informationLecture Topics. Announcements. Today: Advanced Scheduling (Stallings, chapter ) Next: Deadlock (Stallings, chapter
Lecture Topics Today: Advanced Scheduling (Stallings, chapter 10.1-10.4) Next: Deadlock (Stallings, chapter 6.1-6.6) 1 Announcements Exam #2 returned today Self-Study Exercise #10 Project #8 (due 11/16)
More informationSybase Adaptive Server Enterprise on Linux
Sybase Adaptive Server Enterprise on Linux A Technical White Paper May 2003 Information Anywhere EXECUTIVE OVERVIEW ARCHITECTURE OF ASE Dynamic Performance Security Mission-Critical Computing Advanced
More informationHigh Performance Computing. Leopold Grinberg T. J. Watson IBM Research Center, USA
High Performance Computing Leopold Grinberg T. J. Watson IBM Research Center, USA High Performance Computing Why do we need HPC? High Performance Computing Amazon can ship products within hours would it
More informationPatternFinder is a tool that finds non-overlapping or overlapping patterns in any input sequence.
PatternFinder is a tool that finds non-overlapping or overlapping patterns in any input sequence. Pattern Finder Input Parameters: USAGE: PatternDetective.exe [ -help /? -f [filename] -min -max [minimum
More informationMARUTHI SCHOOL OF BANKING (MSB)
MARUTHI SCHOOL OF BANKING (MSB) SO IT - OPERATING SYSTEM(2017) 1. is mainly responsible for allocating the resources as per process requirement? 1.RAM 2.Compiler 3.Operating Systems 4.Software 2.Which
More informationScaling Out Python* To HPC and Big Data
Scaling Out Python* To HPC and Big Data Sergey Maidanov Software Engineering Manager for Intel Distribution for Python* What Problems We Solve: Scalable Performance Make Python usable beyond prototyping
More informationSHARCNET Workshop on Parallel Computing. Hugh Merz Laurentian University May 2008
SHARCNET Workshop on Parallel Computing Hugh Merz Laurentian University May 2008 What is Parallel Computing? A computational method that utilizes multiple processing elements to solve a problem in tandem
More informationPerformance Analysis of Parallel Scientific Applications In Eclipse
Performance Analysis of Parallel Scientific Applications In Eclipse EclipseCon 2015 Wyatt Spear, University of Oregon wspear@cs.uoregon.edu Supercomputing Big systems solving big problems Performance gains
More informationIntroduction to PICO Parallel & Production Enviroment
Introduction to PICO Parallel & Production Enviroment Mirko Cestari m.cestari@cineca.it Alessandro Marani a.marani@cineca.it Domenico Guida d.guida@cineca.it Nicola Spallanzani n.spallanzani@cineca.it
More informationSCALABLE HYBRID PROTOTYPE
SCALABLE HYBRID PROTOTYPE Scalable Hybrid Prototype Part of the PRACE Technology Evaluation Objectives Enabling key applications on new architectures Familiarizing users and providing a research platform
More informationInformatica Developer Tips for Troubleshooting Common Issues PowerCenter 8 Standard Edition. Eugene Gonzalez Support Enablement Manager, Informatica
Informatica Developer Tips for Troubleshooting Common Issues PowerCenter 8 Standard Edition Eugene Gonzalez Support Enablement Manager, Informatica 1 Agenda Troubleshooting PowerCenter issues require a
More informationGenius Quick Start Guide
Genius Quick Start Guide Overview of the system Genius consists of a total of 116 nodes with 2 Skylake Xeon Gold 6140 processors. Each with 18 cores, at least 192GB of memory and 800 GB of local SSD disk.
More informationIllinois Proposal Considerations Greg Bauer
- 2016 Greg Bauer Support model Blue Waters provides traditional Partner Consulting as part of its User Services. Standard service requests for assistance with porting, debugging, allocation issues, and
More informationOvercoming the Memory System Challenge in Dataflow Processing. Darren Jones, Wave Computing Drew Wingard, Sonics
Overcoming the Memory System Challenge in Dataflow Processing Darren Jones, Wave Computing Drew Wingard, Sonics Current Technology Limits Deep Learning Performance Deep Learning Dataflow Graph Existing
More informationProgramming Techniques for Supercomputers. HPC RRZE University Erlangen-Nürnberg Sommersemester 2018
Programming Techniques for Supercomputers HPC Services @ RRZE University Erlangen-Nürnberg Sommersemester 2018 Outline Login to RRZE s Emmy cluster Basic environment Some guidelines First Assignment 2
More informationHPC Architectures. Types of resource currently in use
HPC Architectures Types of resource currently in use Reusing this material This work is licensed under a Creative Commons Attribution- NonCommercial-ShareAlike 4.0 International License. http://creativecommons.org/licenses/by-nc-sa/4.0/deed.en_us
More informationIntel Performance Libraries
Intel Performance Libraries Powerful Mathematical Library Intel Math Kernel Library (Intel MKL) Energy Science & Research Engineering Design Financial Analytics Signal Processing Digital Content Creation
More informationOverview of High Performance Input/Output on LRZ HPC systems. Christoph Biardzki Richard Patra Reinhold Bader
Overview of High Performance Input/Output on LRZ HPC systems Christoph Biardzki Richard Patra Reinhold Bader Agenda Choosing the right file system Storage subsystems at LRZ Introduction to parallel file
More informationPerformance Analysis of BLAS Libraries in SuperLU_DIST for SuperLU_MCDT (Multi Core Distributed) Development
Available online at www.prace-ri.eu Partnership for Advanced Computing in Europe Performance Analysis of BLAS Libraries in SuperLU_DIST for SuperLU_MCDT (Multi Core Distributed) Development M. Serdar Celebi
More informationRHRK-Seminar. High Performance Computing with the Cluster Elwetritsch - II. Course instructor : Dr. Josef Schüle, RHRK
RHRK-Seminar High Performance Computing with the Cluster Elwetritsch - II Course instructor : Dr. Josef Schüle, RHRK Overview Course I Login to cluster SSH RDP / NX Desktop Environments GNOME (default)
More informationIntroduction to Parallel Programming
Introduction to Parallel Programming January 14, 2015 www.cac.cornell.edu What is Parallel Programming? Theoretically a very simple concept Use more than one processor to complete a task Operationally
More informationCode optimization. Geert Jan Bex
Code optimization Geert Jan Bex (geertjan.bex@uhasselt.be) License: this presentation is released under the Creative Commons, see http://creativecommons.org/publicdomain/zero/1.0/ 1 CPU 2 Vectorization
More informationMaximizing Memory Performance for ANSYS Simulations
Maximizing Memory Performance for ANSYS Simulations By Alex Pickard, 2018-11-19 Memory or RAM is an important aspect of configuring computers for high performance computing (HPC) simulation work. The performance
More informationNUMA-aware OpenMP Programming
NUMA-aware OpenMP Programming Dirk Schmidl IT Center, RWTH Aachen University Member of the HPC Group schmidl@itc.rwth-aachen.de Christian Terboven IT Center, RWTH Aachen University Deputy lead of the HPC
More informationUsing MPI One-sided Communication to Accelerate Bioinformatics Applications
Using MPI One-sided Communication to Accelerate Bioinformatics Applications Hao Wang (hwang121@vt.edu) Department of Computer Science, Virginia Tech Next-Generation Sequencing (NGS) Data Analysis NGS Data
More informationAN INTRODUCTION TO UNIX
AN INTRODUCTION TO UNIX Paul Johnson School of Mathematics September 18, 2011 OUTLINE 1 INTRODUTION Unix Common Tasks 2 THE UNIX FILESYSTEM Moving around Copying, deleting File Permissions 3 SUMMARY OUTLINE
More informationStaying Out of the Swamp
Staying Out of the Swamp Perforce User Conference 2001 Richard E. Baum Introduction Perforce runs well when given proper resources. CPU requirements are quite small. A server s I/O bandwidth is generally
More informationSackler Course BMSC-GA 4448 High Performance Computing in Biomedical Informatics. Class 2: Friday February 14 th, :30PM 5:30PM AGENDA
Sackler Course BMSC-GA 4448 High Performance Computing in Biomedical Informatics Class 2: Friday February 14 th, 2014 2:30PM 5:30PM AGENDA Recap 1 st class & Homework discussion. Fundamentals of Parallel
More informationThe MOSIX Scalable Cluster Computing for Linux. mosix.org
The MOSIX Scalable Cluster Computing for Linux Prof. Amnon Barak Computer Science Hebrew University http://www. mosix.org 1 Presentation overview Part I : Why computing clusters (slide 3-7) Part II : What
More informationCerebro Quick Start Guide
Cerebro Quick Start Guide Overview of the system Cerebro consists of a total of 64 Ivy Bridge processors E5-4650 v2 with 10 cores each, 14 TB of memory and 24 TB of local disk. Table 1 shows the hardware
More informationA Study of High Performance Computing and the Cray SV1 Supercomputer. Michael Sullivan TJHSST Class of 2004
A Study of High Performance Computing and the Cray SV1 Supercomputer Michael Sullivan TJHSST Class of 2004 June 2004 0.1 Introduction A supercomputer is a device for turning compute-bound problems into
More informationOperating Systems Design Exam 2 Review: Fall 2010
Operating Systems Design Exam 2 Review: Fall 2010 Paul Krzyzanowski pxk@cs.rutgers.edu 1 1. Why could adding more memory to a computer make it run faster? If processes don t have their working sets in
More informationThe Use of Cloud Computing Resources in an HPC Environment
The Use of Cloud Computing Resources in an HPC Environment Bill, Labate, UCLA Office of Information Technology Prakashan Korambath, UCLA Institute for Digital Research & Education Cloud computing becomes
More informationBeeGFS Solid, fast and made in Europe
David Ramírez Alvarez HPC INTEGRATOR MANAGER WWW.SIE.ES dramirez@sie.es ADMINTECH 2016 BeeGFS Solid, fast and made in Europe www.beegfs.com Thanks to Sven for info!!!, CEO, ThinkParQ What is BeeGFS? BeeGFS
More informationMaximizing NFS Scalability
Maximizing NFS Scalability on Dell Servers and Storage in High-Performance Computing Environments Popular because of its maturity and ease of use, the Network File System (NFS) can be used in high-performance
More informationMIGRATING TO THE SHARED COMPUTING CLUSTER (SCC) SCV Staff Boston University Scientific Computing and Visualization
MIGRATING TO THE SHARED COMPUTING CLUSTER (SCC) SCV Staff Boston University Scientific Computing and Visualization 2 Glenn Bresnahan Director, SCV MGHPCC Buy-in Program Kadin Tseng HPC Programmer/Consultant
More informationECE519 Advanced Operating Systems
IT 540 Operating Systems ECE519 Advanced Operating Systems Prof. Dr. Hasan Hüseyin BALIK (10 th Week) (Advanced) Operating Systems 10. Multiprocessor, Multicore and Real-Time Scheduling 10. Outline Multiprocessor
More informationWorkpackage 5: High Performance Mathematical Computing
Clément Pernet: Workpackage 5 1 Brussels, April 26, 2017 Workpackage 5: High Performance Mathematical Computing Clément Pernet First OpenDreamKit Project review Brussels, April 26, 2017 Clément Pernet:
More informationParallel Processing. Majid AlMeshari John W. Conklin. Science Advisory Committee Meeting September 3, 2010 Stanford University
Parallel Processing Majid AlMeshari John W. Conklin 1 Outline Challenge Requirements Resources Approach Status Tools for Processing 2 Challenge A computationally intensive algorithm is applied on a huge
More informationCSC 2405: Computer Systems II
CSC 2405: Computer Systems II Dr. Mirela Damian http://www.csc.villanova.edu/~mdamian/csc2405/ Spring 2016 Course Goals: Look under the hood Help you learn what happens under the hood of computer systems
More informationChapter 1: Introduction
Chapter 1: Introduction Silberschatz, Galvin and Gagne 2009 Chapter 1: Introduction What Operating Systems Do Computer-System Organization Computer-System Architecture Operating-System Structure Operating-System
More information6.1 Multiprocessor Computing Environment
6 Parallel Computing 6.1 Multiprocessor Computing Environment The high-performance computing environment used in this book for optimization of very large building structures is the Origin 2000 multiprocessor,
More information