Cornell Theory Center 1
|
|
- Amanda Robertson
- 5 years ago
- Views:
Transcription
1 Cornell Theory Center Cornell Theory Center (CTC) is a high-performance computing and interdisciplinary research center at Cornell University. Scientific and engineering research projects supported by CTC represent a variety of disciplines, including biology, behavioral and social sciences, computer science, engineering, geosciences, mathematics, physical sciences, and business. CTC is unique among HPC centers in that it is a showcase for Windows-based high-performance computing. In 1999, we successfully moved our research community from a traditional, proprietary supercomputer running UNIX to an industry-standard cluster complex. Microsoft and Dell are major supporters of CTC, which is directed by computer scientist Thomas F. Coleman. 3/12/03 Cornell Theory Center 1 Case Studies Plant breeders and molecular biologists generating detailed information on the genetic makeup of crops e.g. rice and wheat Management model for manufacturing that helps managers find best balance between capacity, inventory, and subcontracting New stochastic path approach for modeling protein dynamics in more meaningful time scales New social science research made possible by developing new statistical methods for analyzing census data Simulation of crack growth and strength evaluation of airplane components 3/12/03 Cornell Theory Center 2 Compute Power Over 700 Processors Giganet Switch (100 GB/sec) interconnect 100 Mb switched Ethernet 1-4 GB Ram/Node, GB Disk/Node Serial, Dual, and Quad processors Dell PowerEdge Servers Pentium processors, 600 Mhz 2.4Ghz 256 KB - 2 MB Cache/Processor (SMP) Tivoli Backup System Microsoft Distributed File System Fileservers, 5 Terabytes Storage 11/ processor cluster in top gigaflops over Ghz Linpack benchmark 3/12/03 Cornell Theory Center 3 Cornell Theory Center 1
2 Visualization Consultation on how to visualize data Creation of animations, videos, publication graphics Independent study opportunities for undergraduate students Training in the use of OpenDX Video production suite Three-wall Windows CAVE: Virtual reality environment (show VRML files) Write your own with OpenGL or OpenInventor Free by appointment with Computer Lab of Visualization Workstations /12/03 Cornell Theory Center 4 Software See online list for software available on the login nodes, various types of compute nodes and the Visualization Collaboratory. This list is updated as new packages are added. Software is added by researcher's request if feasible. C/C++ compilers: Microsoft Visual C++, Intel C/C++, Gnu gcc/g++ Fortran compilers: Intel Fortran, PGF77, Compaq Visual Fortran Libraries: MPI/PRO parallel programming, Numerical Libraries Applications software: Computational chemistry, Data analysis, Finite element modeling, Fluid dynamics, Visualization tools, /12/03 Cornell Theory Center 5 Allocations Research allocation: Submit proposals at any time. Undergraduate or graduate students must have their requests submitted by a faculty advisor. Non-renewable six-month Exploratory: Allocations of 100 hours. Simplified application. Charging is calculated using wall-clock hours; i.e. you will be charged for the time nodes are reserved for your use, not for the time you are actually using them. Wall-clock hours on all of the node types are charged the same flat wall-clock rate, regardless of number of processors. Parallelize! Charges for compute nodes only, not for login nodes or the collaboratory machines /12/03 Cornell Theory Center 6 Cornell Theory Center 2
3 Access Connect to a login node (ctclogin1, ctclogin2, ) to compile and submit jobs. The login nodes are SSH servers, download an SSH client. Login.bat is run automatically. For full windows desktop functionality from your machine, use a terminal server, Microsoft Terminal Server Client for Windows, Citrix ICA for non-windows. For command line access, use telnet: ssh or telnet ctclogin1.tc.cornell.edu 3/12/03 Cornell Theory Center 7 Development Environment Programs can be compiled on your Windows machine, in the Collaboratory in Rhodes Hall, or from one of the login nodes reached via Terminal Server or telnet. Work from the command line or in a GUI environment. GUI Development environment: Start Programs Microsoft Visual Studio.NET Microsoft Visual Studio.NET Command line: Start Run cmd (bring up a command line window) setup_intelc32 (set environment variables using setup script **) cl hello.c (compile serial) mpicc hello.c (compile MPI) ** To see the full list of setup scripts, issue: call "H:\CTC Tools\setup_list.bat 3/12/03 Cornell Theory Center 8 Compute Environment Batch Scheduler Login ctclogin1 ctclogin2 CCS Compute ctclogin H: Fileserver 3/12/03 Cornell Theory Center 9 Cornell Theory Center 3
4 Serial Batch Job Script 1. Prepare your executable 2. Write a batch script file 3. From one of the login nodes, submit the script to the scheduler 4. Check the status of the queue, your job, etc The remainder of the steps can be done by hand or script: 5. When your job has started: copy files to the local drive run the program 6. When your job has ended: Copy any output files back to H: clean up the drive end the job Login ctclogin1 ctclogin2 CCS Compute ctclogin H: 3/12/03 Cornell Theory Center 10 Parallel Batch Job Script The main batch script runs on the Master node Start the script with the four required REM CCS keyword commands Create the machines file on of the master node Use mpirun to run setup.bat on all nodes listed in the machines file Login Node Create a working subfolder on the compute node drive Copy the executable (and data files, if necessary) to each node Copy any files needed only on the Master node Move the machines file from to \your_userid H: Use mpirun to run the MPI program executable on all nodes listed in the machines file Your_program.exe runs on all the nodes in the machines file Use mpirun to run cleanup.bat on all nodes listed in the machines file Copy the output files from each node to H: Cleanup the working subdirectory on each node Release the nodes Master Node Worker Node Worker Node Worker Node 3/12/03 Cornell Theory Center 11 Parallel Job on a Single Node (page 1 of 2) REM CCS account = your_userid REM CCS type = batch REM CCS nodes = 1 REM CCS minutes = 10 REM CCS requirements = 1@development mkdir %USERNAME% cd %USERNAME% del /Q /S \%USERNAME%\*.* Call machinemaker (in CTC Tools, therefore already in your path) to produce a machines list, listing the node(s) your job has been allocated. Note: if you have more than one job running, you can assure that you'll create the machines file for the right jobid by issuing machinemaker jobid call machinemaker 3/12/03 Cornell Theory Center 12 Cornell Theory Center 4
5 Parallel Job on a Single Node (page 2 of 2) REM Copy the executable and data files to the node copy H:\users\%USERNAME%\lab\wavesend.exe \%USERNAME% copy H:\users\%USERNAME%\lab\wave.in \%USERNAME% Run the MPI program with mpirun. Set -np to the number of tasks. Set -wd to the working directory in the case where your files are not in your current directory. mpirun -wd \%USERNAME% -np 2 wavesend.exe 1>wave.out 2>wave.err REM Copy any output created by the master task back to the H drive. copy \%USERNAME%\wave.* H:\users\%USERNAME%\lab REM Cleanup the working subdirectory on each node del /Q /S \%USERNAME%\*.* REM Release the nodes ccrelease 3/12/03 Cornell Theory Center 13 Parallel Job on Multiple (page 1 of 4) REM CCS account = your_userid REM CCS type = batch REM CCS nodes = 4 REM CCS minutes = 10 REM CCS requirements = 4@development REM Move to the T drive call machinemaker Run machinemaker to create a file called "machines", which contains a list of machines allocated to this job REM Set the communication mode to TCP or VIA set MPI_COMM=VIA 3/12/03 Cornell Theory Center 14 Parallel Job on Multiple (page 2 of 4) Use mpirun to call a batch script called setup.bat. Setup.bat to copy the executable, data files, etc onto the nodes listed in the machines file. When calling setup.bat, set -np to the number of nodes, not the number of tasks. mpirun -np 4 H:\users\%USERNAME%\lab\ex3setup.bat REM setup.bat REM Create a working subdirectory on each node mkdir your_userid cd your_userid del /Q /S \your_userid\*.* REM Copy the executable (and data files, if necessary) to each node copy H:\users\your_userid\lab\wavesend.exe \your_userid 3/12/03 Cornell Theory Center 15 Cornell Theory Center 5
6 Parallel Job on Multiple (page 3 of 4) REM The input file is only needed on the master node copy H:\users\%USERNAME%\lab\wave.in \%USERNAME% Move the machines file from to \your_userid. It couldn't be put there earlier. cd %USERNAME% move \machines \%USERNAME% Run the MPI program with mpirun. Set -np to the number of tasks. Set -wd to the working directory used by mpirun. mpirun -wd \%USERNAME%\ -np 8 wavesend.exe 1>wave.out 2>wave.err Copy any output created by the master task back to H:. copy \%USERNAME%\prog.* H:\users\%USERNAME%\lab 3/12/03 Cornell Theory Center 16 Parallel Job on Multiple (page 4 of 4) Next another script (cleanup.bat) is run all nodes to both copy any other output files back to the H drive and to clean up the subdirectory. When calling cleanup.bat, set -np to the number of nodes,not the number of tasks. mpirun -np 4 H:\users\%USERNAME%\lab\ex3cleanup.bat REM Release the nodes ccrelease REM In this example, only the master node has output files. copy \your_userid\wave.* H:\users\your_userid\lab REM Cleanup the working subdirectory on each node del /Q /S \your_userid\*.* 3/12/03 Cornell Theory Center 17 MCP Fileserver (H:) Standard copy: Fileserver is accessed N times Copy with mcp Fileserver is accessed 1 time 3/12/03 Cornell Theory Center 18 Cornell Theory Center 6
7 More Information: Consulting - Send questions about use of the Velocity Cluster to consult@tc.cornell.edu or call and ask for a consultant. Can answer specific questions, help with problems, discuss strategies and code planning. Strategic users (faculty with proven application, suitable for scaling) have access to more in-depth long-term consulting help /12/03 Cornell Theory Center 19 Cornell Theory Center 7
Introduction to the SHARCNET Environment May-25 Pre-(summer)school webinar Speaker: Alex Razoumov University of Ontario Institute of Technology
Introduction to the SHARCNET Environment 2010-May-25 Pre-(summer)school webinar Speaker: Alex Razoumov University of Ontario Institute of Technology available hardware and software resources our web portal
More informationCluster Network Products
Cluster Network Products Cluster interconnects include, among others: Gigabit Ethernet Myrinet Quadrics InfiniBand 1 Interconnects in Top500 list 11/2009 2 Interconnects in Top500 list 11/2008 3 Cluster
More informationMinnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.
Minnesota Supercomputing Institute Introduction to MSI for Physical Scientists Michael Milligan MSI Scientific Computing Consultant Goals Introduction to MSI resources Show you how to access our systems
More informationDistributed Memory Programming With MPI Computer Lab Exercises
Distributed Memory Programming With MPI Computer Lab Exercises Advanced Computational Science II John Burkardt Department of Scientific Computing Florida State University http://people.sc.fsu.edu/ jburkardt/classes/acs2
More informationIntroduction to PICO Parallel & Production Enviroment
Introduction to PICO Parallel & Production Enviroment Mirko Cestari m.cestari@cineca.it Alessandro Marani a.marani@cineca.it Domenico Guida d.guida@cineca.it Nicola Spallanzani n.spallanzani@cineca.it
More informationUAntwerpen, 24 June 2016
Tier-1b Info Session UAntwerpen, 24 June 2016 VSC HPC environment Tier - 0 47 PF Tier -1 623 TF Tier -2 510 Tf 16,240 CPU cores 128/256 GB memory/node IB EDR interconnect Tier -3 HOPPER/TURING STEVIN THINKING/CEREBRO
More informationReduces latency and buffer overhead. Messaging occurs at a speed close to the processors being directly connected. Less error detection
Switching Operational modes: Store-and-forward: Each switch receives an entire packet before it forwards it onto the next switch - useful in a general purpose network (I.e. a LAN). usually, there is a
More informationOBTAINING AN ACCOUNT:
HPC Usage Policies The IIA High Performance Computing (HPC) System is managed by the Computer Management Committee. The User Policies here were developed by the Committee. The user policies below aim to
More informationBefore We Start. Sign in hpcxx account slips Windows Users: Download PuTTY. Google PuTTY First result Save putty.exe to Desktop
Before We Start Sign in hpcxx account slips Windows Users: Download PuTTY Google PuTTY First result Save putty.exe to Desktop Research Computing at Virginia Tech Advanced Research Computing Compute Resources
More informationMinnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.
Minnesota Supercomputing Institute Introduction to MSI Systems Andrew Gustafson The Machines at MSI Machine Type: Cluster Source: http://en.wikipedia.org/wiki/cluster_%28computing%29 Machine Type: Cluster
More informationName Department/Research Area Have you used the Linux command line?
Please log in with HawkID (IOWA domain) Macs are available at stations as marked To switch between the Windows and the Mac systems, press scroll lock twice 9/27/2018 1 Ben Rogers ITS-Research Services
More informationOur new HPC-Cluster An overview
Our new HPC-Cluster An overview Christian Hagen Universität Regensburg Regensburg, 15.05.2009 Outline 1 Layout 2 Hardware 3 Software 4 Getting an account 5 Compiling 6 Queueing system 7 Parallelization
More informationTech Computer Center Documentation
Tech Computer Center Documentation Release 0 TCC Doc February 17, 2014 Contents 1 TCC s User Documentation 1 1.1 TCC SGI Altix ICE Cluster User s Guide................................ 1 i ii CHAPTER 1
More informationEffective Use of CCV Resources
Effective Use of CCV Resources Mark Howison User Services & Support This talk... Assumes you have some familiarity with a Unix shell Provides examples and best practices for typical usage of CCV systems
More informationIntroduction to High Performance Computing Using Sapelo2 at GACRC
Introduction to High Performance Computing Using Sapelo2 at GACRC Georgia Advanced Computing Resource Center University of Georgia Suchitra Pakala pakala@uga.edu 1 Outline High Performance Computing (HPC)
More informationIntroduction to GALILEO
Introduction to GALILEO Parallel & production environment Mirko Cestari m.cestari@cineca.it Alessandro Marani a.marani@cineca.it Domenico Guida d.guida@cineca.it Maurizio Cremonesi m.cremonesi@cineca.it
More informationBIRLA INSTITUTE OF TECHNOLOGY & SCIENCE, PILANI July, 2006
BACKUP PLANNING AND IMPLEMENTATION FOR ANUPAM SUPERCOMPUTER USING ROBOTIC AUTOLOADERS BY Aalap Tripathy 2004P3PS208 B.E. (Hons) Electrical & Electronics Prepared in partial fulfillment of the Practice
More informationIntroduction to High Performance Computing at UEA. Chris Collins Head of Research and Specialist Computing ITCS
Introduction to High Performance Computing at UEA. Chris Collins Head of Research and Specialist Computing ITCS Introduction to High Performance Computing High Performance Computing at UEA http://rscs.uea.ac.uk/hpc/
More informationHigh Performance Computing (HPC) Prepared By: Abdussamad Muntahi Muhammad Rahman
High Performance Computing (HPC) Prepared By: Abdussamad Muntahi Muhammad Rahman 1 2 Introduction to High Performance Computing (HPC) Introduction High-speed computing. Originally pertaining only to supercomputers
More informationHPCF Cray Phase 2. User Test period. Cristian Simarro User Support. ECMWF April 18, 2016
HPCF Cray Phase 2 User Test period Cristian Simarro User Support advisory@ecmwf.int ECMWF April 18, 2016 Content Introduction Upgrade timeline Changes Hardware Software Steps for the testing on CCB Possible
More informationIntroduction to HPC Programming 4. C and FORTRAN compilers; make, configure, cmake. Valentin Pavlov
Introduction to HPC Programming 4. C and FORTRAN compilers; make, configure, cmake Valentin Pavlov About these lectures This is the fourth of series of six introductory lectures discussing
More informationMeteorology 5344, Fall 2017 Computational Fluid Dynamics Dr. M. Xue. Computer Problem #l: Optimization Exercises
Meteorology 5344, Fall 2017 Computational Fluid Dynamics Dr. M. Xue Computer Problem #l: Optimization Exercises Due Thursday, September 19 Updated in evening of Sept 6 th. Exercise 1. This exercise is
More informationHow to Use a Supercomputer - A Boot Camp
How to Use a Supercomputer - A Boot Camp Shelley Knuth Peter Ruprecht shelley.knuth@colorado.edu peter.ruprecht@colorado.edu www.rc.colorado.edu Outline Today we will discuss: Who Research Computing is
More informationThe cluster system. Introduction 22th February Jan Saalbach Scientific Computing Group
The cluster system Introduction 22th February 2018 Jan Saalbach Scientific Computing Group cluster-help@luis.uni-hannover.de Contents 1 General information about the compute cluster 2 Available computing
More informationThe Supercomputing Facility for Bioinformatics & Computational Biology, IIT Delhi
Supercomputing Facility for Bioinformatics & Computational Biology, IIT Delhi The Supercomputing Facility for Bioinformatics & Computational Biology, IIT Delhi User Manual Dr. B. Jayaram (Professor of
More informationMinnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.
Minnesota Supercomputing Institute Introduction to Job Submission and Scheduling Andrew Gustafson Interacting with MSI Systems Connecting to MSI SSH is the most reliable connection method Linux and Mac
More informationShared Memory Programming With OpenMP Exercise Instructions
Shared Memory Programming With OpenMP Exercise Instructions John Burkardt Interdisciplinary Center for Applied Mathematics & Information Technology Department Virginia Tech... Advanced Computational Science
More informationClusters. Rob Kunz and Justin Watson. Penn State Applied Research Laboratory
Clusters Rob Kunz and Justin Watson Penn State Applied Research Laboratory rfk102@psu.edu Contents Beowulf Cluster History Hardware Elements Networking Software Performance & Scalability Infrastructure
More informationOutline. March 5, 2012 CIRMMT - McGill University 2
Outline CLUMEQ, Calcul Quebec and Compute Canada Research Support Objectives and Focal Points CLUMEQ Site at McGill ETS Key Specifications and Status CLUMEQ HPC Support Staff at McGill Getting Started
More informationOur Workshop Environment
Our Workshop Environment John Urbanic Parallel Computing Scientist Pittsburgh Supercomputing Center Copyright 2016 Our Environment This Week Your laptops or workstations: only used for portal access Bridges
More informationIntroduction to Discovery.
Introduction to Discovery http://discovery.dartmouth.edu The Discovery Cluster 2 Agenda What is a cluster and why use it Overview of computer hardware in cluster Help Available to Discovery Users Logging
More informationIntroduction to Discovery.
Introduction to Discovery http://discovery.dartmouth.edu The Discovery Cluster 2 Agenda What is a cluster and why use it Overview of computer hardware in cluster Help Available to Discovery Users Logging
More informationCluster Clonetroop: HowTo 2014
2014/02/25 16:53 1/13 Cluster Clonetroop: HowTo 2014 Cluster Clonetroop: HowTo 2014 This section contains information about how to access, compile and execute jobs on Clonetroop, Laboratori de Càlcul Numeric's
More informationIntroduction to GALILEO
November 27, 2016 Introduction to GALILEO Parallel & production environment Mirko Cestari m.cestari@cineca.it Alessandro Marani a.marani@cineca.it SuperComputing Applications and Innovation Department
More informationIntroduction to High Performance Computing at UEA. Chris Collins Head of Research and Specialist Computing ITCS
Introduction to High Performance Computing at UEA. Chris Collins Head of Research and Specialist Computing ITCS Introduction to High Performance Computing High Performance Computing at UEA http://rscs.uea.ac.uk/hpc/
More informationIntroduction to High-Performance Computing
Introduction to High-Performance Computing Dr. Axel Kohlmeyer Associate Dean for Scientific Computing, CST Associate Director, Institute for Computational Science Assistant Vice President for High-Performance
More informationOur Workshop Environment
Our Workshop Environment John Urbanic Parallel Computing Scientist Pittsburgh Supercomputing Center Copyright 2017 Our Environment This Week Your laptops or workstations: only used for portal access Bridges
More informationOur Workshop Environment
Our Workshop Environment John Urbanic Parallel Computing Scientist Pittsburgh Supercomputing Center Copyright 2017 Our Environment This Week Your laptops or workstations: only used for portal access Bridges
More informationGROMACS Performance Benchmark and Profiling. September 2012
GROMACS Performance Benchmark and Profiling September 2012 Note The following research was performed under the HPC Advisory Council activities Participating vendors: AMD, Dell, Mellanox Compute resource
More informationUL HPC Monitoring in practice: why, what, how, where to look
C. Parisot UL HPC Monitoring in practice: why, what, how, where to look 1 / 22 What is HPC? Best Practices Getting Fast & Efficient UL HPC Monitoring in practice: why, what, how, where to look Clément
More informationIntroduction to the NCAR HPC Systems. 25 May 2018 Consulting Services Group Brian Vanderwende
Introduction to the NCAR HPC Systems 25 May 2018 Consulting Services Group Brian Vanderwende Topics to cover Overview of the NCAR cluster resources Basic tasks in the HPC environment Accessing pre-built
More informationWindows-HPC Environment at RWTH Aachen University
Windows-HPC Environment at RWTH Aachen University Christian Terboven, Samuel Sarholz {terboven, sarholz}@rz.rwth-aachen.de Center for Computing and Communication RWTH Aachen University PPCES 2009 March
More informationPACE. Instructional Cluster Environment (ICE) Orientation. Research Scientist, PACE
PACE Instructional Cluster Environment (ICE) Orientation Mehmet (Memo) Belgin, PhD Research Scientist, PACE www.pace.gatech.edu What is PACE A Partnership for an Advanced Computing Environment Provides
More informationTools for Intel Xeon Phi: VTune & Advisor Dr. Fabio Baruffa - LRZ,
Tools for Intel Xeon Phi: VTune & Advisor Dr. Fabio Baruffa - fabio.baruffa@lrz.de LRZ, 27.6.- 29.6.2016 Architecture Overview Intel Xeon Processor Intel Xeon Phi Coprocessor, 1st generation Intel Xeon
More informationPACE. Instructional Cluster Environment (ICE) Orientation. Mehmet (Memo) Belgin, PhD Research Scientist, PACE
PACE Instructional Cluster Environment (ICE) Orientation Mehmet (Memo) Belgin, PhD www.pace.gatech.edu Research Scientist, PACE What is PACE A Partnership for an Advanced Computing Environment Provides
More informationOur Workshop Environment
Our Workshop Environment John Urbanic Parallel Computing Scientist Pittsburgh Supercomputing Center Copyright 2018 Our Environment This Week Your laptops or workstations: only used for portal access Bridges
More informationExperiences in Optimizing a $250K Cluster for High- Performance Computing Applications
Experiences in Optimizing a $250K Cluster for High- Performance Computing Applications Kevin Brandstatter Dan Gordon Jason DiBabbo Ben Walters Alex Ballmer Lauren Ribordy Ioan Raicu Illinois Institute
More informationIntroduction to GALILEO
Introduction to GALILEO Parallel & production environment Mirko Cestari m.cestari@cineca.it Alessandro Marani a.marani@cineca.it Alessandro Grottesi a.grottesi@cineca.it SuperComputing Applications and
More informationBeginner's Guide for UK IBM systems
Beginner's Guide for UK IBM systems This document is intended to provide some basic guidelines for those who already had certain programming knowledge with high level computer languages (e.g. Fortran,
More informationIntel Manycore Testing Lab (MTL) - Linux Getting Started Guide
Intel Manycore Testing Lab (MTL) - Linux Getting Started Guide Introduction What are the intended uses of the MTL? The MTL is prioritized for supporting the Intel Academic Community for the testing, validation
More informationIntroduction to HPC Resources and Linux
Introduction to HPC Resources and Linux Burak Himmetoglu Enterprise Technology Services & Center for Scientific Computing e-mail: bhimmetoglu@ucsb.edu Paul Weakliem California Nanosystems Institute & Center
More informationFor Dr Landau s PHYS8602 course
For Dr Landau s PHYS8602 course Shan-Ho Tsai (shtsai@uga.edu) Georgia Advanced Computing Resource Center - GACRC January 7, 2019 You will be given a student account on the GACRC s Teaching cluster. Your
More informationSuperMike-II Launch Workshop. System Overview and Allocations
: System Overview and Allocations Dr Jim Lupo CCT Computational Enablement jalupo@cct.lsu.edu SuperMike-II: Serious Heterogeneous Computing Power System Hardware SuperMike provides 442 nodes, 221TB of
More informationSupercomputing in Plain English Exercise #6: MPI Point to Point
Supercomputing in Plain English Exercise #6: MPI Point to Point In this exercise, we ll use the same conventions and commands as in Exercises #1, #2, #3, #4 and #5. You should refer back to the Exercise
More informationLS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance
11 th International LS-DYNA Users Conference Computing Technology LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance Gilad Shainer 1, Tong Liu 2, Jeff Layton
More informationOpenMP Exercises. These exercises will introduce you to using OpenMP for parallel programming. There are four exercises:
OpenMP Exercises These exercises will introduce you to using OpenMP for parallel programming. There are four exercises: 1. OMP Hello World 2. Worksharing Loop 3. OMP Functions 4. Hand-coding vs. MKL To
More informationInterpolation calculation using cloud computing
Interpolation calculation using cloud computing MPI in Amazon Web Services Eduard Vesel Faculty of Natural Sciences Matej Bel University Banská Bystrica, Slovakia edd.ves@gmail.com At present time, when
More informationProgramming Techniques for Supercomputers. HPC RRZE University Erlangen-Nürnberg Sommersemester 2018
Programming Techniques for Supercomputers HPC Services @ RRZE University Erlangen-Nürnberg Sommersemester 2018 Outline Login to RRZE s Emmy cluster Basic environment Some guidelines First Assignment 2
More informationOur Workshop Environment
Our Workshop Environment John Urbanic Parallel Computing Scientist Pittsburgh Supercomputing Center Copyright 2017 Our Environment This Week Your laptops or workstations: only used for portal access Bridges
More informationIntroduction to HPC Using zcluster at GACRC
Introduction to HPC Using zcluster at GACRC On-class PBIO/BINF8350 Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? What
More informationIntroduction to HPC Using zcluster at GACRC
Introduction to HPC Using zcluster at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? What is HPC Concept? What is
More informationUniversity of California, Riverside. Computing and Communications. Computational UCR. March 28, Introduction 2
University of California, Riverside Computing and Communications Computational Clusters @ UCR March 28, 2008 1 Introduction 2 2 Campus Datacenter Co-Location Facility 2 3 Three Cluster Models 2 3.1 Departmentally
More informationUsing the computational resources at the GACRC
An introduction to zcluster Georgia Advanced Computing Resource Center (GACRC) University of Georgia Dr. Landau s PHYS4601/6601 course - Spring 2017 What is GACRC? Georgia Advanced Computing Resource Center
More informationTo connect to the cluster, simply use a SSH or SFTP client to connect to:
RIT Computer Engineering Cluster The RIT Computer Engineering cluster contains 12 computers for parallel programming using MPI. One computer, phoenix.ce.rit.edu, serves as the master controller or head
More informationBioinformatics Facility at the Biotechnology/Bioservices Center
Bioinformatics Facility at the Biotechnology/Bioservices Center Co-Heads : J.P. Gogarten, Paul Lewis Facility Scientist : Pascal Lapierre Hardware/Software Manager: Jeff Lary Mandate of the Facility: To
More informationHigh Performance Beowulf Cluster Environment User Manual
High Performance Beowulf Cluster Environment User Manual Version 3.1c 2 This guide is intended for cluster users who want a quick introduction to the Compusys Beowulf Cluster Environment. It explains how
More informationHigh performance Linux clusters - design/configuration/maintenance
High performance Linux clusters - design/configuration/maintenance Werner Scholz Theory Group Roy Chantrell Jan. 3rd, 2002 Outline Motivation Installation at TU Vienna Stability, reliability Performance
More informationInstallation and Test of Molecular Dynamics Simulation Packages on SGI Altix and Hydra-Cluster at JKU Linz
Installation and Test of Molecular Dynamics Simulation Packages on SGI Altix and Hydra-Cluster at JKU Linz Rene Kobler May 25, 25 Contents 1 Abstract 2 2 Status 2 3 Changelog 2 4 Installation Notes 3 4.1
More informationNERSC. National Energy Research Scientific Computing Center
NERSC National Energy Research Scientific Computing Center Established 1974, first unclassified supercomputer center Original mission: to enable computational science as a complement to magnetically controlled
More informationAccelerating Simulink Optimization, Code Generation & Test Automation Through Parallelization
Accelerating Simulink Optimization, Code Generation & Test Automation Through Parallelization Ryan Chladny Application Engineering May 13 th, 2014 2014 The MathWorks, Inc. 1 Design Challenge: Electric
More informationCS/Math 471: Intro. to Scientific Computing
CS/Math 471: Intro. to Scientific Computing Getting Started with High Performance Computing Matthew Fricke, PhD Center for Advanced Research Computing Table of contents 1. The Center for Advanced Research
More informationHow to run applications on Aziz supercomputer. Mohammad Rafi System Administrator Fujitsu Technology Solutions
How to run applications on Aziz supercomputer Mohammad Rafi System Administrator Fujitsu Technology Solutions Agenda Overview Compute Nodes Storage Infrastructure Servers Cluster Stack Environment Modules
More informationSharpen Exercise: Using HPC resources and running parallel applications
Sharpen Exercise: Using HPC resources and running parallel applications Andrew Turner, Dominic Sloan-Murphy, David Henty, Adrian Jackson Contents 1 Aims 2 2 Introduction 2 3 Instructions 3 3.1 Log into
More informationUsing Intel VTune Amplifier XE for High Performance Computing
Using Intel VTune Amplifier XE for High Performance Computing Vladimir Tsymbal Performance, Analysis and Threading Lab 1 The Majority of all HPC-Systems are Clusters Interconnect I/O I/O... I/O I/O Message
More informationHigh Performance Computing. What is it used for and why?
High Performance Computing What is it used for and why? Overview What is it used for? Drivers for HPC Examples of usage Why do you need to learn the basics? Hardware layout and structure matters Serial
More informationOur Workshop Environment
Our Workshop Environment John Urbanic Parallel Computing Scientist Pittsburgh Supercomputing Center Copyright 2018 Our Environment Today Your laptops or workstations: only used for portal access Bridges
More informationACEnet for CS6702 Ross Dickson, Computational Research Consultant 29 Sep 2009
ACEnet for CS6702 Ross Dickson, Computational Research Consultant 29 Sep 2009 What is ACEnet? Shared resource......for research computing... physics, chemistry, oceanography, biology, math, engineering,
More informationExercises: Abel/Colossus and SLURM
Exercises: Abel/Colossus and SLURM November 08, 2016 Sabry Razick The Research Computing Services Group, USIT Topics Get access Running a simple job Job script Running a simple job -- qlogin Customize
More informationUoW HPC Quick Start. Information Technology Services University of Wollongong. ( Last updated on October 10, 2011)
UoW HPC Quick Start Information Technology Services University of Wollongong ( Last updated on October 10, 2011) 1 Contents 1 Logging into the HPC Cluster 3 1.1 From within the UoW campus.......................
More informationShared Memory Programming With OpenMP Computer Lab Exercises
Shared Memory Programming With OpenMP Computer Lab Exercises Advanced Computational Science II John Burkardt Department of Scientific Computing Florida State University http://people.sc.fsu.edu/ jburkardt/presentations/fsu
More informationTo connect to the cluster, simply use a SSH or SFTP client to connect to:
RIT Computer Engineering Cluster The RIT Computer Engineering cluster contains 12 computers for parallel programming using MPI. One computer, cluster-head.ce.rit.edu, serves as the master controller or
More informationIntroduction to Discovery.
Introduction to Discovery http://discovery.dartmouth.edu March 2014 The Discovery Cluster 2 Agenda Resource overview Logging on to the cluster with ssh Transferring files to and from the cluster The Environment
More informationIntroduction to BioHPC
Introduction to BioHPC New User Training [web] [email] portal.biohpc.swmed.edu biohpc-help@utsouthwestern.edu 1 Updated for 2018-03-07 Overview Today we re going to cover: What is BioHPC? How do I access
More informationXSEDE New User Tutorial
April 2, 2014 XSEDE New User Tutorial Jay Alameda National Center for Supercomputing Applications XSEDE Training Survey Make sure you sign the sign in sheet! At the end of the module, I will ask you to
More informationCENTER FOR HIGH PERFORMANCE COMPUTING. Overview of CHPC. Martin Čuma, PhD. Center for High Performance Computing
Overview of CHPC Martin Čuma, PhD Center for High Performance Computing m.cuma@utah.edu Spring 2014 Overview CHPC Services HPC Clusters Specialized computing resources Access and Security Batch (PBS and
More informationSharpen Exercise: Using HPC resources and running parallel applications
Sharpen Exercise: Using HPC resources and running parallel applications Contents 1 Aims 2 2 Introduction 2 3 Instructions 3 3.1 Log into ARCHER frontend nodes and run commands.... 3 3.2 Download and extract
More informationHigh Performance Computing (HPC) Using zcluster at GACRC
High Performance Computing (HPC) Using zcluster at GACRC On-class STAT8060 Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC?
More informationHands-On Workshop bwunicluster June 29th 2015
Hands-On Workshop bwunicluster June 29th 2015 Agenda Welcome Introduction to bwhpc and the bwunicluster Modules - Software Environment Management Job Submission and Monitoring Interactive Work and Remote
More informationIntroduction to HPC Using zcluster at GACRC On-Class GENE 4220
Introduction to HPC Using zcluster at GACRC On-Class GENE 4220 Georgia Advanced Computing Resource Center University of Georgia Suchitra Pakala pakala@uga.edu Slides courtesy: Zhoufei Hou 1 OVERVIEW GACRC
More informationIntroduction to NCAR HPC. 25 May 2017 Consulting Services Group Brian Vanderwende
Introduction to NCAR HPC 25 May 2017 Consulting Services Group Brian Vanderwende Topics we will cover Technical overview of our HPC systems The NCAR computing environment Accessing software on Cheyenne
More informationGetting started with the CEES Grid
Getting started with the CEES Grid October, 2013 CEES HPC Manager: Dennis Michael, dennis@stanford.edu, 723-2014, Mitchell Building room 415. Please see our web site at http://cees.stanford.edu. Account
More informationWVU RESEARCH COMPUTING INTRODUCTION. Introduction to WVU s Research Computing Services
WVU RESEARCH COMPUTING INTRODUCTION Introduction to WVU s Research Computing Services WHO ARE WE? Division of Information Technology Services Funded through WVU Research Corporation Provide centralized
More informationDesigning a Cluster for a Small Research Group
Designing a Cluster for a Small Research Group Jim Phillips, John Stone, Tim Skirvin Low-cost Linux Clusters for Biomolecular Simulations Using NAMD Outline Why and why not clusters? Consider your Users
More informationIntroduction to Cluster Computing
Introduction to Cluster Computing Prabhaker Mateti Wright State University Dayton, Ohio, USA Overview High performance computing High throughput computing NOW, HPC, and HTC Parallel algorithms Software
More informationGROMACS Performance Benchmark and Profiling. August 2011
GROMACS Performance Benchmark and Profiling August 2011 Note The following research was performed under the HPC Advisory Council activities Participating vendors: Intel, Dell, Mellanox Compute resource
More informationSlurm and Abel job scripts. Katerina Michalickova The Research Computing Services Group SUF/USIT October 23, 2012
Slurm and Abel job scripts Katerina Michalickova The Research Computing Services Group SUF/USIT October 23, 2012 Abel in numbers Nodes - 600+ Cores - 10000+ (1 node->2 processors->16 cores) Total memory
More informationIntroduction to HPC Using zcluster at GACRC
Introduction to HPC Using zcluster at GACRC On-class STAT8330 Georgia Advanced Computing Resource Center University of Georgia Suchitra Pakala pakala@uga.edu Slides courtesy: Zhoufei Hou 1 Outline What
More informationIntroduc)on to Hyades
Introduc)on to Hyades Shawfeng Dong Department of Astronomy & Astrophysics, UCSSC Hyades 1 Hardware Architecture 2 Accessing Hyades 3 Compu)ng Environment 4 Compiling Codes 5 Running Jobs 6 Visualiza)on
More informationNAMD Performance Benchmark and Profiling. January 2015
NAMD Performance Benchmark and Profiling January 2015 2 Note The following research was performed under the HPC Advisory Council activities Participating vendors: Intel, Dell, Mellanox Compute resource
More informationWorking on the NewRiver Cluster
Working on the NewRiver Cluster CMDA3634: Computer Science Foundations for Computational Modeling and Data Analytics 22 February 2018 NewRiver is a computing cluster provided by Virginia Tech s Advanced
More information