Introduction to BioHPC

Similar documents
Introduction to BioHPC

Introduction to BioHPC New User Training

Introduction to BioHPC New User Training

Introduction to BioHPC

Welcome to the Introduc/on to BioHPC training session. My name is David Trudgian, and I m one of the Computa/onal Scien/sts in the BioHPC team.

Workstations & Thin Clients

The BioHPC Nucleus Cluster & Future Developments

Visualization on BioHPC

Introduction to High-Performance Computing (HPC)

Using CLC Genomics Workbench on Turing

Using MFA with the Pulse Client

AHC IE Data Shelter User Guide

MICROSOFT OFFICE Desktop Applications. Student User Guide Overview

Junos Pulse Installation (SSL VPN)

Quick Start Guide. Table of Contents

DPHremote.ucsf.edu for Webconnect Users

Graphical Access to IU's Supercomputers with Karst Desktop Beta

This guide provides all of the information necessary to connect to MoFo resources from outside of the office.

DPHremote.ucsf.edu for Webconnect Users

Zadara Enterprise Storage in

The Rockefeller University I NFORMATION T ECHNOLOGY E DUCATION & T RAINING. VPN Web Portal Usage Guide

If necessary, you can also download and install the Cisco Webex Meetings desktop app manually.

Cloud Compute. Backup Portal User Guide

Introduction. Kevin Miles. Paul Henderson. Rick Stillings. Essex Scales. Director of Research Support. Systems Engineer.

IBM Cloud Client Technical Engagement Education Network Columbus, Ohio

CLC Genomics Workbench. Setup and User Guide

Robert Bukowski Jaroslaw Pillardy 6/27/2011

WAM!NET Submission Icons. Help Guide. March 2015

Click Studios. Passwordstate. Remote Session Launcher. Installation Instructions

User Guide for Client Remote Access. Version 1.2

AN INTRODUCTION TO CLUSTER COMPUTING

DNA Sequence Bioinformatics Analysis with the Galaxy Platform

From using an External Harddrive, to a Google Cloud Drive; there is no one way to backup data.

Copyright 2013

Introduction to Discovery.

Goverlan Remote Control v7 vs. Microsoft Remote Assistance

docalpha Recognition Station

HIGH PERFORMANCE COMPUTING (PLATFORMS) SECURITY AND OPERATIONS

Remote Workspace. Nubo End User Guide. Version: 1.3 Date: June, Copyright 2017 by Nubo Inc. All rights reserved. Page 1


Job Aid: Citrix Receiver Upgrade

Booting a Galaxy Instance

Document for connecting an ipad/iphone to the University VPN System

Print Audit 6. Print Audit 6 Documentation Apr :07. Version: Date:

XSEDE New User/Allocation Mini-Tutorial

Introduction to HPC Resources and Linux

goremote.carolinas.org

Hollins University VPN

Skyline Healthcare Getting Started Guide for New Users

Using Cartesius and Lisa. Zheng Meyer-Zhao - Consultant Clustercomputing

RU-VPN2 - GlobalProtect Installation for Windows

XSEDE New User Training. Ritu Arora November 14, 2014

Getting Started Accessing Okta All Employees

Our Workshop Environment

Introduction to GACRC Teaching Cluster

Creating Avid Projects, and Media Sharing with Small Tree TitaniumZ

Remote Desktop Services

Name Department/Research Area Have you used the Linux command line?

XSEDE New User Tutorial

For Dr Landau s PHYS8602 course

New Dropbox Users (don t have a Dropbox account set up with your Exeter account)

Remote Access with Imprivata Two-factor Authentication

capvpn Instructions for Windows

docalpha 5.0 Recognition Station User Guide

NIAS2018 The Crowd Compass Attendee Hub Mobile App FAQ

Deployment User Guide

pinremote Manual Version 4.0

LENS Server Maintenance Guide JZ 2017/07/28

iround FAQs Account Questions iround Q. Why can t I log into iround?

Duo End User Education Templates

Introduction to GACRC Teaching Cluster PHYS8602

Getting Started with eggplant Functional

Lifespan Guide for installing and using Multi-Factor Authentication (MFA)

CMTH/TYC Linux Cluster Overview. Éamonn Murray 1 September 2017

Paperspace. Deployment Guide. Cloud VDI. 20 Jay St. Suite 312 Brooklyn, NY Technical Whitepaper

Microsoft Remote Desktop Installation and Setup for Analysis VMs

SONOTON storage server

Numascale Analytics Appliance in the Cloud

Agent and Agent Browser. Updated Friday, January 26, Autotask Corporation

INTRODUCTION TO THE CLUSTER

VPN/RDP Laptop and Workstation Usage instructions:

Remote Access to Unix Machines

VPN Guide. Pre-Requisites. Before connecting to the SGUL VPN, you must ensure the following:

UoW HPC Quick Start. Information Technology Services University of Wollongong. ( Last updated on October 10, 2011)

JHU Economics August 24, Galaxy How To SSH and RDP

Texas Division How to Login and Register for My IT Support and ServiceNow

Introduction to GACRC Teaching Cluster

Remote Access User Guide for Mac OS (Citrix Instructions)

ENROLLING FOR YOUR SYKES HOME TRAINING

Backup and Restore. About Backup and Restore

Access to hosted Systems at ERPsourcing AG using SSL-VPN

TFS Self-Service Password Reset Portal Getting Started Guide

WDC RDS Connection for Android Users

Overcoming Remote Desktop Challenges with

Accessing the CAES Office of Communication File Sharing Services

SAP GUI 7.30 for Windows Computer

End User Manual. December 2014 V1.0

FEPS. SSH Access with Two-Factor Authentication. RSA Key-pairs

Linux for Biologists Part 2

Administrator Guide. Flexible Storage

Transcription:

Introduction to BioHPC New User Training [web] [email] portal.biohpc.swmed.edu biohpc-help@utsouthwestern.edu 1 Updated for 2017-01-04

Overview Today we re going to cover: What is BioHPC? How do I access BioHPC resources? How can I be a good user? (some basic rules) How do I get effective help? If you remember only one thing. If you have any question, ask us via biohpc-help@utsouthwestern.edu 2

What is HPC, and why do we need it? High-performance computing (HPC) is the use of parallel processing for running advanced application programs efficiently, reliably and quickly. Any computing that isn t possible on a standard system PROBLEMS Huge Datasets Complex Algorithms Difficult / inefficient software BioHPC SOLUTIONS Batch HPC jobs Interactive GUI sessions Visualization with GPUs Windows sessions on the cluster Wide range of software Easy web access to services 3

What is BioHPC? - An Overview BioHPC is: A 148-node compute cluster. >3.5 Petabyte (3,500 Terabytes) of storage across various systems. Large number of installed software packages. A network of thin-client and workstation machines. Cloud Services to access these facilities easily. A dedicated team to help you efficiently use these resources for your research. 4

Who is BioHPC? Liqiang Wang Director, 13 years experience in IT infrastructure, HPC. Yi Du Computational Scientist, experience in parallel software design, large-scale data analysis. David Trudgian Computational Scientist, Ph.D. in Computer Science and 10 years experience in bioinformatics (focus on machine learning for sequence classification & computational proteomics). Ross Bateman Technical Support Specialist, experienced in maintaining user systems and troubleshooting. Wei Guo Computational Scientist, Ph.D. in Materials Science and Engineering, experience in HPC for complex simulations. Long Lu Computational Scientist, MS in CS. Biology and Chemistry, Gene sequencing and materials science Yingfei Chen Computational Scientist, PHD in Chemical and Biological Engineering, protein sequences and tertiary structures We are biohpc-help@utsouthwestern.edu https://portal.biohpc.swmed.edu/content/about/staff/ 5

What is BioHPC? Nucleus Computer Cluster Nucleus is our compute cluster 148 nodes 128GB, 256GB, 256GBv1, 384GB, GPU CPU cores: 6816 GPU cores: 19968 Memory: 34TB Network: 100Gb/s core, 56Gb/s per node (internal) 40Gb/s(to campus) Login via ssh to nucleus.biohpc.swmed.edu or use web portal. 6

What can we do using BioHPC? Run any computationally intensive work Linux HPC Jobs GPU Visualization Interactive Sessions Windows with GPU Visualization 7

BioHPC Storage Standard Users As a BioHPC user, you will have access to: BioHPC Cluster /home2/username /project/department/group/username /work/department/username 50 GB / user 5 TB / per group* 5 TB / per user BioHPC File Exchange (web-interface) 50 GB https://cloud.biohpc.swmed.edu 50 GB / user, local storage BioHPC Lamella Cloud Storage (web-interface), on campus only, private cloud https://lamella.biohpc.swmed.edu 100 GB / user, local storage Gateway to BioHPC Cluster (via FTP, SAMBA or WebDAV*) * Can be increased on PI request with Dept. Chair approval. 8

BioHPC Storage Core Users Some core facilities provide access to BioHPC for their users to transfer data etc. The core decides the amount and type of storage to provide to their users, e.g. TIBIR WBMF Core: /project/tibir/wbmf_core/<username> 250GB / core user This is also your home directory No separate home2 or work space Storage allocation and usage is at the discretion of the core, not BioHPC. 9

What is BioHPC? Lamella Storage Gateway Lamella is our storage gateway access your files easily Web Interface Windows / Mac drive mounts (SMB /WebDav) lamella.biohpc.swmed.edu FTP 10

BioHPC Storage Backup Since January we perform mirror back up on: /home2 Twice weekly (Mon & Wed) 2 copies home2 usage counts 3x against lab storage allocation /work Weekly (Fri->Sat) 1 copy work usage counts 2x against lab storage allocation Excludes some large users Mirror backups Copy of your files at the time the backup runs. No old versions. /project Incremental backup of specific locations at request of PI. 11

What is BioHPC? - Software Wide range of packages available as modules. You can ask biohpc-help@utsouthwestern.edu for additions/upgrades etc. 12

What is BioHPC? Thin Client & Workstation Systems Desktop computers directly connected to the BioHPC systems. Run same version of Linux as the cluster, but with a graphical desktop. Login with BioHPC details, direct access to storage like on cluster. Same software available as on cluster. Will make up a distributed compute resource in future, using up to 50% of CPU to run distributed jobs. Thin client is less powerful but cheaper and smaller. Training: Thin Clients and Workstations Check Calendar. 13

What is BioHPC? Cloud Services A big focus at BioHPC is easy-access to our systems. Our cloud services provide web-based access to resources, with only a browser. All accessible via portal.biohpc.swmed.edu 14

Okay, sounds great. But how do I use all of this? 15

Hands on BioHPC -- 1. Portal Website Walkthrough portal.biohpc.swmed.edu 16

Hands on BioHPC 2. Manage Files with Lamella / Cloud Storage Gateway Cloud storage gateway web-based. https://lamella.biohpc.swmed.edu 100GB separate space + Mount /home / project /work Internal https://cloud.biohpc.swmed.edu 50GB space External file transfer Accessible from internet 17

Hands on BioHPC 2. Manage Files with Lamella / Cloud Storage Gateway File Sharing Only you can open, never expires Any user can open, valid before expiration date Lamella cloud storage https://lamella.biohpc.swmed.edu : sharing with user inside UTSW File Exchange https://cloud.biohpc.swmed.edu : sharing with user outside UTSW 18

Hands on BioHPC 2. Setting up Lamella to access home, project, and work space https://lamella.biohpc.swmed.edu Log-in credentials, save in session uses the BioHPC login credentials and only saved in the user session, giving increased security. The drawbacks are that sharing is disabled, as lamella has no access to the cluster storage credentials. Username and password mechanism requires a manually-defined username and password. Remember to click the gear icon and enable sharing. username password BioHPC Endosome/Lysosome project work home For home leave blank For private project space: department/lab/user For lab shared project space: department/lab/shared 19

Hands on BioHPC 2. Accessing BioHPC Storage Directly from Windows Computer -> Map Network Drive Folder is: \\lamella.biohpc.swmed.edu\username (home dir) \\lamella.biohpc.swmed.edu\project \\lamella.biohpc.swmed.edu\work Check Connect using different credentials Enter your BioHPC username and password when prompted. 20

Hands on BioHPC 2. Accessing BioHPC Storage Directly from Mac OSX Finder -> Go -> Connect to Server Folder is: smb://lamella.biohpc.swmed.edu/username (home dir) smb://lamella.biohpc.swmed.edu/project smb://lamella.biohpc.swmed.edu/work Enter your BioHPC username and password when prompted. 21

Hands on BioHPC 3. Web Job Script Generator https://portal.biohpc.swmed.edu -> Cloud Services -> Web Job Submission 22

Hands on BioHPC. 4 Web Visualization: Graphical Interactive Session via Web Portal / VNC https://portal.biohpc.swmed.edu -> Cloud Services -> Web Visualization Connects to GUI running on a cluster node. WebGPU sessions have access to GPU card for 3D rendering. 23

Hands on BioHPC 5. Software Modules module list module avail module load <module name> module unload <module name> module help <module name> module H Show loaded modules Show available modules Load a module Unload a module Help notes for a module Help for the module command 24

Hands on BioHPC 6. SSH Cluster Login via the Web Portal https://portal.biohpc.swmed.edu -> Cloud Services -> Nucleus Web Terminal Connects to the login node, not a cluster node w 25

Hands on BioHPC 7. Connecting from Home Windows - Follow the IR VPN instructions at: http://www.utsouthwestern.net/intranet/administration/information-resources/network/vpn/ Mac Try the IR instructions first. If they don t work: On Campus Go -> Connect to Server Server Address: smb://swnas.swmed.org/data/installs Connect VPN Client (Juniper) -> Juniper Mac VPN Client Installer ->JunosPulse.dmg Install the software from in the.dmg file. You cannot test it on campus. At Home Start Junos Pulse and add a connection to server utswra.swmed.edu When connecting must enter a secondary password, which is obtained using the key icon in the Duo Mobile twofactor authentication smartphone app. Or type push to get a push notification on your phone. We can help surgery session, or NL05.136 26

How To Be a Good User HPC Systems are crowded, shared resources Co-operation is necessary. The BioHPC team has a difficult job to do: Balance the requirements of a diverse group of users, running very different types of jobs. Make sure user actions don t adversely affect others using the systems. Keep the environment secure. Ensure resources are being used efficiently. Web-based Cloud-Services are designed to avoid problems. 27

All we ask is 1. If you have any question, or are unsure about something please ask us. biohpc-help@utsouthwestern.edu 2. When running jobs on the cluster, request the least amount of resources you know you need. Job times / memory limit / smallest node that will work etc. Up to a 2x margin of safety is appropriate. 3. Make reasonable attempts to use the resources efficiently. Run multiple small tasks on a node if you can. Cancel / close any jobs or sessions you no longer need. 4. Keep notes in case you need our help troubleshooting Keep old versions of scripts and job files 28

Currently Enforced Policy Don t run complex things on the login node. (web terminal or nucleus.biohpc.swmed.edu) Maximum of 16 nodes in use concurrently by any single user. 2 GPU node max per user. Interactive use of cluster nodes using the web visualization or remotegui/remotegpu scripts only*. You cannot SSH to a computer node not allocated to you. 29

Getting Effective Help Email the ticket system: biohpc-help@utsouthwestern.edu What is the problem? Provide any error message, and diagnostic output you have When did it happen? What time? Cluster or client? What job id? How did you run it? What did you run, what parameters, what do they mean? Any unusual circumstances? Have you compiled your own software? Do you customize startup scripts? Can we look at your scripts and data? Tell us if you are happy for us to access your scripts/data to help troubleshoot. 30

Next Steps New users wait for confirmation your account is activated. Spend some time experimenting with our systems, read the guides. Check the training schedule and attend relevant sessions. Join us for coffee now part of 2 nd Wednesday training sessions. 31