Enabling web-based interactive notebooks on geographically distributed HPC resources. Alexandre Beche
|
|
- Lydia Stokes
- 5 years ago
- Views:
Transcription
1 Enabling web-based interactive notebooks on geographically distributed HPC resources Alexandre Beche
2 Outlines 1. Context 2. Interactive notebook running on cluster(s) 3. Advanced use cases 4. Summary
3 Context
4 Blue Brain Project Host Institution Main International Collaborations Ecole Polytechnique Fédérale de Lausanne (EPFL) Switzerland (CSCS, CERN) Israel (HUJI) Director USA (Yale, ANL, OLCF, Allen Brain) Henry Markram Spain (UPM) Co-Directors Saudi Arabia (KAUST) Sean Hill, Felix Schürmann Europe (HBP) Team today ~ 100 scientists, engineers & staff Timeline 2005 founded at EPFL 2011/2012 ETH Board funding Swiss National Research Infrastructure Geneva
5 Data-Driven Modeling & Simulation! I 11 electrical types ~ 2 mm thick 55 morphological types 13 excitatory & 42 inhibitory m-types 31,000 neurons 111,700 neurons/mm3 Excitatory to inhibitory neuron ratio of 86:14 % 346 m of axon 211 m of dendrites Maximum branch order of m-types: Ion channel density distribution profiles: 6 ite a m D ite on Markram et al, Cell en dr Ax 6 s 8 dr Uniform en IV So 17 morpho-electrical types n bap & EPSP attenuation for 207 Ax o 24 Inhibitory 13 HH type ion channel models s Excitatory II/III 207 morpho-electrical types D V 0.63 synapses/mm3 Extrinsic to intrinsic synapse ratio of 75:25 % 3025 possible synaptic pathways 2258 viable synaptic pathways 664 excitatory pathways 1594 inhibitory pathways 600 intra-laminar pathways 1658 inter-laminar pathways Mean synapses/connection ns 274 ns ry to bi hi In ci ta to Ex 100 µm ry h. In Ex c. 4.3 VI 6 synapse types 207 synaptomes Space clamp corrected synaptic conductances for 607 pathways The per synapse conductance of 1.5 ns for connections between L5TTPCs is the highest in the microrcircuit Mean conductance per synapse: 0.85 ns for excitatory & 0.66 ns for inhibitory synapses Total conductance in a single neuron is 971 ns Ø 80 authors Ø Joint effort between computer and neuroscientists Ø Reproducible work Ø Extensible
6 HPC Today s Infrastructure Elastic Compute Visualization and analysis Web services SW development Continuous Integration Continuous Deployment IaaS 4 nodes Viz & Analytics 36 nodes CEPH IaaS 12 nodes NetApp CEPH Viz & Analytics 14 nodes GPFS GVA Production HPC Model development Reconstruction Simulation SW development Systems Research Memory extension Application coupling Interactive supercomputing Reproducibility GPFS 4 Compute Racks 4096 Compute Nodes 5D CN torus 0.8 PF/s peak 64TB DRAM 4.2 PB GPFS storage IBM BlueGene/Q 64 IONodes 3D torus 128TB Flash Linux IBM BlueGene Active Storage
7 HPC Today s Infrastructure Elastic Compute Visualization and analysis Web services SW development Continuous Integration Continuous Deployment IaaS 4 nodes Viz & Analytics 36 nodes CEPH IaaS 12 nodes NetApp CEPH Viz & Analytics 14 nodes GPFS GVA Production HPC Infrastructure renewal ongoing4 Compute Racks 4096 Compute Nodes New HPC system planned for Q D CN torus GPFS 0.8 PF/s peak 64TB DRAM 4.2 PB GPFS storage Model development Reconstruction Simulation SW development Systems Research Memory extension Application coupling Interactive supercomputing Reproducibility IBM BlueGene/Q 64 IONodes 3D torus 128TB Flash Linux IBM BlueGene Active Storage
8 Interactive notebook running on cluster(s)
9 Problem description Scientist wants to run a web-based interactive notebook 1) Connect to a cluster frontend 2) Get an allocation ssh salloc
10 Problem description Scientist wants to run a web-based interactive notebook 1) Connect to a cluster frontend 2) Get an allocation 3) Run the Jupyter notebook ssh salloc
11 Problem description Scientist wants to run a web-based interactive notebook 1) Connect to a cluster frontend 2) Get an allocation 3) Run the Jupyter notebook 4) Connect to the notebook http ssh salloc Jupyter notebook
12 Problem description Scientist wants to run a web-based interactive notebook 1) Connect to a cluster frontend 2) Get an allocation 3) Run the Jupyter notebook 4) Connect to the notebook 5) Share the URL with colleagues http ssh salloc Jupyter notebook http
13 Problem description Scientist wants to run a web-based interactive notebook 1) Connect to a cluster frontend 2) Get an allocation 3) Run the Jupyter notebook 4) Connect to the notebook 5) Share the URL with colleagues http ssh salloc Arbitrary code can now be run on behalf of the user... And access to data Jupyter notebook http
14 JupyterHub Multi-user server for Jupyter notebooks Authenticate users Spawn single-user Jupyter notebook Proxy user traffic to notebook securely
15 JupyterHub Multi-user server for Jupyter notebooks Authenticate users Spawn single-user Jupyter notebook Proxy user traffic to notebook securely https Authenticator: Spawner: pam + sssd custom slurm spawner
16 Basic approach Based on JupyterHub configuration file Custom Spawner Slurm options
17 Basic approach Based on JupyterHub configuration file Custom Spawner Slurm options Steps 1. Run as root on the frontend 2. Resolve uid from logged username 3. Submit job with #SBATCH --uid=$user 4. Job connect back to JupyterHub 5. Gracefully fails if no resources available
18 Basic approach Based on JupyterHub configuration file Custom Spawner Slurm options Steps 1. Run as root on the frontend 2. Resolve uid from logged username 3. Submit job with #SBATCH --uid=$user 4. Job connect back to JupyterHub 5. Gracefully fails if no resources available Limitations Single instance per user Hardcoded parameters No Kerberos credentials (FS required)
19 Additional constraints BBP infrastructure is multi-sites (2 Slurm clusters) Frontend Geneva Frontend Lugano
20 Additional constraints BBP infrastructure is multi-sites (2 Slurm clusters) User needs access to Shared File Systems (Kerberos required) Frontend Frontend Geneva Lugano GPFS NFS Kerberos GPFS
21 Additional constraints BBP infrastructure is multi-sites (2 Slurm clusters) User needs access to Shared File Systems (Kerberos required) Frontend Frontend JupyterHub can t run on the frontend anymore...and should probably handle Kerberos AuthN Geneva Lugano GPFS NFS Kerberos GPFS
22 Site agnostic approach JupyterHub is installed out of any cluster And should securely submit a job (ssh as user + sbatch) JupyterHub Frontend Frontend Geneva Lugano GPFS NFS Kerberos GPFS
23 Authentication mechanism Step 1. Authentication KerberosPAMAuthenticator Username + password => kerberos token
24 Authentication mechanism Step 1. Authentication KerberosPAMAuthenticator Username + password => kerberos token Step 2. Job Submission Kerberos token forwarded using ssh (to frontend) auks (to compute)
25 Authentication mechanism Step 1. Authentication KerberosPAMAuthenticator Username + password => kerberos token Step 2. Job Submission Kerberos token forwarded using ssh (to frontend) auks (to compute) Step 3. Access notebook
26 Site agnostic approach JupyterHub Frontend ssh as the user + sbatch Frontend Geneva Lugano GPFS NFS Kerberos GPFS Limitation: User can no longer collaboratively work on a notebook
27 Advance use cases: Extending JupyterHub with custom kernels and modules
28 JupyterHub profiles BBP extension to abstract environment complexity and improve reproducibility
29 JupyterHub profiles BBP extension to abstract environment complexity and improve reproducibility Arbitrary name Optimal resources Environment: Python environment Set of modules
30 Custom virtual environments
31 Kernels Custom virtual environments
32 Kernels Custom virtual environments
33 Integration with Spark
34 Integration with Spark Wrapper around sbatch
35 Integration with Spark Wrapper around sbatch Spark code
36 Summary
37 Limitations and solutions JupyterHub does not support multiple singleuser process ØNon blocking limitation (and may be soon supported upstream) JupyterHub notebook are user specific (not collaborative) Ø copy on open notebooks Sharing and versioning of the notebook ØNbgallery (
38 Summary Ø Provide a secure way to interact with remote web-based interactive notebooks Ø Extend the tool to lower entry barrier to scientists and improve environment reproducibility Ø Integrate the tools with other technology to leverage parallel computing such as Spark
39 We are hiring Devops HPC specialists Storage specialists
40 Acknowledgements BBP core services & HPC teams
Automated Configuration and Administration of a Storage-class Memory System to Support Supercomputer-based Scientific Workflows
Automated Configuration and Administration of a Storage-class Memory System to Support Supercomputer-based Scientific Workflows J. Bernard 1, P. Morjan 2, B. Hagley 3, F. Delalondre 1, F. Schürmann 1,
More informationTeraflops of Jupyter: A Notebook Based Analysis Portal at BNL
Teraflops of Jupyter: A Notebook Based Analysis Portal at BNL Ofer Rind Spring HEPiX, Madison, WI May 17,2018 In collaboration with: Doug Benjamin, Costin Caramarcu, Zhihua Dong, Will Strecker-Kellogg,
More informationHigh Performance Computing Course Notes Grid Computing I
High Performance Computing Course Notes 2008-2009 2009 Grid Computing I Resource Demands Even as computer power, data storage, and communication continue to improve exponentially, resource capacities are
More informationAn Efficient Parallel Load-balancing Framework for Orthogonal Decomposition of Geometrical Data
ISC06 OVERVIEW SORT-BALANCE-SPLIT RESULTS An Efficient Parallel Load-balancing Framework for Orthogonal Decomposition of Geometrical Data Bruno R. C. Magalhães Farhan Tauheed Thomas Heinis Anastasia Ailamaki
More informationSCITAS. Scientific IT and Application support unit at EPFL. created in February today : 5 system engineers + 6 application experts SCITAS
SCITAS Scientific IT and Application support unit at EPFL created in February 2014 today : 5 system engineers + 6 application experts SCITAS SCITAS activities responsible for the central EPFL computing
More informationEmploying HPC DEEP-EST for HEP Data Analysis. Viktor Khristenko (CERN, DEEP-EST), Maria Girone (CERN)
Employing HPC DEEP-EST for HEP Data Analysis Viktor Khristenko (CERN, DEEP-EST), Maria Girone (CERN) 1 Outline The DEEP-EST Project Goals and Motivation HEP Data Analysis on HPC with Apache Spark on HPC
More informationXSEDE New User Training. Ritu Arora November 14, 2014
XSEDE New User Training Ritu Arora Email: rauta@tacc.utexas.edu November 14, 2014 1 Objectives Provide a brief overview of XSEDE Computational, Visualization and Storage Resources Extended Collaborative
More informationDeploying the TeraGrid PKI
Deploying the TeraGrid PKI Grid Forum Korea Winter Workshop December 1, 2003 Jim Basney Senior Research Scientist National Center for Supercomputing Applications University of Illinois jbasney@ncsa.uiuc.edu
More informationThe GISandbox: A Science Gateway For Geospatial Computing. Davide Del Vento, Eric Shook, Andrea Zonca
The GISandbox: A Science Gateway For Geospatial Computing Davide Del Vento, Eric Shook, Andrea Zonca 1 Paleoscape Model and Human Origins Simulate Climate and Vegetation during the Last Glacial Maximum
More informationPLAN-E Workshop Switzerland. Welcome! September 8, 2016
PLAN-E Workshop Switzerland Welcome! September 8, 2016 The Swiss National Supercomputing Centre Driving innovation in computational research in Switzerland Michele De Lorenzi (CSCS) PLAN-E September 8,
More informationSLURM Operation on Cray XT and XE
SLURM Operation on Cray XT and XE Morris Jette jette@schedmd.com Contributors and Collaborators This work was supported by the Oak Ridge National Laboratory Extreme Scale Systems Center. Swiss National
More informationTransient Compute ARC as Cloud Front-End
Digital Infrastructures for Research 2016 2016-09-29, 11:30, Cracow 30 min slot AEC ALBERT EINSTEIN CENTER FOR FUNDAMENTAL PHYSICS Transient Compute ARC as Cloud Front-End Sigve Haug, AEC-LHEP University
More informationKerberos & HPC Batch systems. Matthieu Hautreux (CEA/DAM/DIF)
Kerberos & HPC Batch systems Matthieu Hautreux (CEA/DAM/DIF) matthieu.hautreux@cea.fr Outline Kerberos authentication HPC site environment Kerberos & HPC systems AUKS From HPC site to HPC Grid environment
More informationCray XC Scalability and the Aries Network Tony Ford
Cray XC Scalability and the Aries Network Tony Ford June 29, 2017 Exascale Scalability Which scalability metrics are important for Exascale? Performance (obviously!) What are the contributing factors?
More informationReproducibility and Extensibility in Scientific Research. Jessica Forde
Reproducibility and Extensibility in Scientific Research Jessica Forde Project Jupyter @projectjupyter @mybinderteam Project Jupyter IPython Jupyter Notebook Architecture of JupyterHub Overview The problem
More informationLeveraging the InCommon Federation to access the NSF TeraGrid
Leveraging the InCommon Federation to access the NSF TeraGrid Jim Basney Senior Research Scientist National Center for Supercomputing Applications University of Illinois at Urbana-Champaign jbasney@ncsa.uiuc.edu
More informationSUG Breakout Session: OSC OnDemand App Development
SUG Breakout Session: OSC OnDemand App Development Basil Mohamed Gohar Web and Interface Applications Manager Eric Franz Senior Engineer & Technical Lead This work is supported by the National Science
More informationFacilitating Collaborative Analysis in SWAN
Facilitating Collaborative Analysis in SWAN E. Tejedor, D. Castro, D. Piparo, P. Mato E. Bocchi, J. Moscicki, M. Lamanna, P. Kothuri https://swan.cern.ch July 11th, 2018 CHEP 2018, Sofia (Bulgaria) Introduction
More informationAAI in EGI Current status
AAI in EGI Current status Peter Solagna EGI.eu Operations Manager www.egi.eu EGI-Engage is co-funded by the Horizon 2020 Framework Programme of the European Union under grant number 654142 User authentication
More informationLatest from the Lab: What's New Machine Learning Sam Buhler - Machine Learning Product/Offering Manager
Latest from the Lab: What's New Machine Learning Sam Buhler - Machine Learning Product/Offering Manager Please Note IBM s statements regarding its plans, directions, and intent are subject to change or
More informationIntroduction to Joker Cyber Infrastructure Architecture Team CIA.NMSU.EDU
Introduction to Joker Cyber Infrastructure Architecture Team CIA.NMSU.EDU What is Joker? NMSU s supercomputer. 238 core computer cluster. Intel E-5 Xeon CPUs and Nvidia K-40 GPUs. InfiniBand innerconnect.
More informationEnabling a SuperFacility with Software Defined Networking
Enabling a SuperFacility with Software Defined Networking Shane Canon Tina Declerck, Brent Draney, Jason Lee, David Paul, David Skinner May 2017 CUG 2017-1 - SuperFacility - Defined Combining the capabilities
More informationXcalar Installation Guide
Xcalar Installation Guide Publication date: 2018-03-16 www.xcalar.com Copyright 2018 Xcalar, Inc. All rights reserved. Table of Contents Xcalar installation overview 5 Audience 5 Overview of the Xcalar
More informationOutline. March 5, 2012 CIRMMT - McGill University 2
Outline CLUMEQ, Calcul Quebec and Compute Canada Research Support Objectives and Focal Points CLUMEQ Site at McGill ETS Key Specifications and Status CLUMEQ HPC Support Staff at McGill Getting Started
More informationVC3. Virtual Clusters for Community Computation. DOE NGNS PI Meeting September 27-28, 2017
VC3 Virtual Clusters for Community Computation DOE NGNS PI Meeting September 27-28, 2017 Douglas Thain, University of Notre Dame Rob Gardner, University of Chicago John Hover, Brookhaven National Lab A
More informationElastiCluster Automated provisioning of computational clusters in the cloud
ElastiCluster Automated provisioning of computational clusters in the cloud Riccardo Murri (with contributions from Antonio Messina, Nicolas Bär, Sergio Maffioletti, and Sigve
More informationOpenDreamKit. Computational environments for research and education Min Ragan-Kelley. Simula Research Lab
OpenDreamKit Computational environments for research and education Min Ragan-Kelley Simula Research Lab OpenDreamKit H2020 project Virtual Research Environments 16 Institutions Generic (Jupyter, SageMath)
More informationShort Talk: System abstractions to facilitate data movement in supercomputers with deep memory and interconnect hierarchy
Short Talk: System abstractions to facilitate data movement in supercomputers with deep memory and interconnect hierarchy François Tessier, Venkatram Vishwanath Argonne National Laboratory, USA July 19,
More informationRemote & Collaborative Visualization. Texas Advanced Computing Center
Remote & Collaborative Visualization Texas Advanced Computing Center TACC Remote Visualization Systems Longhorn NSF XD Dell Visualization Cluster 256 nodes, each 8 cores, 48 GB (or 144 GB) memory, 2 NVIDIA
More informationNow SAML takes it all:
Now SAML takes it all: Federation of non Web-based Services in the State of Baden-Württemberg Sebastian Labitzke Karlsruhe Institute of Technology (KIT) Steinbuch Centre for Computing (SCC) labitzke@kit.edu
More informationAllowing Users to Run Services at the OLCF with Kubernetes
Allowing Users to Run Services at the OLCF with Kubernetes Jason Kincl Senior HPC Systems Engineer Ryan Adamson Senior HPC Security Engineer This work was supported by the Oak Ridge Leadership Computing
More informationCloudExpo November 2017 Tomer Levi
CloudExpo November 2017 Tomer Levi About me Full Stack Engineer @ Intel s Advanced Analytics group. Artificial Intelligence unit at Intel. Responsible for (1) Radical improvement of critical processes
More informationParsl: Developing Interactive Parallel Workflows in Python using Parsl
Parsl: Developing Interactive Parallel Workflows in Python using Parsl Kyle Chard (chard@uchicago.edu) Yadu Babuji, Anna Woodard, Zhuozhao Li, Ben Clifford, Ian Foster, Dan Katz, Mike Wilde, Justin Wozniak
More informationThe RAMDISK Storage Accelerator
The RAMDISK Storage Accelerator A Method of Accelerating I/O Performance on HPC Systems Using RAMDISKs Tim Wickberg, Christopher D. Carothers wickbt@rpi.edu, chrisc@cs.rpi.edu Rensselaer Polytechnic Institute
More informationPart2: Let s pick one cloud IaaS middleware: OpenStack. Sergio Maffioletti
S3IT: Service and Support for Science IT Cloud middleware Part2: Let s pick one cloud IaaS middleware: OpenStack Sergio Maffioletti S3IT: Service and Support for Science IT, University of Zurich http://www.s3it.uzh.ch/
More informationHigh Performance Computing Course Notes HPC Fundamentals
High Performance Computing Course Notes 2008-2009 2009 HPC Fundamentals Introduction What is High Performance Computing (HPC)? Difficult to define - it s a moving target. Later 1980s, a supercomputer performs
More informationDARPA-BAA Hierarchical Identify Verify Exploit (HIVE) Frequently Asked Questions (FAQ) August 18, 2016
DARPA-BAA-16-52 Hierarchical Identify Verify Exploit (HIVE) Frequently Asked Questions (FAQ) August 18, 2016 DARPA-BAA-16-52 Hierarchical Identify Verify Exploit (HIVE) Frequently Asked Questions (FAQ)
More informationTools for Handling Big Data and Compu5ng Demands. Humani5es and Social Science Scholars
Tools for Handling Big Data and Compu5ng Demands Humani5es and Social Science Scholars Outline Overview of Compute Canada and WestGrid Focus on Humani5es and Social Sciences The Resource Alloca5on Compe55on
More informationIntroduction to BioHPC
Introduction to BioHPC New User Training [web] [email] portal.biohpc.swmed.edu biohpc-help@utsouthwestern.edu 1 Updated for 2015-06-03 Overview Today we re going to cover: What is BioHPC? How do I access
More informationBlue Gene/Q. Hardware Overview Michael Stephan. Mitglied der Helmholtz-Gemeinschaft
Blue Gene/Q Hardware Overview 02.02.2015 Michael Stephan Blue Gene/Q: Design goals System-on-Chip (SoC) design Processor comprises both processing cores and network Optimal performance / watt ratio Small
More informationSlurm Overview. Brian Christiansen, Marshall Garey, Isaac Hartung SchedMD SC17. Copyright 2017 SchedMD LLC
Slurm Overview Brian Christiansen, Marshall Garey, Isaac Hartung SchedMD SC17 Outline Roles of a resource manager and job scheduler Slurm description and design goals Slurm architecture and plugins Slurm
More informationParallel I/O on JUQUEEN
Parallel I/O on JUQUEEN 4. Februar 2014, JUQUEEN Porting and Tuning Workshop Mitglied der Helmholtz-Gemeinschaft Wolfgang Frings w.frings@fz-juelich.de Jülich Supercomputing Centre Overview Parallel I/O
More informationSupercomputer and grid infrastructure! in Poland!
Supercomputer and grid infrastructure in Poland Franciszek Rakowski Interdisciplinary Centre for Mathematical and Computational Modelling 12th INCF Nodes Workshop, 16.04.2015 Warsaw, Nencki Institute.
More informationA Laconic HPC with an Orgone Accumulator. Presentation to Multicore World Wellington, February 15-17,
A Laconic HPC with an Orgone Accumulator Presentation to Multicore World 2016 Wellington, February 15-17, 2016 http://levlafayette.com Edward - University of Melbourne Cluster - System Installed and operational
More informationVisualiza(on So-ware and Hardware for In- Silico Brain Research. Stefan Eilemann Visualiza0on Team Lead Blue Brain Project, EPFL
Visualiza(on So-ware and Hardware for In- Silico Brain Research Stefan Eilemann Visualiza0on Team Lead Blue Brain Project, EPFL Blue Brain Project / Human Brain Project BBP: Swiss na0onal research project
More informationShared Services Canada Environment and Climate Change Canada HPC Renewal Project
Shared Services Canada Environment and Climate Change Canada HPC Renewal Project CUG 2017 Redmond, WA, USA Deric Sullivan Alain St-Denis & Luc Corbeil May 2017 Background: SSC's HPC Renewal for ECCC Environment
More informationHPC IN EUROPE. Organisation of public HPC resources
HPC IN EUROPE Organisation of public HPC resources Context Focus on publicly-funded HPC resources provided primarily to enable scientific research and development at European universities and other publicly-funded
More informationIBM Data Science Experience (DSX) Partner Application Validation Quick Guide
IBM Data Science Experience (DSX) Partner Application Validation Quick Guide VERSION: 1.6 DATE: Sept 27, 2017 EDITOR: D. Rangarao Table of Contents 1 Overview of the Application Validation Process... 3
More informationManaging Grid Credentials
Managing Grid Credentials Jim Basney http://www.ncsa.uiuc.edu/~jbasney/ Senior Research Scientist Grid and Security Technologies National Center for Supercomputing Applications
More informationIntegration of CASA with Jupyter for efficient remote processing Aard Keimpema
Astronomy ESFRI & Research Infrastructure Cluster ASTERICS - 653477 Integration of CASA with Jupyter for efficient remote processing Aard Keimpema (keimpema@jive.eu) H2020-Astronomy ESFRI and Research
More informationIBM Data Science Experience (DSX) Partner Application Validation Quick Guide
IBM Data Science Experience (DSX) Partner Application Validation Quick Guide VERSION: 2.0 DATE: Feb 15, 2018 EDITOR: D. Rangarao Table of Contents 1 Overview of the Application Validation Process... 3
More informationNERSC Site Update. National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory. Richard Gerber
NERSC Site Update National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory Richard Gerber NERSC Senior Science Advisor High Performance Computing Department Head Cori
More informationUsing Two-Factor Authentication to Connect to a Kerberos-enabled Informatica Domain
Using Two-Factor Authentication to Connect to a Kerberos-enabled Informatica Domain Copyright Informatica LLC 2016, 2018. Informatica LLC. No part of this document may be reproduced or transmitted in any
More informationCornell Red Cloud: Campus-based Hybrid Cloud. Steven Lee Cornell University Center for Advanced Computing
Cornell Red Cloud: Campus-based Hybrid Cloud Steven Lee Cornell University Center for Advanced Computing shl1@cornell.edu Cornell Center for Advanced Computing (CAC) Profile CAC mission, impact on research
More informationLeveraging Globus Identity for the Grid. Suchandra Thapa GlobusWorld, April 22, 2016 Chicago
Leveraging Globus Identity for the Grid Suchandra Thapa GlobusWorld, April 22, 2016 Chicago Open Science Grid Helps researchers speed up their research using high throughput computing methods Helps campus
More information2017 Resource Allocations Competition Results
2017 Resource Allocations Competition Results Table of Contents Executive Summary...3 Computational Resources...5 CPU Allocations...5 GPU Allocations...6 Cloud Allocations...6 Storage Resources...6 Acceptance
More informationMULTITHERMAN: Out-of-band High-Resolution HPC Power and Performance Monitoring Support for Big-Data Analysis
MULTITHERMAN: Out-of-band High-Resolution HPC Power and Performance Monitoring Support for Big-Data Analysis EU H2020 FETHPC project ANTAREX (g.a. 671623) EU FP7 ERC Project MULTITHERMAN (g.a.291125) HPC
More informationCONFIGURING IBM STORWIZE. for Metadata Framework 6.3
CONFIGURING IBM STORWIZE for Metadata Framework 6.3 Publishing Information Software version 6.3.160 Document version 4 Publication date May 22, 2017 Copyright 2005-2017 Varonis Systems Inc. All rights
More informationDuo End User Education Templates
Duo End User Education Email Templates Table of Contents Email Communication Best Practices Email Templates Email #1 - Duo is coming soon, no immediate action required. Email #2 - Duo is coming on ,
More informationLoad Balancing Censornet USS Gateway. Deployment Guide v Copyright Loadbalancer.org
Load Balancing Censornet USS Gateway Deployment Guide v1.0.0 Copyright Loadbalancer.org Table of Contents 1. About this Guide...3 2. Loadbalancer.org Appliances Supported...3 3. Loadbalancer.org Software
More informationLeveraging the Globus Platform in your Web Applications. GlobusWorld April 26, 2018 Greg Nawrocki
Leveraging the Globus Platform in your Web Applications GlobusWorld April 26, 2018 Greg Nawrocki greg@globus.org Topics and Goals Platform Overview Why expose the APIs A quick touch of the Globus Auth
More informationThe LGI Pilot job portal. EGI Technical Forum 20 September 2011 Jan Just Keijser Willem van Engen Mark Somers
The LGI Pilot job portal EGI Technical Forum 20 September 2011 Jan Just Keijser Willem van Engen Mark Somers Outline What? Why? How? Pro's and Cons What's next? Credits 2 What is LGI? LGI Project Server
More informationTranscode and Add Pulse Video Analytics to Video Content on Cisco Show and Share
Transcode and Add Pulse Video Analytics to Video Content on Cisco Show and Share Revised: February 26, 2013 When Cisco MXE 3500 integration is enabled in Cisco Show and Share, videos uploaded to Cisco
More informationIntroduction to Slurm
Introduction to Slurm Tim Wickberg SchedMD Slurm User Group Meeting 2017 Outline Roles of resource manager and job scheduler Slurm description and design goals Slurm architecture and plugins Slurm configuration
More informationSpiNNaker a Neuromorphic Supercomputer. Steve Temple University of Manchester, UK SOS21-21 Mar 2017
SpiNNaker a Neuromorphic Supercomputer Steve Temple University of Manchester, UK SOS21-21 Mar 2017 Outline of talk Introduction Modelling neurons Architecture and technology Principles of operation Summary
More informationAmbry: LinkedIn s Scalable Geo- Distributed Object Store
Ambry: LinkedIn s Scalable Geo- Distributed Object Store Shadi A. Noghabi *, Sriram Subramanian +, Priyesh Narayanan +, Sivabalan Narayanan +, Gopalakrishna Holla +, Mammad Zadeh +, Tianwei Li +, Indranil
More informationArcGIS for Server: Administration and Security. Amr Wahba
ArcGIS for Server: Administration and Security Amr Wahba awahba@esri.com Agenda ArcGIS Server architecture Distributing and scaling components Implementing security Monitoring server logs Automating server
More informationCSCS Site Update. HPC Advisory Council Workshop Colin McMurtrie, Associate Director and Head of HPC Operations.
CSCS Site Update HPC Advisory Council Workshop 2018 Colin McMurtrie, Associate Director and Head of HPC Operations. 9th April 2018 The Swiss National Supercomputing Centre Driving innovation in computational
More informationglite Grid Services Overview
The EPIKH Project (Exchange Programme to advance e-infrastructure Know-How) glite Grid Services Overview Antonio Calanducci INFN Catania Joint GISELA/EPIKH School for Grid Site Administrators Valparaiso,
More informationUsing Cartesius and Lisa. Zheng Meyer-Zhao - Consultant Clustercomputing
Zheng Meyer-Zhao - zheng.meyer-zhao@surfsara.nl Consultant Clustercomputing Outline SURFsara About us What we do Cartesius and Lisa Architectures and Specifications File systems Funding Hands-on Logging
More informationDeploying a Production Gateway with Airavata
Deploying a Production Gateway with Airavata Table of Contents Pre-requisites... 1 Create a Gateway Request... 1 Gateway Deploy Steps... 2 Install Ansible & Python...2 Deploy the Gateway...3 Gateway Configuration...
More informationRHRK-Seminar. High Performance Computing with the Cluster Elwetritsch - II. Course instructor : Dr. Josef Schüle, RHRK
RHRK-Seminar High Performance Computing with the Cluster Elwetritsch - II Course instructor : Dr. Josef Schüle, RHRK Overview Course I Login to cluster SSH RDP / NX Desktop Environments GNOME (default)
More informationResearch Collection. WebParFE A web interface for the high performance parallel finite element solver ParFE. Report. ETH Library
Research Collection Report WebParFE A web interface for the high performance parallel finite element solver ParFE Author(s): Paranjape, Sumit; Kaufmann, Martin; Arbenz, Peter Publication Date: 2009 Permanent
More information1 Bull, 2011 Bull Extreme Computing
1 Bull, 2011 Bull Extreme Computing Table of Contents Overview. Principal concepts. Architecture. Scheduler Policies. 2 Bull, 2011 Bull Extreme Computing SLURM Overview Ares, Gerardo, HPC Team Introduction
More informationIntroduction to BioHPC New User Training
Introduction to BioHPC New User Training [web] [email] portal.biohpc.swmed.edu biohpc-help@utsouthwestern.edu 1 Updated for 2019-02-06 Overview Today we re going to cover: What is BioHPC? How do I access
More informationAdvanced School in High Performance and GRID Computing November Introduction to Grid computing.
1967-14 Advanced School in High Performance and GRID Computing 3-14 November 2008 Introduction to Grid computing. TAFFONI Giuliano Osservatorio Astronomico di Trieste/INAF Via G.B. Tiepolo 11 34131 Trieste
More informationIntroduction to High Performance Computing and an Statistical Genetics Application on the Janus Supercomputer. Purpose
Introduction to High Performance Computing and an Statistical Genetics Application on the Janus Supercomputer Daniel Yorgov Department of Mathematical & Statistical Sciences, University of Colorado Denver
More informationBatch Services at CERN: Status and Future Evolution
Batch Services at CERN: Status and Future Evolution Helge Meinhard, CERN-IT Platform and Engineering Services Group Leader HTCondor Week 20 May 2015 20-May-2015 CERN batch status and evolution - Helge
More informationLeonhard: a new cluster for Big Data at ETH
Leonhard: a new cluster for Big Data at ETH Bernd Rinn, Head of Scientific IT Services Olivier Byrde, Group leader High Performance Computing Bernd Rinn & Olivier Byrde 2017-02-15 1 Agenda Welcome address
More informationThe Cambridge Bio-Medical-Cloud An OpenStack platform for medical analytics and biomedical research
The Cambridge Bio-Medical-Cloud An OpenStack platform for medical analytics and biomedical research Dr Paul Calleja Director of Research Computing University of Cambridge Global leader in science & technology
More informationSAS Grid Manager and Kerberos Authentication
SAS Grid Manager and Kerberos Authentication Learn the considerations for implementing Kerberos authentication so you can submit workload to SAS Grid Manager. SAS Grid Manager and Kerberos Authentication
More informationFederated Services and Data Management in PRACE
Federated Services and Data Management in PRACE G. Erbacci, (CINECA and PRACE) Collaborative Services for e-infrastructure Commons e-irg Workshop, Bratislava 15-16 November 2016 Partnership for Advanced
More informationIntroduction to BioHPC
Introduction to BioHPC New User Training [web] [email] portal.biohpc.swmed.edu biohpc-help@utsouthwestern.edu 1 Updated for 2018-03-07 Overview Today we re going to cover: What is BioHPC? How do I access
More informationLeveraging the Globus Platform in your Web Applications
Leveraging the Globus Platform in your Web Applications Steve Tuecke tuecke@uchicago.edu NCAR September 5, 2018 Globus serves as A platform for building science gateways, web portals and other applications
More informationHigh Performance Computing Course Notes Course Administration
High Performance Computing Course Notes 2009-2010 2010 Course Administration Contacts details Dr. Ligang He Home page: http://www.dcs.warwick.ac.uk/~liganghe Email: liganghe@dcs.warwick.ac.uk Office hours:
More informationOur Workshop Environment
Our Workshop Environment John Urbanic Parallel Computing Scientist Pittsburgh Supercomputing Center Copyright 2018 Our Environment Today Your laptops or workstations: only used for portal access Bridges
More informationFederated XDMoD Requirements
Federated XDMoD Requirements Date Version Person Change 2016-04-08 1.0 draft XMS Team Initial version Summary Definitions Assumptions Data Collection Local XDMoD Installation Module Support Data Federation
More informationXSEDE Infrastructure as a Service Use Cases
XSEDE Infrastructure as a Service Use Cases October 30, 2016 Version 2.2 This document outlines typical XSEDE user needs in the Infrastructure as a Service (IaaS) area. IAAS 1 through IAAS 4 describe how
More informationAnsible Tower Quick Setup Guide
Ansible Tower Quick Setup Guide Release Ansible Tower 2.4.5 Red Hat, Inc. Jun 06, 2017 CONTENTS 1 Quick Start 2 2 Login as a Superuser 3 3 Import a License 4 4 Examine the Tower Dashboard 6 5 The Setup
More informationRENKU - Reproduce, Reuse, Recycle Research. Rok Roškar and the SDSC Renku team
RENKU - Reproduce, Reuse, Recycle Research Rok Roškar and the SDSC Renku team Renku-Reana workshop @ CERN 26.06.2018 Goals of Renku 1. Provide the means to create reproducible data science 2. Facilitate
More informationNote: Who is Dr. Who? You may notice that YARN says you are logged in as dr.who. This is what is displayed when user
Run a YARN Job Exercise Dir: ~/labs/exercises/yarn Data Files: /smartbuy/kb In this exercise you will submit an application to the YARN cluster, and monitor the application using both the Hue Job Browser
More informationAdding Cloud Based Interactive Compute Capabilities to Globus Endpoints
Adding Cloud Based Interactive Compute Capabilities to Globus Endpoints Ben Galewsky Research Programmer, National Center for Supercomputing Applications bengal1@illinois.edu http://www.nationaldataservice.org/
More informationApplications Software Example
Applications Software Example How to run an application on Cluster? Rooh Khurram Supercomputing Laboratory King Abdullah University of Science and Technology (KAUST), Saudi Arabia Cluster Training: Applications
More informationvcloud Director Administrator's Guide
vcloud Director 5.5 This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more recent editions of
More informationContinuous Integration using Docker & Jenkins
Jenkins LinuxCon Europe 2014 October 13-15, 2014 Mattias Giese Solutions Architect giese@b1-systems.de - Linux/Open Source Consulting, Training, Support & Development Introducing B1 Systems founded in
More informationGPFS for Life Sciences at NERSC
GPFS for Life Sciences at NERSC A NERSC & JGI collaborative effort Jason Hick, Rei Lee, Ravi Cheema, and Kjiersten Fagnan GPFS User Group meeting May 20, 2015-1 - Overview of Bioinformatics - 2 - A High-level
More informationQuick Start Guide. Table of Contents
Quick Start Guide Table of Contents Account Registration... 2 Signup Request... 2 Account Activation... 4 Running FLOW-3D on POD... 9 Launching the GUI... 9 Running Simulations... 11 Collaborating with
More informationThe ASCI/DOD Scalable I/O History and Strategy Run Time Systems and Scalable I/O Team Gary Grider CCN-8 Los Alamos National Laboratory LAUR
The ASCI/DOD Scalable I/O History and Strategy Run Time Systems and Scalable I/O Team Gary Grider CCN-8 Los Alamos National Laboratory LAUR 042787 05/2004 Parallel File Systems and Parallel I/O Why - From
More informationAn Overview of the Architecture of Juno: CHPC s New JupyterHub Service By Luan Truong, CHPC, University of Utah
An Overview of the Architecture of Juno: CHPC s New JupyterHub Service By Luan Truong, CHPC, University of Utah Introduction Jupyter notebooks have emerged as a popular and open-source web application
More informationFuncX: A Function Serving Platform for HPC. Ryan Chard 28 Jan 2019
FuncX: A Function Serving Platform for HPC Ryan Chard 28 Jan 2019 Outline - Motivation FuncX: FaaS for HPC Implementation status Preliminary applications - Machine learning inference Automating analysis
More information