Enabling web-based interactive notebooks on geographically distributed HPC resources. Alexandre Beche

Size: px
Start display at page:

Download "Enabling web-based interactive notebooks on geographically distributed HPC resources. Alexandre Beche"

Transcription

1 Enabling web-based interactive notebooks on geographically distributed HPC resources Alexandre Beche

2 Outlines 1. Context 2. Interactive notebook running on cluster(s) 3. Advanced use cases 4. Summary

3 Context

4 Blue Brain Project Host Institution Main International Collaborations Ecole Polytechnique Fédérale de Lausanne (EPFL) Switzerland (CSCS, CERN) Israel (HUJI) Director USA (Yale, ANL, OLCF, Allen Brain) Henry Markram Spain (UPM) Co-Directors Saudi Arabia (KAUST) Sean Hill, Felix Schürmann Europe (HBP) Team today ~ 100 scientists, engineers & staff Timeline 2005 founded at EPFL 2011/2012 ETH Board funding Swiss National Research Infrastructure Geneva

5 Data-Driven Modeling & Simulation! I 11 electrical types ~ 2 mm thick 55 morphological types 13 excitatory & 42 inhibitory m-types 31,000 neurons 111,700 neurons/mm3 Excitatory to inhibitory neuron ratio of 86:14 % 346 m of axon 211 m of dendrites Maximum branch order of m-types: Ion channel density distribution profiles: 6 ite a m D ite on Markram et al, Cell en dr Ax 6 s 8 dr Uniform en IV So 17 morpho-electrical types n bap & EPSP attenuation for 207 Ax o 24 Inhibitory 13 HH type ion channel models s Excitatory II/III 207 morpho-electrical types D V 0.63 synapses/mm3 Extrinsic to intrinsic synapse ratio of 75:25 % 3025 possible synaptic pathways 2258 viable synaptic pathways 664 excitatory pathways 1594 inhibitory pathways 600 intra-laminar pathways 1658 inter-laminar pathways Mean synapses/connection ns 274 ns ry to bi hi In ci ta to Ex 100 µm ry h. In Ex c. 4.3 VI 6 synapse types 207 synaptomes Space clamp corrected synaptic conductances for 607 pathways The per synapse conductance of 1.5 ns for connections between L5TTPCs is the highest in the microrcircuit Mean conductance per synapse: 0.85 ns for excitatory & 0.66 ns for inhibitory synapses Total conductance in a single neuron is 971 ns Ø 80 authors Ø Joint effort between computer and neuroscientists Ø Reproducible work Ø Extensible

6 HPC Today s Infrastructure Elastic Compute Visualization and analysis Web services SW development Continuous Integration Continuous Deployment IaaS 4 nodes Viz & Analytics 36 nodes CEPH IaaS 12 nodes NetApp CEPH Viz & Analytics 14 nodes GPFS GVA Production HPC Model development Reconstruction Simulation SW development Systems Research Memory extension Application coupling Interactive supercomputing Reproducibility GPFS 4 Compute Racks 4096 Compute Nodes 5D CN torus 0.8 PF/s peak 64TB DRAM 4.2 PB GPFS storage IBM BlueGene/Q 64 IONodes 3D torus 128TB Flash Linux IBM BlueGene Active Storage

7 HPC Today s Infrastructure Elastic Compute Visualization and analysis Web services SW development Continuous Integration Continuous Deployment IaaS 4 nodes Viz & Analytics 36 nodes CEPH IaaS 12 nodes NetApp CEPH Viz & Analytics 14 nodes GPFS GVA Production HPC Infrastructure renewal ongoing4 Compute Racks 4096 Compute Nodes New HPC system planned for Q D CN torus GPFS 0.8 PF/s peak 64TB DRAM 4.2 PB GPFS storage Model development Reconstruction Simulation SW development Systems Research Memory extension Application coupling Interactive supercomputing Reproducibility IBM BlueGene/Q 64 IONodes 3D torus 128TB Flash Linux IBM BlueGene Active Storage

8 Interactive notebook running on cluster(s)

9 Problem description Scientist wants to run a web-based interactive notebook 1) Connect to a cluster frontend 2) Get an allocation ssh salloc

10 Problem description Scientist wants to run a web-based interactive notebook 1) Connect to a cluster frontend 2) Get an allocation 3) Run the Jupyter notebook ssh salloc

11 Problem description Scientist wants to run a web-based interactive notebook 1) Connect to a cluster frontend 2) Get an allocation 3) Run the Jupyter notebook 4) Connect to the notebook http ssh salloc Jupyter notebook

12 Problem description Scientist wants to run a web-based interactive notebook 1) Connect to a cluster frontend 2) Get an allocation 3) Run the Jupyter notebook 4) Connect to the notebook 5) Share the URL with colleagues http ssh salloc Jupyter notebook http

13 Problem description Scientist wants to run a web-based interactive notebook 1) Connect to a cluster frontend 2) Get an allocation 3) Run the Jupyter notebook 4) Connect to the notebook 5) Share the URL with colleagues http ssh salloc Arbitrary code can now be run on behalf of the user... And access to data Jupyter notebook http

14 JupyterHub Multi-user server for Jupyter notebooks Authenticate users Spawn single-user Jupyter notebook Proxy user traffic to notebook securely

15 JupyterHub Multi-user server for Jupyter notebooks Authenticate users Spawn single-user Jupyter notebook Proxy user traffic to notebook securely https Authenticator: Spawner: pam + sssd custom slurm spawner

16 Basic approach Based on JupyterHub configuration file Custom Spawner Slurm options

17 Basic approach Based on JupyterHub configuration file Custom Spawner Slurm options Steps 1. Run as root on the frontend 2. Resolve uid from logged username 3. Submit job with #SBATCH --uid=$user 4. Job connect back to JupyterHub 5. Gracefully fails if no resources available

18 Basic approach Based on JupyterHub configuration file Custom Spawner Slurm options Steps 1. Run as root on the frontend 2. Resolve uid from logged username 3. Submit job with #SBATCH --uid=$user 4. Job connect back to JupyterHub 5. Gracefully fails if no resources available Limitations Single instance per user Hardcoded parameters No Kerberos credentials (FS required)

19 Additional constraints BBP infrastructure is multi-sites (2 Slurm clusters) Frontend Geneva Frontend Lugano

20 Additional constraints BBP infrastructure is multi-sites (2 Slurm clusters) User needs access to Shared File Systems (Kerberos required) Frontend Frontend Geneva Lugano GPFS NFS Kerberos GPFS

21 Additional constraints BBP infrastructure is multi-sites (2 Slurm clusters) User needs access to Shared File Systems (Kerberos required) Frontend Frontend JupyterHub can t run on the frontend anymore...and should probably handle Kerberos AuthN Geneva Lugano GPFS NFS Kerberos GPFS

22 Site agnostic approach JupyterHub is installed out of any cluster And should securely submit a job (ssh as user + sbatch) JupyterHub Frontend Frontend Geneva Lugano GPFS NFS Kerberos GPFS

23 Authentication mechanism Step 1. Authentication KerberosPAMAuthenticator Username + password => kerberos token

24 Authentication mechanism Step 1. Authentication KerberosPAMAuthenticator Username + password => kerberos token Step 2. Job Submission Kerberos token forwarded using ssh (to frontend) auks (to compute)

25 Authentication mechanism Step 1. Authentication KerberosPAMAuthenticator Username + password => kerberos token Step 2. Job Submission Kerberos token forwarded using ssh (to frontend) auks (to compute) Step 3. Access notebook

26 Site agnostic approach JupyterHub Frontend ssh as the user + sbatch Frontend Geneva Lugano GPFS NFS Kerberos GPFS Limitation: User can no longer collaboratively work on a notebook

27 Advance use cases: Extending JupyterHub with custom kernels and modules

28 JupyterHub profiles BBP extension to abstract environment complexity and improve reproducibility

29 JupyterHub profiles BBP extension to abstract environment complexity and improve reproducibility Arbitrary name Optimal resources Environment: Python environment Set of modules

30 Custom virtual environments

31 Kernels Custom virtual environments

32 Kernels Custom virtual environments

33 Integration with Spark

34 Integration with Spark Wrapper around sbatch

35 Integration with Spark Wrapper around sbatch Spark code

36 Summary

37 Limitations and solutions JupyterHub does not support multiple singleuser process ØNon blocking limitation (and may be soon supported upstream) JupyterHub notebook are user specific (not collaborative) Ø copy on open notebooks Sharing and versioning of the notebook ØNbgallery (

38 Summary Ø Provide a secure way to interact with remote web-based interactive notebooks Ø Extend the tool to lower entry barrier to scientists and improve environment reproducibility Ø Integrate the tools with other technology to leverage parallel computing such as Spark

39 We are hiring Devops HPC specialists Storage specialists

40 Acknowledgements BBP core services & HPC teams

Automated Configuration and Administration of a Storage-class Memory System to Support Supercomputer-based Scientific Workflows

Automated Configuration and Administration of a Storage-class Memory System to Support Supercomputer-based Scientific Workflows Automated Configuration and Administration of a Storage-class Memory System to Support Supercomputer-based Scientific Workflows J. Bernard 1, P. Morjan 2, B. Hagley 3, F. Delalondre 1, F. Schürmann 1,

More information

Teraflops of Jupyter: A Notebook Based Analysis Portal at BNL

Teraflops of Jupyter: A Notebook Based Analysis Portal at BNL Teraflops of Jupyter: A Notebook Based Analysis Portal at BNL Ofer Rind Spring HEPiX, Madison, WI May 17,2018 In collaboration with: Doug Benjamin, Costin Caramarcu, Zhihua Dong, Will Strecker-Kellogg,

More information

High Performance Computing Course Notes Grid Computing I

High Performance Computing Course Notes Grid Computing I High Performance Computing Course Notes 2008-2009 2009 Grid Computing I Resource Demands Even as computer power, data storage, and communication continue to improve exponentially, resource capacities are

More information

An Efficient Parallel Load-balancing Framework for Orthogonal Decomposition of Geometrical Data

An Efficient Parallel Load-balancing Framework for Orthogonal Decomposition of Geometrical Data ISC06 OVERVIEW SORT-BALANCE-SPLIT RESULTS An Efficient Parallel Load-balancing Framework for Orthogonal Decomposition of Geometrical Data Bruno R. C. Magalhães Farhan Tauheed Thomas Heinis Anastasia Ailamaki

More information

SCITAS. Scientific IT and Application support unit at EPFL. created in February today : 5 system engineers + 6 application experts SCITAS

SCITAS. Scientific IT and Application support unit at EPFL. created in February today : 5 system engineers + 6 application experts SCITAS SCITAS Scientific IT and Application support unit at EPFL created in February 2014 today : 5 system engineers + 6 application experts SCITAS SCITAS activities responsible for the central EPFL computing

More information

Employing HPC DEEP-EST for HEP Data Analysis. Viktor Khristenko (CERN, DEEP-EST), Maria Girone (CERN)

Employing HPC DEEP-EST for HEP Data Analysis. Viktor Khristenko (CERN, DEEP-EST), Maria Girone (CERN) Employing HPC DEEP-EST for HEP Data Analysis Viktor Khristenko (CERN, DEEP-EST), Maria Girone (CERN) 1 Outline The DEEP-EST Project Goals and Motivation HEP Data Analysis on HPC with Apache Spark on HPC

More information

XSEDE New User Training. Ritu Arora November 14, 2014

XSEDE New User Training. Ritu Arora   November 14, 2014 XSEDE New User Training Ritu Arora Email: rauta@tacc.utexas.edu November 14, 2014 1 Objectives Provide a brief overview of XSEDE Computational, Visualization and Storage Resources Extended Collaborative

More information

Deploying the TeraGrid PKI

Deploying the TeraGrid PKI Deploying the TeraGrid PKI Grid Forum Korea Winter Workshop December 1, 2003 Jim Basney Senior Research Scientist National Center for Supercomputing Applications University of Illinois jbasney@ncsa.uiuc.edu

More information

The GISandbox: A Science Gateway For Geospatial Computing. Davide Del Vento, Eric Shook, Andrea Zonca

The GISandbox: A Science Gateway For Geospatial Computing. Davide Del Vento, Eric Shook, Andrea Zonca The GISandbox: A Science Gateway For Geospatial Computing Davide Del Vento, Eric Shook, Andrea Zonca 1 Paleoscape Model and Human Origins Simulate Climate and Vegetation during the Last Glacial Maximum

More information

PLAN-E Workshop Switzerland. Welcome! September 8, 2016

PLAN-E Workshop Switzerland. Welcome! September 8, 2016 PLAN-E Workshop Switzerland Welcome! September 8, 2016 The Swiss National Supercomputing Centre Driving innovation in computational research in Switzerland Michele De Lorenzi (CSCS) PLAN-E September 8,

More information

SLURM Operation on Cray XT and XE

SLURM Operation on Cray XT and XE SLURM Operation on Cray XT and XE Morris Jette jette@schedmd.com Contributors and Collaborators This work was supported by the Oak Ridge National Laboratory Extreme Scale Systems Center. Swiss National

More information

Transient Compute ARC as Cloud Front-End

Transient Compute ARC as Cloud Front-End Digital Infrastructures for Research 2016 2016-09-29, 11:30, Cracow 30 min slot AEC ALBERT EINSTEIN CENTER FOR FUNDAMENTAL PHYSICS Transient Compute ARC as Cloud Front-End Sigve Haug, AEC-LHEP University

More information

Kerberos & HPC Batch systems. Matthieu Hautreux (CEA/DAM/DIF)

Kerberos & HPC Batch systems. Matthieu Hautreux (CEA/DAM/DIF) Kerberos & HPC Batch systems Matthieu Hautreux (CEA/DAM/DIF) matthieu.hautreux@cea.fr Outline Kerberos authentication HPC site environment Kerberos & HPC systems AUKS From HPC site to HPC Grid environment

More information

Cray XC Scalability and the Aries Network Tony Ford

Cray XC Scalability and the Aries Network Tony Ford Cray XC Scalability and the Aries Network Tony Ford June 29, 2017 Exascale Scalability Which scalability metrics are important for Exascale? Performance (obviously!) What are the contributing factors?

More information

Reproducibility and Extensibility in Scientific Research. Jessica Forde

Reproducibility and Extensibility in Scientific Research. Jessica Forde Reproducibility and Extensibility in Scientific Research Jessica Forde Project Jupyter @projectjupyter @mybinderteam Project Jupyter IPython Jupyter Notebook Architecture of JupyterHub Overview The problem

More information

Leveraging the InCommon Federation to access the NSF TeraGrid

Leveraging the InCommon Federation to access the NSF TeraGrid Leveraging the InCommon Federation to access the NSF TeraGrid Jim Basney Senior Research Scientist National Center for Supercomputing Applications University of Illinois at Urbana-Champaign jbasney@ncsa.uiuc.edu

More information

SUG Breakout Session: OSC OnDemand App Development

SUG Breakout Session: OSC OnDemand App Development SUG Breakout Session: OSC OnDemand App Development Basil Mohamed Gohar Web and Interface Applications Manager Eric Franz Senior Engineer & Technical Lead This work is supported by the National Science

More information

Facilitating Collaborative Analysis in SWAN

Facilitating Collaborative Analysis in SWAN Facilitating Collaborative Analysis in SWAN E. Tejedor, D. Castro, D. Piparo, P. Mato E. Bocchi, J. Moscicki, M. Lamanna, P. Kothuri https://swan.cern.ch July 11th, 2018 CHEP 2018, Sofia (Bulgaria) Introduction

More information

AAI in EGI Current status

AAI in EGI Current status AAI in EGI Current status Peter Solagna EGI.eu Operations Manager www.egi.eu EGI-Engage is co-funded by the Horizon 2020 Framework Programme of the European Union under grant number 654142 User authentication

More information

Latest from the Lab: What's New Machine Learning Sam Buhler - Machine Learning Product/Offering Manager

Latest from the Lab: What's New Machine Learning Sam Buhler - Machine Learning Product/Offering Manager Latest from the Lab: What's New Machine Learning Sam Buhler - Machine Learning Product/Offering Manager Please Note IBM s statements regarding its plans, directions, and intent are subject to change or

More information

Introduction to Joker Cyber Infrastructure Architecture Team CIA.NMSU.EDU

Introduction to Joker Cyber Infrastructure Architecture Team CIA.NMSU.EDU Introduction to Joker Cyber Infrastructure Architecture Team CIA.NMSU.EDU What is Joker? NMSU s supercomputer. 238 core computer cluster. Intel E-5 Xeon CPUs and Nvidia K-40 GPUs. InfiniBand innerconnect.

More information

Enabling a SuperFacility with Software Defined Networking

Enabling a SuperFacility with Software Defined Networking Enabling a SuperFacility with Software Defined Networking Shane Canon Tina Declerck, Brent Draney, Jason Lee, David Paul, David Skinner May 2017 CUG 2017-1 - SuperFacility - Defined Combining the capabilities

More information

Xcalar Installation Guide

Xcalar Installation Guide Xcalar Installation Guide Publication date: 2018-03-16 www.xcalar.com Copyright 2018 Xcalar, Inc. All rights reserved. Table of Contents Xcalar installation overview 5 Audience 5 Overview of the Xcalar

More information

Outline. March 5, 2012 CIRMMT - McGill University 2

Outline. March 5, 2012 CIRMMT - McGill University 2 Outline CLUMEQ, Calcul Quebec and Compute Canada Research Support Objectives and Focal Points CLUMEQ Site at McGill ETS Key Specifications and Status CLUMEQ HPC Support Staff at McGill Getting Started

More information

VC3. Virtual Clusters for Community Computation. DOE NGNS PI Meeting September 27-28, 2017

VC3. Virtual Clusters for Community Computation. DOE NGNS PI Meeting September 27-28, 2017 VC3 Virtual Clusters for Community Computation DOE NGNS PI Meeting September 27-28, 2017 Douglas Thain, University of Notre Dame Rob Gardner, University of Chicago John Hover, Brookhaven National Lab A

More information

ElastiCluster Automated provisioning of computational clusters in the cloud

ElastiCluster Automated provisioning of computational clusters in the cloud ElastiCluster Automated provisioning of computational clusters in the cloud Riccardo Murri (with contributions from Antonio Messina, Nicolas Bär, Sergio Maffioletti, and Sigve

More information

OpenDreamKit. Computational environments for research and education Min Ragan-Kelley. Simula Research Lab

OpenDreamKit. Computational environments for research and education Min Ragan-Kelley. Simula Research Lab OpenDreamKit Computational environments for research and education Min Ragan-Kelley Simula Research Lab OpenDreamKit H2020 project Virtual Research Environments 16 Institutions Generic (Jupyter, SageMath)

More information

Short Talk: System abstractions to facilitate data movement in supercomputers with deep memory and interconnect hierarchy

Short Talk: System abstractions to facilitate data movement in supercomputers with deep memory and interconnect hierarchy Short Talk: System abstractions to facilitate data movement in supercomputers with deep memory and interconnect hierarchy François Tessier, Venkatram Vishwanath Argonne National Laboratory, USA July 19,

More information

Remote & Collaborative Visualization. Texas Advanced Computing Center

Remote & Collaborative Visualization. Texas Advanced Computing Center Remote & Collaborative Visualization Texas Advanced Computing Center TACC Remote Visualization Systems Longhorn NSF XD Dell Visualization Cluster 256 nodes, each 8 cores, 48 GB (or 144 GB) memory, 2 NVIDIA

More information

Now SAML takes it all:

Now SAML takes it all: Now SAML takes it all: Federation of non Web-based Services in the State of Baden-Württemberg Sebastian Labitzke Karlsruhe Institute of Technology (KIT) Steinbuch Centre for Computing (SCC) labitzke@kit.edu

More information

Allowing Users to Run Services at the OLCF with Kubernetes

Allowing Users to Run Services at the OLCF with Kubernetes Allowing Users to Run Services at the OLCF with Kubernetes Jason Kincl Senior HPC Systems Engineer Ryan Adamson Senior HPC Security Engineer This work was supported by the Oak Ridge Leadership Computing

More information

CloudExpo November 2017 Tomer Levi

CloudExpo November 2017 Tomer Levi CloudExpo November 2017 Tomer Levi About me Full Stack Engineer @ Intel s Advanced Analytics group. Artificial Intelligence unit at Intel. Responsible for (1) Radical improvement of critical processes

More information

Parsl: Developing Interactive Parallel Workflows in Python using Parsl

Parsl: Developing Interactive Parallel Workflows in Python using Parsl Parsl: Developing Interactive Parallel Workflows in Python using Parsl Kyle Chard (chard@uchicago.edu) Yadu Babuji, Anna Woodard, Zhuozhao Li, Ben Clifford, Ian Foster, Dan Katz, Mike Wilde, Justin Wozniak

More information

The RAMDISK Storage Accelerator

The RAMDISK Storage Accelerator The RAMDISK Storage Accelerator A Method of Accelerating I/O Performance on HPC Systems Using RAMDISKs Tim Wickberg, Christopher D. Carothers wickbt@rpi.edu, chrisc@cs.rpi.edu Rensselaer Polytechnic Institute

More information

Part2: Let s pick one cloud IaaS middleware: OpenStack. Sergio Maffioletti

Part2: Let s pick one cloud IaaS middleware: OpenStack. Sergio Maffioletti S3IT: Service and Support for Science IT Cloud middleware Part2: Let s pick one cloud IaaS middleware: OpenStack Sergio Maffioletti S3IT: Service and Support for Science IT, University of Zurich http://www.s3it.uzh.ch/

More information

High Performance Computing Course Notes HPC Fundamentals

High Performance Computing Course Notes HPC Fundamentals High Performance Computing Course Notes 2008-2009 2009 HPC Fundamentals Introduction What is High Performance Computing (HPC)? Difficult to define - it s a moving target. Later 1980s, a supercomputer performs

More information

DARPA-BAA Hierarchical Identify Verify Exploit (HIVE) Frequently Asked Questions (FAQ) August 18, 2016

DARPA-BAA Hierarchical Identify Verify Exploit (HIVE) Frequently Asked Questions (FAQ) August 18, 2016 DARPA-BAA-16-52 Hierarchical Identify Verify Exploit (HIVE) Frequently Asked Questions (FAQ) August 18, 2016 DARPA-BAA-16-52 Hierarchical Identify Verify Exploit (HIVE) Frequently Asked Questions (FAQ)

More information

Tools for Handling Big Data and Compu5ng Demands. Humani5es and Social Science Scholars

Tools for Handling Big Data and Compu5ng Demands. Humani5es and Social Science Scholars Tools for Handling Big Data and Compu5ng Demands Humani5es and Social Science Scholars Outline Overview of Compute Canada and WestGrid Focus on Humani5es and Social Sciences The Resource Alloca5on Compe55on

More information

Introduction to BioHPC

Introduction to BioHPC Introduction to BioHPC New User Training [web] [email] portal.biohpc.swmed.edu biohpc-help@utsouthwestern.edu 1 Updated for 2015-06-03 Overview Today we re going to cover: What is BioHPC? How do I access

More information

Blue Gene/Q. Hardware Overview Michael Stephan. Mitglied der Helmholtz-Gemeinschaft

Blue Gene/Q. Hardware Overview Michael Stephan. Mitglied der Helmholtz-Gemeinschaft Blue Gene/Q Hardware Overview 02.02.2015 Michael Stephan Blue Gene/Q: Design goals System-on-Chip (SoC) design Processor comprises both processing cores and network Optimal performance / watt ratio Small

More information

Slurm Overview. Brian Christiansen, Marshall Garey, Isaac Hartung SchedMD SC17. Copyright 2017 SchedMD LLC

Slurm Overview. Brian Christiansen, Marshall Garey, Isaac Hartung SchedMD SC17. Copyright 2017 SchedMD LLC Slurm Overview Brian Christiansen, Marshall Garey, Isaac Hartung SchedMD SC17 Outline Roles of a resource manager and job scheduler Slurm description and design goals Slurm architecture and plugins Slurm

More information

Parallel I/O on JUQUEEN

Parallel I/O on JUQUEEN Parallel I/O on JUQUEEN 4. Februar 2014, JUQUEEN Porting and Tuning Workshop Mitglied der Helmholtz-Gemeinschaft Wolfgang Frings w.frings@fz-juelich.de Jülich Supercomputing Centre Overview Parallel I/O

More information

Supercomputer and grid infrastructure! in Poland!

Supercomputer and grid infrastructure! in Poland! Supercomputer and grid infrastructure in Poland Franciszek Rakowski Interdisciplinary Centre for Mathematical and Computational Modelling 12th INCF Nodes Workshop, 16.04.2015 Warsaw, Nencki Institute.

More information

A Laconic HPC with an Orgone Accumulator. Presentation to Multicore World Wellington, February 15-17,

A Laconic HPC with an Orgone Accumulator. Presentation to Multicore World Wellington, February 15-17, A Laconic HPC with an Orgone Accumulator Presentation to Multicore World 2016 Wellington, February 15-17, 2016 http://levlafayette.com Edward - University of Melbourne Cluster - System Installed and operational

More information

Visualiza(on So-ware and Hardware for In- Silico Brain Research. Stefan Eilemann Visualiza0on Team Lead Blue Brain Project, EPFL

Visualiza(on So-ware and Hardware for In- Silico Brain Research. Stefan Eilemann Visualiza0on Team Lead Blue Brain Project, EPFL Visualiza(on So-ware and Hardware for In- Silico Brain Research Stefan Eilemann Visualiza0on Team Lead Blue Brain Project, EPFL Blue Brain Project / Human Brain Project BBP: Swiss na0onal research project

More information

Shared Services Canada Environment and Climate Change Canada HPC Renewal Project

Shared Services Canada Environment and Climate Change Canada HPC Renewal Project Shared Services Canada Environment and Climate Change Canada HPC Renewal Project CUG 2017 Redmond, WA, USA Deric Sullivan Alain St-Denis & Luc Corbeil May 2017 Background: SSC's HPC Renewal for ECCC Environment

More information

HPC IN EUROPE. Organisation of public HPC resources

HPC IN EUROPE. Organisation of public HPC resources HPC IN EUROPE Organisation of public HPC resources Context Focus on publicly-funded HPC resources provided primarily to enable scientific research and development at European universities and other publicly-funded

More information

IBM Data Science Experience (DSX) Partner Application Validation Quick Guide

IBM Data Science Experience (DSX) Partner Application Validation Quick Guide IBM Data Science Experience (DSX) Partner Application Validation Quick Guide VERSION: 1.6 DATE: Sept 27, 2017 EDITOR: D. Rangarao Table of Contents 1 Overview of the Application Validation Process... 3

More information

Managing Grid Credentials

Managing Grid Credentials Managing Grid Credentials Jim Basney http://www.ncsa.uiuc.edu/~jbasney/ Senior Research Scientist Grid and Security Technologies National Center for Supercomputing Applications

More information

Integration of CASA with Jupyter for efficient remote processing Aard Keimpema

Integration of CASA with Jupyter for efficient remote processing Aard Keimpema Astronomy ESFRI & Research Infrastructure Cluster ASTERICS - 653477 Integration of CASA with Jupyter for efficient remote processing Aard Keimpema (keimpema@jive.eu) H2020-Astronomy ESFRI and Research

More information

IBM Data Science Experience (DSX) Partner Application Validation Quick Guide

IBM Data Science Experience (DSX) Partner Application Validation Quick Guide IBM Data Science Experience (DSX) Partner Application Validation Quick Guide VERSION: 2.0 DATE: Feb 15, 2018 EDITOR: D. Rangarao Table of Contents 1 Overview of the Application Validation Process... 3

More information

NERSC Site Update. National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory. Richard Gerber

NERSC Site Update. National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory. Richard Gerber NERSC Site Update National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory Richard Gerber NERSC Senior Science Advisor High Performance Computing Department Head Cori

More information

Using Two-Factor Authentication to Connect to a Kerberos-enabled Informatica Domain

Using Two-Factor Authentication to Connect to a Kerberos-enabled Informatica Domain Using Two-Factor Authentication to Connect to a Kerberos-enabled Informatica Domain Copyright Informatica LLC 2016, 2018. Informatica LLC. No part of this document may be reproduced or transmitted in any

More information

Cornell Red Cloud: Campus-based Hybrid Cloud. Steven Lee Cornell University Center for Advanced Computing

Cornell Red Cloud: Campus-based Hybrid Cloud. Steven Lee Cornell University Center for Advanced Computing Cornell Red Cloud: Campus-based Hybrid Cloud Steven Lee Cornell University Center for Advanced Computing shl1@cornell.edu Cornell Center for Advanced Computing (CAC) Profile CAC mission, impact on research

More information

Leveraging Globus Identity for the Grid. Suchandra Thapa GlobusWorld, April 22, 2016 Chicago

Leveraging Globus Identity for the Grid. Suchandra Thapa GlobusWorld, April 22, 2016 Chicago Leveraging Globus Identity for the Grid Suchandra Thapa GlobusWorld, April 22, 2016 Chicago Open Science Grid Helps researchers speed up their research using high throughput computing methods Helps campus

More information

2017 Resource Allocations Competition Results

2017 Resource Allocations Competition Results 2017 Resource Allocations Competition Results Table of Contents Executive Summary...3 Computational Resources...5 CPU Allocations...5 GPU Allocations...6 Cloud Allocations...6 Storage Resources...6 Acceptance

More information

MULTITHERMAN: Out-of-band High-Resolution HPC Power and Performance Monitoring Support for Big-Data Analysis

MULTITHERMAN: Out-of-band High-Resolution HPC Power and Performance Monitoring Support for Big-Data Analysis MULTITHERMAN: Out-of-band High-Resolution HPC Power and Performance Monitoring Support for Big-Data Analysis EU H2020 FETHPC project ANTAREX (g.a. 671623) EU FP7 ERC Project MULTITHERMAN (g.a.291125) HPC

More information

CONFIGURING IBM STORWIZE. for Metadata Framework 6.3

CONFIGURING IBM STORWIZE. for Metadata Framework 6.3 CONFIGURING IBM STORWIZE for Metadata Framework 6.3 Publishing Information Software version 6.3.160 Document version 4 Publication date May 22, 2017 Copyright 2005-2017 Varonis Systems Inc. All rights

More information

Duo End User Education Templates

Duo End User Education  Templates Duo End User Education Email Templates Table of Contents Email Communication Best Practices Email Templates Email #1 - Duo is coming soon, no immediate action required. Email #2 - Duo is coming on ,

More information

Load Balancing Censornet USS Gateway. Deployment Guide v Copyright Loadbalancer.org

Load Balancing Censornet USS Gateway. Deployment Guide v Copyright Loadbalancer.org Load Balancing Censornet USS Gateway Deployment Guide v1.0.0 Copyright Loadbalancer.org Table of Contents 1. About this Guide...3 2. Loadbalancer.org Appliances Supported...3 3. Loadbalancer.org Software

More information

Leveraging the Globus Platform in your Web Applications. GlobusWorld April 26, 2018 Greg Nawrocki

Leveraging the Globus Platform in your Web Applications. GlobusWorld April 26, 2018 Greg Nawrocki Leveraging the Globus Platform in your Web Applications GlobusWorld April 26, 2018 Greg Nawrocki greg@globus.org Topics and Goals Platform Overview Why expose the APIs A quick touch of the Globus Auth

More information

The LGI Pilot job portal. EGI Technical Forum 20 September 2011 Jan Just Keijser Willem van Engen Mark Somers

The LGI Pilot job portal. EGI Technical Forum 20 September 2011 Jan Just Keijser Willem van Engen Mark Somers The LGI Pilot job portal EGI Technical Forum 20 September 2011 Jan Just Keijser Willem van Engen Mark Somers Outline What? Why? How? Pro's and Cons What's next? Credits 2 What is LGI? LGI Project Server

More information

Transcode and Add Pulse Video Analytics to Video Content on Cisco Show and Share

Transcode and Add Pulse Video Analytics to Video Content on Cisco Show and Share Transcode and Add Pulse Video Analytics to Video Content on Cisco Show and Share Revised: February 26, 2013 When Cisco MXE 3500 integration is enabled in Cisco Show and Share, videos uploaded to Cisco

More information

Introduction to Slurm

Introduction to Slurm Introduction to Slurm Tim Wickberg SchedMD Slurm User Group Meeting 2017 Outline Roles of resource manager and job scheduler Slurm description and design goals Slurm architecture and plugins Slurm configuration

More information

SpiNNaker a Neuromorphic Supercomputer. Steve Temple University of Manchester, UK SOS21-21 Mar 2017

SpiNNaker a Neuromorphic Supercomputer. Steve Temple University of Manchester, UK SOS21-21 Mar 2017 SpiNNaker a Neuromorphic Supercomputer Steve Temple University of Manchester, UK SOS21-21 Mar 2017 Outline of talk Introduction Modelling neurons Architecture and technology Principles of operation Summary

More information

Ambry: LinkedIn s Scalable Geo- Distributed Object Store

Ambry: LinkedIn s Scalable Geo- Distributed Object Store Ambry: LinkedIn s Scalable Geo- Distributed Object Store Shadi A. Noghabi *, Sriram Subramanian +, Priyesh Narayanan +, Sivabalan Narayanan +, Gopalakrishna Holla +, Mammad Zadeh +, Tianwei Li +, Indranil

More information

ArcGIS for Server: Administration and Security. Amr Wahba

ArcGIS for Server: Administration and Security. Amr Wahba ArcGIS for Server: Administration and Security Amr Wahba awahba@esri.com Agenda ArcGIS Server architecture Distributing and scaling components Implementing security Monitoring server logs Automating server

More information

CSCS Site Update. HPC Advisory Council Workshop Colin McMurtrie, Associate Director and Head of HPC Operations.

CSCS Site Update. HPC Advisory Council Workshop Colin McMurtrie, Associate Director and Head of HPC Operations. CSCS Site Update HPC Advisory Council Workshop 2018 Colin McMurtrie, Associate Director and Head of HPC Operations. 9th April 2018 The Swiss National Supercomputing Centre Driving innovation in computational

More information

glite Grid Services Overview

glite Grid Services Overview The EPIKH Project (Exchange Programme to advance e-infrastructure Know-How) glite Grid Services Overview Antonio Calanducci INFN Catania Joint GISELA/EPIKH School for Grid Site Administrators Valparaiso,

More information

Using Cartesius and Lisa. Zheng Meyer-Zhao - Consultant Clustercomputing

Using Cartesius and Lisa. Zheng Meyer-Zhao - Consultant Clustercomputing Zheng Meyer-Zhao - zheng.meyer-zhao@surfsara.nl Consultant Clustercomputing Outline SURFsara About us What we do Cartesius and Lisa Architectures and Specifications File systems Funding Hands-on Logging

More information

Deploying a Production Gateway with Airavata

Deploying a Production Gateway with Airavata Deploying a Production Gateway with Airavata Table of Contents Pre-requisites... 1 Create a Gateway Request... 1 Gateway Deploy Steps... 2 Install Ansible & Python...2 Deploy the Gateway...3 Gateway Configuration...

More information

RHRK-Seminar. High Performance Computing with the Cluster Elwetritsch - II. Course instructor : Dr. Josef Schüle, RHRK

RHRK-Seminar. High Performance Computing with the Cluster Elwetritsch - II. Course instructor : Dr. Josef Schüle, RHRK RHRK-Seminar High Performance Computing with the Cluster Elwetritsch - II Course instructor : Dr. Josef Schüle, RHRK Overview Course I Login to cluster SSH RDP / NX Desktop Environments GNOME (default)

More information

Research Collection. WebParFE A web interface for the high performance parallel finite element solver ParFE. Report. ETH Library

Research Collection. WebParFE A web interface for the high performance parallel finite element solver ParFE. Report. ETH Library Research Collection Report WebParFE A web interface for the high performance parallel finite element solver ParFE Author(s): Paranjape, Sumit; Kaufmann, Martin; Arbenz, Peter Publication Date: 2009 Permanent

More information

1 Bull, 2011 Bull Extreme Computing

1 Bull, 2011 Bull Extreme Computing 1 Bull, 2011 Bull Extreme Computing Table of Contents Overview. Principal concepts. Architecture. Scheduler Policies. 2 Bull, 2011 Bull Extreme Computing SLURM Overview Ares, Gerardo, HPC Team Introduction

More information

Introduction to BioHPC New User Training

Introduction to BioHPC New User Training Introduction to BioHPC New User Training [web] [email] portal.biohpc.swmed.edu biohpc-help@utsouthwestern.edu 1 Updated for 2019-02-06 Overview Today we re going to cover: What is BioHPC? How do I access

More information

Advanced School in High Performance and GRID Computing November Introduction to Grid computing.

Advanced School in High Performance and GRID Computing November Introduction to Grid computing. 1967-14 Advanced School in High Performance and GRID Computing 3-14 November 2008 Introduction to Grid computing. TAFFONI Giuliano Osservatorio Astronomico di Trieste/INAF Via G.B. Tiepolo 11 34131 Trieste

More information

Introduction to High Performance Computing and an Statistical Genetics Application on the Janus Supercomputer. Purpose

Introduction to High Performance Computing and an Statistical Genetics Application on the Janus Supercomputer. Purpose Introduction to High Performance Computing and an Statistical Genetics Application on the Janus Supercomputer Daniel Yorgov Department of Mathematical & Statistical Sciences, University of Colorado Denver

More information

Batch Services at CERN: Status and Future Evolution

Batch Services at CERN: Status and Future Evolution Batch Services at CERN: Status and Future Evolution Helge Meinhard, CERN-IT Platform and Engineering Services Group Leader HTCondor Week 20 May 2015 20-May-2015 CERN batch status and evolution - Helge

More information

Leonhard: a new cluster for Big Data at ETH

Leonhard: a new cluster for Big Data at ETH Leonhard: a new cluster for Big Data at ETH Bernd Rinn, Head of Scientific IT Services Olivier Byrde, Group leader High Performance Computing Bernd Rinn & Olivier Byrde 2017-02-15 1 Agenda Welcome address

More information

The Cambridge Bio-Medical-Cloud An OpenStack platform for medical analytics and biomedical research

The Cambridge Bio-Medical-Cloud An OpenStack platform for medical analytics and biomedical research The Cambridge Bio-Medical-Cloud An OpenStack platform for medical analytics and biomedical research Dr Paul Calleja Director of Research Computing University of Cambridge Global leader in science & technology

More information

SAS Grid Manager and Kerberos Authentication

SAS Grid Manager and Kerberos Authentication SAS Grid Manager and Kerberos Authentication Learn the considerations for implementing Kerberos authentication so you can submit workload to SAS Grid Manager. SAS Grid Manager and Kerberos Authentication

More information

Federated Services and Data Management in PRACE

Federated Services and Data Management in PRACE Federated Services and Data Management in PRACE G. Erbacci, (CINECA and PRACE) Collaborative Services for e-infrastructure Commons e-irg Workshop, Bratislava 15-16 November 2016 Partnership for Advanced

More information

Introduction to BioHPC

Introduction to BioHPC Introduction to BioHPC New User Training [web] [email] portal.biohpc.swmed.edu biohpc-help@utsouthwestern.edu 1 Updated for 2018-03-07 Overview Today we re going to cover: What is BioHPC? How do I access

More information

Leveraging the Globus Platform in your Web Applications

Leveraging the Globus Platform in your Web Applications Leveraging the Globus Platform in your Web Applications Steve Tuecke tuecke@uchicago.edu NCAR September 5, 2018 Globus serves as A platform for building science gateways, web portals and other applications

More information

High Performance Computing Course Notes Course Administration

High Performance Computing Course Notes Course Administration High Performance Computing Course Notes 2009-2010 2010 Course Administration Contacts details Dr. Ligang He Home page: http://www.dcs.warwick.ac.uk/~liganghe Email: liganghe@dcs.warwick.ac.uk Office hours:

More information

Our Workshop Environment

Our Workshop Environment Our Workshop Environment John Urbanic Parallel Computing Scientist Pittsburgh Supercomputing Center Copyright 2018 Our Environment Today Your laptops or workstations: only used for portal access Bridges

More information

Federated XDMoD Requirements

Federated XDMoD Requirements Federated XDMoD Requirements Date Version Person Change 2016-04-08 1.0 draft XMS Team Initial version Summary Definitions Assumptions Data Collection Local XDMoD Installation Module Support Data Federation

More information

XSEDE Infrastructure as a Service Use Cases

XSEDE Infrastructure as a Service Use Cases XSEDE Infrastructure as a Service Use Cases October 30, 2016 Version 2.2 This document outlines typical XSEDE user needs in the Infrastructure as a Service (IaaS) area. IAAS 1 through IAAS 4 describe how

More information

Ansible Tower Quick Setup Guide

Ansible Tower Quick Setup Guide Ansible Tower Quick Setup Guide Release Ansible Tower 2.4.5 Red Hat, Inc. Jun 06, 2017 CONTENTS 1 Quick Start 2 2 Login as a Superuser 3 3 Import a License 4 4 Examine the Tower Dashboard 6 5 The Setup

More information

RENKU - Reproduce, Reuse, Recycle Research. Rok Roškar and the SDSC Renku team

RENKU - Reproduce, Reuse, Recycle Research. Rok Roškar and the SDSC Renku team RENKU - Reproduce, Reuse, Recycle Research Rok Roškar and the SDSC Renku team Renku-Reana workshop @ CERN 26.06.2018 Goals of Renku 1. Provide the means to create reproducible data science 2. Facilitate

More information

Note: Who is Dr. Who? You may notice that YARN says you are logged in as dr.who. This is what is displayed when user

Note: Who is Dr. Who? You may notice that YARN says you are logged in as dr.who. This is what is displayed when user Run a YARN Job Exercise Dir: ~/labs/exercises/yarn Data Files: /smartbuy/kb In this exercise you will submit an application to the YARN cluster, and monitor the application using both the Hue Job Browser

More information

Adding Cloud Based Interactive Compute Capabilities to Globus Endpoints

Adding Cloud Based Interactive Compute Capabilities to Globus Endpoints Adding Cloud Based Interactive Compute Capabilities to Globus Endpoints Ben Galewsky Research Programmer, National Center for Supercomputing Applications bengal1@illinois.edu http://www.nationaldataservice.org/

More information

Applications Software Example

Applications Software Example Applications Software Example How to run an application on Cluster? Rooh Khurram Supercomputing Laboratory King Abdullah University of Science and Technology (KAUST), Saudi Arabia Cluster Training: Applications

More information

vcloud Director Administrator's Guide

vcloud Director Administrator's Guide vcloud Director 5.5 This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more recent editions of

More information

Continuous Integration using Docker & Jenkins

Continuous Integration using Docker & Jenkins Jenkins LinuxCon Europe 2014 October 13-15, 2014 Mattias Giese Solutions Architect giese@b1-systems.de - Linux/Open Source Consulting, Training, Support & Development Introducing B1 Systems founded in

More information

GPFS for Life Sciences at NERSC

GPFS for Life Sciences at NERSC GPFS for Life Sciences at NERSC A NERSC & JGI collaborative effort Jason Hick, Rei Lee, Ravi Cheema, and Kjiersten Fagnan GPFS User Group meeting May 20, 2015-1 - Overview of Bioinformatics - 2 - A High-level

More information

Quick Start Guide. Table of Contents

Quick Start Guide. Table of Contents Quick Start Guide Table of Contents Account Registration... 2 Signup Request... 2 Account Activation... 4 Running FLOW-3D on POD... 9 Launching the GUI... 9 Running Simulations... 11 Collaborating with

More information

The ASCI/DOD Scalable I/O History and Strategy Run Time Systems and Scalable I/O Team Gary Grider CCN-8 Los Alamos National Laboratory LAUR

The ASCI/DOD Scalable I/O History and Strategy Run Time Systems and Scalable I/O Team Gary Grider CCN-8 Los Alamos National Laboratory LAUR The ASCI/DOD Scalable I/O History and Strategy Run Time Systems and Scalable I/O Team Gary Grider CCN-8 Los Alamos National Laboratory LAUR 042787 05/2004 Parallel File Systems and Parallel I/O Why - From

More information

An Overview of the Architecture of Juno: CHPC s New JupyterHub Service By Luan Truong, CHPC, University of Utah

An Overview of the Architecture of Juno: CHPC s New JupyterHub Service By Luan Truong, CHPC, University of Utah An Overview of the Architecture of Juno: CHPC s New JupyterHub Service By Luan Truong, CHPC, University of Utah Introduction Jupyter notebooks have emerged as a popular and open-source web application

More information

FuncX: A Function Serving Platform for HPC. Ryan Chard 28 Jan 2019

FuncX: A Function Serving Platform for HPC. Ryan Chard 28 Jan 2019 FuncX: A Function Serving Platform for HPC Ryan Chard 28 Jan 2019 Outline - Motivation FuncX: FaaS for HPC Implementation status Preliminary applications - Machine learning inference Automating analysis

More information