Interactive Supercomputing for State-of-the-art Biomolecular Simulation

Similar documents
Using AWS EC2 GPU Instances for Computational Microscopy and Biomolecular Simulation

GPU-Accelerated Analysis of Large Biomolecular Complexes

VMD: Immersive Molecular Visualization and Interactive Ray Tracing for Domes, Panoramic Theaters, and Head Mounted Displays

Harnessing GPUs to Probe Biomolecular Machines at Atomic Detail

Proteins and Mesoscale Data: Visualization of Molecular Dynamics

Visualization of Energy Conversion Processes in a Light Harvesting Organelle at Atomic Detail

GPU-Accelerated Visualization and Analysis of Petascale Molecular Dynamics Simulations

S5386 Publication-Quality Ray Tracing of Molecular Graphics with OptiX

Early Experiences Porting the NAMD and VMD Molecular Simulation and Analysis Software to GPU-Accelerated OpenPOWER Platforms

S6253 VMD: Petascale Molecular Visualization and Analysis with Remote Video Streaming

S4410: Visualization and Analysis of Petascale Molecular Simulations with VMD

Using Accelerator Directives to Adapt Science Applications for State-of-the-Art HPC Architectures

S6261 VMD+OptiX: Streaming Interactive Ray Tracing from Remote GPU Clusters to Your VR Headset

Frontiers of Molecular Visualization: Interactive Ray Tracing, Panoramic Displays, VR HMDs, and Remote Visualization. John E.

VMD: Interactive Publication-Quality Molecular Ray Tracing with OptiX

In-Situ Visualization and Analysis of Petascale Molecular Dynamics Simulations with VMD

High Performance Molecular Visualization: In-Situ and Parallel Rendering with EGL

S8665 VMD: Biomolecular Visualization from Atoms to Cells Using Ray Tracing, Rasterization, and VR

Visualizing Biomolecular Complexes on x86 and KNL Platforms: Integrating VMD and OSPRay

S8665 VMD: Biomolecular Visualization from Atoms to Cells Using Ray Tracing, Rasterization, and VR

S8709 Accelerating Molecular Modeling on Desktop and Pre-Exascale Supercomputers

Advances In Biomolecular Simulation with NAMD and VMD

Visualization Challenges and Opportunities Posed by Petascale Molecular Dynamics Simulations

Improving NAMD Performance on Volta GPUs

Molecular Visualization and Simulation in VR

Immersive Out-of-Core Visualization of Large-Size and Long-Timescale Molecular Dynamics Trajectories

S5371 VMD: Visualization and Analysis of Biomolecular Complexes with GPU Computing

Using GPUs to Supercharge Visualization and Analysis of Molecular Dynamics Simulations with VMD

GPU-Accelerated Analysis of Petascale Molecular Dynamics Simulations

VMD: Preparation and Analysis of Molecular and Cellular Simulations

Visualization and Analysis of Petascale Molecular Dynamics Simulations

Early Experiences Scaling VMD Molecular Visualization and Analysis Jobs on Blue Waters

Analysis and Visualization Algorithms in VMD

PERFORMANCE PORTABILITY WITH OPENACC. Jeff Larkin, NVIDIA, November 2015

OPTIMIZING HPC SIMULATION AND VISUALIZATION CODE USING NVIDIA NSIGHT SYSTEMS

NVIDIA Update and Directions on GPU Acceleration for Earth System Models

Faster, Cheaper, Better: Biomolecular Simulation with NAMD, VMD, and CUDA

High Performance Computing with Accelerators

Accelerating Molecular Modeling Applications with Graphics Processors

The challenges of new, efficient computer architectures, and how they can be met with a scalable software development strategy.! Thomas C.

Programming for Hybrid Architectures Today and in the Future

Bridging the Gap Between High Quality and High Performance for HPC Visualization

PLAN-E Workshop Switzerland. Welcome! September 8, 2016

ENABLING NEW SCIENCE GPU SOLUTIONS

Simulating Life at the Atomic Scale

NAMD GPU Performance Benchmark. March 2011

Multilevel Summation of Electrostatic Potentials Using GPUs

COLLEGE OF IMAGING ARTS AND SCIENCES. Medical Illustration

NAMD, CUDA, and Clusters: Taking GPU Molecular Dynamics Beyond the Deskop

Accelerating Molecular Modeling Applications with GPU Computing

Scaling in a Heterogeneous Environment with GPUs: GPU Architecture, Concepts, and Strategies

SCIENTIFIC VISUALIZATION IN HPC

High Performance Molecular Visualization and Analysis on GPUs

Piz Daint: Application driven co-design of a supercomputer based on Cray s adaptive system design

MICROWAY S NVIDIA TESLA V100 GPU SOLUTIONS GUIDE

HETEROGENEOUS HPC, ARCHITECTURAL OPTIMIZATION, AND NVLINK STEVE OBERLIN CTO, TESLA ACCELERATED COMPUTING NVIDIA

Advanced Visualization for Chemistry

Cinematic Scientific Visualization in Houdini

ANSYS Improvements to Engineering Productivity with HPC and GPU-Accelerated Simulation

Using GPUs to compute the multilevel summation of electrostatic forces

VMD Key Features of Recent Release

It s a Multicore World. John Urbanic Pittsburgh Supercomputing Center

GPU Clusters for High- Performance Computing Jeremy Enos Innovative Systems Laboratory

NAMD-QM/MM TUTORIAL. Unix/MacOSX Version. NAMD-QM/MM Suite Developers: Marcelo C. R. Melo, Rafael C. Bernardi

CUDA Experiences: Over-Optimization and Future HPC

OpenPOWER Performance

John E. Stone. S03: High Performance Computing with CUDA. Heterogeneous GPU Computing for Molecular Modeling

IBM J. Res. & Dev.Acun et al.: Scalable Molecular Dynamics with NAMD on the Summit System Page 1

Preparing GPU-Accelerated Applications for the Summit Supercomputer

Abhinav Bhatele, Laxmikant V. Kale University of Illinois at Urbana Champaign Sameer Kumar IBM T. J. Watson Research Center

Lecture 1: Introduction and Computational Thinking

Illinois Proposal Considerations Greg Bauer

GPU Histogramming: Radial Distribution Functions

Debugging CUDA Applications with Allinea DDT. Ian Lumb Sr. Systems Engineer, Allinea Software Inc.

An Introduction to the SPEC High Performance Group and their Benchmark Suites

Stan Posey, NVIDIA, Santa Clara, CA, USA

Quantifying the Impact of GPUs on Performance and Energy Efficiency in HPC Clusters

Petascale Multiscale Simulations of Biomolecular Systems. John Grime Voth Group Argonne National Laboratory / University of Chicago

INTRODUCTION TO OPENACC. Analyzing and Parallelizing with OpenACC, Feb 22, 2017

SIGHT. Benjamin Hernandez, PhD Advanced Data and Workflow(s) Group

Irregular Graph Algorithms on Parallel Processing Systems

ACCELERATED COMPUTING: THE PATH FORWARD. Jen-Hsun Huang, Co-Founder and CEO, NVIDIA SC15 Nov. 16, 2015

IBM CORAL HPC System Solution

GPU-Accelerated Molecular Visualization and Analysis with VMD

It s a Multicore World. John Urbanic Pittsburgh Supercomputing Center Parallel Computing Scientist

Introduction to FREE National Resources for Scientific Computing. Dana Brunson. Jeff Pummill

Steve Scott, Tesla CTO SC 11 November 15, 2011

Cross Teaching Parallelism and Ray Tracing: A Project based Approach to Teaching Applied Parallel Computing

Experiences Developing and Maintaining Scientific Applications on GPU-Accelerated Platforms

WHAT S NEW IN CUDA 8. Siddharth Sharma, Oct 2016

Extreme-scale Graph Analysis on Blue Waters

Exploratory Visualization of Petascale Particle Data in Nvidia DGX-1

GPU ACCELERATED COMPUTING. 1 st AlsaCalcul GPU Challenge, 14-Jun-2016, Strasbourg Frédéric Parienté, Tesla Accelerated Computing, NVIDIA Corporation

Performance Study of Popular Computational Chemistry Software Packages on Cray HPC Systems

IMPORTANT: Visit for the most up-to-date schedule.

Supercomputer and grid infrastructure! in Poland!

Broadening the Use of Scalable Kernels in NAMD/VMD

Harnessing GPU speed to accelerate LAMMPS particle simulations

Titan - Early Experience with the Titan System at Oak Ridge National Laboratory

Recent Advances in Heterogeneous Computing using Charm++

Transcription:

Interactive Supercomputing for State-of-the-art Biomolecular Simulation John E. Stone Theoretical and Computational Biophysics Group Beckman Institute for Advanced Science and Technology University of Illinois at Urbana-Champaign http://www.ks.uiuc.edu/research/namd/ http://www.ks.uiuc.edu/research/vmd/ Interactivity in Supercomputing BoF, 12:15pm-1:15pm, Rooms 503-504, Colorado Convention Center Denver, CO, Tuesday Nov 14 th, 2017

VMD Visual Molecular Dynamics Visualization and analysis of: Molecular dynamics simulations Lattice cell simulations Quantum chemistry calculations Sequence information User extensible scripting and plugins http://www.ks.uiuc.edu/research/vmd/ Cell-Scale Modeling MD Simulation

Goal: A Computational Microscope Study the molecular machines in living cells Ribosome: target for antibiotics Poliovirus

How Does Interactivity in HPC Benefit Scientific Productivity Until 2008, MD simulation performance was the major limiting factor on biomolecular modeling productivity, maybe 95% factor In 2017: Batch MD sim code performance is likely only a 50% factor on overall scientific productivity An extremely simplified outline: Months or years of science team effort go into preparing state-of-the-art biomolecular simulations w/ experimental structures before MD sims begin During the initial and early phases of a simulation campaign, hundreds of modeling problems must be identified and addressed Interactivity is a fundamental requirement during these phases After sim campaign has produced required outputs, extensive analytical work begins, and science teams then make key decisions about outcomes, potentially conducting further sim studies, developing manuscripts with detailed analyses, figures, and movies of dynamics of systems under study

Proto-Cell Modeling Project Initial model of minimal cell membrane and proteins Candidate for our ORNL Summit early science project ~1 billion atoms presently with explicit water, ions, etc. Main structure file used for modeling and simulation tasks is currently ~60GB Initial simulation preparation requires many repeated build / viz / measure / simulate cycles MD trajectory output : ~12GB/frame Need to run all tools required for science campaign at remote supercomputer sites: Optimize work-efficiency of science team Not just NAMD and VMD, but also other tools Old approach of repeatedly transferring files back and forth between UIUC lab and remote supercomputers is untenable at this scale!!!

What Do We Need? Science campaign Bullpen interactive compute/viz Interactive resources supporting the human-intensive parts of science team activities Specific examples: Interactive remote visualization supporting use of arbitrary science tools, including those that are not traditional HPC tools (e.g. CryoEM and X-Ray crystallography pkgs) Special batch queues, or queue policies that support the initial and early phase of simulation bring-up activities, so that meso-scale interactive calculations can be performed at a mouse click to provide rapid-turnaround of results in interactive modeling sessions to correct flaws in molecular models, simulation parameters, etc. A few (10s or less) HPC systems supporting very large memory capacities and dense GPU counts, to run the interactive supercomputing job front end

VMD Petascale Visualization and Analysis Analyze/visualize large trajectories too large to transfer off-site: User-defined parallel analysis operations, data types Parallel rendering, movie making Supports GPU-accelerated Cray XK7 nodes for both visualization and analysis: GPU accelerated trajectory analysis w/ CUDA OpenGL and GPU ray tracing for visualization and movie rendering Parallel I/O rates up to 275 GB/sec on 8192 Cray XE6 nodes can read in 231 TB in 15 minutes! Parallel VMD currently available on: ORNL Titan, NCSA Blue Waters, Indiana Big Red II, CSCS Piz Daint, and similar systems NCSA Blue Waters Hybrid Cray XE6 / XK7 22,640 XE6 dual-opteron CPU nodes 4,224 XK7 nodes w/ Telsa K20X GPUs

Molecular Dynamics Flexible Fitting (MDFF) X-ray crystallography MDFF Electron microscopy APS at Argonne ORNL Titan FEI microscope Molecular dynamics-based refinement and validation for sub-5a cryoelectron microscopy maps. A. Singharoy, I. Teo, R. McGreevy, J. E. Stone, J. Zhao, Beckman and Institute, K. University Schulten. of Illinois elife at Urbana-Champaign 2016;10.7554/eLife.16105 - www.ks.uiuc.edu

GPUs Can Reduce MDFF Trajectory Analysis Runtimes from Hours to Minutes GPUs enable laptops and desktop workstations to handle tasks that would have previously required a cluster, or a very long wait GPU-accelerated petascale supercomputers enable analyses that were previously impractical, allowing detailed study of very large structures such as viruses GPU-accelerated MDFF Cross Correlation Timeline Regions with poor fit Regions with good fit

Parallel MDFF Cross Correlation Analysis on Cray XK7 Rabbit Hemorrhagic Disease Virus (RHDV) Traj. frames 10,000 Relative CC Structure component selections Single-node XK7 (projected) 720 336 hours (14 days) 128-node XK7 2048-node XK7 3.2 hours 105x speedup 19.5 minutes 1035x speedup Calculation of 7M CCs would take 5 years using serial CPU algorithm! RHDV colored by relative CC Stone et al., Faraday Discuss., 169:265-283, 2014. Time

VMD Tesla P100+V100 Cross Correlation Performance Rabbit Hemorrhagic Disease Virus: 702K atoms, 6.5Å resolution Latest results on Pascal P100, first runs on Volta V100 Hardware platform Runtime, Speedup vs. Chimera, VMD+GPU Chimera Xeon E5-2687W (2 socket) [1] 15.860s, 1x VMD-CUDA IBM Power8 + 1x Tesla K40 [2] 0.488s, 32x 0.9x VMD-CUDA Intel Xeon E5-2687W + 1x Quadro K6000 [1,2] 0.458s, 35x 1.0x VMD-CUDA Intel Xeon E5-2698v3 + 1x Tesla P100 0.090s, 176x 5.1x VMD-CUDA IBM Power8 Minsky + 1x Tesla P100 0.080s, 198x 5.7x VMD-CUDA Intel Xeon E5-2697Av4 + 1x Tesla V100 0.050s, 317x 9.2x [1] GPU-Accelerated Analysis and Visualization of Large Structures Solved by Molecular Dynamics Flexible Fitting. J. E. Stone, R. McGreevy, B. Isralewitz, and K. Schulten. Faraday Discussions 169:265-283, 2014. [2] Early Experiences Porting the NAMD and VMD Molecular Simulation and Analysis Software to GPU- Accelerated OpenPOWER Platforms. J. E. Stone, A.-P. Hynninen, J. C. Phillips, K. Schulten. International Workshop on OpenPOWER for HPC (IWOPH'16), LNCS 9945, pp. 188-206, 2016.

Clustering Analysis of Molecular Dynamics Trajectories GPU-Accelerated Molecular Dynamics Clustering Analysis with OpenACC. J.E. Stone, J.R. Perilla, C. K. Cassidy, and K. Schulten. In, Robert Farber, ed., Parallel Programming with OpenACC, Morgan Kaufmann, Chapter 11, pp. 215-240, 2016.

Interactive Ray Tracing Remote visualization tasks on very large macromolecular complexes Large-scale parallel rendering: in situ or post hoc visualization tasks GPU-Accelerated Molecular Visualization on Petascale Supercomputing Platforms. J. E. Stone, K. L. Vandivort, and K. Schulten. UltraVis 13, pp. 6:1-6:8, 2013. Visualization of Energy Conversion Processes in a Light Harvesting Organelle at Atomic Detail. M. Sener, et al. SC'14 Visualization and Data Analytics Showcase, 2014. Chemical Visualization of Human Pathogens: the Retroviral Capsids. J. R. Perilla, B.-C. Goh, J. E. Stone, and K. Schulten. SC'15 Visualization and Data Analytics Showcase, 2015. Atomic Detail Visualization of Photosynthetic Membranes with GPU-Accelerated Ray Tracing. J. E. Stone et al., J. Parallel Computing, 55:17-27, 2016. Immersive Molecular Visualization with Omnidirectional Stereoscopic Ray Tracing and Remote Rendering J. E. Stone, W. R. Sherman, and K. HPDAV, IPDPSW, pp. 1048-1057, 2016. VMD/OptiX GPU Ray Tracing of all-atom Chromatophore w/ lipids.

VMD 1.9.3 supports EGL for in-situ and parallel rendering on clouds, clusters, and supercomputers Eliminate dependency on windowing systems Simplified deployment of parallel VMD builds supporting off-screen rendering Maintains 100% of VMD OpenGL shaders and rendering features Support high-quality vendorsupported commercial OpenGL implementations in HPC systems that were previously limited to Mesa Poliovirus

EGL Is Supported Now! Cloud+Workstations with most recent NVIDIA drivers VMD on HPC systems w/ latest Tesla P100 GPUs: Cray XC50, CSCS Piz Daint, driver 375.39 IBM OpenPOWER, drivers 375.66 and later support both GLX and EGL

When I was a young man, my goal was to look with mathematical and computational means at the inside of cells, one atom at a time, to decipher how living systems work. That is what I strived for and I never deflected from this goal. Klaus Schulten

Acknowledgements Theoretical and Computational Biophysics Group, University of Illinois at Urbana-Champaign Funding: NIH support: P41GM104601 DOE INCITE, ORNL Titan: DE-AC05-00OR22725 NSF Blue Waters: NSF OCI 07-25070, PRAC The Computational Microscope, ACI-1238993, ACI-1440026