In-Situ Visualization and Analysis of Petascale Molecular Dynamics Simulations with VMD

Similar documents
GPU-Accelerated Analysis of Petascale Molecular Dynamics Simulations

Faster, Cheaper, Better: Biomolecular Simulation with NAMD, VMD, and CUDA

GPU-Accelerated Visualization and Analysis of Petascale Molecular Dynamics Simulations

S5386 Publication-Quality Ray Tracing of Molecular Graphics with OptiX

Visualization and Analysis of Petascale Molecular Dynamics Simulations

Early Experiences Scaling VMD Molecular Visualization and Analysis Jobs on Blue Waters

John E. Stone. S03: High Performance Computing with CUDA. Heterogeneous GPU Computing for Molecular Modeling

High Performance Molecular Visualization and Analysis on GPUs

Analysis and Visualization Algorithms in VMD

S4410: Visualization and Analysis of Petascale Molecular Simulations with VMD

VMD: Immersive Molecular Visualization and Interactive Ray Tracing for Domes, Panoramic Theaters, and Head Mounted Displays

Immersive Out-of-Core Visualization of Large-Size and Long-Timescale Molecular Dynamics Trajectories

Early Experiences Porting the NAMD and VMD Molecular Simulation and Analysis Software to GPU-Accelerated OpenPOWER Platforms

S6261 VMD+OptiX: Streaming Interactive Ray Tracing from Remote GPU Clusters to Your VR Headset

Harnessing GPUs to Probe Biomolecular Machines at Atomic Detail

GPU-Accelerated Analysis of Large Biomolecular Complexes

Broadening the Use of Scalable Kernels in NAMD/VMD

High Performance Molecular Visualization: In-Situ and Parallel Rendering with EGL

Using GPUs to Supercharge Visualization and Analysis of Molecular Dynamics Simulations with VMD

Accelerating Molecular Modeling Applications with Graphics Processors

High Performance Molecular Simulation, Visualization, and Analysis on GPUs

Proteins and Mesoscale Data: Visualization of Molecular Dynamics

GPU-Accelerated Molecular Visualization and Analysis with VMD

Experiences Developing and Maintaining Scientific Applications on GPU-Accelerated Platforms

Accelerating Molecular Modeling Applications with GPU Computing

Interactive Supercomputing for State-of-the-art Biomolecular Simulation

Using Accelerator Directives to Adapt Science Applications for State-of-the-Art HPC Architectures

S8665 VMD: Biomolecular Visualization from Atoms to Cells Using Ray Tracing, Rasterization, and VR

S6253 VMD: Petascale Molecular Visualization and Analysis with Remote Video Streaming

Visualizing Biomolecular Complexes on x86 and KNL Platforms: Integrating VMD and OSPRay

Molecular Visualization and Simulation in VR

Accelerating Molecular Modeling Applications with GPU Computing

S8665 VMD: Biomolecular Visualization from Atoms to Cells Using Ray Tracing, Rasterization, and VR

Programming for Hybrid Architectures Today and in the Future

S5371 VMD: Visualization and Analysis of Biomolecular Complexes with GPU Computing

High Performance Molecular Visualization and Analysis with GPU Computing

VMD: Interactive Publication-Quality Molecular Ray Tracing with OptiX

Multilevel Summation of Electrostatic Potentials Using GPUs

Accelerating Scientific Applications with GPUs

Visualization Challenges and Opportunities Posed by Petascale Molecular Dynamics Simulations

Visualization of Energy Conversion Processes in a Light Harvesting Organelle at Atomic Detail

Using GPUs to compute the multilevel summation of electrostatic forces

GPU Acceleration of Molecular Modeling Applications

ECE 498AL. Lecture 21-22: Application Performance Case Studies: Molecular Visualization and Analysis

Accelerating Computational Biology by 100x Using CUDA. John Stone Theoretical and Computational Biophysics Group, University of Illinois

Frontiers of Molecular Visualization: Interactive Ray Tracing, Panoramic Displays, VR HMDs, and Remote Visualization. John E.

GPU Histogramming: Radial Distribution Functions

Case Study: Molecular Modeling Applications John E. Stone

Improving NAMD Performance on Volta GPUs

Simulating Life at the Atomic Scale

S8709 Accelerating Molecular Modeling on Desktop and Pre-Exascale Supercomputers

Scaling in a Heterogeneous Environment with GPUs: GPU Architecture, Concepts, and Strategies

Using AWS EC2 GPU Instances for Computational Microscopy and Biomolecular Simulation

CUDA Applications I. John E. Stone

B. Evaluation and Exploration of Next Generation Systems for Applicability and Performance (Volodymyr Kindratenko, Guochun Shi)

NAMD, CUDA, and Clusters: Taking GPU Molecular Dynamics Beyond the Deskop

VMD Key Features of Recent Release

Advances In Biomolecular Simulation with NAMD and VMD

OPTIMIZING HPC SIMULATION AND VISUALIZATION CODE USING NVIDIA NSIGHT SYSTEMS

S05: High Performance Computing with CUDA. Case Study: Molecular Visualization and Analysis

Bridging the Gap Between High Quality and High Performance for HPC Visualization

ENABLING NEW SCIENCE GPU SOLUTIONS

GPU Particle-Grid Methods: Electrostatics

Quantifying the Impact of GPUs on Performance and Energy Efficiency in HPC Clusters

Petascale Multiscale Simulations of Biomolecular Systems. John Grime Voth Group Argonne National Laboratory / University of Chicago

NAMD Serial and Parallel Performance

GPU Accelerated Visualization and Analysis in VMD

The Parallel Revolution in Computational Science and Engineering

High Performance Computing with Accelerators

Experience with NAMD on GPU-Accelerated Clusters

Dancing Proteins: 3-D Visualization of Protein Structure and Dynamics on Next-Generation Graphics Hardware

Abhinav Bhatele, Laxmikant V. Kale University of Illinois at Urbana Champaign Sameer Kumar IBM T. J. Watson Research Center

NAMD GPU Performance Benchmark. March 2011

CUDA Experiences: Over-Optimization and Future HPC

VSCSE Summer School Lecture 8: Application Case Study Accelerating Molecular Dynamics Experimentation

Evaluation and Exploration of Next Generation Systems for Applicability and Performance Volodymyr Kindratenko Guochun Shi

GPU Accelerated Visualization and Analysis in VMD and Recent NAMD Developments

Fast Visualization of Gaussian Density Surfaces for Molecular Dynamics and Particle System Trajectories

Debugging CUDA Applications with Allinea DDT. Ian Lumb Sr. Systems Engineer, Allinea Software Inc.

Accelerating Biomolecular Modeling with CUDA and GPU Clusters

IMPORTANT: Visit for the most up-to-date schedule.

Overview and Power Monitoring of NCSA-UIUC- NVIDIA EcoG Computational Cluster

Software and Performance Engineering for numerical codes on GPU clusters

Fast Molecular Electrostatics Algorithms on GPUs

NAMD at Extreme Scale. Presented by: Eric Bohm Team: Eric Bohm, Chao Mei, Osman Sarood, David Kunzman, Yanhua, Sun, Jim Phillips, John Stone, LV Kale

Graphics Processor Acceleration and YOU

2009: The GPU Computing Tipping Point. Jen-Hsun Huang, CEO

A Parallel-Object Programming Model for PetaFLOPS Machines and BlueGene/Cyclops Gengbin Zheng, Arun Singla, Joshua Unger, Laxmikant Kalé

GPU Acceleration of Particle Advection Workloads in a Parallel, Distributed Memory Setting

IBM J. Res. & Dev.Acun et al.: Scalable Molecular Dynamics with NAMD on the Summit System Page 1

OpenCL: Molecular Modeling on Heterogeneous Computing Systems

Accelerating NAMD with Graphics Processors

NAMD-QM/MM TUTORIAL. Unix/MacOSX Version. NAMD-QM/MM Suite Developers: Marcelo C. R. Melo, Rafael C. Bernardi

Parallel Geospatial Data Management for Multi-Scale Environmental Data Analysis on GPUs DOE Visiting Faculty Program Project Report

PERFORMANCE PORTABILITY WITH OPENACC. Jeff Larkin, NVIDIA, November 2015

Scheduling Strategies for HPC as a Service (HPCaaS) for Bio-Science Applications

Journal of Computational Physics

GPU-accelerated data expansion for the Marching Cubes algorithm

John E. Stone. Intel HPC Developer Conference, Sheraton Hotel Sunday, Nov 13 th, 2016, Salt Lake City, UT

GPU Clusters for High- Performance Computing Jeremy Enos Innovative Systems Laboratory

Transcription:

In-Situ Visualization and Analysis of Petascale Molecular Dynamics Simulations with VMD John Stone Theoretical and Computational Biophysics Group Beckman Institute for Advanced Science and Technology University of Illinois at Urbana-Champaign Research/vmd/ Accelerated HPC Symposium, San Jose Convention Center, San Jose, CA, May 17, 2012

VMD Visual Molecular Dynamics Visualization and analysis of: molecular dynamics simulations quantum chemistry calculations particle systems and whole cells sequence data User extensible w/ scripting and plugins Research/vmd/ Poliovirus Ribosome Sequences Electrons in Vibrating Buckyball Cellular Tomography, Cryo-electron Microscopy Whole Cell Simulations

Molecular Visualization and Analysis Challenges for Petascale Simulations Very large structures (10M to over 100M atoms) 12-bytes per atom per trajectory frame 100M atom trajectory frame: 1200MB! Long-timescale simulations produce huge trajectories MD integration timesteps are on the femtosecond timescale (10-15 sec) but many important biological processes occur on microsecond to millisecond timescales Even storing trajectory frames infrequently, resulting trajectories frequently contain millions of frames Petabytes of data to analyze, far too large to move Viz and analysis must be done primarily on the supercomputer where the data already resides

Approaches for Visualization and Analysis of Petascale Molecular Simulations with VMD Abandon conventional approaches, e.g. bulk download of trajectory data to remote viz/analysis machines In-place processing of trajectories on the machine running the simulations Use remote visualization techniques: Split-mode VMD with remote frontend instance, and back-end viz/analysis engine running in parallel on supercomputer Large-scale parallel analysis and visualization via distributed memory MPI version of VMD Exploit GPUs and other accelerators to increase per-node analytical capabilities, e.g. NCSA Blue Waters Cray XK6 In-situ on-the-fly viz/analysis and event detection through direct communication with running MD simulation

Improved Support for Large Datasets in VMD New structure building tools, file formats, and data structures enable VMD to operate efficiently up to 150M atoms Up to 30% more memory efficient Analysis routines optimized for large structures, up to 20x faster for calculations on 100M atom complexes where molecular structure traversal can represent a significant amount of runtime New and revised graphical representations support smooth trajectory animation for multi-million atom complexes; VMD remains interactive even when displaying surface reps for 20M atom membrane patch Uses multi-core CPUs and GPUs for the most demanding computations 20M atoms: membrane patch and solvent

New Interactive Display & Analysis of Terabytes of Data: Out-of-Core Trajectory I/O w/ Solid State Disks Timesteps loaded on-the-fly (out-of-core) Eliminates memory capacity limitations, even for multi-terabyte trajectory files High performance achieved by new trajectory file formats, optimized data structures, and efficient I/O Analyze long trajectories significantly faster 450MB/sec to 4GB/sec A DVD movie per second! New SSD Trajectory File Format 2x Faster vs. Existing Formats Immersive out-of-core visualization of large-size and long-timescale molecular dynamics trajectories. J. Stone, K. Vandivort, and K. Schulten. Lecture Notes in Computer Science, 6939:1-12, 2011. Commodity SSD, SSD RAID

VMD Out-of-Core Trajectory I/O Performance: SSD-Optimized Trajectory Format, 8-SSD RAID Ribosome w/ solvent 3M atoms 3 frames/sec w/ HD 60 frames/sec w/ SSDs Membrane patch w/ solvent 20M atoms 0.4 frames/sec w/ HD 8 frames/sec w/ SSDs New SSD Trajectory File Format 2x Faster vs. Existing Formats VMD I/O rate ~2.1 GB/sec w/ 8 SSDs

Parallel VMD Analysis w/ MPI Analyze trajectory frames, structures, or sequences in parallel supercomputers: Parallelize user-written analysis scripts with minimum difficulty Parallel analysis of independent trajectory frames Parallel structural analysis using custom parallel reductions Parallel rendering, movie making Dynamic load balancing: Recently tested with up to 15,360 CPU cores Supports GPU-accelerated clusters and supercomputers Sequence/Structure Data, Trajectory Frames, etc VMD VMD VMD Gathered Results Data-Parallel Analysis in VMD

Molecular Structure Data and Global VMD State Scene Graph Graphical Representations DrawMolecule Non-Molecular Geometry User Interface Subsystem Tcl/Python Scripting Mouse + Windows VR Tools Interactive MD TachyonDisplayDevice OptiXDisplayDevice DisplayDevice OpenGLRenderer Display Subsystem Windowed OpenGL CAVE FreeVR

GPU Accelerated Trajectory Analysis and Visualization in VMD GPU-Accelerated Feature Molecular orbital display Radial distribution function Electrostatic field calculation Molecular surface display Ion placement MDFF density map synthesis Implicit ligand sampling Root mean squared fluctuation Radius of gyration Close contact determination Dipole moment calculation Speedup vs. single CPU core 120x 92x 44x 40x 26x 26x 25x 25x 21x 20x 15x

NCSA Blue Waters Early Science System Cray XK6 nodes w/ NVIDIA Tesla X2090

Time-Averaged Electrostatics Analysis on NCSA Blue Waters Early Science System NCSA Blue Waters Node Type Cray XE6 Compute Node: 32 CPU cores (2xAMD 6200 CPUs) Cray XK6 GPU-accelerated Compute Node: 16 CPU cores + NVIDIA X2090 (Fermi) GPU Speedup for GPU XK6 nodes vs. CPU XE6 nodes Seconds per trajectory frame for one compute node 9.33 2.25 GPU nodes are 4.15x faster overall Preliminary performance for VMD time-averaged electrostatics w/ Multilevel Summation Method on the NCSA Blue Waters Early Science System

In-Situ Visualization and Analysis with VMD Early prototype and testing phase VMD supports live socket connection to running MD code Custom user-written analysis scripts are triggered by callbacks as incoming frames arrive Separate threads handle async. network I/O between MD code and master VMD instance, MPI broadcast or decomposition among peer VMD nodes Perform real-time analysis processing of incoming frames to find rare events Store the most interesting timesteps Build summary analyses useful for accelerating interactive Timeline displays, and subsequent detailed batch mode analysis runs Live, running MD simulation, e.g. NAMD running on thousands of compute nodes VMD VMD VMD Data-Parallel Analysis, Visualization, in VMD Store only interesting trajectory frames

Acknowledgements Theoretical and Computational Biophysics Group, University of Illinois at Urbana- Champaign NCSA Blue Waters Team NCSA Innovative Systems Lab NVIDIA CUDA Center of Excellence, University of Illinois at Urbana-Champaign The CUDA team at NVIDIA NIH support: P41-RR005969

GPU Computing Publications Research/gpu/ Fast Visualization of Gaussian Density Surfaces for Molecular Dynamics and Particle System Trajectories. M. Krone, J. Stone, T. Ertl, and K. Schulten. In proceedings EuroVis 2012, 2012. (In-press) Immersive Out-of-Core Visualization of Large-Size and Long- Timescale Molecular Dynamics Trajectories. J. Stone, K. Vandivort, and K. Schulten. G. Bebis et al. (Eds.): 7th International Symposium on Visual Computing (ISVC 2011), LNCS 6939, pp. 1-12, 2011. Fast Analysis of Molecular Dynamics Trajectories with Graphics Processing Units Radial Distribution Functions. B. Levine, J. Stone, and A. Kohlmeyer. J. Comp. Physics, 230(9):3556-3569, 2011.

GPU Computing Publications Research/gpu/ Quantifying the Impact of GPUs on Performance and Energy Efficiency in HPC Clusters. J. Enos, C. Steffen, J. Fullop, M. Showerman, G. Shi, K. Esler, V. Kindratenko, J. Stone, J Phillips. International Conference on Green Computing, pp. 317-324, 2010. GPU-accelerated molecular modeling coming of age. J. Stone, D. Hardy, I. Ufimtsev, K. Schulten. J. Molecular Graphics and Modeling, 29:116-125, 2010. OpenCL: A Parallel Programming Standard for Heterogeneous Computing. J. Stone, D. Gohara, G. Shi. Computing in Science and Engineering, 12(3):66-73, 2010. An Asymmetric Distributed Shared Memory Model for Heterogeneous Computing Systems. I. Gelado, J. Stone, J. Cabezas, S. Patel, N. Navarro, W. Hwu. ASPLOS 10: Proceedings of the 15 th International Conference on Architectural Support for Programming Languages and Operating Systems, pp. 347-358, 2010.

GPU Computing Publications Research/gpu/ GPU Clusters for High Performance Computing. V. Kindratenko, J. Enos, G. Shi, M. Showerman, G. Arnold, J. Stone, J. Phillips, W. Hwu. Workshop on Parallel Programming on Accelerator Clusters (PPAC), In Proceedings IEEE Cluster 2009, pp. 1-8, Aug. 2009. Long time-scale simulations of in vivo diffusion using GPU hardware. E. Roberts, J. Stone, L. Sepulveda, W. Hwu, Z. Luthey-Schulten. In IPDPS 09: Proceedings of the 2009 IEEE International Symposium on Parallel & Distributed Computing, pp. 1-8, 2009. High Performance Computation and Interactive Display of Molecular Orbitals on GPUs and Multi-core CPUs. J. Stone, J. Saam, D. Hardy, K. Vandivort, W. Hwu, K. Schulten, 2nd Workshop on General-Purpose Computation on Graphics Pricessing Units (GPGPU-2), ACM International Conference Proceeding Series, volume 383, pp. 9-18, 2009. Probing Biomolecular Machines with Graphics Processors. J. Phillips, J. Stone. Communications of the ACM, 52(10):34-41, 2009. Multilevel summation of electrostatic potentials using graphics processing units. D. Hardy, J. Stone, K. Schulten. J. Parallel Computing, 35:164-177, 2009.

GPU Computing Publications Research/gpu/ Adapting a message-driven parallel application to GPU-accelerated clusters. J. Phillips, J. Stone, K. Schulten. Proceedings of the 2008 ACM/IEEE Conference on Supercomputing, IEEE Press, 2008. GPU acceleration of cutoff pair potentials for molecular modeling applications. C. Rodrigues, D. Hardy, J. Stone, K. Schulten, and W. Hwu. Proceedings of the 2008 Conference On Computing Frontiers, pp. 273-282, 2008. GPU computing. J. Owens, M. Houston, D. Luebke, S. Green, J. Stone, J. Phillips. Proceedings of the IEEE, 96:879-899, 2008. Accelerating molecular modeling applications with graphics processors. J. Stone, J. Phillips, P. Freddolino, D. Hardy, L. Trabuco, K. Schulten. J. Comp. Chem., 28:2618-2640, 2007. Continuous fluorescence microphotolysis and correlation spectroscopy. A. Arkhipov, J. Hüve, M. Kahms, R. Peters, K. Schulten. Biophysical Journal, 93:4006-4017, 2007.