Application Example Running on Top of GPI-Space Integrating D/C
|
|
- Johnathan Chase
- 5 years ago
- Views:
Transcription
1 Application Example Running on Top of GPI-Space Integrating D/C Tiberiu Rotaru Fraunhofer ITWM This project is funded from the European Union s Horizon 2020 Research and Innovation programme under Grant Agreement no
2 GPI-Space Software platform intended to help the domain developers to build specific complex parallel applications. The main idea behind GPI-Space is separating the domain specific programming layer from the HPC software infrastructure. Application domains: seismic imaging, financial services, life sciences, mechanical engineering.
3 The Distributed Runtime System of GPI-Space User Orchestrator Agent Agent Worker Worker Worker Worker Virtual Memory Manager Event-driven architecture (asynchronous communication). The components may have multiple masters, slaves or subscribers. Can be deployed in dynamically configurable topologies. The workers may have assigned specific capabilities (e.g. I/O, compute, etc.). The tasks may have requirements. The workers may join and leave the system at any time without affecting the execution of the workflow or other workers.
4 The virtual memory layer of GPI-Space Used for storing data shared between tasks. There is no direct communication between tasks or workers. Features: The storage layer can be easily changed, depending on the application s memory requirements (e.g. in-memory or on disk) Tolerance to worker failures: no impact on other workers nor on the execution of the workflow Shortcomings of the current virtual memory layer: lack of a caching mechanism not an independent software module that can be developed and tested standalone
5 Directory/Cache client-server architecture Node 0 Node 1 Node 2 Node 3 Segment 0 Segment 0 Segment 0 Segment 0 GASPI MPI BeeGFS Segment 1 Segment 1 Segment 1 Original Data Server Cache Cache Cache Cache Cache Cache Scratch Copies Client Client Client Client Client Client Client Client Multiple segments, caches and clients. The caches may be shared between clients. The local servers coordinate with each other for carrying out operations with memory ranges in a consistent way. External programs may connect to an already running D/C. Tolerance to client failures (client and server in different processes).
6 D/C prototype implementation A working prototype implementation of D/C API is available at: Used C++11 advanced features (variadic templates, lambda expressions, asynchronous launching, futures, etc.) The prototype provides implementation for GASPI, MPI and BeeGFS segments. Other segment implementations can be added without modifying the D/C existing implementation. Different allocation policies for segments and data are supported. Direct coupling with applications possible: e.g. Jacobi iteration for solving the Laplace equation on a 2D mesh of processes (tested with GASPI and MPI segments). Intended to be used by the task-based runtimes, primarily.
7 Integration of GPI-Space with the Directory/Cache Goals: Replace the existing virtual memory layer with D/C and achieve the same functionality while ensuring that all tests are working fine. Take advantage of using the D/C features. Scenario: use a shared cache per node and a private cache per worker. Changes required at multiple levels: Runtime API: methods for allocating and freeing segments or data. Workflow engine: workflow description language, parser, module loader. Bootstrapping: on each node, before starting the workers, do: start one or multiple D/C server instances create shared caches Workers: Create at startup D/C inter-process clients and private caches Inform the masters about the caches used Master agent: use D/C API transfer costs for scheduling.
8 Memory buffers In GPI-Space the transfers are managed by the runtime. The buffers used in transfers are described in the workflow. There are either const or mutable buffers. A buffer in GPI-Space equates to a local range in the D/C. The buffers may have attributes (e.g. if it is read-only or not). <memory-buffer name="seek_table" read-only="true"> <size> ${size_of_seek_table} </size> </memory-buffer> <memory-buffer name="data read-only= false"> <size> ${data_buffer.size} </size> </memory-buffer>
9 Managed memory transfers The workflow parser generates lists of get and put operations per task, which are executed by the runtime, when executing a task, using one of the D/C API methods. <memory-get> <global> ${range.handle} := ${seek_table.handle}; ${range.offset} := 0UL; ${range.size} := ${size_of_seek_table}; stack_push (List(), ${range}); </global> <local> ${range.buffer} := "seek_table"; ${range.offset} := 0UL; ${range.size} := ${size_of_seek_table}; stack_push (List(), ${range}); </local> </memory-get> <memory-put> <global> ${range.handle} := ${data.handle}; ${range.offset} := 0UL; ${range.size} := ${size_of_data}; stack_push (List(), ${range} </global> <local> ${range.buffer} := "data"; ${range.offset} := 0UL; ${range.size} := ${data_buffer.size}; stack_push (List(), ${range}); </local> </memory-put>
10 The renderer Splotch Splotch: publicly available rendering software used for exploration and visual discovery in particle-based datasets produced by astronomical observations or numerical simulations. Development started at Max-Planck Institute for Astrophysics München. Several academic institutions are contributing. The rendering algorithm optimizes the ray-tracing calculation by ordering the particles according to a depth function. The algorithm produces high-quality imagery (realistic 3D impressions) and works with very large-scale data sets. Original version: OpenMP/MPI hybrid implementation. Problem: low parallel efficiency (only 26% on the SuperMUC cluster at Leibniz Supercomputer Centre)
11 Example: formation and evolution of a galaxy cluster The upper panels show a region of the universe at an early time. It is rich of filamentary structures, and later, when the universe evolves, almost all these structures are going to collapse to form a single, prominent galaxy cluster. The lower panels show the same region at a later time, where the two main progenitors of the final galaxy cluster are going to merge. From: Splotch: visualizing cosmological simulations (New Journal of Physics, Volume 10, Issue 12, pp (2008)).
12 First version of Splotch implemented on top of GPI-Space Task-based implementation instead of hybrid OpenMP/MPI implementation. The data-flow was re-organized s.t. to allow an overlay of the rendering part with the non-parallel parts of the algorithm (e.g. I/O or the aggregation of partial images). User is required to provide implementation only for three interface functions, which can be developed, tested and compiled without GPI-Space as a dynamic library. GPI-Space provides automatic support for scheduling, dynamic load balancing and fault-tolerance. Used GPI-2 for data transfers (low latencies and high bandwidth)
13 Workflow of the parallelized Splotch algorithm The Splotch algorithm operational scenario consists of a number of stages: a) read data from one or more files, b) process data (e.g. for normalization), c) render data and d) save the final image.
14 Splotch parameters For each simulation Splotch uses a parameter file (contains color, intensity, brightness, palette, etc). Scene file: contains scene descriptions (e.g. camera positions) The input contains specific data for a number of scenes. Splotch produces for each scene a picture (frame). The workflow combines 4 types of tasks: load, make_picture, reduce, store. Invariants: the number of make_picture, reduce and store tasks is proportional to the number of scenes.
15 GPI-Space implementation vs the original version Test data: data-set with 5 billion particles (167 GByte) frames to render (fly-through the data-set at a fixed time step). Output: 161 GByte of picture data. Data chunks of 2 Gbyte size make_picture tasks, reduce tasks, 3245 store tasks 32 Gbyte memory used on each node On 128 nodes: OpenMP/ MPI-Splotch has the same runtime as on 32 nodes The GPI-Space version outperforms the OpenMP/MPI version by a factor of 10.
16 Running Splotch on top of GPI-Space integrating D/C In Splotch, scene_parameters and the seek_table are constant parameters. The node-local shared cache is used for storing const local ranges. The private cache of a worker is used for storing mutable local ranges of assigned tasks. Policies used: evenly_distributed for segments and fill_from_left for data allocations. The application required minor modifications w.r.t segment and data allocation for running on GPI-Space integrating D/C (the D/C API is called under the hood).
17 Allocating memory for the global data by the runtime The D/C API is called under the hood by the runtime, the application developer is not required to call explicitly D/C API methods. gspc::scoped_vmem_segment_and_allocation const allocation_scene_parameters (drts.alloc_and_fill ( gspc::vmem::gaspi_segment_description(), memory_for_scene_parameters_in_bytes, scenes_parameter.data() ) ); gspc::scoped_vmem_segment_and_allocation const allocation_seek_table (drts.alloc_and_fill ( gspc::vmem::gaspi_segment_description(), seek_table.size(), seek_table.data() ) ); gspc::scoped_vmem_segment_and_allocation const allocation_data (drts.alloc ( gspc::vmem::gaspi_segment_description(), size_of_data ) ); gspc::scoped_vmem_segment_and_allocation const allocation_picture (drts.alloc ( gspc::vmem::gaspi_segment_description(), memory_for_pictures ) );
18 Splotch running on top of GPI-Space + D/C on 30 nodes (100 scenes)
19 Output validation Tests performed with various numbers of nodes and scenes. The reference output contains 3425 frames ( 161 GB of data). In all cases the output was compared against the reference output produced by the original application. The produced frames are identical to those in the reference output (up to a very small tolerance error). This proves that the Directory/Cache API operations are correctly implemented, resulting in no data corruption. Note: the correctness of data transfers was also previously checked with the Jacobi 2D example implemented directly on top of D/C.
20 Execution times for varying numbers of scenes and nodes #Scenes #Tasks Execution times of simulations with 80, 90, and 100 scenes on sets of 20 to 30 cluster nodes, each with 16 Intel Xeon E cores.
21 Evaluating the cache reuse Total number of cache misses compared with the number of cache hits per rank for a simulation with 20 nodes and 100 scenes. The maximum number of data transfers from the global memory into a node-local shared cache is the number of scenes plus one. As shown in the graph this limit is never exceeded (i.e. 101)
22 Improvement in terms of absolut times The cache reuse leads to improvement of the execution times (compared to the situation when no caching is performed). The improvement depends on the size of the cached data and the number of const transfers. The gain is higher when the size of cached data is larger.
23 Splotch on GPI-Space integrating D/C vs Splotch on GPI- Space using the current virtual memory layer The version using the current virtual memory layer is up to 2,2% faster than the one over GPI-Space + D/C. Not too bad, taking into account that the D/C implementation is just a prototype and the current virtual memory layer is highly optimized, being used for years already. The D/C prototype implementation can be further optimized. number of scenes with D/C current difference in % ,38% ,19% ,02% ,13% ,12%
24 Conclusions and future work The results produced by the application (Splotch) are correct, which proves that the API is correctly implemented The application scales and it benefits from automatic caching. The architectural design of D/C allows GPI-Space to preserve important features such as: easy switching between storage layers, depending on the application requirements and tolerance to worker failures. Long term goal: replace the current version of the virtual memory manager with an optimized implementation of D/C. In order to exploit the full potential of the API, further changes touching multiple layers of the GPI-Space architecture are planned.
THOUGHTS ABOUT THE FUTURE OF I/O
THOUGHTS ABOUT THE FUTURE OF I/O Dagstuhl Seminar Challenges and Opportunities of User-Level File Systems for HPC Franz-Josef Pfreundt, May 2017 Deep Learning I/O Challenges Memory Centric Computing :
More informationTowards Exascale Programming Models HPC Summit, Prague Erwin Laure, KTH
Towards Exascale Programming Models HPC Summit, Prague Erwin Laure, KTH 1 Exascale Programming Models With the evolution of HPC architecture towards exascale, new approaches for programming these machines
More informationSplotch: High Performance Visualization using MPI, OpenMP and CUDA
Splotch: High Performance Visualization using MPI, OpenMP and CUDA Klaus Dolag (Munich University Observatory) Martin Reinecke (MPA, Garching) Claudio Gheller (CSCS, Switzerland), Marzia Rivi (CINECA,
More informationSSS: An Implementation of Key-value Store based MapReduce Framework. Hirotaka Ogawa (AIST, Japan) Hidemoto Nakada Ryousei Takano Tomohiro Kudoh
SSS: An Implementation of Key-value Store based MapReduce Framework Hirotaka Ogawa (AIST, Japan) Hidemoto Nakada Ryousei Takano Tomohiro Kudoh MapReduce A promising programming tool for implementing largescale
More informationParallel Computing Using OpenMP/MPI. Presented by - Jyotsna 29/01/2008
Parallel Computing Using OpenMP/MPI Presented by - Jyotsna 29/01/2008 Serial Computing Serially solving a problem Parallel Computing Parallelly solving a problem Parallel Computer Memory Architecture Shared
More informationGetting Performance from OpenMP Programs on NUMA Architectures
Getting Performance from OpenMP Programs on NUMA Architectures Christian Terboven, RWTH Aachen University terboven@itc.rwth-aachen.de EU H2020 Centre of Excellence (CoE) 1 October 2015 31 March 2018 Grant
More informationExtreme I/O Scaling with HDF5
Extreme I/O Scaling with HDF5 Quincey Koziol Director of Core Software Development and HPC The HDF Group koziol@hdfgroup.org July 15, 2012 XSEDE 12 - Extreme Scaling Workshop 1 Outline Brief overview of
More informationPerformance optimization of the Smoothed Particle Hydrodynamics code Gadget3 on 2nd generation Intel Xeon Phi
Performance optimization of the Smoothed Particle Hydrodynamics code Gadget3 on 2nd generation Intel Xeon Phi Dr. Luigi Iapichino Leibniz Supercomputing Centre Supercomputing 2017 Intel booth, Nerve Center
More informationTechnologies for High Performance Data Analytics
Technologies for High Performance Data Analytics Dr. Jens Krüger Fraunhofer ITWM 1 Fraunhofer ITWM n Institute for Industrial Mathematics n Located in Kaiserslautern, Germany n Staff: ~ 240 employees +
More informationEvaluation and Improvements of Programming Models for the Intel SCC Many-core Processor
Evaluation and Improvements of Programming Models for the Intel SCC Many-core Processor Carsten Clauss, Stefan Lankes, Pablo Reble, Thomas Bemmerl International Workshop on New Algorithms and Programming
More informationMunara Tolubaeva Technical Consulting Engineer. 3D XPoint is a trademark of Intel Corporation in the U.S. and/or other countries.
Munara Tolubaeva Technical Consulting Engineer 3D XPoint is a trademark of Intel Corporation in the U.S. and/or other countries. notices and disclaimers Intel technologies features and benefits depend
More informationApproaches to I/O Scalability Challenges in the ECMWF Forecasting System
Approaches to I/O Scalability Challenges in the ECMWF Forecasting System PASC 16, June 9 2016 Florian Rathgeber, Simon Smart, Tiago Quintino, Baudouin Raoult, Stephan Siemen, Peter Bauer Development Section,
More informationParallel Mesh Partitioning in Alya
Available online at www.prace-ri.eu Partnership for Advanced Computing in Europe Parallel Mesh Partitioning in Alya A. Artigues a *** and G. Houzeaux a* a Barcelona Supercomputing Center ***antoni.artigues@bsc.es
More informationA Scalable Adaptive Mesh Refinement Framework For Parallel Astrophysics Applications
A Scalable Adaptive Mesh Refinement Framework For Parallel Astrophysics Applications James Bordner, Michael L. Norman San Diego Supercomputer Center University of California, San Diego 15th SIAM Conference
More informationMOHA: Many-Task Computing Framework on Hadoop
Apache: Big Data North America 2017 @ Miami MOHA: Many-Task Computing Framework on Hadoop Soonwook Hwang Korea Institute of Science and Technology Information May 18, 2017 Table of Contents Introduction
More informationMERCED CLUSTER BASICS Multi-Environment Research Computer for Exploration and Discovery A Centerpiece for Computational Science at UC Merced
MERCED CLUSTER BASICS Multi-Environment Research Computer for Exploration and Discovery A Centerpiece for Computational Science at UC Merced Sarvani Chadalapaka HPC Administrator University of California
More informationA Container On a Virtual Machine On an HPC? Presentation to HPC Advisory Council. Perth, July 31-Aug 01, 2017
A Container On a Virtual Machine On an HPC? Presentation to HPC Advisory Council Perth, July 31-Aug 01, 2017 http://levlafayette.com Necessary and Sufficient Definitions High Performance Computing: High
More informationTowards modernisation of the Gadget code on many-core architectures Fabio Baruffa, Luigi Iapichino (LRZ)
Towards modernisation of the Gadget code on many-core architectures Fabio Baruffa, Luigi Iapichino (LRZ) Overview Modernising P-Gadget3 for the Intel Xeon Phi : code features, challenges and strategy for
More informationGPI-2: a PGAS API for asynchronous and scalable parallel applications
GPI-2: a PGAS API for asynchronous and scalable parallel applications Rui Machado CC-HPC, Fraunhofer ITWM Barcelona, 13 Jan. 2014 1 Fraunhofer ITWM CC-HPC Fraunhofer Institute for Industrial Mathematics
More informationExploiting peer group concept for adaptive and highly available services
Computing in High Energy and Nuclear Physics, 24-28 March 2003 La Jolla California 1 Exploiting peer group concept for adaptive and highly available services Muhammad Asif Jan Centre for European Nuclear
More informationOpenACC 2.6 Proposed Features
OpenACC 2.6 Proposed Features OpenACC.org June, 2017 1 Introduction This document summarizes features and changes being proposed for the next version of the OpenACC Application Programming Interface, tentatively
More informationAdaptive Cluster Computing using JavaSpaces
Adaptive Cluster Computing using JavaSpaces Jyoti Batheja and Manish Parashar The Applied Software Systems Lab. ECE Department, Rutgers University Outline Background Introduction Related Work Summary of
More informationPerformance of the 3D-Combustion Simulation Code RECOM-AIOLOS on IBM POWER8 Architecture. Alexander Berreth. Markus Bühler, Benedikt Anlauf
PADC Anual Workshop 20 Performance of the 3D-Combustion Simulation Code RECOM-AIOLOS on IBM POWER8 Architecture Alexander Berreth RECOM Services GmbH, Stuttgart Markus Bühler, Benedikt Anlauf IBM Deutschland
More informationFVM - How to program the Multi-Core FVM instead of MPI
FVM - How to program the Multi-Core FVM instead of MPI DLR, 15. October 2009 Dr. Mirko Rahn Competence Center High Performance Computing and Visualization Fraunhofer Institut for Industrial Mathematics
More informationBei Wang, Dmitry Prohorov and Carlos Rosales
Bei Wang, Dmitry Prohorov and Carlos Rosales Aspects of Application Performance What are the Aspects of Performance Intel Hardware Features Omni-Path Architecture MCDRAM 3D XPoint Many-core Xeon Phi AVX-512
More informationChallenges of Big Data Movement in support of the ESA Copernicus program and global research collaborations
APAN Cloud WG Challenges of Big Data Movement in support of the ESA Copernicus program and global research collaborations Lift off NCI and Copernicus The National Computational Infrastructure (NCI) in
More informationAutoTune Workshop. Michael Gerndt Technische Universität München
AutoTune Workshop Michael Gerndt Technische Universität München AutoTune Project Automatic Online Tuning of HPC Applications High PERFORMANCE Computing HPC application developers Compute centers: Energy
More informationPLB-HeC: A Profile-based Load-Balancing Algorithm for Heterogeneous CPU-GPU Clusters
PLB-HeC: A Profile-based Load-Balancing Algorithm for Heterogeneous CPU-GPU Clusters IEEE CLUSTER 2015 Chicago, IL, USA Luis Sant Ana 1, Daniel Cordeiro 2, Raphael Camargo 1 1 Federal University of ABC,
More information6.1 Multiprocessor Computing Environment
6 Parallel Computing 6.1 Multiprocessor Computing Environment The high-performance computing environment used in this book for optimization of very large building structures is the Origin 2000 multiprocessor,
More informationMost real programs operate somewhere between task and data parallelism. Our solution also lies in this set.
for Windows Azure and HPC Cluster 1. Introduction In parallel computing systems computations are executed simultaneously, wholly or in part. This approach is based on the partitioning of a big task into
More informationIBM High Performance Computing Toolkit
IBM High Performance Computing Toolkit Pidad D'Souza (pidsouza@in.ibm.com) IBM, India Software Labs Top 500 : Application areas (November 2011) Systems Performance Source : http://www.top500.org/charts/list/34/apparea
More informationAUTOMATIC SMT THREADING
AUTOMATIC SMT THREADING FOR OPENMP APPLICATIONS ON THE INTEL XEON PHI CO-PROCESSOR WIM HEIRMAN 1,2 TREVOR E. CARLSON 1 KENZO VAN CRAEYNEST 1 IBRAHIM HUR 2 AAMER JALEEL 2 LIEVEN EECKHOUT 1 1 GHENT UNIVERSITY
More informationCafeGPI. Single-Sided Communication for Scalable Deep Learning
CafeGPI Single-Sided Communication for Scalable Deep Learning Janis Keuper itwm.fraunhofer.de/ml Competence Center High Performance Computing Fraunhofer ITWM, Kaiserslautern, Germany Deep Neural Networks
More informationCloud Programming. Programming Environment Oct 29, 2015 Osamu Tatebe
Cloud Programming Programming Environment Oct 29, 2015 Osamu Tatebe Cloud Computing Only required amount of CPU and storage can be used anytime from anywhere via network Availability, throughput, reliability
More informationImproving Hadoop MapReduce Performance on Supercomputers with JVM Reuse
Thanh-Chung Dao 1 Improving Hadoop MapReduce Performance on Supercomputers with JVM Reuse Thanh-Chung Dao and Shigeru Chiba The University of Tokyo Thanh-Chung Dao 2 Supercomputers Expensive clusters Multi-core
More informationEnergy Efficiency Tuning: READEX. Madhura Kumaraswamy Technische Universität München
Energy Efficiency Tuning: READEX Madhura Kumaraswamy Technische Universität München Project Overview READEX Starting date: 1. September 2015 Duration: 3 years Runtime Exploitation of Application Dynamism
More informationMotivation Goal Idea Proposition for users Study
Exploring Tradeoffs Between Power and Performance for a Scientific Visualization Algorithm Stephanie Labasan Computer and Information Science University of Oregon 23 November 2015 Overview Motivation:
More informationPeta-Scale Simulations with the HPC Software Framework walberla:
Peta-Scale Simulations with the HPC Software Framework walberla: Massively Parallel AMR for the Lattice Boltzmann Method SIAM PP 2016, Paris April 15, 2016 Florian Schornbaum, Christian Godenschwager,
More informationsimulation framework for piecewise regular grids
WALBERLA, an ultra-scalable multiphysics simulation framework for piecewise regular grids ParCo 2015, Edinburgh September 3rd, 2015 Christian Godenschwager, Florian Schornbaum, Martin Bauer, Harald Köstler
More informationECE 7650 Scalable and Secure Internet Services and Architecture ---- A Systems Perspective
ECE 7650 Scalable and Secure Internet Services and Architecture ---- A Systems Perspective Part II: Data Center Software Architecture: Topic 3: Programming Models Piccolo: Building Fast, Distributed Programs
More informationEmploying HPC DEEP-EST for HEP Data Analysis. Viktor Khristenko (CERN, DEEP-EST), Maria Girone (CERN)
Employing HPC DEEP-EST for HEP Data Analysis Viktor Khristenko (CERN, DEEP-EST), Maria Girone (CERN) 1 Outline The DEEP-EST Project Goals and Motivation HEP Data Analysis on HPC with Apache Spark on HPC
More informationPOP CoE: Understanding applications and how to prepare for exascale
POP CoE: Understanding applications and how to prepare for exascale Jesus Labarta (BSC) EU H2020 Center of Excellence (CoE) Lecce, May 17 th 2018 5 th ENES HPC workshop POP objective Promote methodologies
More informationDynamic Load-Balanced Multicast for Data-Intensive Applications on Clouds 1
Dynamic Load-Balanced Multicast for Data-Intensive Applications on Clouds 1 Contents: Introduction Multicast on parallel distributed systems Multicast on P2P systems Multicast on clouds High performance
More informationGreen Supercomputing
Green Supercomputing On the Energy Consumption of Modern E-Science Prof. Dr. Thomas Ludwig German Climate Computing Centre Hamburg, Germany ludwig@dkrz.de Outline DKRZ 2013 and Climate Science The Exascale
More informationLeveraging Flash in HPC Systems
Leveraging Flash in HPC Systems IEEE MSST June 3, 2015 This work was performed under the auspices of the U.S. Department of Energy by under Contract DE-AC52-07NA27344. Lawrence Livermore National Security,
More informationAlgorithms, System and Data Centre Optimisation for Energy Efficient HPC
2015-09-14 Algorithms, System and Data Centre Optimisation for Energy Efficient HPC Vincent Heuveline URZ Computing Centre of Heidelberg University EMCL Engineering Mathematics and Computing Lab 1 Energy
More informationLarge-scale Ultrasound Simulations Using the Hybrid OpenMP/MPI Decomposition
Large-scale Ultrasound Simulations Using the Hybrid OpenMP/MPI Decomposition Jiri Jaros*, Vojtech Nikl*, Bradley E. Treeby *Department of Compute Systems, Brno University of Technology Department of Medical
More informationTOOLS FOR IMPROVING CROSS-PLATFORM SOFTWARE DEVELOPMENT
TOOLS FOR IMPROVING CROSS-PLATFORM SOFTWARE DEVELOPMENT Eric Kelmelis 28 March 2018 OVERVIEW BACKGROUND Evolution of processing hardware CROSS-PLATFORM KERNEL DEVELOPMENT Write once, target multiple hardware
More informationDICE: a Model-Driven DevOps Framework for Big Data
DICE: a Model-Driven DevOps Framework for Big Data Giuliano Casale Imperial College London DICE Horizon 2020 Project Grant Agreement no. 644869 http://www.dice-h2020.eu Funded by the Horizon 2020 Framework
More informationThe Google File System
The Google File System Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung December 2003 ACM symposium on Operating systems principles Publisher: ACM Nov. 26, 2008 OUTLINE INTRODUCTION DESIGN OVERVIEW
More informationlibhio: Optimizing IO on Cray XC Systems With DataWarp
libhio: Optimizing IO on Cray XC Systems With DataWarp May 9, 2017 Nathan Hjelm Cray Users Group May 9, 2017 Los Alamos National Laboratory LA-UR-17-23841 5/8/2017 1 Outline Background HIO Design Functionality
More informationBIG DATA. Using the Lambda Architecture on a Big Data Platform to Improve Mobile Campaign Management. Author: Sandesh Deshmane
BIG DATA Using the Lambda Architecture on a Big Data Platform to Improve Mobile Campaign Management Author: Sandesh Deshmane Executive Summary Growing data volumes and real time decision making requirements
More informationHPC IN EUROPE. Organisation of public HPC resources
HPC IN EUROPE Organisation of public HPC resources Context Focus on publicly-funded HPC resources provided primarily to enable scientific research and development at European universities and other publicly-funded
More informationEvaluating New Communication Models in the Nek5000 Code for Exascale
Evaluating New Communication Models in the Nek5000 Code for Exascale Ilya Ivanov (KTH), Rui Machado (Fraunhofer), Mirko Rahn (Fraunhofer), Dana Akhmetova (KTH), Erwin Laure (KTH), Jing Gong (KTH), Philipp
More informationOpenNebula on VMware: Cloud Reference Architecture
OpenNebula on VMware: Cloud Reference Architecture Version 1.2, October 2016 Abstract The OpenNebula Cloud Reference Architecture is a blueprint to guide IT architects, consultants, administrators and
More informationBuilding a Secure Data Environment. Maureen Dougherty Center for High-Performance Computing University of Southern California
Building a Secure Data Environment Maureen Dougherty Center for High-Performance Computing University of Southern California Secure Data Environment USC Center for High-Performance Computing (HPC) created
More informationA distributed rendering architecture for ray tracing large scenes on commodity hardware. FlexRender. Bob Somers Zoe J.
FlexRender A distributed rendering architecture for ray tracing large scenes on commodity hardware. GRAPP 2013 Bob Somers Zoe J. Wood Increasing Geometric Complexity Normal Maps artifacts on silhouette
More informationENERGY-EFFICIENT VISUALIZATION PIPELINES A CASE STUDY IN CLIMATE SIMULATION
ENERGY-EFFICIENT VISUALIZATION PIPELINES A CASE STUDY IN CLIMATE SIMULATION Vignesh Adhinarayanan Ph.D. (CS) Student Synergy Lab, Virginia Tech INTRODUCTION Supercomputers are constrained by power Power
More informationTrafficDB: HERE s High Performance Shared-Memory Data Store Ricardo Fernandes, Piotr Zaczkowski, Bernd Göttler, Conor Ettinoffe, and Anis Moussa
TrafficDB: HERE s High Performance Shared-Memory Data Store Ricardo Fernandes, Piotr Zaczkowski, Bernd Göttler, Conor Ettinoffe, and Anis Moussa EPL646: Advanced Topics in Databases Christos Hadjistyllis
More informationDavid R. Mackay, Ph.D. Libraries play an important role in threading software to run faster on Intel multi-core platforms.
Whitepaper Introduction A Library Based Approach to Threading for Performance David R. Mackay, Ph.D. Libraries play an important role in threading software to run faster on Intel multi-core platforms.
More informationAcuSolve Performance Benchmark and Profiling. October 2011
AcuSolve Performance Benchmark and Profiling October 2011 Note The following research was performed under the HPC Advisory Council activities Participating vendors: Intel, Dell, Mellanox, Altair Compute
More informationMicrosoft Windows HPC Server 2008 R2 for the Cluster Developer
50291B - Version: 1 02 May 2018 Microsoft Windows HPC Server 2008 R2 for the Cluster Developer Microsoft Windows HPC Server 2008 R2 for the Cluster Developer 50291B - Version: 1 5 days Course Description:
More informationMapReduce. U of Toronto, 2014
MapReduce U of Toronto, 2014 http://www.google.org/flutrends/ca/ (2012) Average Searches Per Day: 5,134,000,000 2 Motivation Process lots of data Google processed about 24 petabytes of data per day in
More informationDigital Single Market Technologies and Public Service Modernisation Package -DSM. Grazyna Wojcieszko DG CONNECT
Digital Single Market Technologies and Public Service Modernisation Package -DSM Grazyna Wojcieszko DG CONNECT Business and the Digital Universe, IDC, 2012 Tapping into the potential of data Data is not
More informationFunctional Partitioning to Optimize End-to-End Performance on Many-core Architectures
Functional Partitioning to Optimize End-to-End Performance on Many-core Architectures Min Li, Sudharshan S. Vazhkudai, Ali R. Butt, Fei Meng, Xiaosong Ma, Youngjae Kim,Christian Engelmann, and Galen Shipman
More informationAll routines were built with VS2010 compiler, OpenMP 2.0 and TBB 3.0 libraries were used to implement parallel versions of programs.
technologies for multi-core numeric computation In order to compare ConcRT, OpenMP and TBB technologies, we implemented a few algorithms from different areas of numeric computation and compared their performance
More informationComputing architectures Part 2 TMA4280 Introduction to Supercomputing
Computing architectures Part 2 TMA4280 Introduction to Supercomputing NTNU, IMF January 16. 2017 1 Supercomputing What is the motivation for Supercomputing? Solve complex problems fast and accurately:
More informationHaLoop Efficient Iterative Data Processing on Large Clusters
HaLoop Efficient Iterative Data Processing on Large Clusters Yingyi Bu, Bill Howe, Magdalena Balazinska, and Michael D. Ernst University of Washington Department of Computer Science & Engineering Presented
More informationStreaming Massive Environments From Zero to 200MPH
FORZA MOTORSPORT From Zero to 200MPH Chris Tector (Software Architect Turn 10 Studios) Turn 10 Internal studio at Microsoft Game Studios - we make Forza Motorsport Around 70 full time staff 2 Why am I
More informationREADEX Runtime Exploitation of Application Dynamism for Energyefficient
READEX Runtime Exploitation of Application Dynamism for Energyefficient exascale computing EnA-HPC @ ISC 17 Robert Schöne TUD Project Motivation Applications exhibit dynamic behaviour Changing resource
More informationOn the Use of Burst Buffers for Accelerating Data-Intensive Scientific Workflows
On the Use of Burst Buffers for Accelerating Data-Intensive Scientific Workflows Rafael Ferreira da Silva, Scott Callaghan, Ewa Deelman 12 th Workflows in Support of Large-Scale Science (WORKS) SuperComputing
More informationECMWF's Next Generation IO for the IFS Model and Product Generation
ECMWF's Next Generation IO for the IFS Model and Product Generation Future workflow adaptations Tiago Quintino, B. Raoult, S. Smart, A. Bonanni, F. Rathgeber, P. Bauer ECMWF tiago.quintino@ecmwf.int ECMWF
More informationUsing Industry Standards to Exploit the Advantages and Resolve the Challenges of Multicore Technology
Using Industry Standards to Exploit the Advantages and Resolve the Challenges of Multicore Technology September 19, 2007 Markus Levy, EEMBC and Multicore Association Enabling the Multicore Ecosystem Multicore
More informationHigh performance Computing and O&G Challenges
High performance Computing and O&G Challenges 2 Seismic exploration challenges High Performance Computing and O&G challenges Worldwide Context Seismic,sub-surface imaging Computing Power needs Accelerating
More informationEfficient AMG on Hybrid GPU Clusters. ScicomP Jiri Kraus, Malte Förster, Thomas Brandes, Thomas Soddemann. Fraunhofer SCAI
Efficient AMG on Hybrid GPU Clusters ScicomP 2012 Jiri Kraus, Malte Förster, Thomas Brandes, Thomas Soddemann Fraunhofer SCAI Illustration: Darin McInnis Motivation Sparse iterative solvers benefit from
More informationMaxwell: a 64-FPGA Supercomputer
Maxwell: a 64-FPGA Supercomputer Copyright 2007, the University of Edinburgh Dr Rob Baxter Software Development Group Manager, EPCC R.Baxter@epcc.ed.ac.uk +44 131 651 3579 Outline The FHPCA Why build Maxwell?
More informationIntroduction to OpenMP. OpenMP basics OpenMP directives, clauses, and library routines
Introduction to OpenMP Introduction OpenMP basics OpenMP directives, clauses, and library routines What is OpenMP? What does OpenMP stands for? What does OpenMP stands for? Open specifications for Multi
More informationIntroduction to GPU hardware and to CUDA
Introduction to GPU hardware and to CUDA Philip Blakely Laboratory for Scientific Computing, University of Cambridge Philip Blakely (LSC) GPU introduction 1 / 35 Course outline Introduction to GPU hardware
More informationA Breakthrough in Non-Volatile Memory Technology FUJITSU LIMITED
A Breakthrough in Non-Volatile Memory Technology & 0 2018 FUJITSU LIMITED IT needs to accelerate time-to-market Situation: End users and applications need instant access to data to progress faster and
More informationCME 213 S PRING Eric Darve
CME 213 S PRING 2017 Eric Darve Summary of previous lectures Pthreads: low-level multi-threaded programming OpenMP: simplified interface based on #pragma, adapted to scientific computing OpenMP for and
More informationEnzo-P / Cello. Formation of the First Galaxies. San Diego Supercomputer Center. Department of Physics and Astronomy
Enzo-P / Cello Formation of the First Galaxies James Bordner 1 Michael L. Norman 1 Brian O Shea 2 1 University of California, San Diego San Diego Supercomputer Center 2 Michigan State University Department
More informationGood Practices in Parallel and Scientific Software. Gabriel Pedraza Ferreira
Good Practices in Parallel and Scientific Software Gabriel Pedraza Ferreira gpedraza@uis.edu.co Parallel Software Development Research Universities Supercomputing Centers Oil & Gas 2004 Time Present CAE
More informationA Characterization of Shared Data Access Patterns in UPC Programs
IBM T.J. Watson Research Center A Characterization of Shared Data Access Patterns in UPC Programs Christopher Barton, Calin Cascaval, Jose Nelson Amaral LCPC `06 November 2, 2006 Outline Motivation Overview
More informationNVIDIA Think about Computing as Heterogeneous One Leo Liao, 1/29/2106, NTU
NVIDIA Think about Computing as Heterogeneous One Leo Liao, 1/29/2106, NTU GPGPU opens the door for co-design HPC, moreover middleware-support embedded system designs to harness the power of GPUaccelerated
More informationBlueGene/L. Computer Science, University of Warwick. Source: IBM
BlueGene/L Source: IBM 1 BlueGene/L networking BlueGene system employs various network types. Central is the torus interconnection network: 3D torus with wrap-around. Each node connects to six neighbours
More informationTopology and affinity aware hierarchical and distributed load-balancing in Charm++
Topology and affinity aware hierarchical and distributed load-balancing in Charm++ Emmanuel Jeannot, Guillaume Mercier, François Tessier Inria - IPB - LaBRI - University of Bordeaux - Argonne National
More informationRAMSES on the GPU: An OpenACC-Based Approach
RAMSES on the GPU: An OpenACC-Based Approach Claudio Gheller (ETHZ-CSCS) Giacomo Rosilho de Souza (EPFL Lausanne) Romain Teyssier (University of Zurich) Markus Wetzstein (ETHZ-CSCS) PRACE-2IP project EU
More informationAchieving Efficient Strong Scaling with PETSc Using Hybrid MPI/OpenMP Optimisation
Achieving Efficient Strong Scaling with PETSc Using Hybrid MPI/OpenMP Optimisation Michael Lange 1 Gerard Gorman 1 Michele Weiland 2 Lawrence Mitchell 2 Xiaohu Guo 3 James Southern 4 1 AMCG, Imperial College
More informationMapReduce-II. September 2013 Alberto Abelló & Oscar Romero 1
MapReduce-II September 2013 Alberto Abelló & Oscar Romero 1 Knowledge objectives 1. Enumerate the different kind of processes in the MapReduce framework 2. Explain the information kept in the master 3.
More informationRAIDIX Data Storage Solution. Clustered Data Storage Based on the RAIDIX Software and GPFS File System
RAIDIX Data Storage Solution Clustered Data Storage Based on the RAIDIX Software and GPFS File System 2017 Contents Synopsis... 2 Introduction... 3 Challenges and the Solution... 4 Solution Architecture...
More informationIntroduction to Grid Computing
Milestone 2 Include the names of the papers You only have a page be selective about what you include Be specific; summarize the authors contributions, not just what the paper is about. You might be able
More informationThe MPI Message-passing Standard Practical use and implementation (I) SPD Course 2/03/2010 Massimo Coppola
The MPI Message-passing Standard Practical use and implementation (I) SPD Course 2/03/2010 Massimo Coppola What is MPI MPI: Message Passing Interface a standard defining a communication library that allows
More informationVisualizing Biomolecular Complexes on x86 and KNL Platforms: Integrating VMD and OSPRay
Visualizing Biomolecular Complexes on x86 and KNL Platforms: Integrating VMD and OSPRay John E. Stone Theoretical and Computational Biophysics Group Beckman Institute for Advanced Science and Technology
More informationThe V-Ray installer will automatically remove any previous installations. This will allow you to seamlessly upgrade from previous versions.
Overview This guide is to serve as a reference for installing V-Ray for 3ds Max and 3ds Max Design. Please refer to the Glossary page for specific terms used in this document. If you experience any problems,
More informationPreviously... contour or image rendering in 2D
Volume Rendering Visualisation Lecture 10 Taku Komura Institute for Perception, Action & Behaviour School of Informatics Volume Rendering 1 Previously... contour or image rendering in 2D 2D Contour line
More informationIOS: A Middleware for Decentralized Distributed Computing
IOS: A Middleware for Decentralized Distributed Computing Boleslaw Szymanski Kaoutar El Maghraoui, Carlos Varela Department of Computer Science Rensselaer Polytechnic Institute http://www.cs.rpi.edu/wwc
More informationVarys. Efficient Coflow Scheduling. Mosharaf Chowdhury, Yuan Zhong, Ion Stoica. UC Berkeley
Varys Efficient Coflow Scheduling Mosharaf Chowdhury, Yuan Zhong, Ion Stoica UC Berkeley Communication is Crucial Performance Facebook analytics jobs spend 33% of their runtime in communication 1 As in-memory
More informationScalasca support for Intel Xeon Phi. Brian Wylie & Wolfgang Frings Jülich Supercomputing Centre Forschungszentrum Jülich, Germany
Scalasca support for Intel Xeon Phi Brian Wylie & Wolfgang Frings Jülich Supercomputing Centre Forschungszentrum Jülich, Germany Overview Scalasca performance analysis toolset support for MPI & OpenMP
More informationWorkflow and Direct Communication in the opendiel
Workflow and Direct Communication in the opendiel (Distributive Interoperable Executive Library) By Tanner Curren and Nicholas Moran Mentor: Kwai Wong What is the opendiel? Open Distributive Interoperable
More informationPerformance Optimization for an ARM Cortex-A53 System Using Software Workloads and Cycle Accurate Models. Jason Andrews
Performance Optimization for an ARM Cortex-A53 System Using Software Workloads and Cycle Accurate Models Jason Andrews Agenda System Performance Analysis IP Configuration System Creation Methodology: Create,
More information