Geant4 in a Distributed Computing Environment

Similar documents
Geant4 simulation in a distributed computing environment

MONTE CARLO SIMULATION FOR RADIOTHERAPY IN A DISTRIBUTED COMPUTING ENVIRONMENT

Benchmarks of medical dosimetry simulation on the grid

Verification / Validation Tools

Test & Analysis Project aka statistical testing

Physics data management tools: computational evolutions and benchmarks

Design and performance evaluations of generic programming techniques in a R&D prototype of Geant4 physics

INTRODUCTION TO THE ANAPHE/LHC++ SOFTWARE SUITE

UW-ATLAS Experiences with Condor

Status of KISTI Tier2 Center for ALICE

A new development cycle of the Statistical Toolkit

The GAP project: GPU applications for High Level Trigger and Medical Imaging

I Tier-3 di CMS-Italia: stato e prospettive. Hassen Riahi Claudio Grandi Workshop CCR GRID 2011

Geant4 Computing Performance Benchmarking and Monitoring

The Grid: Processing the Data from the World s Largest Scientific Machine

Ganga - a job management and optimisation tool. A. Maier CERN

Simulation Techniques Using Geant4

Batch Services at CERN: Status and Future Evolution

A Prototype of the CMS Object Oriented Reconstruction and Analysis Framework for the Beam Test Data

Fractals. Investigating task farms and load imbalance

CERN and Scientific Computing

GRIDS INTRODUCTION TO GRID INFRASTRUCTURES. Fabrizio Gagliardi

LHCb Computing Strategy

Storage and I/O requirements of the LHC experiments

arxiv: v1 [physics.ins-det] 19 Oct 2017

System upgrade and future perspective for the operation of Tokyo Tier2 center. T. Nakamura, T. Mashimo, N. Matsui, H. Sakamoto and I.

High Performance Computing from an EU perspective

Scalable Computing: Practice and Experience Volume 10, Number 4, pp

Monitoring of Computing Resource Use of Active Software Releases at ATLAS

Fractals exercise. Investigating task farms and load imbalance

GAUDI - The Software Architecture and Framework for building LHCb data processing applications. Marco Cattaneo, CERN February 2000

Kepler Scientific Workflow and Climate Modeling

Big Computing and the Mitchell Institute for Fundamental Physics and Astronomy. David Toback

CMS HLT production using Grid tools

ATLAS Tracking Detector Upgrade studies using the Fast Simulation Engine

A short introduction to the Worldwide LHC Computing Grid. Maarten Litmaath (CERN)

IN a system of many electronics boards of many different

Improved ATLAS HammerCloud Monitoring for Local Site Administration

Altix Usage and Application Programming

e-science for High-Energy Physics in Korea

Analisi Tier2 e Tier3 Esperienze ai Tier-2 Giacinto Donvito INFN-BARI

Virtualization. A very short summary by Owen Synge

OBJECT ORIENTED DESIGN OF ANTHROPOMORPHIC PHANTOMS AND GEANT4-BASED IMPLEMENTATIONS ABSTRACT

BOSCO Architecture. Derek Weitzel University of Nebraska Lincoln

Ganga The Job Submission Tool. WeiLong Ueng

10 Gbit/s Challenge inside the Openlab framework

DIRAC pilot framework and the DIRAC Workload Management System

Clustering and Reclustering HEP Data in Object Databases

The LHC Computing Grid. Slides mostly by: Dr Ian Bird LCG Project Leader 18 March 2008

A High Availability Solution for GRID Services

An innovative compilation tool-chain for embedded multi-core architectures M. Torquati, Computer Science Departmente, Univ.

Machine Learning for (fast) simulation

Anna Morajko.

NUSGRID a computational grid at NUS

Constant monitoring of multi-site network connectivity at the Tokyo Tier2 center

The GTPC Package: Tracking and Analysis Software for GEM TPCs

CMS users data management service integration and first experiences with its NoSQL data storage

Multi-Channel Clustered Web Application Servers

Workload Management. Stefano Lacaprara. CMS Physics Week, FNAL, 12/16 April Department of Physics INFN and University of Padova

CREATE 2 Control Room Engineering Advanced Toolkit Environment

The Grid. Processing the Data from the World s Largest Scientific Machine II Brazilian LHC Computing Workshop

CMS Simulation Software

Ivane Javakhishvili Tbilisi State University High Energy Physics Institute HEPI TSU

PROOF-Condor integration for ATLAS

Storage Resource Sharing with CASTOR.

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Parallelizing Multiple Group by Query in Shared-nothing Environment: A MapReduce Study Case

An approach to calculate and visualize intraoperative scattered radiation exposure

SPACE SITUATIONAL AWARENESS

Hungarian Supercomputing Grid 1

LHCb Distributed Conditions Database

Bob Jones. EGEE and glite are registered trademarks. egee EGEE-III INFSO-RI

The INFN Tier1. 1. INFN-CNAF, Italy

ATLAS software configuration and build tool optimisation

High-density Grid storage system optimization at ASGC. Shu-Ting Liao ASGC Operation team ISGC 2011

Image-based Monte Carlo calculations for dosimetry

Grid technologies, solutions and concepts in the synchrotron Elettra

Preparing for High-Luminosity LHC. Bob Jones CERN Bob.Jones <at> cern.ch

Simulation Techniques Using Geant4

[workshop welcome graphics]

ALHAD G. APTE, BARC 2nd GARUDA PARTNERS MEET ON 15th & 16th SEPT. 2006

Grid-Based Genetic Algorithm Approach to Colour Image Segmentation

Multi-threaded, discrete event simulation of distributed computing systems

Trivial And Non-Trivial Data Analysis for Geant4

Clouds in High Energy Physics

Benchmarking and accounting for the (private) cloud

Geant4 v9.5. Kernel III. Makoto Asai (SLAC) Geant4 Tutorial Course

Digital Opportunity Index. Michael Minges Telecommunications Management Group, Inc.

CSCS CERN videoconference CFD applications

Outline. Infrastructure and operations architecture. Operations. Services Monitoring and management tools

FAMOS: A Dynamically Configurable System for Fast Simulation and Reconstruction for CMS

Trivial Data Analysis for Geant4 Geant4 v9.3p01

Message Passing Interface (MPI)

Application of Virtualization Technologies & CernVM. Benedikt Hegner CERN

GFS: A Distributed File System with Multi-source Data Access and Replication for Grid Computing

ATLAS DQ2 to Rucio renaming infrastructure

Track pattern-recognition on GPGPUs in the LHCb experiment

Cloud Computing. Summary

A comparison of data-access platforms for BaBar and ALICE analysis computing model at the Italian Tier1

Interoperating AliEn and ARC for a distributed Tier1 in the Nordic countries.

Transcription:

Geant4 in a Distributed Computing Environment S. Guatelli 1, P. Mendez Lorenzo 2, J. Moscicki 2, M.G. Pia 1 1. INFN Genova, Italy, 2. CERN, Geneva, Switzerland Geant4 2005 10 th user conference and collaboration workshop November 3-10, 2005, Bordeaux, France

Vision Problem How to obtain a quick response from a Geant4 simulation Case 1: quick response in few minutes i.e. dosimetry, study the efficiency of detectors Case 2: reasonable time for response from G4 simulations requiring high statistics i.e. medical, space science, high energy physics applications, tests of Geant4 physics models Solution Parallelisation On dedicated pc clusters On the GRID Study a general approach, independent from the specific Geant4 application

Quick response Parallelisation Transparent configuration in sequential or parallel mode Access to distributed computing resources Transparent access to the GRID through an intermediate software layer

Strategy Study the performance of two Geant4 applications as typical examples: Geant4 Brachytherapy application Geant4 IMRT application G4 Brachytherapy application Sequential mode on a Pentium IV, 3 GHz G4 IMRT application Execution time of 20 M events ~ 5 hours Goal: quick response ~ few minutes Execution time of 10 9 events ~ 9 days and half Goal: quick response ~ few hours Parallelisation through DIANE Performance tests On a single CPU On clusters On the GRID Quantitative analysis of the results

Outline Diane overview How to dianize a G4 application Results of performance tests Conclusions

DIANE Overview DIANE R&D Project started in 2001 in CERN/IT with very limited resources collaboration with Geant 4 groups at CERN, INFN, ESA succesful prototypes running on LSF and EDG Developed by J. Moscicki, CERN http://cern.ch/diane Parallel cluster processing make fine tuning and customisation easy transparently using GRID technology application independent

Practical Example Example: simulation with analysis The job is divided into tasks The tasks are executed on worker components Each task produces a file with histograms Job result = sum of histograms produced by tasks Master-worker model Client starts a job Workers perform tasks and produce histograms Master integrates the results

Running in a distributed environment The application developer is shielded from the complexity of underlying technology via DIANE Not affecting the original code of the application standalone and distributed case is the same code Good separation of the subsystems the application does not need to know that it runs in a distributed environment the distributed framework (DIANE) does not need to care about what actions an application performs internally

How to dianize a G4 application Look at the Geant4 extended example: ExDIANE in the parallel directory Completely transparent to the user: same G4 code Documentation at http://www.cern.ch/diane specific for Geant4 applications Installing and compiling DIANE Compiling and running a Geant4 application through DIANE

Test results Study the performance of the execution of the dianized G4Brachy: Test on a single CPU Test on a dedicated farm (60 CPUs) Test on a farm, shared with other users (LSF, CERN) Test on the GRID Tools and libraries: Simulation toolkit: Geant4 7.0.p01 Analysis tools: AIDA 3.2.1 and PI 1.3.3 DIANE: DIANE 1.4.2 CLHEP: 1.9.1.2 G4EMLOW2.3

Results G4Brachy: 1 CPU Test on a single dedicated CPU (Intel, Pentium IV, 3.00 GHz) Execution time with respect to the number of events of the job The overhead of DIANE is negligible in high statistics jobs Ratio = ExecutionTime ExecutionTime DIANE SEQUENTIAL _ MODE with respect to the number of events

Results G4Brachy: farm Dedicated farm : 30 identical biprocessors (Pentium IV, 3 GHz) Thanks to Hurng-Chun Lee (Academia Sinica Grid Computing Center, Taiwan) Thanks to Regional Operation Centre (ROC) Team, Taiwan ExecutionTime Efficiency = ExecutionTime sequential parallel n

Comment The job ends when all the tasks are executed on the workers If the job is split into a higher number of tasks, there is a higher chance that the workers finish the tasks at the same moment Worker number Worker number Time (seconds) Example of a job that can be improved from a performance point of view Time (seconds) Example of a good job balancing

Results G4Brachy: farm (3) Test on LSF cluster of CERN: case of farm shared with other users Ratio = ExecutionTime ExecutionTime PARALLEL _ MODE _ Diane PARALLEL _ MODE _ NoDiane Execution in parallel mode on 5 workers of LSF Comparison Execution in parallel mode on 5 workers of LSF DIANE used as intermediate layer Preliminary! The load of the cluster changes quickly in time The conditions of the test are not reproducible

Results G4Brachy: GRID (1) The load of the GRID changes quickly in time The conditions of the test are not reproducible G4Brachy executed on the GRID on nodes located in Spain, Russia, Italy, Germany, Switzerland ExecutionTime ExecutionTime PARALLEL_ MODE _ Diane PARALLEL_ MODE _ NoDiane 0.3 Worker number Execution on the GRID, without DIANE Execution on the GRID through DIANE, Worker number 20 M events,180 tasks, 30 workers Time (seconds) Without DIANE: - 2 jobs not successfully executed due to set-up problems of the workers Time (seconds) Through DIANE: - All the tasks are executed successfully on 22 workers - Not all the workers are initialized and used: on-going investigation

How the GRID load changes Execution time of G4Brachy in two different conditions of the GRID DIANE used as intermediate layer Worker number Worker number Time (seconds) Time (seconds) 20 M events, 60 workers initialized, 360 tasks

Conclusions General approach to obtain quick response from Geant4 simulations Advantage of using DIANE as intermediate layer in a dedicated farm or GRID Transparency Good separation of the subsystems Good management of the CPU resources DIANE is very advantageous as an intermediate layer to the GRID from a performance point of view A quantitative analysis of the performance results is in progress Submission of this work for publication in IEEE Trans. Nucl. Sci. Acknowledgments to: M. Lamanna (CERN), Hurng-Chun Lee (ASGC, Taiwan), L. Moneta (CERN), A. Pfeiffer (CERN) Thanks to the GRID team of CERN and the Regional Operation Centre Team of Taiwan