Operating two InfiniBand grid clusters over 28 km distance

Similar documents
A Long-distance InfiniBand Interconnection between two Clusters in Production Use

Performance Analysis and Prediction for distributed homogeneous Clusters

bwgrid Treff am URZ Sabine Richling, Heinz Kredel Universitätsrechenzentrum Heidelberg Rechenzentrum Universität Mannheim 29.

A Long-distance InfiniBand Interconnection between two Clusters in Production Use

Hands-On Workshop bwunicluster June 29th 2015

bwfortreff bwhpc user meeting

Lessons learned from Lustre file system operation

Now SAML takes it all:

Datura The new HPC-Plant at Albert Einstein Institute

Virtualization of the ATLAS Tier-2/3 environment on the HPC cluster NEMO

Habanero Operating Committee. January

Rechenzentrum HIGH PERFORMANCE SCIENTIFIC COMPUTING

Sami Saarinen Peter Towers. 11th ECMWF Workshop on the Use of HPC in Meteorology Slide 1

Assistance in Lustre administration

A Case for High Performance Computing with Virtual Machines

The Optimal CPU and Interconnect for an HPC Cluster

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance

My operating system is old but I don't care : I'm using NIX! B.Bzeznik BUX meeting, Vilnius 22/03/2016

bwfdm Communities - a Research Data Management Initiative in the State of Baden-Wuerttemberg

To Infiniband or Not Infiniband, One Site s s Perspective. Steve Woods MCNC

Sun Lustre Storage System Simplifying and Accelerating Lustre Deployments

MOHA: Many-Task Computing Framework on Hadoop

Scalable Ethernet Clos-Switches. Norbert Eicker John von Neumann-Institute for Computing Ferdinand Geier ParTec Cluster Competence Center GmbH

Using the IBM Opteron 1350 at OSC. October 19-20, 2010

Composite Metrics for System Throughput in HPC

DELL EMC ISILON F800 AND H600 I/O PERFORMANCE

Users and utilization of CERIT-SC infrastructure

Parallel & Cluster Computing. cs 6260 professor: elise de doncker by: lina hussein

I/O Monitoring at JSC, SIONlib & Resiliency

How to run applications on Aziz supercomputer. Mohammad Rafi System Administrator Fujitsu Technology Solutions

Cluster Network Products

SUN CUSTOMER READY HPC CLUSTER: REFERENCE CONFIGURATIONS WITH SUN FIRE X4100, X4200, AND X4600 SERVERS Jeff Lu, Systems Group Sun BluePrints OnLine

How то Use HPC Resources Efficiently by a Message Oriented Framework.

Interconnect Your Future

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

ARCHER/RDF Overview. How do they fit together? Andy Turner, EPCC

Feedback on BeeGFS. A Parallel File System for High Performance Computing

HPC projects. Grischa Bolls

HPC Architectures. Types of resource currently in use

System Requirements. PREEvision. System requirements and deployment scenarios Version 7.0 English

Reduces latency and buffer overhead. Messaging occurs at a speed close to the processors being directly connected. Less error detection

Our new HPC-Cluster An overview

Patagonia Cluster Project Research Cluster

Real Application Performance and Beyond

InfoBrief. Platform ROCKS Enterprise Edition Dell Cluster Software Offering. Key Points

Ivane Javakhishvili Tbilisi State University High Energy Physics Institute HEPI TSU

Dell EMC Ready Bundle for HPC Digital Manufacturing Dassault Systѐmes Simulia Abaqus Performance

Session Case study. Planning of different broadband solutions in the last mile for urban and suburban areas. suburban areas

System Requirements. PREEvision. System requirements and deployment scenarios Version 7.0 English

High Performance Computing in C and C++

Communication has significant impact on application performance. Interconnection networks therefore have a vital role in cluster systems.

QLogic TrueScale InfiniBand and Teraflop Simulations

Computing for LHC in Germany

Overview of the Texas Advanced Computing Center. Bill Barth TACC September 12, 2011

HPC In The Cloud? Michael Kleber. July 2, Department of Computer Sciences University of Salzburg, Austria

PART-I (B) (TECHNICAL SPECIFICATIONS & COMPLIANCE SHEET) Supply and installation of High Performance Computing System

The Red Storm System: Architecture, System Update and Performance Analysis

Bright Cluster Manager Advanced HPC cluster management made easy. Martijn de Vries CTO Bright Computing

Supercomputer and grid infrastructure! in Poland!

Performance comparison between a massive SMP machine and clusters

Introduction to Cluster Computing

Advances of parallel computing. Kirill Bogachev May 2016

I/O at JSC. I/O Infrastructure Workloads, Use Case I/O System Usage and Performance SIONlib: Task-Local I/O. Wolfgang Frings

Optimizing Cluster Utilisation with Bright Cluster Manager

UAntwerpen, 24 June 2016

TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 6 th CALL (Tier-0)

2008 International ANSYS Conference

PREEvision. System Requirements. Version 7.5 English

Introduction to High Performance Parallel I/O

TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 11th CALL (T ier-0)

Integration of Cloud and Grid Middleware at DGRZR

InfiniBand and Next Generation Enterprise Networks

NUSGRID a computational grid at NUS

High-density Grid storage system optimization at ASGC. Shu-Ting Liao ASGC Operation team ISGC 2011

Before We Start. Sign in hpcxx account slips Windows Users: Download PuTTY. Google PuTTY First result Save putty.exe to Desktop

APENet: LQCD clusters a la APE

Introduction to High Performance Computing Using Sapelo2 at GACRC

ANALYSIS OF CLUSTER INTERCONNECTION NETWORK TOPOLOGIES

UL HPC Monitoring in practice: why, what, how, where to look

XSEDE New User Training. Ritu Arora November 14, 2014

PREEvision System Requirements. Version 9.0 English

SuperMike-II Launch Workshop. System Overview and Allocations

Grid and Networking in Armenia. Hrachya Astsatryan, Institute for Informatics and Automation Problem, National Academy of Sciences of Armenia

Scaling to Petaflop. Ola Torudbakken Distinguished Engineer. Sun Microsystems, Inc

Determining the MPP LS-DYNA Communication and Computation Costs with the 3-Vehicle Collision Model and the Infiniband Interconnect

High Performance Computing (HPC) Prepared By: Abdussamad Muntahi Muhammad Rahman

Managing complex cluster architectures with Bright Cluster Manager

Advanced Topics in High Performance Scientific Computing [MA5327] Exercise 1

How to Build a Cluster

VREDPro HPC Raytracing Cluster

Technical guide. Windows HPC server 2016 for LS-DYNA How to setup. Reference system setup - v1.0

Modeling Resource Utilization of a Large Data Acquisition System

The creation of a Tier-1 Data Center for the ALICE experiment in the UNAM. Lukas Nellen ICN-UNAM

Dell EMC Ready Bundle for HPC Digital Manufacturing ANSYS Performance

Network Design Considerations for Grid Computing

On-demand provisioning of HEP compute resources on cloud sites and shared HPC centers

Beyond Petascale. Roger Haskin Manager, Parallel File Systems IBM Almaden Research Center

Real Parallel Computers

Optimizing LS-DYNA Productivity in Cluster Environments

Maxwell: a 64-FPGA Supercomputer

Transcription:

Operating two InfiniBand grid clusters over 28 km distance Sabine Richling, Steffen Hau, Heinz Kredel, Hans-Günther Kruse IT-Center University of Heidelberg, Germany IT-Center University of Mannheim, Germany 3PGCIC-2010, Fukuoka, 4. November 2010 Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 1 / 41

Introduction Motivation Circumstances in Baden-Württemberg (BW) Increasing demand for high-performance computing capacities from scientific communities Demands are not high enough to qualify for the top German HPC centers in Jülich, Munich and Stuttgart Grid infrastructure concept for the Universities in Baden-Württemberg Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 2 / 41

Introduction Motivation Special Circumstances in Heidelberg/Mannheim Both IT-centers have a long record of cooperations Both IT-centers are connected by a 10 Gbit dark fibre connection of 28 km (two color lines already used for backup and other services) Connection of the clusters in Heidelberg and Mannheim to ease operation and to enhance utilization Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 3 / 41

Outline Introduction 1 Introduction 2 bwgrid cooperation 3 Interconnection of two bwgrid clusters 4 Cluster operation 5 Performance modeling 6 Summary and Conclusions Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 4 / 41

bwgrid cooperation bwgrid cooperation Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 5 / 41

bwgrid cooperation D-Grid German Grid Initiative (www.d-grid.de) Start: September 2005 Aim: Development and establishment of a reliable and sustainable Grid infrastructure for e-science in Germany Funded by the Federal Ministry of Education and Research (BMBF) with 50 Million Euro Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 6 / 41

bwgrid cooperation bwgrid Community project of the Universities of BW (www.bw-grid.de) Compute clusters at 8 locations: Stuttgart, Ulm (Konstanz), Karlsruhe, Tübingen, Freiburg, Mannheim/Heidelberg, Esslingen Central storage unit in Karlsruhe Distributed system with local administration Access for all D-Grid virtual organizations via at least one middleware supported by D-Grid Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 7 / 41

bwgrid cooperation bwgrid Objectives Verifying the functionality and the benefit of Grid concepts for the HPC community in BW Managing organisational and security problems Development of new cluster and Grid applications Solving license difficulties Enabling the computing centers to specialize Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 8 / 41

bwgrid cooperation bwgrid Access Possibilities Access with local university accounts (via ssh): Access to a local bwgrid cluster only Access with Grid Certificate and VO membership using a Grid middleware (e.g. Globus Toolkit: gsissh, GridFTP or Webservices): Access to all bwgrid resources Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 9 / 41

bwgrid cooperation bwgrid Resources Compute cluster: Mannheim/Heidelberg: 280 nodes Direct Interconnection Frankfurt Karlsruhe: 140 nodes Stuttgart: 420 nodes Tübingen: 140 nodes Mannheim (interconnected to a single cluster) Heidelberg Ulm (Konstanz): 280 nodes Hardware in Ulm Karlsruhe Stuttgart Freiburg: 140 nodes Esslingen Esslingen: 180 nodes more recent Hardware Central storage: Freiburg Tübingen Ulm (joint cluster with Konstanz) München Karlsruhe: 128 TB (with Backup) 256 TB (without Backup) Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 10 / 41

bwgrid Software bwgrid cooperation Common Software: Scientific Linux, Torque/Moab batch system, GNU and Intel compiler suite Central repository for software modules (MPI versions, mathematical libraries, various free software, application software from each bwgrid site) Application areas of bwgrid sites: Freiburg: System Technology, Fluid Mechanics Karlsruhe: Engineering, Compiler & Tools Heidelberg: Mathematics, Neuroscience Mannheim: Business Administration, Economics, Computer Algebra Stuttgart: Automotive simulations, Particle simulations Tübingen: Astrophysics, Bioinformatics Ulm: Chemistry, Molecular Dynamics Konstanz: Biochemistry, Theoretical Physics Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 11 / 41

Interconnection of two bwgrid clusters Interconnection of two bwgrid clusters Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 12 / 41

Interconnection of two bwgrid clusters Hardware before Interconnection 10 Blade-Center in Heidelberg and 10 Blade-Center in Mannheim Each Blade-Center contains 14 IBM HS21 XM Blades Each Blade contains 2 Intel Xeon CPUs, 2.8 GHz (each CPU with 4 Cores) 16 GB Memory 140 GB Hard Drive (since January 2009) Gigabit-Ethernet (1 Gbit) Infiniband Network (20 Gbit) 1120 Cores in Heidelberg and 1120 Cores in Mannheim Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 13 / 41

Interconnection of two bwgrid clusters Hardware Bladecenter Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 14 / 41

Interconnection of two bwgrid clusters Hardware Infiniband Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 15 / 41

Interconnection of two bwgrid clusters Interconnection of the bwgrid clusters Proposal in 2008 Acquisition and Assembly until May 2009 Running since July 2009 Infiniband over Ethernet over fibre optics: Longbow adaptor from Obsidian InfiniBand connector (black cable) fibre optic connector (yellow cable) Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 16 / 41

Interconnection of two bwgrid clusters Interconnection of the bwgrid clusters ADVA component: Transformation of white light from Longbow to one color light for the dark fibre connection between IT centers Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 17 / 41

Interconnection of two bwgrid clusters MPI Performance Prospects Measurements for different distances (HLRS, Stuttgart, Germany) Bandwidth 900-1000 MB/sec for up to 50-60 km Latency is not published Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 18 / 41

Interconnection of two bwgrid clusters MPI Performance Latency Local: 2 µsec Interconnection: 145 µsec 1e+06 100000 local MA-HD IMB 3.2 PingPong time [µsec] 10000 1000 100 10 1 1 10 100 1e3 1e4 1e5 1e6 buffer size [bytes] 1e7 1e8 1e9 Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 19 / 41

Interconnection of two bwgrid clusters MPI Performance Bandwidth Local: 1400 MB/sec Interconnection: 930 MB/sec bandwidth [Mbytes/sec] 1800 1600 1400 1200 1000 800 600 400 200 local MA-HD IMB 3.2 PingPong 0 1 10 100 1e3 1e4 1e5 1e6 buffer size [bytes] 1e7 1e8 1e9 Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 20 / 41

Interconnection of two bwgrid clusters Experiences with Interconnection Network Cable distance MA-HD is 28 km (18 km linear distance in air) Light needs 143 µsec for this distance Latency is high: 145 µsec = Light transit time + 2 µsec local latency Bandwidth is as expected: about 930 MB/sec Local bandwidth 1200-1400 MB/sec Obsidian needs a license for 40 km Obsidian has buffers for larger distances Activation of buffers with license License for 10 km is not sufficient Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 21 / 41

Interconnection of two bwgrid clusters MPI Bandwidth Influence of the Obsidian License IMB 3.2 - PingPong - buffer size 1 GB bandwidth [Mbytes/sec] 1000 800 600 400 200 0 16 Sep 00:00 23 Sep 00:00 30 Sep 00:00 start time [date hour] 07 Oct 00:00 Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 22 / 41

Cluster operation Cluster operation Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 23 / 41

Cluster operation bwgrid Cluster Mannheim/Heidelberg Belwue VORM Benutzer Benutzer PBS Benutzer Benutzer VORM LDAP Admin AD Cluster Mannheim InfiniBand Obsidian + ADVA passwd InfiniBand Cluster Heidelberg Lustre bwfs MA Lustre bwfs HD Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 24 / 41

Cluster operation bwgrid Cluster Mannheim/Heidelberg Overview Two clusters (blue boxes) are connected by InfiniBand (orange lines) Obsidian and ADVA (orange box) represents the 28 km fibre connection bwgrid storage systems (grey boxes) are also connected by Infiniband Access nodes ( Benutzer ) are connected with 10 GBit (light orange lines) to the outside Internet Belwue (BW science net) Access with local accounts from Mannheim ( LDAP ) Access with local accounts from Heidelberg ( AD ) Access with Grid certificates ( VORM ) Ethernet connection between all components is not shown Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 25 / 41

Node Management Cluster operation Compute nodes are booted via PXE and use NFS read-only export as root file system Administration server provides DHCP service for the nodes (MAC-to-IP address configuration file) NFS export for root file system NFS directory for software packages accessible via module utilities queuing and scheduling system Node administration (power on/off, execute commands, BIOS update, etc.) with adjusted shell scripts originally developed by HLRS IBM management module (command line interface and Web-GUI) Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 26 / 41

User Management Cluster operation Users should have exclusive access to compute nodes user names and user-ids must be unique replacing passwd with reduced passwd proofed unreliable better is a direct connection to PBS for user authorization via PAM module Authentication at the access nodes directly against directory services: LDAP (MA) and AD (HD) or with D-Grid certificate Combining information from directory services from both universities Prefix ma, hd or mh for group names Adding offsets to group-ids Adding offsets to user-ids Activated user names from MA and HD must be different Activation process Adding a special attribute for the user in the directory service (for authentication) Updating the user database of the cluster (for authorization) Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 27 / 41

Cluster operation Job Management Interconnection (high latency, limited bandwidth) provides enough bandwidth for I/O operations not sufficient for all kinds of MPI jobs Jobs only run on nodes located either in HD or in MA (realized with attributes provides by the queuing system) Before interconnection In Mannheim: mostly single node jobs free nodes In Heidelberg: many MPI jobs long waiting times With interconnection better resource utilization (see Ganglia report) Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 28 / 41

Cluster operation Monitoring Report during activation of the interconnection Number of processes Percent CPU Usage Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 29 / 41

Performance modeling Performance modeling Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 30 / 41

Performance modeling MPI Jobs running across the interconnection How does the interconnection influence the performance? How much bandwidth would be necessary to the improve the performance? How much would such an upgrade cost? Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 31 / 41

Performance modeling Performance modeling Numerical model High-Performance Linpack (HPL) benchmark OpenMPI Intel MKL Model variants Calculations on a single cluster with up to 1025 CPU cores Calculations on the coupled cluster with up to 2048 CPU cores symmetrically distributed Analytical model for the speed-up to analyze the characteristics of the interconnection high latency of 145 µsec limited bandwidth of 930 MB/sec Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 32 / 41

Performance modeling Results for a single cluster speed-up 300 250 200 150 100 50 HPL 1.0a local p/ln(p) 0 0 500 1000 1500 2000 number of processors p n p =40000 n p =30000 n p =20000 n p =10000 n p load parameter (matrix size for HPL) ideal speed-up for perfect parallel programs S ideal (p) = p speed-up for a simple model all CPU configurations have equal probability S simple (p) = p/ ln p Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 33 / 41

Performance modeling Results for coupled cluster speed-up 100 80 60 40 HPL 1.0a MA-HD n p =40000 n p =30000 n p =20000 n p =10000 n p load parameter (matrix size for HPL) for p > 256 reduced speed-up by a factor of 4 compared to single cluster 20 0 0 500 1000 1500 2000 number of processors p for p > 500 constant (decreasing) speed-up Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 34 / 41

Performance modeling Direct comparison of the two cases HPL 1.0a speed-up 100 10 1 local n p =40000 n p =10000 MA-HD n p =40000 n p =10000 n p load parameter (matrix size for HPL) for p < 50 speed-up for coupled cluster is acceptable, applications could run across interconnection effectively (in the case of exclusive usage) 1 10 100 1000 number of processors p Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 35 / 41

Performance modeling Performance modeling Following a performance model developed by Kruse (2009): t c (p): communication time t B (1): processing time for p = 1 Speed-up S c (p) p ln p + tc(p) t B (1) For t c (p) = 0, we receive the result of the simple model: S simple (p) = p/ ln p Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 36 / 41

Performance modeling Performance model for the high latency Modeling t c (p) as a function of the typical communication time between 2 processes t c (2) an the communication topology c(p): t c (p) = t c (2) c(p) Defining a rate r = t c (2) /t A between t c (2) and the computation time for a typical instruction t A = t B (1)/n: Speed-up S c (p) p ln p + r n c(p) Analysis for HPL (n = 2 3 n3 p): for n p = 1000: p/ ln p for small p, decrease for p 30 for n p = 10 000: p/ ln p for p 10 000, decrease for c(p) > 10 6 Analysis does not explain the numerical results. Decrease of speed-up already for smaller p. Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 37 / 41

Performance modeling Performance model including a limited bandwidth Modeling the interconnection as a shared medium for the communication of p processes with a given bandwidth B and average message length < m >: t (2) c = t L + m (B/p) r(p) = t L ta + m t A B p With the measured bandwidth B = 1.5 10 6 and m = 10 6 : Speed-up S c (p) p ln p + 3 4 ( 100 n p ) 3 (1 + 4p)c(p) With assumption c(p) = 1 2 p2 : for n p = 10 000: p/ ln p, decrease for p 50 for n p = 40 000: p/ ln p, decrease for p 250 Speed-up reproduces the measurements. Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 38 / 41

Performance modeling Speed-up of the model including limited bandwidth n p load parameter (matrix size for HPL) limited bandwidth is the performance bottleneck for shared connection between the clusters Doubling the bandwidth: 25 % improvement for n p = 40 000 100 % improvement with a ten-fold bandwidth (in the case of exclusive usage) Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 39 / 41

Summary and Conclusions Summary and Conclusions Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 40 / 41

Summary and Conclusions InfiniBand connection of two compute clusters Network (Obsidian, ADVA and Infiniband switches) is stable and reliable Latency of 145 µsec is very high Bandwidth of 930 MB/sec is as expected Jobs are limited to one site, because MPI jobs would be slow (Interconnection is a shared medium ) Performance model predicts the cost for an improvement of the interconnection Bandwidth sufficient for cluster administration and file I/O on Lustre file systems Interconnection is useful and stable for a Single System Cluster administration Better load balance at both sites due to common PBS Solving organizational issues between two universities is a great challenge Richling, Hau, Kredel, Kruse (URZ/RUM) Operating two grid clusters over 28 km Fukuoka, November 2010 41 / 41