Meta-cluster of distributed computing Dubna-Grid

Similar documents
CLOUDS OF JINR, UNIVERSITY OF SOFIA AND INRNE JOIN TOGETHER

The JINR Tier1 Site Simulation for Research and Development Purposes

JINR cloud infrastructure

LHC Grid Computing in Russia: present and future

The Status and Perspectives of the JINR 10 Gbps Network Infrastructure

The LHC Computing Grid

Andrea Sciabà CERN, Switzerland

Interoperating AliEn and ARC for a distributed Tier1 in the Nordic countries.

Ivane Javakhishvili Tbilisi State University High Energy Physics Institute HEPI TSU

Grid Computing a new tool for science

LCG MCDB Knowledge Base of Monte Carlo Simulated Events

Austrian Federated WLCG Tier-2

Evaluation of the computing resources required for a Nordic research exploitation of the LHC

RUSSIAN DATA INTENSIVE GRID (RDIG): CURRENT STATUS AND PERSPECTIVES TOWARD NATIONAL GRID INITIATIVE

IEPSAS-Kosice: experiences in running LCG site

Support for multiple virtual organizations in the Romanian LCG Federation

University of Johannesburg South Africa. Stavros Lambropoulos Network Engineer

The ATLAS Tier-3 in Geneva and the Trigger Development Facility

ISTITUTO NAZIONALE DI FISICA NUCLEARE

CONTROL SYSTEM OF THE BOOSTER INJECTION POWER SUPPLY

Evolution of the ATLAS PanDA Workload Management System for Exascale Computational Science

Distributing storage of LHC data - in the nordic countries

AGIS: The ATLAS Grid Information System

Federated Data Storage System Prototype based on dcache

STATUS OF PLANS TO USE CONTAINERS IN THE WORLDWIDE LHC COMPUTING GRID

Grid Computing at Ljubljana and Nova Gorica

ATLAS NorduGrid related activities

Development of new security infrastructure design principles for distributed computing systems based on open protocols

EGEE and Interoperation

HP integrated Citrix XenServer Online Help

Implementation and. Oracle VM. Administration Guide. Oracle Press ORACLG. Mc Grauv Hill. Edward Whalen

ANALYSIS OF THE FEATURES OF THE OPTIMAL LOGICAL STRUCTURE OF DISTRIBUTED DATABASES

Use of the virtualized HPC infrastructure of the Novosibirsk Scientific Center for running production analysis for HEP experiments at BINP

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

GRIDS INTRODUCTION TO GRID INFRASTRUCTURES. Fabrizio Gagliardi

the one of the receiver. Apache ActiveMQ is applied as the message broker in WLCG. It is an open-source software written in Java and supporting such

Volunteer Computing at CERN

Federated data storage system prototype for LHC experiments and data intensive science

A distributed tier-1. International Conference on Computing in High Energy and Nuclear Physics (CHEP 07) IOP Publishing. c 2008 IOP Publishing Ltd 1

CernVM-FS beyond LHC computing

A short introduction to the Worldwide LHC Computing Grid. Maarten Litmaath (CERN)

Grids and Security. Ian Neilson Grid Deployment Group CERN. TF-CSIRT London 27 Jan

Data Management for the World s Largest Machine

Virtualizing a Batch. University Grid Center

Provisioning Oracle RAC in a Virtualized Environment, Using Oracle Enterprise Manager

Interfacing Operational Grid Security to Site Security. Eileen Berman Fermi National Accelerator Laboratory

Clouds at other sites T2-type computing

LCG Installation LCFGng

Geant4 on Azure using Docker containers

ATLAS Nightly Build System Upgrade

ELFms industrialisation plans

Ó³ Ÿ , º 5(203) Ä1050. Š Œ œ ƒˆˆ ˆ ˆŠ

Network Services, Cloud Computing and Virtualization

VMware vfabric Data Director Installation Guide

Virtualizing Oracle 11g/R2 RAC Database on Oracle VM: Methods/Tips

High Performance Computing Course Notes Grid Computing I

Experience of Data Grid simulation packages using.

PrepAwayExam. High-efficient Exam Materials are the best high pass-rate Exam Dumps

ALICE Grid Activities in US

30 Nov Dec Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy

MONTE CARLO SIMULATION FOR RADIOTHERAPY IN A DISTRIBUTED COMPUTING ENVIRONMENT

Long Term Data Preservation for CDF at INFN-CNAF

ATLAS COMPUTING AT OU

Singularity in CMS. Over a million containers served

/ sicam SICAM CMIC The smart cube for your distribution network Answers for infrastructure and cities.

CSCS CERN videoconference CFD applications

Benchmarking the ATLAS software through the Kit Validation engine

Exploring cloud storage for scien3fic research

Verification and Diagnostics Framework in ATLAS Trigger/DAQ

10 Gbit/s Challenge inside the Openlab framework

Federated Services for Scientists Thursday, December 9, p.m. EST

RDMS CMS Computing Activities before the LHC start

APPROACHES TO THE AUTOMATED DEPLOYMENT OF THE CLOUD INFRASTRUCTURE OF GEOGRAPHICALLY DISTRIBUTED DATA CENTERS

JINR cloud infrastructure development

Travelling securely on the Grid to the origin of the Universe

SAM at CCIN2P3 configuration issues

Scientific data management

GRID 2008 International Conference on Distributed computing and Grid technologies in science and education 30 June 04 July, 2008, Dubna, Russia

For more information, see "Provision APs for Mesh" on page 6 6. Connect your APs to the network. See "Install the APs" on page 6

ATLAS TDAQ System Administration: evolution and re-design

Tests of PROOF-on-Demand with ATLAS Prodsys2 and first experience with HTTP federation

Monte Carlo Production on the Grid by the H1 Collaboration

ATTACHMENT A SCOPE OF WORK IMPLEMENTATION SERVICES. Cisco Server and NetApp Storage Implementation

Site Report. Stephan Wiesand DESY -DV

Use of containerisation as an alternative to full virtualisation in grid environments.

GENSER, the generator repository in LCG

Linux Automation.

Testing Under Time Pressure. Versatile Test Benches for Avionic Systems

Introduction to Geant4

Clouds in High Energy Physics

Bringing ATLAS production to HPC resources - A use case with the Hydra supercomputer of the Max Planck Society

Improved ATLAS HammerCloud Monitoring for Local Site Administration

The Cisco HCM-F Administrative Interface

Tracking and flavour tagging selection in the ATLAS High Level Trigger

User interface for a computational cluster: resource description approach

Physics Computing at CERN. Helge Meinhard CERN, IT Department OpenLab Student Lecture 21 July 2011

Construction and Application of Cloud Data Center in University

UL TEST REPORT AND PROCEDURE

Experiences with OracleVM 3.3

ATLAS Experiment and GCE

Transcription:

Meta-cluster of distributed computing Dubna-Grid Creation of a multi-purpose town scale network of distributed computing Peter Zrelov, JINR on behalf of the Dubna-Grid Collaboration

AGREEMENT Between - Administration of Dubna, - Joint Institute for Nuclear Research, - International University of Nature, Society and Man DUBNA Subject: Creation of a city multi-purpose informational infrastructure of the new generation based on the Gridtechnology 2

Project Dubna-Grid Participants: JINR Belyakov D.V., Chkhaberidze D. Golub D.C., Dolbilov A.G., Ivanov V.V., Ivanov Valeri V., Kalmikova L.A., Korenkov V.V., Mitsyn V.V., Popov L.A., Strizh T.A., Zrelov P.V. Dubna University Antonov A.N., Kryukov Yu.A., Cheremisina E.N. Administration of Dubna City Dobromislov S.N., Rats A.A., Rybov E.B. Lund University Smirnova O.G. Chicago University Smirnov Yu.S. 3

Main purposes of the project The Dubna-Grid Project is aimed at the creation of a distributed environment of meta-computing on the basis of vacant computing resources of office computers. The project foresees creation of a common pool of accessible nodes of more than 1000 units. Implementation of the project Dubna-Grid will allow: to create a unified computing environment of the city Dubna (meta-cluster) on the basis of the resources of scientific and educational institutions, in particular, JINR subdivisions, University Dubna, secondary schools and other organizations; to create a segment of the international Grid-infrastructure, operating in frames of various Grid systems (LGG, NorduGrid and OSG); to provide execution of large-scale computing tasks of JINR and other enterprises of Dubna scientific industrial complex and other organizations; to distribute experience of creation of the city segment of the Grid-infrastructure for creation of similar systems in Russia. 4

Motivation of the participating organizations The Joint Institute for Nuclear Research is interested in additional computing resources, in order to conduct research in frames both the current experiments and the experiments that are being prepared at the LHC. The University Dubna and MIREA are interested in training of young specialists for work with advanced information technologies. The city of Dubna as a City of Science, is interested in creation of informational computational infrastructure for high-technology production and scientific research as well as for using the information technologies for solving tasks in various spheres of the city life. The participants from the universities of Lund and Chicago are interested in joint testing of different Grid systems. 5

Responsibilities of participating organizations Administration of the Dubna city provides an informational educational computer network upgrades the computers in computer classes of secondary schools 6

Responsibilities of the participating organizations JINR develops and organizes a testing exploitation of the test-bed architecture provides pool of office computers and managing servers installs various Grid systems on the test-bed organizes a joint Grid laboratory for research in the field of Grid technologies and trains specialists of a corresponding profile 7

Responsibilities of the participating organizations University Dubna provides 500 computers from the university network creates a server center for the maintenance of the university segment of the meta-cluster works out the courses on Grid technologies and trains specialists in frames of the joint Grid-laboratory 8

Main stages of Project realization development of technology of the meta-cluster construction on the basis of software of virtual machines creation of a meta-cluster prototype integrating resources of various organizations development of mass installation technologies and spreading software to all accessible nodes of the city infrastructure development of the meta-cluster monitoring system installation of various Grid systems at the meta-cluster training of specialists in the field of Grid technologies. 9

Network resources of the Project City Informational-Educational Computer Network JINR Backbone 10

Current State 24 nodes 112 nodes 24 nodes 212 nodes 28 nodes 9 nodes 24 nodes 8 nodes (GridLab) 11

Construction of the system Dubna-Grid meta-cluster includes central server, bridge and computational nodes. Computational nodes represent Virtual PCs emulated by virtual machine technology (VMware). By means of VMware we get two computers instead of one, both have their own memory, IP address etc., but physically they use the same resources - two machines share one processor, memory, Ethernet card etc. 12

Construction of the system (cont d) The following software tools and technologies were used: software support for virtual machines maintenance (VMware), virtual network (VLAN), virtual access to the software and data (AFS), building and loading OS images to the Virtual PCs of the metacluster. On the central server the following software is installed: OS Scientific Linux CERN; a package for support of cluster architecture Warewulf; Ganglia monitoring system; OpenAFS, the batch system Torque with Maui scheduler. 13

Logical scheme of the Dubna-Grid meta-cluster 14

Schematic view of loading the computing node Loading Process Server DHCP TFTP HTTP NFS NTP TORQUE WAREWULF GANGLIA Get IP address Load Kernel Load OS File System Emulated Virtual PC Mount some repositories Synchronized time Get Jobs Real Window s 15

Monitoring of the Dubna-Grid Meta-Cluster 16

Test infrastructure at JINR Tasks: Check-up and test exploitation of new versions of the software for nodes and servers of the metacluster Installation of various Grid-systems Training of specialists in the field of Grid technologies Creation of the test complex is taking place in the framework of the project of the joint Grid laboratory GridLab at LIT JINR 17

The Project Parameters Parameters/ years 2005 2006 2007 2008 A number of Grid nodes involved 200 500 1000 1500 18

How to submit job to the metacluster? Using batch system of the JINR Central Information and Computer Complex (CICC) Using web-interface for some specific tasks (now it is in test mode for the project HIGH Energy Physics WEB at LIT (HEPWEB server) The main aim of this project is to perform Monte-Carlo simulations of high energy physics processes, and to evaluate the main properties of the hadron-hadron and nusleus-nucleus interactions using WEB access to LIT computer resources in the framework of the JINR participation in LCG program (GENSER). http://hepweb.jinr.ru 19

High Energy Physics WEB at LIT 20

HEPWEB: Interfaces to generators 21

Integration of HEPWEB server in DUBNA-GRID enviroment 22

Organization problems Provision of round-the-clock operation of the meta-cluster nodes, requires control over temperature and installation of an improved ventilation or air-conditioning systems Necessity of setting up fire-alarm networks and organization of a round-the-clock monitoring in frames of the whole city. Training of system administrators of computer classes. 23

Conclusion A test-bed of distributed meta-computing environment of Dubna city based on vacant computing resources of office computers of scientific and educational institutions has been created. Mass installation technologies and spreading software to all accessible nodes of the meta-cluster have been developed. Monitoring system of the meta-cluster has been developed. The meta-cluster have been integrated with JINR batch system. First real tasks have been performed (including real tasks for the ATLAS experiment). Integration of HEPWEB server in DUBNA-GRID enviroment is realized. 24